WO2023231478A1 - Audio and video sharing method and device, and computer-readable storage medium - Google Patents

Audio and video sharing method and device, and computer-readable storage medium Download PDF

Info

Publication number
WO2023231478A1
WO2023231478A1 PCT/CN2023/078498 CN2023078498W WO2023231478A1 WO 2023231478 A1 WO2023231478 A1 WO 2023231478A1 CN 2023078498 W CN2023078498 W CN 2023078498W WO 2023231478 A1 WO2023231478 A1 WO 2023231478A1
Authority
WO
WIPO (PCT)
Prior art keywords
audio
video
receiving end
stream
application
Prior art date
Application number
PCT/CN2023/078498
Other languages
French (fr)
Chinese (zh)
Inventor
杨海城
黄图斌
董桥桥
钱宇
严敏之
Original Assignee
中兴通讯股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 中兴通讯股份有限公司 filed Critical 中兴通讯股份有限公司
Publication of WO2023231478A1 publication Critical patent/WO2023231478A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/07User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail characterised by the inclusion of specific contents
    • H04L51/10Multimedia information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8547Content authoring involving timestamps for synchronizing content
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems

Definitions

  • the present disclosure relates to the field of communication technology, and in particular to an audio and video sharing method, device and computer-readable storage medium.
  • the main purpose of the present disclosure is to provide an audio and video sharing method, device and storage medium, aiming to solve the existing technical problem of being unable to share audio and video files that synchronize audio and video during audio and video calls.
  • the present disclosure provides an audio and video sharing method.
  • the audio and video sharing method includes: obtaining the audio stream and video stream in the audio and video to be shared, and parsing the video stream into video frames; and determining based on the timestamp information of the audio and video to be shared. Audio information corresponding to each video frame; when receiving an audio and video sharing instruction, the video frame and the audio information corresponding to the video frame are sent to the receiving end to share the audio and video to be shared.
  • the present disclosure also provides an audio and video sharing device.
  • the audio and video sharing device includes a processor, a memory, and an audio and video sharing program stored on the memory and executable by the processor, wherein the audio and video sharing program is executed by the processor. When, implement the steps of the above audio and video sharing method.
  • the present disclosure also provides a computer-readable storage medium.
  • An audio and video sharing program is stored on the computer-readable storage medium.
  • the audio and video sharing program is executed by a processor, the steps of the above audio and video sharing method are implemented. .
  • Figure 1 is a schematic diagram of the hardware structure of the audio and video sharing device involved in the embodiment of the present disclosure
  • Figure 2 is a schematic flowchart of the first embodiment of the audio and video sharing method of the present disclosure
  • Figure 3 is a schematic flow chart of the second embodiment of the audio and video sharing method of the present disclosure.
  • Figure 4 is a schematic flowchart of a third embodiment of the audio and video sharing method of the present disclosure.
  • Figure 5 is a schematic flowchart of the fourth embodiment of the audio and video sharing method of the present disclosure.
  • FIG. 6 is a functional module diagram of the first embodiment of the audio and video sharing device of the present disclosure.
  • the audio and video sharing method involved in the embodiments of the present disclosure is mainly applied to audio and video sharing devices.
  • the audio and video sharing devices may be devices with display and processing functions such as PCs, portable computers, and mobile terminals.
  • FIG. 1 is a schematic diagram of the hardware structure of the audio and video sharing device involved in the embodiment of the present disclosure.
  • the audio and video sharing device may include a processor 1001 (such as a CPU), a communication bus 1002, a user interface 1003, a network interface 1004, and a memory 1005.
  • the communication bus 1002 is used to realize connection and communication between these components;
  • the user interface 1003 can include a display screen (Display) and an input unit such as a keyboard (Keyboard);
  • the network interface 1004 can optionally include a standard wired interface and a wireless interface.
  • the memory 1005 can be a high-speed RAM memory or a stable memory (non-volatile memory), such as a disk memory.
  • the memory 1005 can optionally be a storage device independent of the aforementioned processor 1001 .
  • Figure 1 does not constitute a limitation on the audio and video sharing equipment, and may include more or less components than shown in the figure, or combine certain components, or or different component arrangements.
  • the memory 1005 as a computer-readable storage medium in Figure 1 can include an operating system, a network communication module, and an audio and video sharing program.
  • the network communication module is mainly used to connect to the server and perform data communication with the server; and the processor 1001 can call the audio and video sharing program stored in the memory 1005 and execute the audio and video sharing method provided by the embodiment of the present disclosure.
  • Embodiments of the present disclosure provide an audio and video sharing method.
  • Figure 2 is a schematic flowchart of the first embodiment of the audio and video sharing method of the present disclosure.
  • the audio and video sharing method includes the following steps:
  • Step S10 obtain the audio stream and video stream in the audio and video to be shared, and parse the video stream into video frames;
  • the audio and video to be shared can be played through the video player in the mobile terminal; when the video player reads the audio and video files, it will transcode and decode the video stream and audio stream in the audio and video files respectively; Convert the format of the audio and video files into the code stream corresponding to the video player, and then decode and output the transcoded audio and video stream.
  • step S10 specifically includes:
  • the audio and video files to be shared are read through the currently running video player, and the audio and video files to be shared are parsed into video frames corresponding to the audio stream and the video stream.
  • an audio and video file is compressed in one or more formats, that is, video encoding and audio encoding.
  • the purpose of encoding is to reduce the amount of data and facilitate storage and transmission.
  • the audio and video files include audio and video
  • the audio stream and the video stream are compressed separately
  • the decoding is also different, so the audio stream and the video stream need to be decoded separately.
  • the audio and video files need to be parsed into standard encapsulation format data, such as parsing the data transmitted by the PTMP protocol and outputting data in FLV format.
  • the audio stream and the video stream are compressed separately, they are bundled together for transmission during the transmission process. Therefore, when parsing audio and video files, first To separate the audio stream and the video stream, that is, demultiplex or decapsulate, the input encapsulated format data is separated into audio stream compression encoding data and video stream compression encoding data.
  • encapsulation formats such as MP4, MKV, RMVB, TS, FLV, AVI, etc. Its function is to put the compressed and encoded video data and audio data together in a certain format. For example, after decapsulating data in FLV format, an H.264-encoded video stream and an AAC-encoded audio stream are output.
  • the audio/video compressed coded data is decoded into uncompressed audio/video original data through a decoder, wherein the decoded output of the video compressed coded data is uncompressed color data, such as YUV, RGB, etc., while the audio compressed coded data is The decoded output of the encoded data is uncompressed audio sample data, such as PCM data.
  • the video stream is parsed into a series of video frames according to the specified size through the parsing module of the video player, and the video frames are numbered to obtain the frame number of each video frame.
  • both parties to the call will negotiate the picture size, frame rate, etc. of the video call.
  • the parsing module can parse audio and video files (such as .mp4 files) into static video frames of corresponding sizes (such as 480 ⁇ 640) according to the picture size of the video call.
  • the video player synchronizes the decoded video data and audio data based on the parameter information obtained during the processing of the decapsulation module, and plays them through the graphics card and sound card of the sharing end.
  • Step S20 Determine the audio information corresponding to each video frame based on the timestamp information of the audio and video to be shared;
  • the display timestamp and decoding timestamp when parsing the audio and video to be shared, there will be a display timestamp and a decoding timestamp.
  • the display timestamp and decoding timestamp of the audio are the same.
  • the decoding order of the video will be different due to the existence of image compression.
  • the display order is not the same, that is, the display timestamp and decoding timestamp are different.
  • the display timestamp refers to the display order of data decoded from the data packet;
  • the decoding timestamp refers to the decoding order of the data packet.
  • the relative playback time of the video stream and the audio stream is mostly calculated to determine the difference between the two.
  • the delay method can be used to catch up.
  • the video stream lags far behind the audio stream, it needs to catch up by losing frames.
  • This method cannot achieve more precise synchronization adjustment, and there will still be differences in sound and picture.
  • the audio streams and video streams will not be transmitted or played immediately. Instead, the timestamps of the parsed audio streams and video streams will be aligned to determine each The audio information corresponding to the video frame.
  • a reference clock such as the system clock of the processor
  • each frame Video frames and each piece of audio information will also have timestamp information, which can determine the corresponding relationship between video frames and audio information.
  • Step S30 When receiving an audio and video sharing instruction, the video frame and the audio information corresponding to the video frame are sent to the receiving end to share the audio and video to be shared.
  • the audio and video files can be sent to the receiving end, and then played and displayed at the receiving end.
  • the playback speeds of the video stream and the audio stream are inconsistent, the audio and video files played at the receiving end.
  • video frames and audio information corresponding to the same timestamp information are synchronously output to the receiving end according to the timestamp order.
  • the audio and video files received by the receiving end are Audio stream and video frame files corrected and synchronized based on timestamp information.
  • the sharing end first plays the video frame and the corresponding audio information synchronously according to the synchronization signal (that is, the frame number of the currently played video frame), thereby outputting the synchronized audio and video stream, and the receiving end receives the process.
  • Synchronized audio and video streams can play audio and video files with synchronized audio and video.
  • the call type for establishing the audio and video call may be an IMS call type or an OTT call type.
  • IMS is also called IP multimedia subsystem, which can implement voice services under packet switching networks, including VoNR, VoLTE, VoWiFi and other types.
  • VoNR is a call solution based on pure 5G access, which enables both voice and data services to be carried on the 5G network.
  • VoLTE is an IP data transmission technology that does not require a 2G/3G network. All services are carried on the 4G network, which can realize the unification of data and voice services under the same network. VoLTE is an end-to-end voice solution built on the 4G network under all-IP conditions. Compared with traditional calls, VoLTE high-definition call technology has fast connection times and close to zero call drop rates. LTE’s spectrum utilization efficiency is far superior to traditional calls. Standard.
  • VoWiFi is a voice service provided through WiFi networks. Users can make calls without mobile signals.
  • VoWiFi is a complementary technology to VoLTE. Compared with traditional calling services, VoWiFi prioritizes the use of WiFi networking to implement calling functions and automatically and seamlessly switches between mobile networks and WiFi networks. Users can make calls in different locations without special settings. More importantly, using WiFi networking overcomes the problem of poor signal indoors or in the basement. In places with weak network coverage or interference, you can make or receive calls as long as you can connect to WiFi.
  • OTT refers to Internet companies that go beyond operators (telecom, mobile, linkage) to develop Internet-based value-added services such as video, social networking, games, and data services, including WeChat, Skype and other types.
  • WeChat is an application that provides instant messaging services for smart terminals. It can quickly send voice messages, videos, pictures, text, etc. across communication operators and operating system platforms through the network.
  • Skype is an instant messaging software that has the functions required for IM, such as video chat, multi-person voice conferencing, multi-person chat, file transfer, file chat, etc.
  • the sharing terminal can switch the current display interface, display the pictures collected by the terminal camera in real time, and switch the current audio device to the call audio device to collect the user's Voice information, and send the real-time collected pictures and voice information to the receiving end through the call data channel until the audio and video call ends.
  • the present disclosure provides an audio and video sharing method.
  • the method obtains the audio stream and video stream in the audio and video to be shared, and parses the video stream into video frames; according to the timestamp information of the audio and video to be shared, The audio information corresponding to each video frame is determined in the audio stream; based on the call data channel or new data channel of the audio and video call, each video frame, and the audio information corresponding to each video frame, all the audio information corresponding to each video frame is synchronously sent to the receiving end. Audio and video to be shared.
  • the present disclosure parses the video stream of the audio and video file into video frames, and parses the timestamp information corresponding to the video stream.
  • the timestamp information of audio and video can be determined, so the audio information corresponding to each video frame can be determined to achieve precise synchronization of video stream and audio stream; during audio and video calls, by establishing a data channel for audio and video calls or establishing a data channel for transmitting data
  • the newly created data channel synchronously transmits the video frame and the audio information corresponding to the current video frame.
  • the receiving end can receive the accurately synchronized video stream and audio stream, that is, the receiving end receives the audio and video files with synchronized audio and video, improving user experience, solving the current problem of Technical issue that audio and video files synchronized with audio and video cannot be shared during video calls.
  • Figure 3 is a schematic flow chart of a second embodiment of the audio and video sharing method of the present disclosure.
  • step S30 specifically includes:
  • Step S31 determine the current video frame in the video frame, and synchronously output the current video frame and the audio information corresponding to the current video frame according to the frame number of the current video frame, to obtain synchronized audio and video flow;
  • the video stream and audio stream with time stamp alignment are simultaneously played through the video player on the sharing end.
  • the video stream is further parsed into video frames, and the video frames are numbered to obtain each video frame. frame number.
  • the video frame corresponding to the frame number is synchronously output according to the frame number corresponding to the currently played video stream, and the audio stream is synchronously output to the audio front end while being played. , thereby achieving synchronous output of video frames and audio streams, and obtaining synchronous output of audio and video streams.
  • Step S32 Send the synchronized audio and video stream to the receiving end based on the data transmission channel, so that the receiving end plays the audio and video to be shared synchronously.
  • the synchronously output audio and video streams are sent to the receiving end by establishing a call data channel for the audio and video calls or creating a new data channel, and the receiving end can call the corresponding target application (such as video playback devices, audio and video applications, etc.) to directly play the synchronized audio and video streams.
  • the target application such as video playback devices, audio and video applications, etc.
  • step S32 specifically includes:
  • the synchronized audio and video stream is sent to the receiving end through a transmission data channel in the data transmission channel.
  • a non-Bootstrap data channel when establishing a VoNR+ video call, can be established to transmit audio streams and video frames to the receiving end; a Bootstrap data channel can then be established to transmit the target application for playing audio and video files to the receiving end. Corresponding application data information.
  • the method before sending the application data information corresponding to the target application to the receiving end through the application data channel, the method further includes:
  • the local application data information is sent to the receiving end based on the application data channel.
  • the sharing end can send a SIP UPDATE message carrying the dcmap field to the network side, and then establish a data channel with the network data server or the receiving end; the sharing end can pre-save the audio and video to be shared in the network data server.
  • the sharing terminal when the sharing terminal and the receiving terminal establish a data connection, the receiving terminal simultaneously establishes a data channel with the network data server; when receiving an audio and video file sharing instruction, the sharing terminal displays the network data according to the sharing instruction. server, and queries the target application in the network data server. If the target application exists, directly points to the location of the target application according to the query result, and sends the application data information corresponding to the target application through the data channel between the receiving end and the network data server. ; If the target application does not exist in the network data server, the query will show no results.
  • the sharing end can choose to close the network data server and query the target application locally.
  • the sharing end can select audio and video to be shared that are pre-stored in the network data server and transmit them to the receiving end through the data channel of the audio and video call, or directly transmit them through the data channel between the network data server and the receiving end.
  • the sharing end can select audio and video files and applications in the network data server, or can choose audio and video files and applications stored locally on the sharing end.
  • the dcmap field represents the need to establish a data channel.
  • the data channel can be a Bootstrap data channel used to transmit data channel applications, or a non-Bootstrap data channel used to transmit data information required by data channel applications.
  • the audio, video and application data information to be shared are not in the network data server, you can select locally stored audio and video files or streaming videos, as well as the corresponding applications, and transmit them to the receiving end through the data channel established for the call. .
  • the application data information is sent to the receiving end through the network data server.
  • the sharing end can select the corresponding audio and video file to be shared in the network data server, and complete the editing of the audio and video file to be shared on the sharing end. Operations such as analysis and synchronization; then the synchronized audio stream and video frame are sent to the receiving end through the data channel, and the corresponding video frame and audio stream are played on the receiving end through real-time transmission of the video frame number currently played by the sharing end. Realize audio and video at both ends of the call Synchronous playback of video files.
  • the receiving end can directly obtain the audio and video files and applications from the network data server, open the audio and video files through the application, and the sharing end transmits the frame number of the currently played video frame in real time. Play the corresponding video frames and audio streams, thereby achieving synchronous playback of audio and video files at both ends of the call.
  • the sharing end can replace the output video frame with the picture currently collected by the camera, and output the audio stream to the audio device collected by the microphone.
  • the video corresponding to the frame number is synchronously output.
  • Frame and audio information, and through the call data channel of the audio and video call, the video frame and audio information played by the current call are transmitted to the receiving end, and played synchronously at the receiving end.
  • the sharing end can directly send the audio and video files outputted synchronously with the video frames and audio streams to the receiving end.
  • the receiving end can save the audio and video streams locally.
  • the audio and video files can include video frame files and audio stream files; then during the audio and video call, the sender can synchronously send the frame number of the currently played audio frame to the receiving end, and the receiving end calls
  • the corresponding playback application plays the audio and video file, and plays the current video frame corresponding to the frame number in the audio and video file according to the received frame number, and based on the timestamp information of the current video frame, synchronously plays the audio corresponding to the same timestamp information. information, you can achieve synchronous playback of the same audio and video files between the sharing end and the receiving end.
  • the sharing end adjusts the playback speed of audio and video files (such as fast forward, rewind, or pause, etc.)
  • the refresh speed of the frame number received by the receiving end will also Synchronous adjustment
  • the receiving end will also realize synchronous adjustment of the audio and video file playback speed, further ensuring the synchronous playback of the same audio and video files between the sharing end and the receiving end.
  • step S32 it also includes:
  • a data channel is established with the receiving end based on the call data channel as the data transmission channel.
  • the current audio and video call is VoNR+video call
  • you can create a new data call Channel is used as a data transmission channel to transmit audio and video streams or application data information; otherwise, the call data channel of the current audio and video call can be directly used as a data transmission channel to transmit audio and video streams to the receiving end.
  • a non-Bootstrap data channel when establishing a VoNR+ video call, can be established to transmit audio streams and video frames to the receiving end; a Bootstrap data channel can then be established to transmit the target application for playing audio and video files to the receiving end. Corresponding application data information.
  • VoNR+ 5G New Call
  • VoNR+ refers to a new data transmission channel based on 5G VoNR multimedia real-time communication, providing users with richer real-time interactive services in addition to high-definition audio and video, and establishing multimedia real-time communication capabilities.
  • Platform-centered, unified and open network architecture enables agile development and rapid deployment of innovative services on the basis of compatibility with existing services.
  • VoNR+ is a real-time communication network architecture based on VoNR (5GNR), which can quickly integrate new business forms to meet people's diverse communication needs.
  • the interactive channel that carries real-time interactive information is wider, the interactive content types are more, and the interactive forms are richer.
  • the sharing end first transmits the application data information for playing the audio and video to be shared to the receiving end through the data channel of Bootstrap.
  • the receiving end generates the corresponding target application through the application data information, which can be a temporary application. (such as a small program, etc.), or it can be an application download link, through which the receiving end downloads the corresponding media playback application.
  • the sharing end after completing the transmission of the target application, simultaneously plays the parsed and synchronized audio stream and video stream, and synchronously transmits the audio stream and video frame to the receiving end through a non-Bootstrap data channel, and will share The frame number of the current video frame currently played by the end is transmitted to the receiving end in real time.
  • the receiving end synchronously plays the audio and video files transmitted by the sharing end through the target application based on the frame number of the current video frame.
  • the Bootstrap data channel is a data channel with a Stream ID less than 1000. It is used for the terminal to obtain HTML web pages from the network side (that is, the data channel application defined in the 3GPP TS26.114 specification.
  • the application generally includes HTML, JavaScript scripts, CSS, etc. ).
  • Figure 4 is a schematic flowchart of a third embodiment of the audio and video sharing method of the present disclosure.
  • step S30 it specifically includes:
  • Step S031 Determine the target audio channel corresponding to the audio information according to the data type of the audio information
  • Step S032 Switch the first audio channel corresponding to the video player and the second audio channel corresponding to the audio and video call to the target audio channel, so as to play the audio information through the target audio channel while simultaneously providing the audio information to the receiver.
  • the terminal sends the audio information synchronously.
  • the sharing end when the sharing end transmits the audio stream and video frame, it first needs to synchronize the audio stream and video frame through the timestamp information. After the synchronization is completed, the audio stream and video frame will not be transmitted immediately. Use the video player to play the audio stream and video frame simultaneously, so that the audio stream and video frame start playing at the same starting time. While playing, the audio stream and video frame are synchronously transmitted to the receiving end so that the receiving end can play synchronously. Synchronized audio and video streams.
  • the audio stream is transmitted to the receiving end while playing.
  • the audio stream In order to avoid the problem of audio and video being out of sync due to transmission delays caused by operations such as transcoding, the audio stream must be transmitted before the sharing end starts playing the audio stream. audio channel, and then switch the audio data channels of the video player and call data channel to this audio channel. While playing the audio stream, the audio stream can be output synchronously, avoiding the transcoding process, and the audio output can be realized.
  • the timestamps of streaming and playing audio streams are synchronized, thereby achieving synchronous output of audio streams and video frames. For example, before playing, you can switch the audio channel of the video player to the audio channel corresponding to the call mode. When playing the audio and video files through the video player, the audio stream can be output and the audio stream data can be transmitted at the same time. to the receiving end and played through the call audio channel of the receiving end.
  • Figure 5 is a schematic flowchart of a fourth embodiment of the audio and video sharing method of the present disclosure.
  • step S30 specifically includes:
  • Step S310 Determine an application whitelist corresponding to the application to which the audio and video to be shared belongs based on the application information to which the audio and video to be shared belongs, and determine whether the application to which the audio and video call belongs belongs to the application whitelist;
  • Step S320 If the application to which the audio and video call belongs belongs to the application whitelist, the audio and video to be shared are sent to the receiving end based on the respective video frames and the audio information corresponding to the respective video frames.
  • OTT audio and video call such as WeChat, Skype, etc.
  • the OTT audio and video call is a call type that transcends operators and is based on the Internet
  • the audio and video to be shared comes from a short video application (such as Douyin, Kuaishou, etc.), it is necessary to obtain the shareable application whitelist of the application.
  • a short video application such as Douyin, Kuaishou, etc.
  • the whitelist the information that can be used for sharing will be displayed. Call application derived from the audio and video files of the short video application.
  • the audio and video to be shared can be shared to the receiving end directly through the calling application; and when the calling application is not in the whitelist, you need to use the calling application to share the audio and video to be shared with the receiving end. Only by setting the permissions of the calling application and adding the calling application to the whitelist can the audio and video to be shared be shared.
  • the sharing terminal needs to share an audio and video file in Douyin. At this time, the sharing terminal needs to determine whether WeChat has the permission to share the audio and video file. If WeChat is not selected in the whitelist, then It means that WeChat does not have the sharing permission, so you need to set the permissions so that WeChat can share Douyin audio and video files.
  • embodiments of the present disclosure also provide an audio and video sharing device.
  • FIG. 6 is a schematic diagram of functional modules of the audio and video sharing device according to the first embodiment of the present disclosure.
  • the audio and video sharing device includes:
  • the audio and video parsing module 10 is used to obtain the audio stream and video stream in the audio and video to be shared, and parse the video stream into video frames;
  • the audio and video stream synchronization module 20 is used to determine the audio information corresponding to each video frame based on the timestamp information of the audio and video to be shared;
  • the audio and video sending module 30 is configured to, when receiving an audio and video sharing instruction, send the video frame and the audio information corresponding to the video frame to the receiving end to share the audio and video to be shared.
  • the audio and video sending module 30 specifically includes:
  • An audio and video stream synchronization output unit is used to determine the current video frame in the video frame, and synchronously output the current video frame and the audio information corresponding to the current video frame according to the frame number of the current video frame. , obtain synchronized audio and video streams;
  • the audio and video stream sending unit is configured to send the synchronized audio and video stream to the receiving end based on the data transmission channel, so that the receiving end can synchronously play the audio and video to be shared.
  • the audio and video stream sending unit specifically includes:
  • the application data information sending subunit is configured to send application data information corresponding to the target application to the receiving end through the application data channel in the data transmission channel; wherein the target application is used to play the audio and video stream;
  • the audio and video stream sending subunit is configured to send the synchronized audio and video stream to the receiving end through the transmission data channel in the data transmission channel.
  • the audio and video sharing device also includes a network data server module.
  • the network data server module specifically includes:
  • a network server query unit used to determine whether the application data information exists in the network data server
  • a local application information sending unit is configured to send the local application data information to the receiving end based on the application data channel when the application data information does not exist in the network data server.
  • a network-side application sending unit is configured to send the application data information to the receiving end through the network data server if the application data information exists in the network data server.
  • the audio and video sharing device also includes a data transmission channel determination module.
  • the data transmission channel determination module specifically includes:
  • An audio and video call type judgment unit is used to establish an audio and video call with the receiving end. If the audio and video call is not the VoNR+ call type, use the call data channel of the audio and video call as the data transmission channel;
  • the audio and video analysis module 10 includes:
  • the audio and video file parsing unit is used to read the audio and video to be shared through the currently running video player, and parse the audio and video file to be shared into the audio stream and the video frame corresponding to the video stream.
  • the audio and video sharing device includes an audio channel switching module, and the audio channel switching module specifically includes:
  • a target audio channel determination unit configured to determine the target audio channel corresponding to the audio information according to the data type of the audio information
  • An audio channel switching unit is used to switch the first audio channel corresponding to the video player and the second audio channel corresponding to the audio and video call to the target audio channel, so as to play the audio information through the target audio channel while playing The audio information is synchronously sent to the receiving end.
  • the audio and video sharing device includes a whitelist module
  • the whitelist module specifically includes:
  • An application whitelist determination unit is configured to determine based on the application information to which the audio and video to be shared belongs.
  • An audio and video file sending unit configured to, if the application to which the audio and video call belongs belongs to the application whitelist, send the to-be-received file to the receiving end based on the respective video frames and the audio information corresponding to the respective video frames. Share audio and video.
  • Each module in the above-mentioned audio and video sharing device corresponds to each step in the above-mentioned audio and video sharing method embodiment, and their functions and implementation processes will not be described in detail here.
  • embodiments of the present disclosure also provide a computer-readable storage medium.
  • the computer-readable storage medium of the present disclosure stores an audio and video sharing program.
  • the audio and video sharing program is executed by a processor, the steps of the above audio and video sharing method are implemented.
  • the terms “include”, “comprising” or any other variation thereof are intended to cover a non-exclusive inclusion, such that a process, method, article or system that includes a list of elements not only includes those elements, but It also includes other elements not expressly listed or that are inherent to the process, method, article or system. Without further limitation, an element qualified by the statement “comprises a" does not exclude the presence of additional identical elements in the process, method, article, or system that includes that element.
  • the present disclosure may be used in numerous general purpose or special purpose computer system environments or configurations. For example: personal computers, server computers, handheld or portable devices, tablet devices, multiprocessor systems, microprocessor-based systems, set-top boxes, programmable consumer electronics devices, network PCs, minicomputers, mainframe computers, including Distributed computing environment for any of the above systems or devices, etc.
  • the present disclosure may be described in the general context of computer-executable instructions, such as program modules, being executed by a computer.
  • program modules include routines, programs, objects, components, data structures, etc. that perform specific tasks or implement specific abstract data types.
  • the present disclosure may also be practiced in distributed computing environments where tasks are performed by remote processing devices connected through a communications network.
  • program modules may be located in both local and remote computer storage media including storage devices.
  • the embodiment method can be implemented by means of software plus the necessary general hardware platform. Of course, it can also be implemented by hardware, but in many cases the former is a better implementation method.
  • the technical solution of the present disclosure can be embodied in the form of a software product in nature or in part that contributes to the existing technology.
  • the computer software product is stored in a storage medium (such as ROM/RAM) as mentioned above. , magnetic disk, optical disk), including several instructions to cause a terminal device (which can be a mobile phone, computer, server, air conditioner, or network device, etc.) to execute the methods described in various embodiments of the present disclosure.

Abstract

The present disclosure relates to communication technology. Provided in the present disclosure are an audio and video sharing method and device, and a computer-readable storage medium. The method comprises: acquiring an audio stream and a video stream in audio and video to be shared, and parsing the video stream into video frames; determining audio information corresponding to each video frame according to timestamp information of said audio and video; and when an audio and video sharing instruction has been received, sending to a receiving end the video frames and the audio information corresponding to the video frames, so as to share said audio and video. By means of the method in the present disclosure, audio information corresponding to each video frame is determined according to timestamp information of an audio and video file, thus realizing accurate synchronization of a video stream and an audio stream; and according to an audio and video sharing instruction, the video frames and the audio information corresponding to the current video frames are synchronously transmitted, and the audio and video file with synchronized audio and picture is received at a receiving end, thus improving the user experience.

Description

音视频共享方法、设备及计算机可读存储介质Audio and video sharing method, device and computer-readable storage medium
相关申请的交叉引用Cross-references to related applications
本公开基于2022年5月31日提交的发明名称为“音视频共享方法、设备及计算机可读存储介质”的中国专利申请CN202210606704.8,并且要求该专利申请的优先权,通过引用将其所公开的内容全部并入本公开。This disclosure is based on the Chinese patent application CN202210606704.8 with the invention title "Audio and Video Sharing Method, Device and Computer-Readable Storage Medium" submitted on May 31, 2022, and claims the priority of this patent application, and all its contents are incorporated by reference. The entire disclosure is incorporated into this disclosure.
技术领域Technical field
本公开涉及通信技术领域,尤其涉及一种音视频共享方法、设备及计算机可读存储介质。The present disclosure relates to the field of communication technology, and in particular to an audio and video sharing method, device and computer-readable storage medium.
背景技术Background technique
随着通信技术的不断发展,网络通信对用户的日常生活影响越来越大。人们在日常生活中常使用网络通信跨越空间的限制进行及时地交流,实现信息的快速传递。而基于网络通信的视频共享功能,可用于向共享端共享本端的信息。然而,目前还无法实现在通话过程中同步共享音视频文件,接收端需要在本地查看发送端发送的音视频文件,且音视频文件在输出时存在音画不同步的问题,降低了用户体验。因此,如何解决现有音视频通话时不能共享音画同步的音视频文件成为了目前亟待解决的技术问题。With the continuous development of communication technology, network communication has an increasing impact on users' daily lives. In daily life, people often use network communication to communicate in a timely manner across space constraints and achieve rapid transmission of information. The video sharing function based on network communication can be used to share local information to the sharing terminal. However, it is currently not possible to share audio and video files synchronously during a call. The receiving end needs to view the audio and video files sent by the sending end locally, and there is a problem of audio and video being out of sync when outputting the audio and video files, which reduces the user experience. Therefore, how to solve the problem of being unable to share audio and video files synchronized with audio and video during existing audio and video calls has become an urgent technical problem that needs to be solved.
发明内容Contents of the invention
本公开的主要目的在于提供一种音视频共享方法、设备及存储介质,旨在解决现有音视频通话时不能共享音画同步的音视频文件的技术问题。The main purpose of the present disclosure is to provide an audio and video sharing method, device and storage medium, aiming to solve the existing technical problem of being unable to share audio and video files that synchronize audio and video during audio and video calls.
本公开提供一种音视频共享方法,该音视频共享方法包括:获取待分享音视频中的音频流和视频流,并将视频流解析成视频帧;根据待分享音视频的时间戳信息,确定各视频帧对应的音频信息;在接收到音视频分享的指令时,向接收端发送视频帧以及视频帧对应的音频信息,以进行待分享音视频的共享。The present disclosure provides an audio and video sharing method. The audio and video sharing method includes: obtaining the audio stream and video stream in the audio and video to be shared, and parsing the video stream into video frames; and determining based on the timestamp information of the audio and video to be shared. Audio information corresponding to each video frame; when receiving an audio and video sharing instruction, the video frame and the audio information corresponding to the video frame are sent to the receiving end to share the audio and video to be shared.
本公开还提供一种音视频共享设备,该音视频共享设备包括处理器、存储器、以及存储在存储器上并可被处理器执行的音视频共享程序,其中音视频共享程序被所述处理器执行时,实现如上述的音视频共享方法的步骤。 The present disclosure also provides an audio and video sharing device. The audio and video sharing device includes a processor, a memory, and an audio and video sharing program stored on the memory and executable by the processor, wherein the audio and video sharing program is executed by the processor. When, implement the steps of the above audio and video sharing method.
本公开还提供一种计算机可读存储介质,所述计算机可读存储介质上存储有音视频共享程序,其中所述音视频共享程序被处理器执行时,实现如上述的音视频共享方法的步骤。The present disclosure also provides a computer-readable storage medium. An audio and video sharing program is stored on the computer-readable storage medium. When the audio and video sharing program is executed by a processor, the steps of the above audio and video sharing method are implemented. .
附图说明Description of the drawings
图1为本公开实施例方案中涉及的音视频共享设备的硬件结构示意图;Figure 1 is a schematic diagram of the hardware structure of the audio and video sharing device involved in the embodiment of the present disclosure;
图2为本公开音视频共享方法第一实施例的流程示意图;Figure 2 is a schematic flowchart of the first embodiment of the audio and video sharing method of the present disclosure;
图3为本公开音视频共享方法第二实施例的流程示意图;Figure 3 is a schematic flow chart of the second embodiment of the audio and video sharing method of the present disclosure;
图4为本公开音视频共享方法第三实施例的流程示意图;Figure 4 is a schematic flowchart of a third embodiment of the audio and video sharing method of the present disclosure;
图5为本公开音视频共享方法第四实施例的流程示意图;Figure 5 is a schematic flowchart of the fourth embodiment of the audio and video sharing method of the present disclosure;
图6为本公开音视频共享装置第一实施例的功能模块示意图。FIG. 6 is a functional module diagram of the first embodiment of the audio and video sharing device of the present disclosure.
本公开目的的实现、功能特点及优点将结合实施例,参照附图做进一步说明。The realization of the purpose, functional features and advantages of the present disclosure will be further described with reference to the embodiments and the accompanying drawings.
具体实施方式Detailed ways
应当理解,此处所描述的具体实施例仅仅用以解释本公开,并不用于限定本公开。It should be understood that the specific embodiments described here are only used to explain the present disclosure and are not used to limit the present disclosure.
本公开实施例涉及的音视频共享方法主要应用于音视频共享设备,该音视频共享设备可以是PC、便携计算机、移动终端等具有显示和处理功能的设备。The audio and video sharing method involved in the embodiments of the present disclosure is mainly applied to audio and video sharing devices. The audio and video sharing devices may be devices with display and processing functions such as PCs, portable computers, and mobile terminals.
参照图1,图1为本公开实施例方案中涉及的音视频共享设备的硬件结构示意图。本公开实施例中,音视频共享设备可以包括处理器1001(例如CPU),通信总线1002,用户接口1003,网络接口1004,存储器1005。其中,通信总线1002用于实现这些组件之间的连接通信;用户接口1003可以包括显示屏(Display)、输入单元比如键盘(Keyboard);网络接口1004可选的可以包括标准的有线接口、无线接口(如WI-FI接口);存储器1005可以是高速RAM存储器,也可以是稳定的存储器(non-volatile memory),例如磁盘存储器,存储器1005可选的还可以是独立于前述处理器1001的存储装置。Referring to Figure 1, Figure 1 is a schematic diagram of the hardware structure of the audio and video sharing device involved in the embodiment of the present disclosure. In the embodiment of the present disclosure, the audio and video sharing device may include a processor 1001 (such as a CPU), a communication bus 1002, a user interface 1003, a network interface 1004, and a memory 1005. Among them, the communication bus 1002 is used to realize connection and communication between these components; the user interface 1003 can include a display screen (Display) and an input unit such as a keyboard (Keyboard); the network interface 1004 can optionally include a standard wired interface and a wireless interface. (such as WI-FI interface); the memory 1005 can be a high-speed RAM memory or a stable memory (non-volatile memory), such as a disk memory. The memory 1005 can optionally be a storage device independent of the aforementioned processor 1001 .
本领域技术人员可以理解,图1中示出的硬件结构并不构成对音视频共享设备的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或 者不同的部件布置。Those skilled in the art can understand that the hardware structure shown in Figure 1 does not constitute a limitation on the audio and video sharing equipment, and may include more or less components than shown in the figure, or combine certain components, or or different component arrangements.
继续参照图1,图1中作为一种计算机可读存储介质的存储器1005可以包括操作系统、网络通信模块以及音视频共享程序。Continuing to refer to Figure 1, the memory 1005 as a computer-readable storage medium in Figure 1 can include an operating system, a network communication module, and an audio and video sharing program.
在图1中,网络通信模块主要用于连接服务器,与服务器进行数据通信;而处理器1001可以调用存储器1005中存储的音视频共享程序,并执行本公开实施例提供的音视频共享方法。In Figure 1, the network communication module is mainly used to connect to the server and perform data communication with the server; and the processor 1001 can call the audio and video sharing program stored in the memory 1005 and execute the audio and video sharing method provided by the embodiment of the present disclosure.
本公开实施例提供了一种音视频共享方法。Embodiments of the present disclosure provide an audio and video sharing method.
参照图2,图2为本公开音视频共享方法第一实施例的流程示意图。Referring to Figure 2, Figure 2 is a schematic flowchart of the first embodiment of the audio and video sharing method of the present disclosure.
本实施例中,所述音视频共享方法包括以下步骤:In this embodiment, the audio and video sharing method includes the following steps:
步骤S10,获取待分享音视频中的音频流和视频流,并将所述视频流解析成视频帧;Step S10, obtain the audio stream and video stream in the audio and video to be shared, and parse the video stream into video frames;
本实施例中,可以通过移动终端内的视频播放器播放待分享音视频;视频播放器在读取音视频文件时,会对音视频文件中的视频流和音频流分别进行转码和解码;将音视频文件的格式转换成视频播放器相对应的码流,再将转码过后的音视频流进行解码输出。In this embodiment, the audio and video to be shared can be played through the video player in the mobile terminal; when the video player reads the audio and video files, it will transcode and decode the video stream and audio stream in the audio and video files respectively; Convert the format of the audio and video files into the code stream corresponding to the video player, and then decode and output the transcoded audio and video stream.
进一步地,所述步骤S10具体包括:Further, the step S10 specifically includes:
通过当前正在运行的视频播放器,读取所述待分享音视频,将所述待分享音视频文件解析成所述音频流和所述视频流对应的视频帧。The audio and video files to be shared are read through the currently running video player, and the audio and video files to be shared are parsed into video frames corresponding to the audio stream and the video stream.
本实施例中,一个音视频文件是经过某种或者多种格式压缩的,也即视频编码和音频编码,编码是为了减少数据量,便于存储和传输。在打开一个音视频文件时,需要判断当前视频播放器是否拥有该音视频文件的压缩格式的解析协议,如果没有,则需要更换成可解析的视频播放器,或者将音视频文件转换成该视频播放器可解析的压缩格式。In this embodiment, an audio and video file is compressed in one or more formats, that is, video encoding and audio encoding. The purpose of encoding is to reduce the amount of data and facilitate storage and transmission. When opening an audio and video file, you need to determine whether the current video player has a parsing protocol for the compression format of the audio and video file. If not, you need to change to a video player that can parse it, or convert the audio and video file into the video. A compression format that the player can parse.
具体地,因为音视频文件中包括音频和视频,而音频流和视频流是分开进行压缩的,因为两者的压缩算法不同,解码也不相同,所以需要对音频流和视频流分别进行解码。Specifically, because the audio and video files include audio and video, and the audio stream and the video stream are compressed separately, because the two compression algorithms are different, the decoding is also different, so the audio stream and the video stream need to be decoded separately.
可以理解地,对于流媒体音视频文件,需要将音视频文件解析成标准的封装格式数据,比如将PTMP协议传输的数据解析,输出FLV格式的数据。Understandably, for streaming audio and video files, the audio and video files need to be parsed into standard encapsulation format data, such as parsing the data transmitted by the PTMP protocol and outputting data in FLV format.
具体实施例中,虽然音频流和视频流是分开进行压缩的,但是在传输过程中两者是捆绑在一起进行传输的。所以在对音视频文件进行解析时,首先 要将音频流和视频流分开,也即解复用或者解封装,将输入的封装格式数据,分离成为音频流压缩编码数据和视频流压缩编码数据。封装格式种类很多,例如MP4、MKV、RMVB、TS、FLV、AVI等等,它的作用就是将已经压缩编码的视频数据和音频数据按照一定的格式放到一起。例如,FLV格式的数据,经过解封装操作后,输出H.264编码的视频码流和AAC编码的音频码流。In a specific embodiment, although the audio stream and the video stream are compressed separately, they are bundled together for transmission during the transmission process. Therefore, when parsing audio and video files, first To separate the audio stream and the video stream, that is, demultiplex or decapsulate, the input encapsulated format data is separated into audio stream compression encoding data and video stream compression encoding data. There are many types of encapsulation formats, such as MP4, MKV, RMVB, TS, FLV, AVI, etc. Its function is to put the compressed and encoded video data and audio data together in a certain format. For example, after decapsulating data in FLV format, an H.264-encoded video stream and an AAC-encoded audio stream are output.
具体实施例中,通过解码器将音频/视频压缩编码数据解码成为非压缩的音频/视频原始数据,其中,视频压缩编码数据解码输出的是非压缩的颜色数据,例如YUV、RGB等,而音频压缩编码数据解码输出的是非压缩的音频抽样数据,例如PCM数据等。In a specific embodiment, the audio/video compressed coded data is decoded into uncompressed audio/video original data through a decoder, wherein the decoded output of the video compressed coded data is uncompressed color data, such as YUV, RGB, etc., while the audio compressed coded data is The decoded output of the encoded data is uncompressed audio sample data, such as PCM data.
具体地,通过视频播放器的解析模块将视频流按照指定尺寸解析成一系列的视频帧,并且对视频帧进行编号,获得每个视频帧的帧号。Specifically, the video stream is parsed into a series of video frames according to the specified size through the parsing module of the video player, and the video frames are numbered to obtain the frame number of each video frame.
具体实施例中,建立IMS视频通话(如VoNR、VoLTE、VoWiFi等)时,通话双方会协商本次视频通话的画面尺寸、帧率等。解析模块可以按照视频通话的画面尺寸,将音视频文件(如.mp4文件)解析成相应尺寸(如480×640)的静态的视频帧。In specific embodiments, when establishing an IMS video call (such as VoNR, VoLTE, VoWiFi, etc.), both parties to the call will negotiate the picture size, frame rate, etc. of the video call. The parsing module can parse audio and video files (such as .mp4 files) into static video frames of corresponding sizes (such as 480×640) according to the picture size of the video call.
具体实施例中,视频播放器根据解封装模块处理过程中获取到的参数信息,同步解码出来的视频数据和音频数据,并通过共享端的显卡和声卡播放出来。In a specific embodiment, the video player synchronizes the decoded video data and audio data based on the parameter information obtained during the processing of the decapsulation module, and plays them through the graphics card and sound card of the sharing end.
步骤S20,根据所述待分享音视频的时间戳信息,确定各视频帧对应的音频信息;Step S20: Determine the audio information corresponding to each video frame based on the timestamp information of the audio and video to be shared;
本实施例中,在解析待分享音视频时,会存在显示时间戳和解码时间戳,音频的音频的显示时间戳和解码时间戳是相同的,视频因为图片压缩的存在,会造成解码顺序与显示顺序并不相同,即显示时间戳和解码时间戳是不相同的。In this embodiment, when parsing the audio and video to be shared, there will be a display timestamp and a decoding timestamp. The display timestamp and decoding timestamp of the audio are the same. The decoding order of the video will be different due to the existence of image compression. The display order is not the same, that is, the display timestamp and decoding timestamp are different.
其中,显示时间戳是指从数据包解码出来的数据的显示顺序;解码时间戳是指数据包的解码顺序。Among them, the display timestamp refers to the display order of data decoded from the data packet; the decoding timestamp refers to the decoding order of the data packet.
具体地,现有技术方案中,对于音频流和视频流的同步,大多是通过计算视频流和音频流的相对播放时间,判断两者的差异,差异较小时可以通过延迟的方法进行追赶。而如果视频流落后音频流很多,就需要采用丢帧的形式进行追赶。这种方法不能实现较为精确的同步调整,仍然会存在音画不同 步的问题;且如果存在丢帧,会导致画面播放衔接不流畅,用户体验较差。Specifically, in the existing technical solutions, for the synchronization of the audio stream and the video stream, the relative playback time of the video stream and the audio stream is mostly calculated to determine the difference between the two. When the difference is small, the delay method can be used to catch up. And if the video stream lags far behind the audio stream, it needs to catch up by losing frames. This method cannot achieve more precise synchronization adjustment, and there will still be differences in sound and picture. There is a synchronization problem; and if there are frame drops, it will cause the screen playback connection to be unsmooth and the user experience to be poor.
本实施例中,在将待分享音视频解析成音频流和视频流之后,不会立即传输或者播放音频流和视频流,而是将解析后音频流和视频流的时间戳对齐,确定每一个视频帧所对应的音频信息。In this embodiment, after the audio and video to be shared are parsed into audio streams and video streams, the audio streams and video streams will not be transmitted or played immediately. Instead, the timestamps of the parsed audio streams and video streams will be aligned to determine each The audio information corresponding to the video frame.
具体地,首先可以选择一个参考时钟(比如处理器的系统时钟);编码时可以依据参考时钟上的给每个音视频数据块都打上时间戳;在解析音频流和视频流时,每一帧视频帧以及每一段音频信息也都会存在时间戳信息,即可确定视频帧与音频信息的对应关系。Specifically, you can first select a reference clock (such as the system clock of the processor); when encoding, you can time stamp each audio and video data block based on the reference clock; when parsing the audio stream and video stream, each frame Video frames and each piece of audio information will also have timestamp information, which can determine the corresponding relationship between video frames and audio information.
步骤S30,在接收到音视频分享的指令时,向接收端发送所述视频帧以及所述视频帧对应的音频信息,以进行所述待分享音视频的共享。Step S30: When receiving an audio and video sharing instruction, the video frame and the audio information corresponding to the video frame are sent to the receiving end to share the audio and video to be shared.
现有技术中,共享音视频文件,可以将音视频文件发送到接收端,在接收端再进行播放显示,但是因为视频流和音频流的播放速度不一致,所以在接收端所播放的音视频文件存在音画不同步的问题。In the existing technology, to share audio and video files, the audio and video files can be sent to the receiving end, and then played and displayed at the receiving end. However, because the playback speeds of the video stream and the audio stream are inconsistent, the audio and video files played at the receiving end There is a problem of audio and video being out of sync.
本实施例中,通过建立音视频通话的数据通道或者新建数据通道,按照时间戳顺序,将对应相同时间戳信息的视频帧和音频信息同步输出到接收端,接收端所接收的音视频文件是根据时间戳信息进行校正和同步的音频流和视频帧文件。且共享端在发送音频流和视频帧之前,先根据同步信号(即当前播放视频帧的帧号)同步播放视频帧以及对应的音频信息,从而输出同步的音视频流,接收端接收到的经过同步的音视频流,即可播放音画同步的音视频文件。In this embodiment, by establishing a data channel for audio and video calls or creating a new data channel, video frames and audio information corresponding to the same timestamp information are synchronously output to the receiving end according to the timestamp order. The audio and video files received by the receiving end are Audio stream and video frame files corrected and synchronized based on timestamp information. And before sending the audio stream and video frame, the sharing end first plays the video frame and the corresponding audio information synchronously according to the synchronization signal (that is, the frame number of the currently played video frame), thereby outputting the synchronized audio and video stream, and the receiving end receives the process. Synchronized audio and video streams can play audio and video files with synchronized audio and video.
本实施例中,建立音视频通话的通话类型可以是IMS通话类型,也可以是OTT通话类型。In this embodiment, the call type for establishing the audio and video call may be an IMS call type or an OTT call type.
其中,IMS又叫IP多媒体子系统,可以在分组交换网络下实现语音业务,包括VoNR、VoLTE、VoWiFi等类型。Among them, IMS is also called IP multimedia subsystem, which can implement voice services under packet switching networks, including VoNR, VoLTE, VoWiFi and other types.
具体地,VoNR是基于纯5G接入的通话解决方案,实现话音业务和数据业务均承载在5G网络。Specifically, VoNR is a call solution based on pure 5G access, which enables both voice and data services to be carried on the 5G network.
具体地,VoLTE是一种IP数据传输技术,无需2G/3G网,全部业务承载于4G网络上,可实现数据与语音业务在同一网络下的统一。VoLTE是架构在4G网络上全IP条件下的端到端语音方案,相较于传统的通话,VoLTE高清通话技术接通快、掉线率接近于零,LTE的频谱利用效率远远优于传统制式。 Specifically, VoLTE is an IP data transmission technology that does not require a 2G/3G network. All services are carried on the 4G network, which can realize the unification of data and voice services under the same network. VoLTE is an end-to-end voice solution built on the 4G network under all-IP conditions. Compared with traditional calls, VoLTE high-definition call technology has fast connection times and close to zero call drop rates. LTE’s spectrum utilization efficiency is far superior to traditional calls. Standard.
具体地,VoWiFi是通过WiFi网络提供的语音业务,用户可以在没有移动信号的条件下拨打电话,VoWiFi是VoLTE的互补技术。相比传统的通话服务,VoWiFi优先利用WiFi连网实现通话功能且自动无缝转换,实现移动网络及WiFi网络间的自动转换,用户无需特别设置就可以在不同地点实现通话。更重要的是,利用WiFi连网克服了室内或地下室信号不良的问题。在网络覆盖较弱或受干扰的地方,只要能连上WiFi,就可拨打或接听电话。Specifically, VoWiFi is a voice service provided through WiFi networks. Users can make calls without mobile signals. VoWiFi is a complementary technology to VoLTE. Compared with traditional calling services, VoWiFi prioritizes the use of WiFi networking to implement calling functions and automatically and seamlessly switches between mobile networks and WiFi networks. Users can make calls in different locations without special settings. More importantly, using WiFi networking overcomes the problem of poor signal indoors or in the basement. In places with weak network coverage or interference, you can make or receive calls as long as you can connect to WiFi.
其中,OTT是指互联网公司越过运营商(电信、移动、联动),发展基于互联网的视频、社交、游戏、数据服务等增值业务,包括微信、Skype等类型。Among them, OTT refers to Internet companies that go beyond operators (telecom, mobile, linkage) to develop Internet-based value-added services such as video, social networking, games, and data services, including WeChat, Skype and other types.
具体地,微信是一个为智能终端提供即时通讯服务的应用程序,能够实现跨通信运营商、跨操作系统平台、通过网络快速发送语音短信、视频、图片和文字等。Specifically, WeChat is an application that provides instant messaging services for smart terminals. It can quickly send voice messages, videos, pictures, text, etc. across communication operators and operating system platforms through the network.
具体地,Skype是一款即时通讯软件,其具备IM所需的功能,比如视频聊天、多人语音会议、多人聊天、传送文件、文件聊天等功能。Specifically, Skype is an instant messaging software that has the functions required for IM, such as video chat, multi-person voice conferencing, multi-person chat, file transfer, file chat, etc.
可以理解地,在完成音视频共享之后,关闭视频播放器,结束音视频共享,共享端可以切换当前显示界面,显示终端相机实时采集的图片,并且切换当前音频设备为通话音频设备,采集用户的语音信息,并且通过通话数据通道将实时采集的图片和语音信息发送到接收端,直至音视频通话结束。Understandably, after completing the audio and video sharing, close the video player and end the audio and video sharing. The sharing terminal can switch the current display interface, display the pictures collected by the terminal camera in real time, and switch the current audio device to the call audio device to collect the user's Voice information, and send the real-time collected pictures and voice information to the receiving end through the call data channel until the audio and video call ends.
本公开提供了一种音视频共享方法,所述方法获取待分享音视频中的音频流和视频流,并将所述视频流解析成视频帧;根据所述待分享音视频的时间戳信息,在所述音频流中确定各个视频帧对应的音频信息;基于音视频通话的通话数据通道或新建数据通道、所述各个视频帧以及所述各个视频帧对应的音频信息,向接收端同步发送所述待分享音视频。通过上述方式,本公开将音视频文件的视频流解析成视频帧,并且将视频流对应的时间戳信息进行解析,视频帧与时间戳信息存在对应关系;而音视频文件的音频流同样对应于音视频的时间戳信息,所以可以确定每个视频帧所对应的音频信息,实现视频流和音频流的精确同步;在音视频通话时,通过建立音视频通话的数据通道或者建立用于传输数据的新建数据通道,将视频帧以及当前视频帧所对应的音频信息进行同步传输,接收端即可接收到精确同步的视频流和音频流,即接收端接收到音画同步的音视频文件,提高用户体验,解决了目前音 视频通话时不能共享音画同步的音视频文件的技术问题。The present disclosure provides an audio and video sharing method. The method obtains the audio stream and video stream in the audio and video to be shared, and parses the video stream into video frames; according to the timestamp information of the audio and video to be shared, The audio information corresponding to each video frame is determined in the audio stream; based on the call data channel or new data channel of the audio and video call, each video frame, and the audio information corresponding to each video frame, all the audio information corresponding to each video frame is synchronously sent to the receiving end. Audio and video to be shared. Through the above method, the present disclosure parses the video stream of the audio and video file into video frames, and parses the timestamp information corresponding to the video stream. There is a corresponding relationship between the video frame and the timestamp information; and the audio stream of the audio and video file also corresponds to The timestamp information of audio and video can be determined, so the audio information corresponding to each video frame can be determined to achieve precise synchronization of video stream and audio stream; during audio and video calls, by establishing a data channel for audio and video calls or establishing a data channel for transmitting data The newly created data channel synchronously transmits the video frame and the audio information corresponding to the current video frame. The receiving end can receive the accurately synchronized video stream and audio stream, that is, the receiving end receives the audio and video files with synchronized audio and video, improving user experience, solving the current problem of Technical issue that audio and video files synchronized with audio and video cannot be shared during video calls.
参照图3,图3为本公开音视频共享方法第二实施例的流程示意图。Referring to Figure 3, Figure 3 is a schematic flow chart of a second embodiment of the audio and video sharing method of the present disclosure.
基于上述图2所示实施例,本实施例中,所述步骤S30,具体包括:Based on the above embodiment shown in Figure 2, in this embodiment, step S30 specifically includes:
步骤S31,在所述视频帧中确定当前视频帧,并根据所述当前视频帧的帧号,同步输出所述当前视频帧以及所述当前视频帧对应的所述音频信息,获得同步的音视频流;Step S31, determine the current video frame in the video frame, and synchronously output the current video frame and the audio information corresponding to the current video frame according to the frame number of the current video frame, to obtain synchronized audio and video flow;
本实施例中,通过共享端的视频播放器同时播放完成时间戳对齐的视频流和音频流,在解析过程中,视频流进一步解析成视频帧,并且对视频帧进行编号,获得每一帧视频帧的帧号。In this embodiment, the video stream and audio stream with time stamp alignment are simultaneously played through the video player on the sharing end. During the parsing process, the video stream is further parsed into video frames, and the video frames are numbered to obtain each video frame. frame number.
进一步地,在通过视频播放器同步播放视频流和音频流时,根据当前播放的视频流对应的帧号,同步输出该帧号对应的视频帧,而音频流在播放的同时同步向音频前端输出,从而实现同步输出的视频帧和音频流,获得同步输出的音视频流。Further, when the video stream and the audio stream are played synchronously through the video player, the video frame corresponding to the frame number is synchronously output according to the frame number corresponding to the currently played video stream, and the audio stream is synchronously output to the audio front end while being played. , thereby achieving synchronous output of video frames and audio streams, and obtaining synchronous output of audio and video streams.
步骤S32,基于数据传输通道,向所述接收端发送所述同步的音视频流,以使所述接收端同步播放所述待分享音视频。Step S32: Send the synchronized audio and video stream to the receiving end based on the data transmission channel, so that the receiving end plays the audio and video to be shared synchronously.
本实施例中,根据音视频通话的类型,通过建立音视频通话的通话数据通道或者新建数据通道,将同步输出的音视频流发送到接收端,接收端可以调用对应的目标应用(如视频播放器、音视频应用等)直接播放经过同步的音视频流。In this embodiment, according to the type of audio and video calls, the synchronously output audio and video streams are sent to the receiving end by establishing a call data channel for the audio and video calls or creating a new data channel, and the receiving end can call the corresponding target application (such as video playback devices, audio and video applications, etc.) to directly play the synchronized audio and video streams.
进一步地,所述步骤S32具体包括:Further, the step S32 specifically includes:
通过所述数据传输通道中的应用数据通道向所述接收端发送目标应用对应的应用数据信息;其中,所述目标应用用于播放所述音视频流;Send application data information corresponding to the target application to the receiving end through the application data channel in the data transmission channel; wherein the target application is used to play the audio and video stream;
通过所述数据传输通道中的传输数据通道向所述接收端发送所述同步的音视频流。The synchronized audio and video stream is sent to the receiving end through a transmission data channel in the data transmission channel.
本实施例中,在建立VoNR+视频通话时,可以通过建立非Bootstrap的数据通道,向接收端传输音频流和视频帧;再建立一个Bootstrap的数据通道,向接收端传输播放音视频文件的目标应用对应的应用数据信息。In this embodiment, when establishing a VoNR+ video call, a non-Bootstrap data channel can be established to transmit audio streams and video frames to the receiving end; a Bootstrap data channel can then be established to transmit the target application for playing audio and video files to the receiving end. Corresponding application data information.
进一步地,所述通过所述应用数据通道向所述接收端发送目标应用对应的应用数据信息之前,还包括:Further, before sending the application data information corresponding to the target application to the receiving end through the application data channel, the method further includes:
判断网络数据服务器中是否存在所述应用数据信息; Determine whether the application data information exists in the network data server;
若所述网络数据服务器中不存在所述应用数据信息,则基于所述应用数据通道,将本地的所述应用数据信息发送至所述接收端。If the application data information does not exist in the network data server, the local application data information is sent to the receiving end based on the application data channel.
本实施例中,共享端可以发送携带dcmap字段的SIP UPDATE信今给网络侧,然后建立与网络数据服务器或接收端的数据通道;共享端可以将待分享音视频预先保存在网络数据服务器中。In this embodiment, the sharing end can send a SIP UPDATE message carrying the dcmap field to the network side, and then establish a data channel with the network data server or the receiving end; the sharing end can pre-save the audio and video to be shared in the network data server.
具体地,当建立音视频通话时,共享端和接收端建立数据连接时,接收端同时与网络数据服务器建立数据通道;当接收到音视频文件共享指令时,共享端根据共享指令,显示网络数据服务器,并且在网络数据服务器中查询目标应用,如果存在所述目标应用,则直接根据查询结果指向目标应用所在位置,并通过接收端与网络数据服务器的数据通道发送该目标应用对应的应用数据信息;如果目标应用不存在于网络数据服务器,则显示查询无结果,共享端可以选择关闭网络数据服务器,在本地中查询目标应用。Specifically, when establishing an audio and video call, when the sharing terminal and the receiving terminal establish a data connection, the receiving terminal simultaneously establishes a data channel with the network data server; when receiving an audio and video file sharing instruction, the sharing terminal displays the network data according to the sharing instruction. server, and queries the target application in the network data server. If the target application exists, directly points to the location of the target application according to the query result, and sends the application data information corresponding to the target application through the data channel between the receiving end and the network data server. ; If the target application does not exist in the network data server, the query will show no results. The sharing end can choose to close the network data server and query the target application locally.
具体地,共享端可以选择预先保存在网络数据服务器中的待分享音视频,通过音视频通话的数据通道传输给接收端,或者通过网络数据服务器与接收端之间的数据通道直接传输。在需要进行音视频文件共享时,共享端可以选择网络数据服务器中的音视频文件和应用,也可以选择存储于共享端本地的音视频文件和应用。Specifically, the sharing end can select audio and video to be shared that are pre-stored in the network data server and transmit them to the receiving end through the data channel of the audio and video call, or directly transmit them through the data channel between the network data server and the receiving end. When audio and video files need to be shared, the sharing end can select audio and video files and applications in the network data server, or can choose audio and video files and applications stored locally on the sharing end.
其中,dcmap字段表征需要建立数据通道,该数据通道可以是用于传输数据通道应用的Bootstrap的数据通道,也可以是用于传输数据通道应用需要的数据信息的非Bootstrap的数据通道。Among them, the dcmap field represents the need to establish a data channel. The data channel can be a Bootstrap data channel used to transmit data channel applications, or a non-Bootstrap data channel used to transmit data information required by data channel applications.
具体实施例中,如果待分享音视频和应用数据信息不在网络数据服务器中,则可以选择本地存储的音视频文件或者流媒体视频,以及相对应的应用,通过建立通话的数据通道传输到接收端。In a specific embodiment, if the audio, video and application data information to be shared are not in the network data server, you can select locally stored audio and video files or streaming videos, as well as the corresponding applications, and transmit them to the receiving end through the data channel established for the call. .
进一步地,若所述网络数据服务器中存在所述应用数据信息,则通过所述网络数据服务器向所述接收端发送所述应用数据信息。Further, if the application data information exists in the network data server, the application data information is sent to the receiving end through the network data server.
具体实施例中,如果待分享音视频和应用数据信息已经预先存储于网络数据服务器中,共享端可以在网络数据服务器中选择对应的待分享音视频文件,在共享端完成待分享音视频文件的解析和同步等操作;再通过数据通道将同步后的音频流和视频帧发送到接收端,并且通过实时传输共享端当前播放的视频帧帧号,在接收端播放对应的视频帧和音频流,实现通话两端音视 频文件的同步播放。In a specific embodiment, if the audio, video and application data information to be shared have been stored in the network data server in advance, the sharing end can select the corresponding audio and video file to be shared in the network data server, and complete the editing of the audio and video file to be shared on the sharing end. Operations such as analysis and synchronization; then the synchronized audio stream and video frame are sent to the receiving end through the data channel, and the corresponding video frame and audio stream are played on the receiving end through real-time transmission of the video frame number currently played by the sharing end. Realize audio and video at both ends of the call Synchronous playback of video files.
具体实施例中,接收端可以直接从网络数据服务器中获取其中的音视频文件和应用,通过该应用打开该音视频文件,共享端通过实时传输当前播放的视频帧帧号,接收端根据该帧号播放对应的视频帧和音频流,从而实现通话两端音视频文件的同步播放。In a specific embodiment, the receiving end can directly obtain the audio and video files and applications from the network data server, open the audio and video files through the application, and the sharing end transmits the frame number of the currently played video frame in real time. Play the corresponding video frames and audio streams, thereby achieving synchronous playback of audio and video files at both ends of the call.
可选地,共享端可以将输出的视频帧替换摄像机当前采集的画面,并且将音频流输出到麦克风采集的音频设备,通过当前播放的视频流对应的帧号,同步输出该帧号对应的视频帧和音频信息,并且通过音视频通话的通话数据通道,将当前通话播放的视频帧和音频信息传输到接收端,在接收端同步播放。Optionally, the sharing end can replace the output video frame with the picture currently collected by the camera, and output the audio stream to the audio device collected by the microphone. Through the frame number corresponding to the currently played video stream, the video corresponding to the frame number is synchronously output. Frame and audio information, and through the call data channel of the audio and video call, the video frame and audio information played by the current call are transmitted to the receiving end, and played synchronously at the receiving end.
作为另一种实施方式,共享端可以将视频帧和音频流同步输出的音视频文件直接发送给接收端,接收端在接收到同步的音视频流之后,可以将该音视频流保存到本地,作为本地的音视频文件,音视频文件中可以包括视频帧文件和音频流文件;然后在音视频通话过程中,发送端可以将当前播放的音频帧的帧号同步发送到接收端,接收端调用对应的播放应用播放该音视频文件,并根据接收到的帧号播放该音视频文件中对应帧号的当前视频帧,并且根据当前视频帧的时间戳信息,同步播放对应相同时间戳信息的音频信息,即可实现共享端和接收端之间相同音视频文件的同步播放。As another implementation, the sharing end can directly send the audio and video files outputted synchronously with the video frames and audio streams to the receiving end. After receiving the synchronized audio and video streams, the receiving end can save the audio and video streams locally. As local audio and video files, the audio and video files can include video frame files and audio stream files; then during the audio and video call, the sender can synchronously send the frame number of the currently played audio frame to the receiving end, and the receiving end calls The corresponding playback application plays the audio and video file, and plays the current video frame corresponding to the frame number in the audio and video file according to the received frame number, and based on the timestamp information of the current video frame, synchronously plays the audio corresponding to the same timestamp information. information, you can achieve synchronous playback of the same audio and video files between the sharing end and the receiving end.
具体地,因为通过帧号进行音视频文件的同步输出,在共享端调整音视频文件的播放速度(如快进、快退或暂停等)时,接收端接收到的帧号的刷新速度也会同步调整,那么在接收端也会实现音视频文件播放速度的同步调整,进一步保证共享端和接收端之间相同音视频文件的同步播放。Specifically, because audio and video files are synchronously output through frame numbers, when the sharing end adjusts the playback speed of audio and video files (such as fast forward, rewind, or pause, etc.), the refresh speed of the frame number received by the receiving end will also Synchronous adjustment, then the receiving end will also realize synchronous adjustment of the audio and video file playback speed, further ensuring the synchronous playback of the same audio and video files between the sharing end and the receiving end.
进一步地,所述步骤S32之前,还包括:Further, before step S32, it also includes:
与接收端建立音视频通话,若所述音视频通话不是所述VoNR+通话类型,则将所述音视频通话的通话数据通道作为所述数据传输通道;Establish an audio and video call with the receiving end. If the audio and video call is not the VoNR+ call type, use the call data channel of the audio and video call as the data transmission channel;
若所述音视频通话是VoNR+通话类型,则基于所述通话数据通道,与接收端建立数据通道,作为所述数据传输通道。If the audio and video call is a VoNR+ call type, a data channel is established with the receiving end based on the call data channel as the data transmission channel.
本实施例中,根据音视频通话的类型,确定是直接通过当前音视频通话的通话数据通道进行数据传输,或者是通过新建数据通道进行数据传输。In this embodiment, according to the type of the audio and video call, it is determined whether data transmission is performed directly through the call data channel of the current audio and video call, or data transmission is performed through a new data channel.
具体地,如果是当前音视频通话是VoNR+视频通话,则可以新建数据通 道,作为数据传输通道,用以传输音视频流或者应用数据信息;否则,可以直接将当前音视频通话的通话数据通道作为数据传输通道,向接收端传输音视频流。Specifically, if the current audio and video call is VoNR+video call, you can create a new data call Channel is used as a data transmission channel to transmit audio and video streams or application data information; otherwise, the call data channel of the current audio and video call can be directly used as a data transmission channel to transmit audio and video streams to the receiving end.
本实施例中,在建立VoNR+视频通话时,可以通过建立非Bootstrap的数据通道,向接收端传输音频流和视频帧;再建立一个Bootstrap的数据通道,向接收端传输播放音视频文件的目标应用对应的应用数据信息。In this embodiment, when establishing a VoNR+ video call, a non-Bootstrap data channel can be established to transmit audio streams and video frames to the receiving end; a Bootstrap data channel can then be established to transmit the target application for playing audio and video files to the receiving end. Corresponding application data information.
其中,VoNR+(5G新通话)是指在5G VoNR多媒体实时通信的基础上搭载新的数据传输通道,为用户提供除高清音视频之外的更丰富的实时交互服务,并建立以多媒体实时通信能力平台为中心,统一开放的网络架构,在兼容现有业务的基础上实现创新业务的敏捷开发和快速部署。VoNR+是基于VoNR(5GNR)基础上的实时通信网络架构,可以快速集成新的业务形态,以满足人们多元化的通信需求。承载实时交互信息的交互的信道更宽,交互的内容类型更多,交互形式更丰富。Among them, VoNR+ (5G New Call) refers to a new data transmission channel based on 5G VoNR multimedia real-time communication, providing users with richer real-time interactive services in addition to high-definition audio and video, and establishing multimedia real-time communication capabilities. Platform-centered, unified and open network architecture enables agile development and rapid deployment of innovative services on the basis of compatibility with existing services. VoNR+ is a real-time communication network architecture based on VoNR (5GNR), which can quickly integrate new business forms to meet people's diverse communication needs. The interactive channel that carries real-time interactive information is wider, the interactive content types are more, and the interactive forms are richer.
具体实施例中,共享端首先通过Bootstrap的数据通道,将用于播放待分享音视频的应用数据信息传输到接收端,接收端通过该应用数据信息生成对应的目标应用,可以是临时的应用程序(如小程序等),也可以是应用下载链接,接收端通过该链接下载相对应的媒体播放应用。In a specific embodiment, the sharing end first transmits the application data information for playing the audio and video to be shared to the receiving end through the data channel of Bootstrap. The receiving end generates the corresponding target application through the application data information, which can be a temporary application. (such as a small program, etc.), or it can be an application download link, through which the receiving end downloads the corresponding media playback application.
具体实施例中,完成目标应用的传输之后,共享端在同时播放经过解析和同步的音频流和视频流时,通过非Bootstrap的数据通道将音频流和视频帧同步传输到接收端,并且将共享端当前播放的当前视频帧的帧号实时传输给接收端,接收端根据当前视频帧的帧号,通过目标应用同步播放共享端传输的音视频文件。In a specific embodiment, after completing the transmission of the target application, the sharing end simultaneously plays the parsed and synchronized audio stream and video stream, and synchronously transmits the audio stream and video frame to the receiving end through a non-Bootstrap data channel, and will share The frame number of the current video frame currently played by the end is transmitted to the receiving end in real time. The receiving end synchronously plays the audio and video files transmitted by the sharing end through the target application based on the frame number of the current video frame.
其中,Bootstrap的数据通道,是Stream ID小于1000的数据通道,用于终端从网络侧获取HTML网页(即3GPP TS26.114规范中定义的数据通道应用,应用中一般包括HTML、JavaScript脚本、CSS等)。Among them, the Bootstrap data channel is a data channel with a Stream ID less than 1000. It is used for the terminal to obtain HTML web pages from the network side (that is, the data channel application defined in the 3GPP TS26.114 specification. The application generally includes HTML, JavaScript scripts, CSS, etc. ).
参照图4,图4为本公开音视频共享方法第三实施例的流程示意图。Referring to Figure 4, Figure 4 is a schematic flowchart of a third embodiment of the audio and video sharing method of the present disclosure.
基于上述图2所示实施例,本实施例中,所述步骤S30之前,具体还包括:Based on the above embodiment shown in Figure 2, in this embodiment, before step S30, it specifically includes:
步骤S031,根据所述音频信息的数据类型,确定所述音频信息对应的目标音频通道; Step S031: Determine the target audio channel corresponding to the audio information according to the data type of the audio information;
步骤S032,将视频播放器对应的第一音频通道以及音视频通话对应的第二音频通道切换为所述目标音频通道,以通过所述目标音频通道在播放所述音频信息的同时向所述接收端同步发送所述音频信息。Step S032: Switch the first audio channel corresponding to the video player and the second audio channel corresponding to the audio and video call to the target audio channel, so as to play the audio information through the target audio channel while simultaneously providing the audio information to the receiver. The terminal sends the audio information synchronously.
本实施例中,共享端在传输音频流和视频帧时,首先需要通过时间戳信息对音频流和视频帧进行同步,同步完成后,并不会立刻对音频流和视频帧进行传输,而是通过视频播放器对音频流和视频帧同时播放,使得音频流和视频帧以相同的起始时间开始播放,播放的同时,向接收端同步传输音频流和视频帧,以使得接收端能够同步播放音画同步的音视频流。In this embodiment, when the sharing end transmits the audio stream and video frame, it first needs to synchronize the audio stream and video frame through the timestamp information. After the synchronization is completed, the audio stream and video frame will not be transmitted immediately. Use the video player to play the audio stream and video frame simultaneously, so that the audio stream and video frame start playing at the same starting time. While playing, the audio stream and video frame are synchronously transmitted to the receiving end so that the receiving end can play synchronously. Synchronized audio and video streams.
进一步地,音频流在播放的同时向接收端进行传输,为了避免因为转码等操作造成传输延时而导致音画不同步的问题,在共享端开始播放音频流之前,先确定传输该音频流的音频通道,然后将视频播放器和通话数据通道的音频数据通道都切换成该音频通道,在播放音频流的同时,即可同步输出该音频流,避免转码的过程,即可实现输出音频流和播放音频流的时间戳同步,进而实现音频流和视频帧的同步输出。比如,可以在播放前,将视频播放器的音频通道切换成对应通话模式外放的音频通道,在通过视频播放器进行音视频文件的播放时,即可输出音频流,同时将音频流数据传输到接收端,并通过接收端的通话音频通道进行播放。Furthermore, the audio stream is transmitted to the receiving end while playing. In order to avoid the problem of audio and video being out of sync due to transmission delays caused by operations such as transcoding, the audio stream must be transmitted before the sharing end starts playing the audio stream. audio channel, and then switch the audio data channels of the video player and call data channel to this audio channel. While playing the audio stream, the audio stream can be output synchronously, avoiding the transcoding process, and the audio output can be realized. The timestamps of streaming and playing audio streams are synchronized, thereby achieving synchronous output of audio streams and video frames. For example, before playing, you can switch the audio channel of the video player to the audio channel corresponding to the call mode. When playing the audio and video files through the video player, the audio stream can be output and the audio stream data can be transmitted at the same time. to the receiving end and played through the call audio channel of the receiving end.
参照图5,图5为本公开音视频共享方法第四实施例的流程示意图。Referring to Figure 5, Figure 5 is a schematic flowchart of a fourth embodiment of the audio and video sharing method of the present disclosure.
基于上述图2所示实施例,本实施例中,所述步骤S30,具体还包括:Based on the above embodiment shown in Figure 2, in this embodiment, step S30 specifically includes:
步骤S310,根据所述待分享音视频所属应用信息,确定所述待分享音视频所属应用对应的应用白名单,并判断所述音视频通话所属应用是否属于所述应用白名单;Step S310: Determine an application whitelist corresponding to the application to which the audio and video to be shared belongs based on the application information to which the audio and video to be shared belongs, and determine whether the application to which the audio and video call belongs belongs to the application whitelist;
步骤S320,若所述音视频通话所属应用属于所述应用白名单,则基于所述各个视频帧以及所述各个视频帧对应的音频信息,向所述接收端发送所述待分享音视频。Step S320: If the application to which the audio and video call belongs belongs to the application whitelist, the audio and video to be shared are sent to the receiving end based on the respective video frames and the audio information corresponding to the respective video frames.
本实施例中,在建立OTT音视频通话(如微信、Skype等)时,因为OTT音视频通话是越过运营商、基于互联网的通话类型,在进行音视频文件共享时,需要对建立OTT音视频通话的应用的权限进行检查。In this embodiment, when establishing an OTT audio and video call (such as WeChat, Skype, etc.), because the OTT audio and video call is a call type that transcends operators and is based on the Internet, when sharing audio and video files, it is necessary to establish an OTT audio and video call. The permissions of the calling app are checked.
具体地,当待分享音视频来源于如短视频应用(如抖音、快手等)时,需要获得该应用的可分享应用白名单,在该白名单中,会显示能够用于分享 来源于该短视频应用的音视频文件的通话应用。Specifically, when the audio and video to be shared comes from a short video application (such as Douyin, Kuaishou, etc.), it is necessary to obtain the shareable application whitelist of the application. In the whitelist, the information that can be used for sharing will be displayed. Call application derived from the audio and video files of the short video application.
具体地,当建立OTT音视频通话的通话应用在该白名单中时,则可以直接通过该通话应用将待分享音视频分享给接收端;而当该通话应用不在白名单中时,则需要通过设置该通话应用的权限,将该通话应用加入到白名单中,才能实现待分享音视频的分享。比如,通过微信建立音视频通话时,共享端需要共享抖音中的某个音视频文件,此时共享端需要判断微信是否具备分享该音视频文件的权限,如果白名单中没有微信选择,则说明微信不具备分享权限,则需要进行权限设置,使得微信能够分享抖音音视频文件。Specifically, when the calling application that establishes the OTT audio and video call is in the whitelist, the audio and video to be shared can be shared to the receiving end directly through the calling application; and when the calling application is not in the whitelist, you need to use the calling application to share the audio and video to be shared with the receiving end. Only by setting the permissions of the calling application and adding the calling application to the whitelist can the audio and video to be shared be shared. For example, when establishing an audio and video call through WeChat, the sharing terminal needs to share an audio and video file in Douyin. At this time, the sharing terminal needs to determine whether WeChat has the permission to share the audio and video file. If WeChat is not selected in the whitelist, then It means that WeChat does not have the sharing permission, so you need to set the permissions so that WeChat can share Douyin audio and video files.
此外,本公开实施例还提供一种音视频共享装置。In addition, embodiments of the present disclosure also provide an audio and video sharing device.
参照图6,图6为本公开音视频共享装置第一实施例的功能模块示意图。Referring to FIG. 6 , FIG. 6 is a schematic diagram of functional modules of the audio and video sharing device according to the first embodiment of the present disclosure.
本实施例中,所述音视频共享装置包括:In this embodiment, the audio and video sharing device includes:
音视频解析模块10,用于获取待分享音视频中的音频流和视频流,并将所述视频流解析成视频帧;The audio and video parsing module 10 is used to obtain the audio stream and video stream in the audio and video to be shared, and parse the video stream into video frames;
音视频流同步模块20,用于根据所述待分享音视频的时间戳信息,确定各视频帧对应的音频信息;The audio and video stream synchronization module 20 is used to determine the audio information corresponding to each video frame based on the timestamp information of the audio and video to be shared;
音视频发送模块30,用于在接收到音视频分享的指令时,向接收端发送所述视频帧以及所述视频帧对应的音频信息,以进行所述待分享音视频的共享。The audio and video sending module 30 is configured to, when receiving an audio and video sharing instruction, send the video frame and the audio information corresponding to the video frame to the receiving end to share the audio and video to be shared.
进一步地,所述音视频发送模块30具体包括:Further, the audio and video sending module 30 specifically includes:
音视频流同步输出单元,用于在所述视频帧中确定当前视频帧,并根据所述当前视频帧的帧号,同步输出所述当前视频帧以及所述当前视频帧对应的所述音频信息,获得同步的音视频流;An audio and video stream synchronization output unit is used to determine the current video frame in the video frame, and synchronously output the current video frame and the audio information corresponding to the current video frame according to the frame number of the current video frame. , obtain synchronized audio and video streams;
音视频流发送单元,用于基于数据传输通道,向所述接收端发送所述同步的音视频流,以使所述接收端同步播放所述待分享音视频。The audio and video stream sending unit is configured to send the synchronized audio and video stream to the receiving end based on the data transmission channel, so that the receiving end can synchronously play the audio and video to be shared.
进一步地,所述音视频流发送单元具体包括:Further, the audio and video stream sending unit specifically includes:
应用数据信息发送子单元,用于通过所述数据传输通道中的应用数据通道向所述接收端发送目标应用对应的应用数据信息;其中,所述目标应用用于播放所述音视频流;The application data information sending subunit is configured to send application data information corresponding to the target application to the receiving end through the application data channel in the data transmission channel; wherein the target application is used to play the audio and video stream;
音视频流发送子单元,用于通过所述数据传输通道中的传输数据通道向所述接收端发送所述同步的音视频流。 The audio and video stream sending subunit is configured to send the synchronized audio and video stream to the receiving end through the transmission data channel in the data transmission channel.
进一步地,所述音视频共享装置还包括网络数据服务器模块,所述网络数据服务器模块具体包括:Further, the audio and video sharing device also includes a network data server module. The network data server module specifically includes:
网络服务器查询单元,用于判断网络数据服务器中是否存在所述应用数据信息;A network server query unit, used to determine whether the application data information exists in the network data server;
本地应用信息发送单元,用于在所述网络数据服务器中不存在所述应用数据信息的情况下,则基于所述应用数据通道,将本地的所述应用数据信息发送至所述接收端。A local application information sending unit is configured to send the local application data information to the receiving end based on the application data channel when the application data information does not exist in the network data server.
网络端应用发送单元,用于若所述网络数据服务器中存在所述应用数据信息,则通过所述网络数据服务器向所述接收端发送所述应用数据信息。A network-side application sending unit is configured to send the application data information to the receiving end through the network data server if the application data information exists in the network data server.
进一步地,所述音视频共享装置还包括数据传输通道确定模块,所述数据传输通道确定模块具体包括:Further, the audio and video sharing device also includes a data transmission channel determination module. The data transmission channel determination module specifically includes:
音视频通话类型判断单元,用于与接收端建立音视频通话,若所述音视频通话不是所述VoNR+通话类型,则将所述音视频通话的通话数据通道作为所述数据传输通道;An audio and video call type judgment unit is used to establish an audio and video call with the receiving end. If the audio and video call is not the VoNR+ call type, use the call data channel of the audio and video call as the data transmission channel;
新建数据通道单元,用于若所述音视频通话是VONR+通话类型,则基于所述通话数据通道,与接收端建立数据通道,作为所述数据传输通道。Create a new data channel unit, used to establish a data channel with the receiving end based on the call data channel as the data transmission channel if the audio and video call is a VONR+ call type.
进一步地,音视频解析模块10包括:Further, the audio and video analysis module 10 includes:
音视频文件解析单元,用于通过当前正在运行的视频播放器,读取所述待分享音视频,将所述待分享音视频文件解析成所述音频流和所述视频流对应的视频帧。The audio and video file parsing unit is used to read the audio and video to be shared through the currently running video player, and parse the audio and video file to be shared into the audio stream and the video frame corresponding to the video stream.
进一步地,所述音视频共享装置包括音频通道切换模块,所述音频通道切换模块具体包括:Further, the audio and video sharing device includes an audio channel switching module, and the audio channel switching module specifically includes:
目标音频通道确定单元,用于根据所述音频信息的数据类型,确定所述音频信息对应的目标音频通道;A target audio channel determination unit, configured to determine the target audio channel corresponding to the audio information according to the data type of the audio information;
音频通道切换单元,用于将视频播放器对应的第一音频通道以及音视频通话对应的第二音频通道切换为所述目标音频通道,以通过所述目标音频通道在播放所述音频信息的同时向所述接收端同步发送所述音频信息。An audio channel switching unit is used to switch the first audio channel corresponding to the video player and the second audio channel corresponding to the audio and video call to the target audio channel, so as to play the audio information through the target audio channel while playing The audio information is synchronously sent to the receiving end.
进一步地,所述音视频共享装置包括白名单模块,所述白名单模块具体包括:Further, the audio and video sharing device includes a whitelist module, and the whitelist module specifically includes:
应用白名单确定单元,用于根据所述待分享音视频所属应用信息,确定 所述待分享音视频所属应用对应的应用白名单,并判断所述音视频通话所属应用是否属于所述应用白名单;An application whitelist determination unit is configured to determine based on the application information to which the audio and video to be shared belongs. The application whitelist corresponding to the application to which the audio and video to be shared belongs, and determine whether the application to which the audio and video call belongs belongs to the application whitelist;
音视频文件发送单元,用于若所述音视频通话所属应用属于所述应用白名单,则基于所述各个视频帧以及所述各个视频帧对应的音频信息,向所述接收端发送所述待分享音视频。An audio and video file sending unit, configured to, if the application to which the audio and video call belongs belongs to the application whitelist, send the to-be-received file to the receiving end based on the respective video frames and the audio information corresponding to the respective video frames. Share audio and video.
其中,上述音视频共享装置中各个模块与上述音视频共享方法实施例中各步骤相对应,其功能和实现过程在此处不再一一赘述。Each module in the above-mentioned audio and video sharing device corresponds to each step in the above-mentioned audio and video sharing method embodiment, and their functions and implementation processes will not be described in detail here.
此外,本公开实施例还提供一种计算机可读存储介质。In addition, embodiments of the present disclosure also provide a computer-readable storage medium.
本公开计算机可读存储介质上存储有音视频共享程序,其中所述音视频共享程序被处理器执行时,实现如上述的音视频共享方法的步骤。The computer-readable storage medium of the present disclosure stores an audio and video sharing program. When the audio and video sharing program is executed by a processor, the steps of the above audio and video sharing method are implemented.
其中,音视频共享程序被执行时所实现的方法可参照本公开音视频共享方法的各个实施例,此处不再赘述。For the method implemented when the audio and video sharing program is executed, reference may be made to various embodiments of the audio and video sharing method of the present disclosure, and details will not be described again here.
需要说明的是,在本文中,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者系统不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者系统所固有的要素。在没有更多限制的情况下,由语句“包括一个......”限定的要素,并不排除在包括该要素的过程、方法、物品或者系统中还存在另外的相同要素。It should be noted that, as used herein, the terms "include", "comprising" or any other variation thereof are intended to cover a non-exclusive inclusion, such that a process, method, article or system that includes a list of elements not only includes those elements, but It also includes other elements not expressly listed or that are inherent to the process, method, article or system. Without further limitation, an element qualified by the statement "comprises a..." does not exclude the presence of additional identical elements in the process, method, article, or system that includes that element.
上述本公开实施例序号仅仅为了描述,不代表实施例的优劣。The above serial numbers of the embodiments of the present disclosure are only for description and do not represent the advantages and disadvantages of the embodiments.
本公开可用于众多通用或专用的计算机系统环境或配置中。例如:个人计算机、服务器计算机、手持设备或便携式设备、平板型设备、多处理器系统、基于微处理器的系统、置顶盒、可编程的消费电子设备、网络PC、小型计算机、大型计算机、包括以上任何系统或设备的分布式计算环境等等。本公开可以在由计算机执行的计算机可执行指令的一般上下文中描述,例如程序模块。一般地,程序模块包括执行特定任务或实现特定抽象数据类型的例程、程序、对象、组件、数据结构等等。也可以在分布式计算环境中实践本公开,在这些分布式计算环境中,由通过通信网络而被连接的远程处理设备来执行任务。在分布式计算环境中,程序模块可以位于包括存储设备在内的本地和远程计算机存储介质中。The present disclosure may be used in numerous general purpose or special purpose computer system environments or configurations. For example: personal computers, server computers, handheld or portable devices, tablet devices, multiprocessor systems, microprocessor-based systems, set-top boxes, programmable consumer electronics devices, network PCs, minicomputers, mainframe computers, including Distributed computing environment for any of the above systems or devices, etc. The present disclosure may be described in the general context of computer-executable instructions, such as program modules, being executed by a computer. Generally, program modules include routines, programs, objects, components, data structures, etc. that perform specific tasks or implement specific abstract data types. The present disclosure may also be practiced in distributed computing environments where tasks are performed by remote processing devices connected through a communications network. In a distributed computing environment, program modules may be located in both local and remote computer storage media including storage devices.
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到上述 实施例方法可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件,但很多情况下前者是更佳的实施方式。基于这样的理解,本公开的技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品存储在如上所述的一个存储介质(如ROM/RAM、磁碟、光盘)中,包括若干指令用以使得一台终端设备(可以是手机,计算机,服务器,空调器,或者网络设备等)执行本公开各个实施例所述的方法。Through the above description of the embodiments, those skilled in the art can clearly understand the above The embodiment method can be implemented by means of software plus the necessary general hardware platform. Of course, it can also be implemented by hardware, but in many cases the former is a better implementation method. Based on this understanding, the technical solution of the present disclosure can be embodied in the form of a software product in nature or in part that contributes to the existing technology. The computer software product is stored in a storage medium (such as ROM/RAM) as mentioned above. , magnetic disk, optical disk), including several instructions to cause a terminal device (which can be a mobile phone, computer, server, air conditioner, or network device, etc.) to execute the methods described in various embodiments of the present disclosure.
以上仅为本公开的优选实施例,并非因此限制本公开的专利范围,凡是利用本公开说明书及附图内容所作的等效结构或等效流程变换,或直接或间接运用在其他相关的技术领域,均同理包括在本公开的专利保护范围内。 The above are only preferred embodiments of the present disclosure, and are not intended to limit the patent scope of the present disclosure. Any equivalent structure or equivalent process transformation made using the contents of the disclosure description and drawings may be directly or indirectly applied in other related technical fields. , are all similarly included in the patent protection scope of this disclosure.

Claims (10)

  1. 一种音视频共享方法,所述方法包括:An audio and video sharing method, the method includes:
    获取待分享音视频中的音频流和视频流,并将所述视频流解析成视频帧;Obtain the audio stream and video stream in the audio and video to be shared, and parse the video stream into video frames;
    根据所述待分享音视频的时间戳信息,确定各视频帧对应的音频信息;Determine the audio information corresponding to each video frame according to the timestamp information of the audio and video to be shared;
    在接收到音视频分享的指令时,向接收端发送所述视频帧以及所述视频帧对应的音频信息,以进行所述待分享音视频的共享。When receiving an instruction to share audio and video, the video frame and the audio information corresponding to the video frame are sent to the receiving end to share the audio and video to be shared.
  2. 如权利要求1所述的音视频共享方法,其中,所述在接收到音视频分享的指令时,向接收端发送所述视频帧以及所述视频帧对应的音频信息,以进行所述待分享音视频的共享,包括:The audio and video sharing method according to claim 1, wherein when receiving an audio and video sharing instruction, the video frame and the audio information corresponding to the video frame are sent to the receiving end for the to-be-shared Audio and video sharing, including:
    在所述视频帧中确定当前视频帧,并根据所述当前视频帧的帧号,同步输出所述当前视频帧以及所述当前视频帧对应的所述音频信息,获得同步的音视频流;Determine the current video frame in the video frame, and synchronously output the current video frame and the audio information corresponding to the current video frame according to the frame number of the current video frame, to obtain a synchronized audio and video stream;
    基于数据传输通道,向所述接收端发送所述同步的音视频流,以使所述接收端同步播放所述待分享音视频。Based on the data transmission channel, the synchronized audio and video stream is sent to the receiving end, so that the receiving end plays the audio and video to be shared synchronously.
  3. 如权利要求2所述的音视频共享方法,其中,所述基于数据传输通道,向所述接收端发送所述同步的音视频流,包括:The audio and video sharing method according to claim 2, wherein the sending the synchronized audio and video stream to the receiving end based on the data transmission channel includes:
    通过所述数据传输通道中的应用数据通道向所述接收端发送目标应用对应的应用数据信息;其中,所述目标应用用于播放所述音视频流;Send application data information corresponding to the target application to the receiving end through the application data channel in the data transmission channel; wherein the target application is used to play the audio and video stream;
    通过所述数据传输通道中的传输数据通道向所述接收端发送所述同步的音视频流。The synchronized audio and video stream is sent to the receiving end through a transmission data channel in the data transmission channel.
  4. 如权利要求3所述的音视频共享方法,其中,所述通过所述数据传输通道中的应用数据通道向所述接收端发送目标应用对应的应用数据信息之前,还包括:The audio and video sharing method according to claim 3, wherein before sending the application data information corresponding to the target application to the receiving end through the application data channel in the data transmission channel, it further includes:
    判断网络数据服务器中是否存在所述应用数据信息;Determine whether the application data information exists in the network data server;
    在所述网络数据服务器中不存在所述应用数据信息的情况下,基于所述应用数据通道,将本地的所述应用数据信息发送至所述接收端。If the application data information does not exist in the network data server, the local application data information is sent to the receiving end based on the application data channel.
    若所述网络数据服务器中存在所述应用数据信息,则通过所述网络数据服务器向所述接收端发送所述应用数据信息。If the application data information exists in the network data server, the application data information is sent to the receiving end through the network data server.
  5. 如权利要求2所述的音视频共享方法,其中,所述基于数据传输通道,向所述接收端发送所述同步的音视频流之前,包括: The audio and video sharing method according to claim 2, wherein before sending the synchronized audio and video stream to the receiving end based on the data transmission channel, the method includes:
    与接收端建立音视频通话,若所述音视频通话不是所述VoNR+通话类型,则将所述音视频通话的通话数据通道作为所述数据传输通道;Establish an audio and video call with the receiving end. If the audio and video call is not the VoNR+ call type, use the call data channel of the audio and video call as the data transmission channel;
    若所述音视频通话是VoNR+通话类型,则基于所述通话数据通道,与接收端建立数据通道,作为所述数据传输通道。If the audio and video call is a VoNR+ call type, a data channel is established with the receiving end based on the call data channel as the data transmission channel.
  6. 如权利要求1所述的音视频共享方法,其中,所述获取待分享音视频中的音频流和视频流,并将所述视频流解析成视频帧,包括:The audio and video sharing method according to claim 1, wherein said obtaining the audio stream and video stream in the audio and video to be shared, and parsing the video stream into video frames includes:
    通过当前正在运行的视频播放器,读取所述待分享音视频,将所述待分享音视频文件解析成所述音频流和所述视频流对应的视频帧。The audio and video files to be shared are read through the currently running video player, and the audio and video files to be shared are parsed into video frames corresponding to the audio stream and the video stream.
  7. 如权利要求1所述的音视频共享方法,其中,所述在接收到音视频分享的指令时,向接收端发送所述视频帧以及所述视频帧对应的音频信息,以进行所述待分享音视频的共享之前,还包括:The audio and video sharing method according to claim 1, wherein when receiving an audio and video sharing instruction, the video frame and the audio information corresponding to the video frame are sent to the receiving end for the to-be-shared Before sharing audio and video, it also includes:
    根据所述音频信息的数据类型,确定所述音频信息对应的目标音频通道;Determine the target audio channel corresponding to the audio information according to the data type of the audio information;
    将视频播放器对应的第一音频通道以及音视频通话对应的第二音频通道切换为所述目标音频通道,以通过所述目标音频通道在播放所述音频信息的同时向所述接收端同步发送所述音频信息。Switch the first audio channel corresponding to the video player and the second audio channel corresponding to the audio and video call to the target audio channel, so as to simultaneously send the audio information to the receiving end through the target audio channel while playing the audio information. the audio information.
  8. 如权利要求1-7中任一项所述的音视频共享方法,其中,所述在接收到音视频分享的指令时,向接收端发送所述视频帧以及所述视频帧对应的音频信息,还包括:The audio and video sharing method according to any one of claims 1 to 7, wherein when receiving an audio and video sharing instruction, the video frame and the audio information corresponding to the video frame are sent to the receiving end, Also includes:
    根据所述待分享音视频所属应用信息,确定所述待分享音视频所属应用对应的应用白名单,并判断所述音视频通话所属应用是否属于所述应用白名单;According to the application information to which the audio and video to be shared belongs, determine the application whitelist corresponding to the application to which the audio and video to be shared belongs, and determine whether the application to which the audio and video call belongs belongs to the application whitelist;
    若所述音视频通话所属应用属于所述应用白名单,则基于所述各个视频帧以及所述各个视频帧对应的音频信息,向所述接收端发送所述待分享音视频。If the application to which the audio and video call belongs belongs to the application whitelist, the audio and video to be shared are sent to the receiving end based on the respective video frames and the audio information corresponding to the respective video frames.
  9. 一种音视频共享设备,其中,所述音视频共享设备包括处理器、存储器、以及存储在所述存储器上并可被所述处理器执行的音视频共享程序,其中所述音视频共享程序被所述处理器执行时,实现如权利要求1至8中任一项所述的音视频共享方法的步骤。An audio and video sharing device, wherein the audio and video sharing device includes a processor, a memory, and an audio and video sharing program stored on the memory and executable by the processor, wherein the audio and video sharing program is When the processor is executed, the steps of the audio and video sharing method according to any one of claims 1 to 8 are implemented.
  10. 一种计算机可读存储介质,其中,所述计算机可读存储介质上存储有音视频共享程序,其中所述音视频共享程序被处理器执行时,实现如权利要求1至8中任一项所述的音视频共享方法的步骤。 A computer-readable storage medium, wherein an audio and video sharing program is stored on the computer-readable storage medium, and when the audio and video sharing program is executed by a processor, the method of any one of claims 1 to 8 is implemented. Steps of the audio and video sharing method described above.
PCT/CN2023/078498 2022-05-31 2023-02-27 Audio and video sharing method and device, and computer-readable storage medium WO2023231478A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210606704.8 2022-05-31
CN202210606704.8A CN117201719A (en) 2022-05-31 2022-05-31 Audio and video sharing method, equipment and computer readable storage medium

Publications (1)

Publication Number Publication Date
WO2023231478A1 true WO2023231478A1 (en) 2023-12-07

Family

ID=88998456

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/078498 WO2023231478A1 (en) 2022-05-31 2023-02-27 Audio and video sharing method and device, and computer-readable storage medium

Country Status (2)

Country Link
CN (1) CN117201719A (en)
WO (1) WO2023231478A1 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117439976B (en) * 2023-12-13 2024-03-26 深圳大数信科技术有限公司 Audio and video call system based on WebRTC

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101279366B1 (en) * 2012-02-23 2013-07-04 (주)제이이모바일 Sharing method for contents by mobile devices
CN103237191A (en) * 2013-04-16 2013-08-07 成都飞视美视频技术有限公司 Method for synchronously pushing audios and videos in video conference
CN105577818A (en) * 2016-01-22 2016-05-11 腾讯科技(深圳)有限公司 Data transmission methods and apparatuses, and vehicle-mounted terminal
US9549152B1 (en) * 2014-06-09 2017-01-17 Google Inc. Application content delivery to multiple computing environments using existing video conferencing solutions
CN106657305A (en) * 2016-12-12 2017-05-10 掌阅科技股份有限公司 Data sharing method, data sharing device, terminal equipment and server
CN114339454A (en) * 2022-03-11 2022-04-12 浙江大华技术股份有限公司 Audio and video synchronization method and device, electronic device and storage medium

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101279366B1 (en) * 2012-02-23 2013-07-04 (주)제이이모바일 Sharing method for contents by mobile devices
CN103237191A (en) * 2013-04-16 2013-08-07 成都飞视美视频技术有限公司 Method for synchronously pushing audios and videos in video conference
US9549152B1 (en) * 2014-06-09 2017-01-17 Google Inc. Application content delivery to multiple computing environments using existing video conferencing solutions
CN105577818A (en) * 2016-01-22 2016-05-11 腾讯科技(深圳)有限公司 Data transmission methods and apparatuses, and vehicle-mounted terminal
CN106657305A (en) * 2016-12-12 2017-05-10 掌阅科技股份有限公司 Data sharing method, data sharing device, terminal equipment and server
CN114339454A (en) * 2022-03-11 2022-04-12 浙江大华技术股份有限公司 Audio and video synchronization method and device, electronic device and storage medium

Also Published As

Publication number Publication date
CN117201719A (en) 2023-12-08

Similar Documents

Publication Publication Date Title
US10187668B2 (en) Method, system and server for live streaming audio-video file
US9591262B2 (en) Flow-control based switched group video chat and real-time interactive broadcast
US10057662B2 (en) Flow controlled based synchronized playback of recorded media
US20060085823A1 (en) Media communications method and apparatus
CN105338425A (en) System and method for realizing video seamless switching between multiple screens
US20090106288A1 (en) Method and system for supporting media data of various coding formats
US20150181003A1 (en) Method and apparatus for transmitting and receiving packets in hybrid transmission service of mmt
RU2504090C2 (en) Method, apparatus and system for making video call
KR20080038251A (en) Method for signaling a device to perform no synchronization or include a synchronization delay on multimedia streams
Boronat et al. HbbTV-compliant platform for hybrid media delivery and synchronization on single-and multi-device scenarios
EP1603046B1 (en) Reception apparatus and information browsing method
CN109495761A (en) Video switching method and device
US20130204973A1 (en) Method for transmitting a scalable http stream for natural reproduction upon the occurrence of expression-switching during http streaming
CN105354002A (en) System and method for implementing video seamless switching among multiple screens
CN111526387B (en) Video processing method and device, electronic equipment and storage medium
CN112954433B (en) Video processing method, device, electronic equipment and storage medium
WO2023231478A1 (en) Audio and video sharing method and device, and computer-readable storage medium
CN108494792A (en) A kind of flash player plays the converting system and its working method of hls video flowings
WO2015180446A1 (en) System and method for maintaining connection channel in multi-device interworking service
CN113194278A (en) Conference control method and device and computer readable storage medium
CN205230019U (en) System for realize video seamless handover between many screens
EP3316593B1 (en) Method and device for implementing synchronous playing
CN113014950A (en) Live broadcast synchronization method and system and electronic equipment
CN108989737B (en) Data playing method and device and electronic equipment
CN108881793B (en) Data processing method and device for video network

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23814677

Country of ref document: EP

Kind code of ref document: A1