CN116567288A - Information generation method and device - Google Patents

Information generation method and device Download PDF

Info

Publication number
CN116567288A
CN116567288A CN202310667883.0A CN202310667883A CN116567288A CN 116567288 A CN116567288 A CN 116567288A CN 202310667883 A CN202310667883 A CN 202310667883A CN 116567288 A CN116567288 A CN 116567288A
Authority
CN
China
Prior art keywords
time length
video
audio
data
sub
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202310667883.0A
Other languages
Chinese (zh)
Inventor
李飞霞
黄思远
李磊
高锐
王左龙
孙晔
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Electronics China R&D Center
Samsung Electronics Co Ltd
Original Assignee
Samsung Electronics China R&D Center
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Samsung Electronics China R&D Center, Samsung Electronics Co Ltd filed Critical Samsung Electronics China R&D Center
Priority to CN202310667883.0A priority Critical patent/CN116567288A/en
Publication of CN116567288A publication Critical patent/CN116567288A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/231Content storage operation, e.g. caching movies for short term storage, replicating data over plural servers, prioritizing data for deletion
    • H04N21/23106Content storage operation, e.g. caching movies for short term storage, replicating data over plural servers, prioritizing data for deletion involving caching operations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/434Disassembling of a multiplex stream, e.g. demultiplexing audio and video streams, extraction of additional data from a video stream; Remultiplexing of multiplex streams; Extraction or processing of SI; Disassembling of packetised elementary stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8547Content authoring involving timestamps for synchronizing content
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Security & Cryptography (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

According to the information generation method and device, the current sub-data is determined from the audio-video data set based on the arrangement sequence of at least one sub-data in the audio-video data set corresponding to the target audio-video file, then the video playing time length and the audio playing time length corresponding to the current sub-data are obtained based on the target audio-video file, the time length difference between the video playing time length and the audio playing time length corresponding to the current sub-data is calculated, the maximum playing time length of a buffer zone corresponding to the current playing equipment is obtained, finally the interleaving result information corresponding to the current sub-data is generated based on the time length difference and the maximum playing time length of the buffer zone, and new current sub-data is determined from the audio-video data set based on the arrangement sequence of at least one sub-data in the audio-video data set continuously, so that interleaving conditions at different positions of the same target audio-video file can be judged respectively.

Description

Information generation method and device
Technical Field
The present disclosure relates to the field of computer technologies, and in particular, to the field of multimedia playing and information processing technologies, and in particular, to an information generating method and apparatus.
Background
With the technical development of the multimedia field, the smooth playing requirements of audio and video are higher and higher, and high-code-rate files and various media playing devices are generated. Because the audio and video synchronization is required when the multimedia file is played, the audio and video synchronization is required, the audio buffer area has the audio data with corresponding time when the video buffer area has the video data, and if the video buffer area has the data, the playing is stopped and a certain amount of audio data is buffered under the condition that the audio buffer area is empty, so that the playing can be resumed.
Therefore, if the playback device sequentially reads the audio and video frames from front to back, if the position intervals of some audio and video packets of the file are large, for example, if there is no audio packet for a long time, the video packets fill the video buffer area, and audio data cannot be read, so that playback cannot be continued; or, for the file with the index table, the audio and video frames are read by jumping the file read pointer back and forth according to an algorithm or a method set in advance, under the condition that the buffer area is not empty and full, the jumping times of the file read pointer at different positions of the file are reduced as much as possible, the more the jumping times are, the slower the data are downloaded and read, the playing performance is affected, the jumping times of the read pointer are greatly dependent on the interleaving condition of the file, and the method for reading the audio and video data frames needs to be correspondingly adjusted according to the interleaving condition of different files. However, the existing interleaving judgment algorithm cannot dynamically adjust the interleaving judgment result according to the buffer size change of the playing device and the change of the playing file code rate, and the existing interleaving judgment algorithm does not respectively judge the interleaving conditions at different file positions, so that the audio and video reading methods at different file positions are updated.
Disclosure of Invention
Embodiments of the present disclosure provide an information generating method, apparatus, electronic device, and computer readable medium.
In a first aspect, embodiments of the present disclosure provide an information generating method, including: determining current sub-data from the audio-video data set based on the arrangement sequence of at least one sub-data in the audio-video data set corresponding to the target audio-video file, wherein the audio-video data set comprises at least one sub-data which is arranged in sequence after the target audio-video file is split; based on the target audio and video file, acquiring video playing time length and audio playing time length corresponding to the current sub-data, and calculating a time length difference value between the video playing time length and the audio playing time length corresponding to the current sub-data; obtaining the maximum playing time length of a buffer area corresponding to current playing equipment; generating interleaving result information corresponding to the current sub-data based on the time length difference value and the maximum playing time length of the buffer area, and continuously determining new current sub-data from the audio-video data set based on the arrangement sequence of at least one sub-data in the audio-video data set.
In some embodiments, generating interleaving result information corresponding to current sub-data based on the time length difference value and the maximum playing time length of the buffer area includes: judging whether the time length difference is larger than the maximum play time length of the buffer area; and generating first interleaving result information representing the current sub-data interleaving condition in response to the fact that the time length difference value is larger than or equal to the maximum playing time length of the buffer area.
In some embodiments, based on the time length difference value and the maximum playing time length of the buffer area, generating interleaving result information corresponding to the current sub-data, and further includes: responding to the fact that the time length difference value is smaller than the maximum playing time length of the buffer area, and obtaining the number of bytes of the target video frame corresponding to the current sub-data; acquiring the total byte number of a video buffer area of current playing equipment; judging whether the byte number of the target video frame is larger than the total byte number of the video buffer area; and generating first interleaving result information representing the current sub-data interleaving condition in response to determining that the target video frame byte number is greater than the total byte number of the video buffer.
In some embodiments, based on the time length difference value and the maximum playing time length of the buffer area, generating interleaving result information corresponding to the current sub-data, and further includes: and generating second interleaving result information representing the current sub-data interleaving condition in response to determining that the number of bytes of the target video frame is less than or equal to the total number of bytes of the video buffer.
In some embodiments, the buffer maximum play time length includes a video maximum play time length of the video buffer; and judging whether the time length difference value is larger than the maximum playing time length of the buffer area, comprising the following steps: judging whether the time length difference is a positive value or not; and in response to determining that the time length difference is a positive value, judging whether the time length difference is larger than the maximum playing time length of the video in the video cache region.
In some embodiments, the buffer maximum play time length includes an audio maximum play time length of the audio buffer; and judging whether the time length difference value is larger than the maximum playing time length of the buffer area, comprising the following steps: and in response to determining that the time length difference value is a negative value, judging whether the absolute value of the time length difference value is larger than the audio maximum playing time length of the audio buffer area.
In some embodiments, the audio-video data set is obtained based on the steps of: acquiring target attribute information corresponding to a target audio/video file and current playing equipment corresponding to the target audio/video file; splitting the target audio and video file based on the target attribute information and the attribute information of the current playing device to obtain at least one piece of sub data corresponding to the target audio and video file; determining the arrangement sequence of at least one piece of sub data based on the target audio/video file; and sequentially arranging the at least one piece of sub data based on the arrangement sequence of the at least one piece of sub data to generate an audio/video data set.
In a second aspect, embodiments of the present disclosure provide an information generating apparatus including: the determining module is configured to determine current sub-data from the audio-video data set based on the arrangement sequence of at least one sub-data in the audio-video data set corresponding to the target audio-video file, wherein the audio-video data set comprises at least one sub-data which is arranged in sequence after the target audio-video file is split; the computing module is configured to acquire video playing time length and audio playing time length corresponding to the current sub-data based on the target audio and video file, and calculate a time length difference value between the video playing time length and the audio playing time length corresponding to the current sub-data; the acquisition module is configured to acquire the maximum playing time length of the buffer area corresponding to the current playing device; the generating module is configured to generate interleaving result information corresponding to the current sub-data based on the time length difference value and the maximum playing time length of the buffer area, and continuously determine new current sub-data from the audio-video data set based on the arrangement sequence of at least one sub-data in the audio-video data set.
In some embodiments, the generating module comprises: the judging unit is configured to judge whether the time length difference value is larger than the maximum play time length of the buffer area; the generating unit is configured to generate first interleaving result information representing the current sub-data interleaving condition in response to the fact that the time length difference value is larger than or equal to the maximum playing time length of the buffer area.
In some embodiments, the generating module further comprises: an acquisition unit; the obtaining unit is configured to obtain the number of bytes of the target video frame corresponding to the current sub-data in response to the fact that the time length difference value is smaller than the maximum playing time length of the buffer; acquiring the total byte number of a video buffer area of current playing equipment; the judging unit is further configured to judge whether the byte number of the target video frame is larger than the total byte number of the video buffer area; the generating unit is further configured to generate first interleaving result information representing the current sub-data interleaving condition in response to determining that the target video frame byte number is greater than the total byte number of the video buffer.
In some embodiments, the generating unit is further configured to: and generating second interleaving result information representing the current sub-data interleaving condition in response to determining that the number of bytes of the target video frame is less than or equal to the total number of bytes of the video buffer.
In some embodiments, the buffer maximum play time length includes a video maximum play time length of the video buffer; and a judgment unit configured to: judging whether the time length difference is a positive value or not; and in response to determining that the time length difference is a positive value, judging whether the time length difference is larger than the maximum playing time length of the video in the video cache region.
In some embodiments, the buffer maximum play time length includes a video maximum play time length of the video buffer; and a judgment unit configured to: and in response to determining that the time length difference value is a negative value, judging whether the absolute value of the time length difference value is larger than the audio maximum playing time length of the audio buffer area.
In some embodiments, the audio-video data set is obtained based on the steps of: acquiring target attribute information corresponding to a target audio/video file and current playing equipment corresponding to the target audio/video file; splitting the target audio and video file based on the target attribute information and the attribute information of the current playing device to obtain at least one piece of sub data corresponding to the target audio and video file; determining the arrangement sequence of at least one piece of sub data based on the target audio/video file; and sequentially arranging the at least one piece of sub data based on the arrangement sequence of the at least one piece of sub data to generate an audio/video data set.
In a third aspect, the present application provides an electronic device comprising one or more processors; and a storage device having one or more programs stored thereon, which when executed by the one or more processors, cause the one or more processors to implement the information generation method as described in any of the implementations of the first aspect.
In a fourth aspect, the present application provides a computer-readable medium having stored thereon a computer program which, when executed by a processor, implements an information generating method as described in any of the implementations of the first aspect.
According to the information generation method and device provided by the embodiment of the disclosure, through determining the current sub-data from the audio-video data set based on the arrangement sequence of at least one sub-data in the audio-video data set corresponding to the target audio-video file, wherein the audio-video data set comprises at least one sub-data which is arranged in sequence after the target audio-video file is split, then based on the target audio-video file, the video playing time length and the audio playing time length corresponding to the current sub-data are acquired, the time length difference between the video playing time length and the audio playing time length corresponding to the current sub-data is calculated, then the buffer zone maximum playing time length corresponding to the current playing equipment is acquired, finally, based on the time length difference and the buffer zone maximum playing time length, interleaving result information corresponding to the current sub-data is generated, and new current sub-data is determined from the audio-video data set based on the arrangement sequence of at least one sub-data, audio-video interleaving judgment can be carried out on each sub-data, so that the interleaving situation of different positions of the same target audio-video file can be respectively judged, the whole target audio-video file can be more accurately judged, and the interleaving result information of the current audio-video file can be more accurately read according to the interleaving result of the interleaving devices can be more dynamically adjusted when the interleaving result information of the current playing equipment is more than the interleaving result information in time of the playing equipment is more than the interleaving device, and the interleaving result information of the interleaving device can be more dynamically adjusted for each time, thereby improving the playing performance of the player.
Drawings
Other features, objects and advantages of the present disclosure will become more apparent upon reading of the detailed description of non-limiting embodiments, made with reference to the following drawings:
FIG. 1 illustrates an exemplary system architecture diagram in which some embodiments of the present disclosure may be applied;
FIG. 2 illustrates a flow chart of one embodiment of an information generation method of the present disclosure;
FIG. 3 shows a schematic diagram of one application scenario of the information generation method of the present disclosure;
FIG. 4 shows a schematic diagram of another embodiment of the information generation method of the present disclosure;
FIG. 5 shows a schematic diagram of yet another embodiment of the information generation method of the present disclosure;
FIG. 6 is a schematic diagram showing the structure of an embodiment of an information generating apparatus of the present disclosure;
fig. 7 shows a schematic structural diagram of an electronic device suitable for use in implementing embodiments of the present disclosure.
Detailed Description
The present disclosure is described in further detail below with reference to the drawings and examples. It is to be understood that the specific embodiments described herein are merely illustrative of the invention and are not limiting of the invention. It should be noted that, for convenience of description, only the portions related to the present invention are shown in the drawings.
It should be noted that, without conflict, the embodiments of the present disclosure and features of the embodiments may be combined with each other. The present disclosure will be described in detail below with reference to the accompanying drawings in conjunction with embodiments.
Fig. 1 illustrates an exemplary system architecture 100 to which an information generating method or information generating apparatus of embodiments of the present disclosure may be applied.
As shown in fig. 1, the system architecture 100 may include servers 101, 102, a network 103, and playback devices 104, 105. The network 103 is used as a medium to provide a communication link between the servers 101, 102 and the playback devices 104, 105. The network 103 may include various connection types, such as wired, wireless communication links, or fiber optic cables, among others.
A user may interact with the servers 101, 102 via the playback devices 104, 105 using the network 103 to receive or transmit information, etc. Various applications, such as video playback applications, data processing applications, etc., may be installed on the playback devices 104, 105.
The playback devices 104, 105 may be hardware or software. When the playback device is hardware, it may be a variety of electronic devices that have a display screen and support communication with a server, including but not limited to smartphones, tablets, laptop and desktop computers, and the like. When the terminal device is software, it can be installed in the above-listed electronic device. Which may be implemented as a plurality of software or software modules, or as a single software or software module. The present invention is not particularly limited herein.
The servers 101 and 102 may receive the acquisition request of the target audio/video file sent by the playing devices 104 and 105, and acquire the target audio/video file according to the acquisition request. The servers 101 and 102 may then obtain an audio/video data set corresponding to the target audio/video file according to the target audio/video file, where the audio/video data set may include at least one sub-data that is sequentially arranged after splitting the target audio/video file, and determine, according to an arrangement sequence of at least one sub-data in the audio/video data set, current sub-data from the audio/video data set. Then, the servers 101 and 102 may obtain the video playing time length and the audio playing time length corresponding to the current sub-data based on the target audio and video file, calculate the time length difference between the video playing time length and the audio playing time length corresponding to the current sub-data, and obtain the maximum playing time length of the buffer corresponding to the current playing device (i.e. the playing devices 104 and 105). Finally, the servers 101 and 102 can generate interleaving result information corresponding to the current sub-data based on the time length difference value and the maximum playing time length of the buffer area, and continuously determine new current sub-data from the audio and video data set based on the arrangement sequence of at least one sub-data in the audio and video data set until interleaving result information corresponding to the last sub-data is generated.
The server may be hardware or software. When the server is hardware, it may be various electronic devices that provide various services to the terminal device. When the server is software, a plurality of software or software modules providing various services to the terminal device may be realized, or a single software or software module providing various services to the terminal device may be realized. The present invention is not particularly limited herein.
It should be noted that the information generating method provided by the embodiments of the present disclosure may be executed by the servers 101, 102. Accordingly, the information generating device is provided in the servers 101, 102.
It should be understood that the number of servers, networks, and playback devices in fig. 1 is merely illustrative. There may be any number of playback devices, networks, and servers, as desired for implementation.
With continued reference to fig. 2, a flow 200 of one embodiment of an information generation method according to the present disclosure is shown. The information generating method may include the steps of:
step 210, determining current sub-data from the audio-video data set based on the arrangement sequence of at least one sub-data in the audio-video data set corresponding to the target audio-video file.
In this step, the execution body of the information generating method (for example, the servers 101 and 102 shown in fig. 1) may receive an acquisition request of a target audio/video file sent by the current playing device, where the acquisition request may be request information generated by the current playing device based on a playing request input by a user, where the acquisition request includes a target audio/video file to be acquired and attribute information of the current playing device, where the attribute information of the current playing device may include device information that characterizes playing performance affecting in the current playing device, and may include a total number of bytes of an audio buffer area and a total number of bytes of a video buffer area of the current playing device. The executing body can acquire the corresponding target audio/video file according to the current playing device acquisition request, split the target audio/video file into at least one piece of sub data, namely the target audio/video file can be directly taken as one piece of sub data without splitting the target audio/video file, the target audio/video file can be split into a plurality of pieces of sub data, the plurality of pieces of sub data are sequentially arranged according to the positions in the target audio/video file, and therefore an audio/video data set corresponding to the target audio/video file is acquired, and the audio/video data set comprises at least one piece of sub data which is sequentially arranged after the target audio/video file is split.
After the executing body obtains the audio/video data set corresponding to the target audio/video file, the executing body can determine the current sub-data from the audio/video data set according to the arrangement sequence of at least one sub-data in the audio/video data set. If the audio-video data set only comprises one piece of sub data, the current sub data can be the only sub data in the audio-video data set; if the audio-video data set includes a plurality of sub-data, the current sub-data may be the first sub-data in the audio-video data set, or may be the next sub-data adjacent to the sub-data that has generated the interleaving result information last time.
In some optional implementations of this embodiment, the audio-video data set in step 210 is obtained based on the following steps:
the method comprises the first step of obtaining target attribute information corresponding to a target audio and video file and current playing equipment corresponding to the target audio and video file.
Specifically, the executing body may receive an acquisition request of a target audio/video file sent by the current playing device, where the acquisition request may be request information generated by the current playing device based on a playing request input by a user, where the acquisition request includes attribute information of the target audio/video file and the current playing device that need to be acquired, where the attribute information of the current playing device may include device information that characterizes playing performance in the current playing device, and may include a total byte number of an audio buffer area and a total byte number of a video buffer area of the current playing device.
The executing body can acquire the corresponding target audio/video file according to the acquisition request of the current playing device, and perform attribute analysis on the target audio/video file to acquire target attribute information corresponding to the target audio/video file, where the target attribute information can include the total number of bytes of the file corresponding to the target audio/video file and the total playing duration.
The executing body can further determine the total playing duration of the current playing device according to the target attribute information corresponding to the target audio/video file, so that the attribute information of the current playing device includes the total byte number of the audio buffer, the total byte number of the video buffer, the total playing duration and the like.
And secondly, splitting the target audio and video file based on the target attribute information and the attribute information of the current playing device to obtain at least one piece of sub data corresponding to the target audio and video file.
Specifically, the executing body may compare the target attribute information corresponding to the target audio/video file with the attribute information of the current playing device, and determine the size. If the total byte number of the audio buffer area and the total byte number of the video buffer area are both larger than the total byte number of the file of the target audio and video file and the total playing time length of the buffer area is longer than the total playing time length of the target audio and video file, the target audio and video file is not split, and the target audio and video file is taken as a whole. Otherwise, splitting the target audio and video file to obtain at least one piece of sub data corresponding to the target audio and video file.
And thirdly, determining the arrangement sequence of at least one piece of sub data based on the target audio/video file.
Specifically, after the execution body may obtain at least one sub-data corresponding to the target audio/video file, the execution body may determine the position of each sub-data according to the target audio/video file, so as to determine the arrangement sequence of at least one sub-data according to the position of each sub-data.
And fourthly, sequentially arranging at least one piece of sub data based on the arrangement sequence of the at least one piece of sub data to generate an audio and video data set.
Specifically, the execution body may sequentially arrange at least one sub-data according to an arrangement order of the at least one sub-data, and generate an audio/video data set including the sequentially arranged at least one sub-data.
In the implementation manner, the target audio/video file is split according to the attribute information based on the target attribute information and the attribute information of the current playing device, so that each part of sub-data can be subjected to interleaving condition judgment from front to back, and therefore each part of sub-data is subjected to audio/video interleaving judgment, and the playing device can timely and dynamically adjust the reading method of the audio/video according to each interleaving judgment result.
Step 220, based on the target audio/video file, obtaining the video playing time length and the audio playing time length corresponding to the current sub-data, and calculating the time length difference between the video playing time length and the audio playing time length corresponding to the current sub-data.
In this step, after the executing body determines the current sub-data, the current position of the current sub-data may be determined according to the target audio/video file, and all video frames and all audio frames from the first portion of sub-data to the current position of the current sub-data may be determined according to the target audio/video file. And then the execution main body can respectively calculate the video playing time length and the audio playing time length corresponding to the current position of the current sub-data according to all the video frames and all the audio frames, namely the video playing time length can represent the total playing time length of all the video frames corresponding to the current position, and the audio playing time length can represent the total playing time length of all the audio frames corresponding to the current position. And then the execution main body can compare the video playing time length and the audio playing time length corresponding to the current sub-data to determine the time length difference between the video playing time length and the audio playing time length corresponding to the current sub-data.
Step 230, obtaining the maximum playing duration of the buffer corresponding to the current playing device.
In this step, the executing body obtains the total number of bytes of the audio buffer area and the total number of bytes of the video buffer area of the current playing device, so as to obtain the total number of bytes of the buffer area, and then divides the total number of bytes of the buffer area by the average code rate of the target audio/video file to obtain the maximum playing duration of the buffer area corresponding to the current playing device, where the average code rate of the target audio/video file can be obtained by dividing the sum of the byte sizes of all video frames or audio frames by the sum of the playing durations of all video frames or audio frames.
As an example, the executing body may obtain the maximum playing duration of the video buffer by dividing the total number of bytes of the video buffer by the average code rate of the video, and the average code rate of the video may be obtained by dividing the sum of the byte sizes of all video frames by the sum of the playing durations of all video frames.
As an example, the executing body may obtain the maximum playing duration of the audio buffer by dividing the total number of bytes of the audio buffer by the average code rate of the audio, and the average code rate of the audio may be obtained by dividing the sum of the byte sizes of all the audio frames by the sum of the playing durations of all the audio frames.
Step 240, generating interleaving result information corresponding to the current sub-data based on the time length difference value and the maximum playing time length of the buffer area, and continuously determining new current sub-data from the audio/video data set based on the arrangement sequence of at least one sub-data in the audio/video data set.
In this step, after the executing body obtains the time length difference between the video playing time length and the audio playing time length corresponding to the current sub-data and the maximum playing time length of the buffer, the time length difference may be compared with the maximum playing time length of the buffer to obtain a comparison result of the time length difference and the maximum playing time length of the buffer, and then interleaving result information corresponding to the current sub-data is generated according to the comparison result, where the interleaving result information represents an audio/video interleaving condition of the current sub-data and may include two interleaving result information of good interleaving condition and poor interleaving condition.
Specifically, if the comparison result indicates that the time length difference is greater than the maximum play time length of the buffer area, generating interleaving result information representing that the interleaving condition of the current sub-data is poor; if the comparison result indicates that the time length difference is smaller than or equal to the maximum playing time length of the buffer area, generating interleaving result information which characterizes that the interleaving condition of the current sub-data is good.
With continued reference to fig. 3, fig. 3 is a schematic diagram of an application scenario of the information generating method according to the present embodiment. The method may be applied to the application scenario of fig. 3, after the server 301 obtains the audio and video data set corresponding to the target audio and video file, the current sub-data may be determined from the audio and video data set based on the arrangement sequence of at least one sub-data in the audio and video data set corresponding to the target audio and video file, and the ith sub-data is determined to be the current sub-data. The server 301 may then obtain the timestamp T of the first video packet v0 = first video frame pts, and obtain the timestamp T of the second video packet v1 = second video frame pts, according to the two above-mentioned time stampsCalculating time duration Inter of video frame v =abs(T v0 -T v1 ). The server 301 may then obtain the timestamp T of the first audio packet a0 = first audio frame pts, and obtain the timestamp T of the second audio packet a1 = second audio frame pts calculating the time duration Inter of an audio frame from the two time stamps a =abs(T a0 -T a1 )。
And, the server 301 determines the number of video frames from the first sub-data to the current sub-data according to the audio/video data set And server 301 may also determine the number of audio frames from the first sub-data to the current sub-data >
The server 301 may then determine the time duration of the video frame based on the time duration of the video frame v And the number of video frames C vi Calculating the video playing time length corresponding to the current sub-data, and according to the time duration Inter of the audio frame a And the number of video frames C ai Calculating the audio playing time length corresponding to the current sub-data, and calculating the time length difference between the video playing time length and the audio playing time length, and Diff D i =C vi *Inter v -C ai *Inter a
The server 301 may further obtain attribute information of the current playing device 302, and obtain a maximum playing time length of a buffer area corresponding to the current playing device by dividing the total byte number of the buffer area by the average code rate of the target audio/video file, where the server 301 calculates an average code of video frames according to the sum of byte sizes of all video frames divided by the sum of playing time lengths of all video framesRate B avg_v = video content avg bitrate, the average bit rate B of the audio frames can also be calculated from the sum of the byte sizes of all audio frames divided by the sum of the playing durations of all audio frames avg_a =audio content avg bitrate。
The server 301 may further generate interleaving result information corresponding to the current sub-data based on the time length difference value and the maximum playing time length of the buffer area, and may send the interleaving result information of the current sub-data to the current playing device 302 together, so that the current playing device 302 dynamically updates the audio/video reading method according to the interleaving condition of the current sub-data, and then continues to determine new current sub-data from the audio/video data set based on the arrangement sequence of at least one sub-data in the audio/video data set, determine the (i+1) th sub-data as new current sub-data, and generate new interleaving result information for the new current sub-data until interleaving result information of the last sub-data is generated.
According to the information generation method and device provided by the embodiment of the disclosure, through determining the current sub-data from the audio-video data set based on the arrangement sequence of at least one sub-data in the audio-video data set corresponding to the target audio-video file, wherein the audio-video data set comprises at least one sub-data which is arranged in sequence after the target audio-video file is split, then based on the target audio-video file, acquiring the video playing time length and the audio playing time length corresponding to the current sub-data, calculating the time length difference between the video playing time length and the audio playing time length corresponding to the current sub-data, then acquiring the maximum playing time length of a buffer area corresponding to the current playing equipment, finally generating interleaving result information corresponding to the current sub-data based on the time length difference and the maximum playing time length of the buffer area, and continuing to determine new current sub-data from the audio-video data set based on the arrangement sequence of at least one sub-data in the audio-video data set, and performing audio-video interleaving judgment on each sub-data to realize respective judgment on the same target audio-video file, so that the whole target audio-video file is more accurately judged, and the interleaving result information of the current audio-video file is more accurately read according to the interleaving result of the current playing equipment and the interleaving result of the audio-video equipment is more than the interleaving result of the current playing equipment is more, and the interleaving result information of the interleaving device is more accurately can be adjusted for each time of the interleaving result of the playing device is more than the interleaving result of the playing device, the method can be applied to the conditions that the same playing file is played on different playing devices, the same playing device plays multimedia files with different code rates, different application programs play the same multimedia files on the same playing device, the application programs set different buffer zone byte sizes according to actual requirements, and the like.
Referring to fig. 4, fig. 4 shows a flow 400 of yet another embodiment of an information generation method. The information generating method may include the steps of:
in step 410, in response to obtaining the audio/video data set corresponding to the target audio/video file, current sub-data is determined from the audio/video data set based on the arrangement order of at least one sub-data in the audio/video data set.
In this step, step 410 is the same as step 210 in the embodiment shown in fig. 2, and is not described here.
Step 420, based on the target audio/video file, obtaining the video playing time length and the audio playing time length corresponding to the current sub-data, and calculating the time length difference between the video playing time length and the audio playing time length corresponding to the current sub-data.
In this step, step 420 is the same as step 220 in the embodiment shown in fig. 2, and is not repeated here.
Step 430, obtaining the maximum playing duration of the buffer corresponding to the current playing device.
In this step, step 430 is the same as step 230 in the embodiment shown in fig. 2, and is not repeated here.
Step 440, determining whether the time difference is greater than the maximum playing time of the buffer.
In this step, after the executing body obtains the time length difference between the video playing time length and the audio playing time length corresponding to the current sub-data and the maximum playing time length of the buffer area, the executing body may compare the time length difference with the maximum playing time length of the buffer area, and determine whether the time length difference is greater than the maximum playing time length of the buffer area.
In response to determining that the difference in time length is greater than or equal to the maximum playing time length of the buffer, step 450 is executed, and in response to determining that the difference in time length is greater than or equal to the maximum playing time length of the buffer, first interleaving result information representing the current sub-data interleaving condition is generated.
In this step, the executing body determines that the interleaving condition of the current sub-data is poor by comparing and determining that the difference of the time lengths is greater than or equal to the maximum playing time length of the buffer area, and generates first interleaving result information representing the interleaving condition of the current sub-data, where the first interleaving result information represents the interleaving condition difference of the current sub-data.
In this embodiment, by using the comparison result of the time length difference and the maximum playing time length of the buffer area, if it is determined that the time length difference is greater than or equal to the maximum playing time length of the buffer area, first interleaving result information representing the interleaving condition of the current sub-data is generated, and the interleaving result information of the audio and video can be dynamically adjusted continuously according to the maximum playing time length of the buffer area of the current playing device, when the same target audio and video file is specific to different playing devices, the interleaving result information has different results, so that the interleaving result information is more accurate and more attached to the playing devices.
With further reference to fig. 4, the information generating method may further include the following steps:
in response to determining that the difference in time length is less than the maximum playing time length of the buffer area, step 460 is executed, and in response to determining that the difference in time length is less than the maximum playing time length of the buffer area, the number of bytes of the target video frame corresponding to the current sub-data is obtained.
In this step, the executing body further determines, by comparing the time length difference value with the maximum playing time length of the buffer area, a continuous video frame closest to the current sub-data from the audio/video data set, where the continuous video frame refers to a section of continuous multiple video frames closest to the current sub-data. And then the execution main body can determine the byte number of the target video frame corresponding to the target video frame according to the acquired target video frame.
Specifically, if the current sub-data is the ith sub-data, the executing body searches for the maximum byte offset position Oend of the consecutive video frames from the video frame closest to the current sub-data i The following are provided:
and, the executing body may look for the minimum byte offset position Ostarti of the consecutive video frames from the video frame closest to the current sub-data, as follows:
The execution body may then shift the maximum byte offset from the end i The difference from the minimum byte offset position Ostarti is determined as the target video frame byte number.
Step 470, obtaining the total byte number of the video buffer of the current playing device.
In this step, the executing body may acquire attribute information of the current playing device, and acquire a total byte number of the video buffer of the current playing device from the attribute information.
In step 480, it is determined whether the number of bytes of the target video frame is greater than the total number of bytes of the video buffer.
In this step, the executing body may compare the number of bytes of the target video frame with the total number of bytes of the video buffer, and determine whether the number of bytes of the target video frame is greater than the total number of bytes of the video buffer.
In response to determining that the number of bytes of the target video frame is greater than the total number of bytes of the video buffer, step 490 is performed to generate first interleaving result information indicative of a current sub-data interleaving condition in response to determining that the number of bytes of the target video frame is greater than the total number of bytes of the video buffer.
In this step, the executing body determines, by comparing, that the number of bytes of the target video frame is greater than the total number of bytes of the video buffer, that the interleaving condition of the current sub-data is poor, and generates first interleaving result information indicating the interleaving condition of the current sub-data, where the first interleaving result information indicates the interleaving condition of the current sub-data.
In response to determining that the number of bytes of the target video frame is greater than the total number of bytes of the video buffer, step 4100 is executed, and in response to determining that the number of bytes of the target video frame is less than or equal to the total number of bytes of the video buffer, second interleaving result information that characterizes a current sub-data interleaving condition is generated.
In this step, the executing body determines that the number of bytes of the target video frame is less than or equal to the total number of bytes of the video buffer area through comparison, and then determines that the interleaving condition of the current sub-data is good, and generates second interleaving result information representing the interleaving condition of the current sub-data, where the second interleaving result information represents that the interleaving condition of the current sub-data is good.
In this implementation manner, when the time length difference is smaller than the maximum playing time length of the buffer area, further, comparing the target video frame byte number with the total byte number of the video buffer area, generating interleaving result information based on the comparison result, so that the interleaving result information is more accurate, performing audio/video interleaving judgment on each piece of sub data according to the maximum playing time length of the buffer area of the current playing device, and performing judgment according to the target video frame byte number and the total byte number of the video buffer area, and continuously and dynamically adjusting the interleaving result information of the audio/video.
With further reference to fig. 5, fig. 5 shows a schematic diagram 500 of yet another embodiment of an information generation method, which may include the steps of:
step 510, in response to obtaining the audio/video data set corresponding to the target audio/video file, determining the current sub-data from the audio/video data set based on the arrangement sequence of at least one sub-data in the audio/video data set.
In this step, step 510 is the same as step 410 in the embodiment shown in fig. 4, and is not repeated here.
Step 520, based on the target audio/video file, obtaining the video playing time length and the audio playing time length corresponding to the current sub-data, and calculating the time length difference between the video playing time length and the audio playing time length corresponding to the current sub-data.
In this step, step 520 is the same as step 420 in the embodiment shown in fig. 4, and is not repeated here.
Step 530, obtaining the video maximum playing time length of the video buffer area and the audio maximum playing time length of the audio buffer area corresponding to the current playing device.
In this step, step 530 is the same as step 430 in the embodiment shown in fig. 4, and is not repeated here.
Step 540, determine whether the duration difference is positive.
In this step, after the executing body obtains the time length difference between the video playing time length and the audio playing time length corresponding to the current sub-data, it is determined whether the time length difference is a positive value.
In response to determining that the difference in duration is positive, step 550 is performed to determine whether the difference in duration is greater than the maximum play duration of the video in the video buffer in response to determining that the difference in duration is positive.
The maximum playing time of the buffer area may include a video maximum playing time of the video buffer area.
In this step, the executing body determines that the difference between the time lengths is a positive value, and compares the difference between the time lengths with the video maximum playing time length of the video buffer, so as to determine whether the difference between the time lengths is greater than the video maximum playing time length of the video buffer.
In response to determining that the difference in time length is greater than or equal to the maximum playing time length of the buffer, step 570 is executed, and in response to determining that the difference in time length is greater than or equal to the maximum playing time length of the buffer, first interleaving result information representing the current sub-data interleaving condition is generated.
In response to determining that the time length difference is less than the buffer maximum play time length, steps 580-5120 are performed.
And the maximum playing time length of the buffer area may include the maximum playing time length of the video of the audio buffer area.
In response to determining that the duration difference is negative, step 560 is performed to determine whether the absolute value of the duration difference is greater than the audio maximum play duration of the audio buffer in response to determining that the duration difference is negative.
In this step, the executing body determines that the difference value of the time length is a negative value, and compares the absolute value of the difference value of the time length with the audio maximum playing time length of the audio buffer, so as to determine whether the absolute value of the difference value of the time length is greater than the audio maximum playing time length of the audio buffer.
In response to determining that the difference in time length is greater than or equal to the maximum playing time length of the buffer, step 570 is executed, and in response to determining that the difference in time length is greater than or equal to the maximum playing time length of the buffer, first interleaving result information representing the current sub-data interleaving condition is generated.
In response to determining that the time length difference is less than the buffer maximum play time length, steps 580-5120 are performed.
In this embodiment, by judging whether the difference between the durations is a positive value, and comparing the difference with the video maximum playing duration of the video buffer area or the audio maximum playing duration of the audio buffer area according to different situations, pertinence under different situations is improved, and interleaving result information is more accurate.
With further reference to fig. 6, as an implementation of the method shown in the foregoing figures, the present application provides an embodiment of an information generating apparatus, where an embodiment of the apparatus corresponds to the embodiment of the method shown in fig. 2, and the apparatus may be specifically applied to various electronic devices.
As shown in fig. 6, the information generating apparatus 600 of the present embodiment includes: a determination module 610, a calculation module 620, an acquisition module 630, and a generation module 640.
The determining module 610 is configured to determine current sub-data from the audio-video data set based on an arrangement sequence of at least one sub-data in the audio-video data set corresponding to the target audio-video file, where the audio-video data set includes at least one sub-data in which the target audio-video file is arranged in sequence after being split;
the calculating module 620 is configured to obtain a video playing time length and an audio playing time length corresponding to the current sub-data based on the target audio and video file, and calculate a time length difference between the video playing time length and the audio playing time length corresponding to the current sub-data;
an obtaining module 630, configured to obtain a maximum playing duration of the buffer area corresponding to the current playing device;
the generating module 640 is configured to generate interleaving result information corresponding to the current sub-data based on the time length difference value and the maximum playing time length of the buffer area, and continuously determine new current sub-data from the audio-video data set based on the arrangement sequence of at least one sub-data in the audio-video data set.
In some optional implementations of the present embodiment, the generating module 640 includes: the judging unit is configured to judge whether the time length difference value is larger than the maximum play time length of the buffer area; the generating unit is configured to generate first interleaving result information representing the current sub-data interleaving condition in response to the fact that the time length difference value is larger than or equal to the maximum playing time length of the buffer area.
In some optional implementations of this embodiment, the generating module 640 further includes: an acquisition unit; the obtaining unit is configured to obtain the number of bytes of the target video frame corresponding to the current sub-data in response to the fact that the time length difference value is smaller than the maximum playing time length of the buffer; acquiring the total byte number of a video buffer area of current playing equipment; the judging unit is further configured to judge whether the byte number of the target video frame is larger than the total byte number of the video buffer area; the generating unit is further configured to generate first interleaving result information representing the current sub-data interleaving condition in response to determining that the target video frame byte number is greater than the total byte number of the video buffer.
In some optional implementations of the present embodiment, the generating unit is further configured to: and generating second interleaving result information representing the current sub-data interleaving condition in response to determining that the number of bytes of the target video frame is less than or equal to the total number of bytes of the video buffer.
In some optional implementations of this embodiment, the maximum playing duration of the buffer area includes a video maximum playing duration of the video buffer area; and a judgment unit configured to: judging whether the time length difference is a positive value or not; and in response to determining that the time length difference is a positive value, judging whether the time length difference is larger than the maximum playing time length of the video in the video cache region.
In some optional implementations of this embodiment, the maximum playing duration of the buffer area includes a video maximum playing duration of the video buffer area; and a judgment unit configured to: and in response to determining that the time length difference value is a negative value, judging whether the absolute value of the time length difference value is larger than the audio maximum playing time length of the audio buffer area.
In some optional implementations of the present embodiment, the audio-video data set is acquired based on the steps of: acquiring target attribute information corresponding to a target audio/video file and current playing equipment corresponding to the target audio/video file; splitting the target audio and video file based on the target attribute information and the attribute information of the current playing device to obtain at least one piece of sub data corresponding to the target audio and video file; determining the arrangement sequence of at least one piece of sub data based on the target audio/video file; and sequentially arranging the at least one piece of sub data based on the arrangement sequence of the at least one piece of sub data to generate an audio/video data set.
According to the information generating device provided by the embodiment of the disclosure, the current sub-data is determined from the audio-video data set based on the arrangement sequence of at least one sub-data in the audio-video data set corresponding to the target audio-video file, wherein the audio-video data set comprises at least one sub-data which is arranged in sequence after the target audio-video file is split, then the video playing time length and the audio playing time length corresponding to the current sub-data are acquired based on the target audio-video file, the time length difference between the video playing time length and the audio playing time length corresponding to the current sub-data is calculated, the buffer zone maximum playing time length corresponding to the current playing equipment is acquired, finally the interleaving result information corresponding to the current sub-data is generated based on the time length difference and the buffer zone maximum playing time length, and new current sub-data is determined from the audio-video data set based on the arrangement sequence of at least one sub-data, the target audio-video file can be split, the audio-video interleaving judgment is carried out on each sub-data, the situation at different positions of the same target audio-video file is respectively judged, the result information of the whole target audio-video file is more accurately, the interleaving result information of the current audio-video file can be more accurately judged, the interleaving result information of the current audio-video data can be more accurately read according to the interleaving result information of the interleaving device in time-interleaving device can be more than the interleaving result information of the playing device in time when the interleaving device is more than the interleaving device is more dynamically, and the interleaving result information of the interleaving device can be more dynamically adjusted, and the interleaving result information of the interleaving device can be more dynamically adjusted, according to the interleaving result of the interleaving result information of the interleaving device can be more interleaving result of the interleaving device can be more corresponding to the audio-interleaving device can be more according to the interleaving result of the audio data can be more, thereby improving the playing performance of the player.
Those skilled in the art will appreciate that the above-described apparatus also includes some other well-known structures, such as a processor, memory, etc., which are not shown in fig. 6 in order to unnecessarily obscure embodiments of the present disclosure.
Referring now to fig. 7, a schematic diagram of an electronic device (e.g., server in fig. 1) 700 suitable for use in implementing embodiments of the present disclosure is shown.
As shown in fig. 7, the electronic device 700 may include a processing means (e.g., a central processor, a graphics processor, etc.) 701, which may perform various appropriate actions and processes according to a program stored in a Read Only Memory (ROM) 702 or a program loaded from a storage means 708 into a Random Access Memory (RAM) 703. In the RAM 703, various programs and data required for the operation of the electronic device 700 are also stored. The processing device 701, the ROM 702, and the RAM 703 are connected to each other through a bus 704. An input/output (I/O) interface 705 is also connected to bus 704.
In general, the following devices may be connected to the I/O interface 705: input devices 706 including, for example, a touch screen, touchpad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, and the like; an output device 707 including, for example, a Liquid Crystal Display (LCD), a speaker, a vibrator, and the like; storage 708 including, for example, magnetic tape, hard disk, etc.; and a communication device 709. The communication means 709 may allow the electronic device 700 to communicate wirelessly or by wire with other devices to exchange data. While fig. 7 shows an electronic device 700 having various means, it is to be understood that not all of the illustrated means are required to be implemented or provided. More or fewer devices may be implemented or provided instead. Each block shown in fig. 7 may represent one device or a plurality of devices as needed.
In particular, according to embodiments of the present disclosure, the processes described above with reference to flowcharts may be implemented as computer software programs. For example, embodiments of the present disclosure include a computer program product comprising a computer program embodied on a computer readable medium, the computer program comprising program code for performing the method shown in the flowcharts. In such an embodiment, the computer program may be downloaded and installed from a network via communication device 709, or installed from storage 708, or installed from ROM 702. The above-described functions defined in the methods of the embodiments of the present disclosure are performed when the computer program is executed by the processing device 701.
It should be noted that, the computer readable medium according to the embodiments of the present disclosure may be a computer readable signal medium or a computer readable storage medium, or any combination of the two. The computer readable storage medium can be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or a combination of any of the foregoing. More specific examples of the computer-readable storage medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In an embodiment of the present disclosure, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. Whereas in embodiments of the present disclosure, the computer-readable signal medium may comprise a data signal propagated in baseband or as part of a carrier wave, with computer-readable program code embodied therein. Such a propagated data signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination of the foregoing. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: electrical wires, fiber optic cables, RF (radio frequency), and the like, or any suitable combination of the foregoing.
The computer readable medium may be contained in the electronic device; or may exist alone without being incorporated into the electronic device. The computer readable medium carries one or more programs which, when executed by the electronic device, cause the electronic device to: determining current sub-data from the audio-video data set based on the arrangement sequence of at least one sub-data in the audio-video data set corresponding to the target audio-video file, wherein the audio-video data set comprises at least one sub-data which is arranged in sequence after the target audio-video file is split; based on the target audio and video file, acquiring video playing time length and audio playing time length corresponding to the current sub-data, and calculating a time length difference value between the video playing time length and the audio playing time length corresponding to the current sub-data; obtaining the maximum playing time length of a buffer area corresponding to current playing equipment; generating interleaving result information corresponding to the current sub-data based on the time length difference value and the maximum playing time length of the buffer area, and continuously determining new current sub-data from the audio-video data set based on the arrangement sequence of at least one sub-data in the audio-video data set.
Computer program code for carrying out operations for embodiments of the present disclosure may be written in one or more programming languages, including an object oriented programming language such as Java, smalltalk, C ++ and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the case of a remote computer, the remote computer may be connected to the user's computer through any kind of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or may be connected to an external computer (for example, through the Internet using an Internet service provider).
The flowcharts and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present disclosure. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The modules involved in the embodiments described in the present application may be implemented by software, or may be implemented by hardware. The described modules may also be provided in a processor, for example, as: a processor includes a determination module, a calculation module, an acquisition module, and a generation module. The names of these modules do not limit the module itself in some cases, for example, the determining module may also be described as "a module for determining the current sub-data from the audio-video data set based on the arrangement order of at least one sub-data in the audio-video data set corresponding to the target audio-video file".
The foregoing description is only of the preferred embodiments of the present disclosure and description of the principles of the technology being employed. It will be appreciated by those skilled in the art that the scope of the invention in the embodiments of the present disclosure is not limited to the specific combination of the above technical features, but encompasses other technical features formed by any combination of the above technical features or their equivalents without departing from the spirit of the invention. Such as the above-described features, are mutually substituted with (but not limited to) the features having similar functions disclosed in the embodiments of the present disclosure.

Claims (16)

1. An information generation method, the method comprising:
determining current sub-data from an audio-video data set based on the arrangement sequence of at least one sub-data in the audio-video data set corresponding to a target audio-video file, wherein the audio-video data set comprises at least one sub-data which is arranged in sequence after the target audio-video file is split;
based on the target audio and video file, acquiring video playing time length and audio playing time length corresponding to the current sub-data, and calculating a time length difference value between the video playing time length and the audio playing time length corresponding to the current sub-data;
obtaining the maximum playing time length of a buffer area corresponding to current playing equipment;
generating interleaving result information corresponding to the current sub-data based on the time length difference value and the maximum playing time length of the buffer area, and continuously determining new current sub-data from the audio-video data set based on the arrangement sequence of at least one sub-data in the audio-video data set.
2. The method of claim 1, wherein the generating interleaving result information corresponding to the current sub-data based on the time length difference value and the buffer maximum playing time length includes:
Judging whether the time length difference value is larger than the maximum play time length of the buffer area or not;
and generating first interleaving result information representing the current sub-data interleaving condition in response to determining that the time length difference value is greater than or equal to the maximum playing time length of the buffer area.
3. The method of claim 2, wherein the generating interleaving result information corresponding to the current sub-data based on the time length difference value and the buffer maximum playing time length further comprises:
responding to the fact that the time length difference value is smaller than the maximum playing time length of the buffer area, and obtaining the number of bytes of the target video frame corresponding to the current sub-data;
acquiring the total byte number of a video buffer area of the current playing device;
judging whether the byte number of the target video frame is larger than the total byte number of the video buffer area;
and generating first interleaving result information representing the current sub-data interleaving condition in response to determining that the target video frame byte number is greater than the total byte number of the video buffer.
4. The method of claim 3, wherein the generating interleaving result information corresponding to the current sub-data based on the time length difference value and the buffer maximum playing time length further comprises:
And generating second interleaving result information representing the current sub-data interleaving condition in response to determining that the target video frame byte number is less than or equal to the total byte number of the video buffer.
5. The method of claim 2, wherein the buffer maximum play time length comprises a video maximum play time length of a video buffer; and determining whether the time length difference is greater than the maximum playing time length of the buffer area, including:
judging whether the time length difference value is a positive value or not;
and in response to determining that the time length difference value is a positive value, judging whether the time length difference value is larger than the video maximum playing time length of the video cache region.
6. The method of claim 5, wherein the buffer maximum play-out duration comprises an audio maximum play-out duration of an audio buffer; and determining whether the time length difference is greater than the maximum playing time length of the buffer area, including:
and in response to determining that the time length difference value is a negative value, judging whether the absolute value of the time length difference value is larger than the audio maximum playing time length of the audio buffer area.
7. The method of claim 1, wherein the audio-video data set is obtained based on:
Acquiring target attribute information corresponding to the target audio and video file and current playing equipment corresponding to the target audio and video file;
splitting the target audio and video file based on the target attribute information and the attribute information of the current playing device to obtain at least one piece of sub data corresponding to the target audio and video file;
determining the arrangement sequence of the at least one piece of sub data based on the target audio/video file;
and sequentially arranging the at least one piece of sub data based on the arrangement sequence of the at least one piece of sub data to generate the audio and video data set.
8. An information generating apparatus, the apparatus comprising:
the determining module is configured to determine current sub-data from the audio-video data set based on the arrangement sequence of at least one sub-data in the audio-video data set corresponding to the target audio-video file, wherein the audio-video data set comprises at least one sub-data which is arranged in sequence after the target audio-video file is split;
the calculating module is configured to acquire the video playing time length and the audio playing time length corresponding to the current sub-data based on the target audio/video file, and calculate a time length difference value between the video playing time length and the audio playing time length corresponding to the current sub-data;
The acquisition module is configured to acquire the maximum playing time length of the buffer area corresponding to the current playing device;
the generating module is configured to generate interleaving result information corresponding to the current sub-data based on the time length difference value and the maximum playing time length of the buffer area, and continuously determine new current sub-data from the audio-video data set based on the arrangement sequence of at least one sub-data in the audio-video data set.
9. The apparatus of claim 8, wherein the generating module comprises:
the judging unit is configured to judge whether the time length difference value is larger than the maximum play time length of the buffer area;
the generating unit is configured to generate first interleaving result information representing the current sub-data interleaving condition in response to determining that the time length difference value is greater than or equal to the maximum playing time length of the buffer.
10. The apparatus of claim 9, wherein the generating module further comprises: an acquisition unit; the method comprises the steps of,
the obtaining unit is configured to obtain the number of bytes of the target video frame corresponding to the current sub-data in response to determining that the time length difference is smaller than the maximum playing time length of the buffer; acquiring the total byte number of a video buffer area of the current playing device;
The judging unit is further configured to judge whether the target video frame byte number is greater than the total byte number of the video buffer;
the generating unit is further configured to generate first interleaving result information representing the current sub-data interleaving condition in response to determining that the target video frame byte number is greater than the total byte number of the video buffer.
11. The apparatus of claim 10, wherein the generating unit is further configured to:
and generating second interleaving result information representing the current sub-data interleaving condition in response to determining that the target video frame byte number is less than or equal to the total byte number of the video buffer.
12. The apparatus of claim 9, wherein the buffer maximum play-out duration comprises a video maximum play-out duration of a video buffer; and the judging unit is further configured to:
judging whether the time length difference value is a positive value or not;
and in response to determining that the time length difference value is a positive value, judging whether the time length difference value is larger than the video maximum playing time length of the video cache region.
13. The apparatus of claim 12, wherein the buffer maximum playout duration comprises a video maximum playout duration of a video buffer; and the judging unit is further configured to:
And in response to determining that the time length difference value is a negative value, judging whether the absolute value of the time length difference value is larger than the audio maximum playing time length of the audio buffer area.
14. The apparatus of claim 8, wherein the audio-video data set is acquired based on:
acquiring target attribute information corresponding to the target audio and video file and current playing equipment corresponding to the target audio and video file;
splitting the target audio and video file based on the target attribute information and the attribute information of the current playing device to obtain at least one piece of sub data corresponding to the target audio and video file;
determining the arrangement sequence of the at least one piece of sub data based on the target audio/video file;
and sequentially arranging the at least one piece of sub data based on the arrangement sequence of the at least one piece of sub data to generate the audio and video data set.
15. An electronic device, comprising:
at least one processor; and
a memory communicatively coupled to the at least one processor; wherein,,
the memory stores instructions executable by the at least one processor to enable the at least one processor to perform the method of any one of claims 1-7.
16. A non-transitory computer readable storage medium storing computer instructions for causing the computer to perform the method of any one of claims 1-7.
CN202310667883.0A 2023-06-06 2023-06-06 Information generation method and device Pending CN116567288A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310667883.0A CN116567288A (en) 2023-06-06 2023-06-06 Information generation method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310667883.0A CN116567288A (en) 2023-06-06 2023-06-06 Information generation method and device

Publications (1)

Publication Number Publication Date
CN116567288A true CN116567288A (en) 2023-08-08

Family

ID=87503583

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310667883.0A Pending CN116567288A (en) 2023-06-06 2023-06-06 Information generation method and device

Country Status (1)

Country Link
CN (1) CN116567288A (en)

Similar Documents

Publication Publication Date Title
JP7214295B2 (en) Distributed system data synchronization method, apparatus, computer program and electronic equipment
US10659832B1 (en) Dynamic bitrate selection for streaming media
CN110213614B (en) Method and device for extracting key frame from video file
CN110581973B (en) Data playback method and device, terminal equipment and storage medium
CN111309962B (en) Method and device for extracting audio clips and electronic equipment
CN110418183B (en) Audio and video synchronization method and device, electronic equipment and readable medium
CN109889922B (en) Method, device, equipment and storage medium for forwarding streaming media data
CN110162394B (en) Media object playing method and device, electronic equipment and storage medium
CN114038465B (en) Voice processing method and device and electronic equipment
JP5170891B2 (en) Internet network device and data transmission method thereof
WO2023197811A1 (en) Video downloading method and apparatus, video transmission method and apparatus, terminal device, server and medium
US20150312369A1 (en) Checkpoints for media buffering
CN109951737B (en) Video processing method, video processing device, electronic equipment and computer-readable storage medium
CN116567288A (en) Information generation method and device
CN111290861A (en) Message processing method and device and electronic equipment
CN114584826B (en) Video dynamic downloading method and device, electronic equipment and storage medium
CN116033199A (en) Multi-device audio and video synchronization method and device, electronic device and storage medium
US20150271440A1 (en) Information processing apparatus, information processing method, program, and information processing system
US9806967B2 (en) Communication device and data processing method
CN112153322B (en) Data distribution method, device, equipment and storage medium
CN112015746B (en) Data real-time processing method, device, medium and electronic equipment
CN112671966B (en) Ear-return time delay detection device, method, electronic equipment and computer readable storage medium
CN115348460B (en) Video preloading method, device, equipment and storage medium
CN114928862B (en) System overhead reduction method and system based on task unloading and service caching
US11457287B2 (en) Method and system for processing video

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination