CN115567757A - Playing processing method, device, equipment and storage medium - Google Patents

Playing processing method, device, equipment and storage medium Download PDF

Info

Publication number
CN115567757A
CN115567757A CN202211152210.3A CN202211152210A CN115567757A CN 115567757 A CN115567757 A CN 115567757A CN 202211152210 A CN202211152210 A CN 202211152210A CN 115567757 A CN115567757 A CN 115567757A
Authority
CN
China
Prior art keywords
playing
frame
target
determining
time
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202211152210.3A
Other languages
Chinese (zh)
Inventor
常炎隆
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Baidu Netcom Science and Technology Co Ltd
Original Assignee
Beijing Baidu Netcom Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Baidu Netcom Science and Technology Co Ltd filed Critical Beijing Baidu Netcom Science and Technology Co Ltd
Priority to CN202211152210.3A priority Critical patent/CN115567757A/en
Publication of CN115567757A publication Critical patent/CN115567757A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47217End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for controlling playback functions for recorded or on-demand content, e.g. using progress bars, mode or play-point indicators or bookmarks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8547Content authoring involving timestamps for synchronizing content

Abstract

The disclosure provides a playing processing method, a playing processing device, playing processing equipment and a storage medium. The method relates to the field of artificial intelligence, in particular to cloud computing and network playing technologies, and can be applied to intelligent cloud scenes. The specific implementation scheme is as follows: in response to detecting that a preset trigger event occurs, determining a first target playing time stamp; determining a target frame tracking mode from the candidate frame tracking modes; determining a playing start frame based on the target frame tracking mode; the playing of the playing start frame is performed based on the first target playing time stamp. According to the scheme disclosed by the invention, the timeliness of playing can be improved.

Description

Playing processing method, device, equipment and storage medium
Technical Field
The disclosure relates to the field of artificial intelligence, in particular to cloud computing and network playing technologies, which can be applied to an intelligent cloud scene.
Background
With the rapid development of networks, the use of networks to play programs is rapidly emerging. In the process of starting or playing live broadcast, situations such as delay of playing pictures or pause of pictures are easy to occur. Therefore, the timeliness of playback still needs to be improved.
Disclosure of Invention
The disclosure provides a playing processing method, a playing processing device, playing processing equipment and a storage medium.
According to a first aspect of the present disclosure, there is provided a playback processing method including:
in response to detecting that a preset trigger event occurs, determining a first target playing timestamp;
determining a target frame tracking mode from the candidate frame tracking modes;
determining a playing start frame based on the target frame tracking mode;
and executing the playing of the playing start frame based on the first target playing time stamp.
According to a second aspect of the present disclosure, there is provided a playback processing apparatus including:
the first determining module is used for determining a first target playing time stamp in response to the detection of the occurrence of a preset triggering event;
the second determining module is used for determining a target frame chasing mode from the candidate frame chasing modes;
a third determining module, configured to determine a play start frame based on the target frame tracking mode;
and the playing control module is used for executing the playing of the playing start frame based on the first target playing time stamp.
According to a third aspect of the present disclosure, there is provided an electronic device comprising:
at least one processor; and
a memory communicatively coupled to the at least one processor; wherein, the first and the second end of the pipe are connected with each other,
the memory stores instructions executable by the at least one processor to enable the at least one processor to perform the method of the first aspect.
According to a fourth aspect of the present disclosure, there is provided a non-transitory computer readable storage medium having stored thereon computer instructions for causing a computer to perform the method provided by the first aspect described above.
According to a fifth aspect of the present disclosure, there is provided a computer program product comprising a computer program which, when executed by a processor, implements the method provided by the first aspect described above.
According to the technical scheme disclosed by the invention, the timeliness of playing can be improved.
The foregoing summary is provided for the purpose of description only and is not intended to be limiting in any way. In addition to the illustrative aspects, embodiments, and features described above, further aspects, embodiments, and features of the present application will be readily apparent by reference to the drawings and following detailed description.
Drawings
In the drawings, like reference numerals refer to the same or similar parts or elements throughout the several views unless otherwise specified. The figures are not necessarily to scale. It is appreciated that these drawings depict only some embodiments in accordance with the disclosure and are therefore not to be considered limiting of its scope.
Fig. 1 is a schematic flow chart diagram of a play processing method according to an embodiment of the present disclosure;
fig. 2 is a schematic diagram of a first scenario determining a first target play time stamp according to an embodiment of the present disclosure;
FIG. 3 is a schematic diagram of determining a first target play time stamp in a second scenario according to an embodiment of the present disclosure;
fig. 4 is a schematic diagram of determining a second target play time stamp for a first scene according to an embodiment of the present disclosure;
fig. 5 is a schematic diagram of determining a second target play time stamp in a second scenario according to an embodiment of the present disclosure;
fig. 6 is a schematic structural diagram of a playback processing apparatus according to an embodiment of the present disclosure;
FIG. 7 is a first scene schematic diagram of a playback process according to an embodiment of the disclosure;
FIG. 8 is a second scenario diagram of a play process according to an embodiment of the present disclosure;
fig. 9 is a block diagram of an electronic device for implementing a play processing method according to an embodiment of the present disclosure.
Detailed Description
Exemplary embodiments of the present disclosure are described below with reference to the accompanying drawings, in which various details of the embodiments of the disclosure are included to assist understanding, and which are to be considered as merely exemplary. Accordingly, it will be recognized by those of ordinary skill in the art that various changes and modifications may be made to the embodiments described herein without departing from the scope of the present disclosure. Also, descriptions of well-known functions and constructions are omitted in the following description for clarity and conciseness.
The terms "first," "second," and "third," etc. in the description embodiments and claims of the present disclosure and the above-described drawings are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. Furthermore, the terms "comprises" and "comprising," as well as any variations thereof, are intended to cover a non-exclusive inclusion, such as a list of steps or elements. A method, system, article, or apparatus is not necessarily limited to those steps or elements explicitly listed, but may include other steps or elements not explicitly listed or inherent to such process, system, article, or apparatus.
Before the technical solutions of the embodiments of the present disclosure are introduced, technical terms that may be used in the present disclosure are further described:
GOP: the English is called Group of Picture, the Chinese name is Group of pictures, and in the video coding sequence, GOP refers to the distance between two I frames.
I frame: english is called Intra Picture, chinese is called Intra-frame. The I-frame is typically the first frame of each GOP, and is intra-frame compressed, with moderate compression, as a reference point for random access, and can be independently decoded and displayed independent of any frame.
P frame: the English is called Prdictive Frame and the Chinese name is forward predictive coding Frame. Coded pictures of transmission data are compressed by substantially reducing temporal redundancy information below previously coded frames in the sequence of pictures, which are coded using inter-frame prediction techniques.
B frame: the English is called Bi-directional interpolated prediction frame, and the Chinese name is bidirectional predictive interpolated frame. The B frame is referred to by the previous I or P frame and the next P frame, and the compression rate is very high, so that more CPU (Central Processing Unit) resources are consumed during decoding, and the B frame is not a reference frame, and the diffusion of the influence of decoding errors is not caused.
PTS: the English is called Play Start Time and the Chinese name is the playing Time stamp.
Fig. 1 is a schematic flow diagram illustrating a playing processing method according to an embodiment of the present disclosure, where the playing processing method may be applied to a playing processing apparatus. The playing processing device is located in a player of the electronic equipment, and the player can be a web page player and can also be a player application. The electronic device player includes, but is not limited to, a stationary device and/or a mobile device. For example, the fixed device includes, but is not limited to, a server, which may be a cloud server or a general server. For example, mobile devices include, but are not limited to: one or more terminals of a mobile phone, a tablet computer and a vehicle-mounted terminal. In some possible implementations, the playback processing method may also be implemented by the processor calling computer-readable instructions stored in the memory. As shown in fig. 1, the playing processing method includes:
s101: in response to detecting that a preset trigger event occurs, determining a first target playing timestamp;
s102: determining a target frame tracking mode from the candidate frame tracking modes;
s103: determining a playing start frame based on the target frame tracking mode;
s104: the playing of the playing start frame is performed based on the first target playing time stamp.
In the embodiment of the present disclosure, the preset trigger event is a preset event that may trigger the determination of the first target play timestamp. Preset triggering events include, but are not limited to: starting broadcasting, network faults in the live broadcasting process, picture blocking in the live broadcasting process and the like.
In the embodiment of the present disclosure, the first target play time stamp is a target time stamp referred to when chasing a frame. The first target playing time stamp can be understood as the ending time point of the frame tracing process.
In the embodiment of the present disclosure, the first target play time stamp may be PTS1.
In the embodiments of the present disclosure, the candidate frame tracking patterns may be increased or decreased according to design requirements, and the present disclosure does not limit the types and the number of the candidate frame tracking patterns.
In the embodiment of the present disclosure, the candidate frame tracking patterns include N frame tracking patterns, where N is a positive integer greater than or equal to 1. The N chase frame modes can be referred to as chase frame mode 1, chase frame mode 2, chase frame mode 3, \ 8230, chase frame mode N.
Illustratively, chase frame mode 1 is: and (4) catching up by using double-speed playing. For example, 2 times fast forward play, 4 times fast forward play, M times fast forward play, and the like, where M is a number greater than 1.
As another example, frame chase pattern 2 is: all frames before the latest I frame before PTS1 are discarded and double-speed fast-forward play starts from the current I frame decoding.
As another example, chase frame mode 3 is: all frames before the most recent I frame before PTS1 are discarded, decoded from the current I frame, but not show play until the rendering play is started by the current decoding elapsed time (which can be denoted as PTS _ decode) = PTS1.
In some embodiments, determining the target chase frame mode from the candidate chase frame modes comprises: and randomly selecting one frame tracking mode from the candidate frame tracking modes as a target frame tracking mode.
In some embodiments, determining the target chase frame pattern from the candidate chase frame patterns comprises: and selecting a frame chasing mode which is adaptive to the frame chasing scene from the candidate frame chasing modes as a target frame chasing mode according to the frame chasing scene corresponding to the preset trigger event. Thus, the frame tracking effect is improved.
In the embodiment of the present disclosure, the play start frame is a frame at which rendering play starts.
Here, how to determine the play start frame based on the target frame tracking mode and how to perform the play of the play start frame based on the first target play time stamp will be described in the following embodiments, and will not be described herein again.
According to the technical scheme of the embodiment of the disclosure, a first target playing time stamp is determined in response to the detection of the occurrence of a preset trigger event; determining a target frame tracking mode from the candidate frame tracking modes; determining a playing start frame based on the target frame tracking mode; performing playing of the playing start frame based on the first target playing timestamp; therefore, the first target playing time stamp and the playing start frame are determined firstly, the playing of the playing start frame is executed based on the first target playing time stamp, the frame tracing mode is adopted, the playing progress can be traced quickly, and therefore the playing timeliness can be improved.
In some embodiments, S103 comprises:
s103a: in the case where the target chase frame mode is the first chase frame mode, one intra-coded frame (I-frame) closest to the first target play time stamp is taken as a decoding start frame, and the intra-coded frame is determined as a play start frame.
Correspondingly, S104 includes:
s104a: and playing the analyzed image frame at a first preset speed from the beginning frame of playing until the first target playing time stamp is reached, and resuming the default speed playing.
Here, the first frame tracking mode may be: all frames before the latest I frame before PTS1 are discarded and double-speed fast-forward play starts from the current I frame decoding.
In the embodiment of the present disclosure, the first preset speed is a speed greater than 1. Specifically, the first preset multiple speed may be determined according to a time length from the current time point to the first target play time stamp. For example, when the time length from the current time point to the first target playing time stamp is greater than a certain threshold, S1-time speed playing can be selected; and when the time length from the current time point to the first target playing time stamp is less than or equal to a certain threshold value, S2-time speed playing can be selected, wherein the value of S2 is less than the value of S1.
Illustratively, two groups of GOPs are obtained, wherein the 1 st group of GOPs is IPPPBBPPPPPP and the 2 nd group of GOPs is IPPPPPPBBB, the 1 st group of GOPs is discarded, decoding and rendering are started from the I frame of the 2 nd group of GOPs, and the playing is fast-forwarded at double speed.
In the embodiment of the present disclosure, the default double speed is the default normal double speed of the player, such as 1 double speed.
Therefore, when the target frame tracking mode is the first frame tracking mode, the analyzed image frame is played at the first preset speed from the beginning of playing the frame, and the default speed is recovered to play until the first target playing time stamp is reached, so that the frame tracking effect can be realized, and the timeliness of playing can be improved.
In some embodiments, S103 comprises:
s103b: and when the target frame tracking mode is the second frame tracking mode, taking one intra-frame coding frame closest to the first target playing time stamp as a decoding start frame, starting decoding from the decoding start frame, and determining a frame decoded at the first target playing time stamp as a playing start frame.
Correspondingly, S104 includes:
s104b: and playing the playing start frame at the time corresponding to the first target playing time stamp.
Here, the second frame tracking mode may be: all frames before the latest I frame before PTS1 are discarded, decoding is performed from the current I frame, and presentation playback is not performed until the rendering playback is started by the current decoding elapsed time (which may be denoted as PTS _ decode) = PTS1.
Illustratively, two groups of GOPs are obtained, the 1 st GOP is IPPPBBPPPPPP, and the 2 nd GOP is IPPPPPPBBB, the 1 st GOP is discarded, decoding is started from the I frame of the 2 nd GOP, but rendering is not performed until the current decoding time (which may be referred to as PTS _ decode) = PTS1 is consumed to start rendering and playing.
Here, playing the play start frame at a time corresponding to the first target play time stamp includes: and playing the playing start frame at the moment corresponding to the first target playing time stamp according to the default double speed.
Therefore, when the target frame tracking mode is the second frame tracking mode, the playing start frame is played at the moment corresponding to the first target playing time stamp after the playing start frame is determined, the frame tracking effect can be achieved, and the playing timeliness is improved.
In some embodiments, S103 comprises:
s103c: and under the condition that the target frame tracking mode is the first frame tracking mode or the second frame tracking mode, in response to detecting that the type of the frame corresponding to the first target playing time stamp is an intra-frame coding frame, determining the frame corresponding to the first target playing time stamp as a decoding start frame, and determining the frame corresponding to the first target playing time stamp as a playing start frame.
Correspondingly, S104 includes:
s104c: and playing the playing start frame at the moment corresponding to the first target playing time stamp.
Here, playing the play start frame at a time corresponding to the first target play time stamp includes: and playing the playing start frame at the time corresponding to the first target playing time stamp according to the default double speed.
Exemplarily, in the case that the target chase frame mode is the first chase frame mode or the second chase frame mode, if the frame corresponding to the PTS1 position is an I frame, the playback is directly switched to this, and the double speed playback is no longer required.
Therefore, under the condition that the type of the frame corresponding to the first target playing time stamp is the intra-frame coding frame, the frame corresponding to the first target playing time stamp is directly switched to be played, the effect of frame tracing can be rapidly achieved, and the playing timeliness is improved.
In some embodiments, S103 comprises:
s103d: and under the condition that the target frame tracking mode is the third frame tracking mode, taking the first intra-frame coding frame as a decoding starting frame, and determining the intra-frame coding frame as a playing starting frame.
Correspondingly, S104 includes:
s104d: and rendering and playing the analyzed image frames according to a second preset speed from the beginning of playing the frame until the first target playing time stamp is reached, and recovering the default speed playing.
Here, the third frame tracking mode may be: catch-up is performed by using double-speed playing. For example, 2 × speed fast forward play, 4 × speed fast forward play, M × speed fast forward play, etc., where M is a number greater than 1.
In the embodiment of the present disclosure, the second preset double speed is a speed greater than 1. Specifically, the second preset multiple speed may be selected according to a time length from the current time point to the first target play time stamp. For example, when the time length from the current time point to the first target playing time stamp is greater than a certain threshold, S3 times speed playing may be selected; and when the time length from the current time point to the first target playing time stamp is less than or equal to a certain threshold value, S4-time speed playing can be selected, wherein the value of S3 is less than the value of S4.
Therefore, when the target frame tracking mode is the third frame tracking mode, the analyzed image frame is rendered and played at the second preset multiple speed from the beginning frame of playing until the first target playing time stamp is reached, and the default multiple speed playing is resumed, so that the frame tracking effect can be realized, and the playing timeliness can be improved.
In some embodiments, S101 comprises:
s101a: determining a playing scene corresponding to a preset trigger event;
s101b: a first target play timestamp is determined based on the play scenario.
In the embodiment of the present disclosure, the playing scene includes, but is not limited to, a start playing scene and a live playing scene.
Here, the requirements of different playing scenes on the timeliness of playing may be different. For example, the timeliness of the start scene is lower than that in the live broadcasting process.
Therefore, the determined first target playing time stamp can be adapted to the playing scene, and the frame tracking effect and the playing timeliness can be improved.
In some embodiments, S101b comprises:
s101b1: under the condition that the playing scene is a first scene, acquiring a first time corresponding to the start of playing and a second time corresponding to the completion of playing;
s101b2: determining a first time duration for starting playing based on the first time and the second time;
s101b3: a first target play timestamp for the first scene is determined based on the first elapsed time length.
In the embodiment of the present disclosure, the first scene may be an initiating scene.
In some embodiments, a first elapsed time of playout is determined based on a first time (denoted t 1) and a second time (denoted t 2)Δt1 cost (ii) a The first time length is increased by delta t1 cost A first target play timestamp for the first scene is determined.
As shown in fig. 2, after receiving a start-up command, the player determines that a preset trigger event occurs, and obtains a first time t1 corresponding to start-up start and a second time t2 corresponding to start-up completion; determining a first target play timestamp PTS1= PTS _ startTime + Δ t1 cost . In general, in the start-up scene, PTS _ startTime =0, and thus PTS1= Δ t1 cost
It should be understood that the schematic diagram shown in fig. 2 is merely exemplary and not limiting, and it is extensible, and that various obvious changes and/or substitutions may be made by those skilled in the art based on the example of fig. 2, and the resulting technical solutions still fall within the disclosure scope of the embodiments of the present disclosure.
Therefore, the first target playing time stamp in the first scene can be quickly determined, and the frame tracking speed of the first scene is improved.
In some embodiments, S101b comprises:
s101b1': under the condition that the playing scene is a second scene, acquiring a third time corresponding to the starting of buffering and a fourth time corresponding to the completion of buffering;
s101b2': determining a second buffered time duration based on the third time and the fourth time;
s101b3': and determining a first target playing time stamp of the second scene based on the occurrence time point and the second time length corresponding to the preset trigger event.
In the embodiment of the present disclosure, the second scene may be a live scene. For example, the live scenes include, but are not limited to, a web break scene occurring during the live process, and a katoon scene occurring during the live process.
For example, if the network is disconnected suddenly in the live broadcast process and then the network is connected, as shown in fig. 3, if the time of the network disconnection is td, buffering is started after the network is connected again, the time corresponding to the start of buffering is t3, and the time corresponding to the completion of buffering is t4, the first target play timestamp PTS1= PTS _ startTime + Δ t2 cost (ii) a Wherein PTS _ startTime = t3-td,Δt2 cost =t4-t3。
in some embodiments, the second time duration Δ t2 of the buffer is determined at the third time t3 and the fourth time t4 cost When the time is less than a certain threshold value, based on the corresponding occurrence time point (denoted as td) of the preset trigger event and the second time length delta t2 cost Determining a first target play timestamp for the second scene, PTS1= PTS _ startTime + Δ t2 cost =(t3-td)+(t4-t3)。
In some embodiments, the second time duration Δ t2 of the buffer is determined at the third time t3 and the fourth time t4 cost When the time is larger than a certain threshold value, the occurrence time point (denoted as td) corresponding to the preset trigger event is ignored, and the second time duration delta t2 is based on cost Determining a first target play timestamp for a second scene, PTS1= Δ t2 cost
It should be understood that the schematic diagram shown in fig. 3 is merely exemplary and not limiting, and it is extensible, and that various obvious changes and/or substitutions may be made by those skilled in the art based on the example of fig. 3, and the resulting technical solutions still fall within the disclosure scope of the embodiments of the present disclosure.
Therefore, the first target playing time stamp in the second scene can be determined quickly, the frame tracking speed of the second scene is improved, and the timeliness of frame tracking is improved.
In some embodiments, the playback processing method further includes:
acquiring the time-consuming duration from the occurrence of a preset trigger event to the time corresponding to the first target playing time stamp;
determining a second target playing time stamp based on the time-consuming duration under the condition that the time-consuming duration is greater than a preset threshold;
determining a second playing start frame based on the target frame tracking mode;
the playing of the second play start frame is performed based on the second target play time stamp.
In the embodiment of the present disclosure, the preset threshold may be set or adjusted according to the frame tracking accuracy.
In the embodiment of the present disclosure, if the first target play time stamp is PTS1, the second target play time stamp is PTS2, and the time duration from the occurrence of the preset trigger event to the time corresponding to the first target play time stamp is denoted as Δ T _ decode, then PTS2= PTS1+ Δ T _ decode.
As shown in fig. 4, after receiving a start-up instruction, the player determines that a preset trigger event occurs, and obtains a first time t1 corresponding to start-up of the start-up and a second time t2 corresponding to completion of the start-up; first target play time stamp PTS1= PTS _ startTime + Δ t1 cost . In general, in the start-up scene, PTS _ startTime =0, pts1= Δ t1 cost . And if the decoding time duration corresponding to the PTS1 is delta T _ decode, the PTS2= PTS1+ delta T _ decode. Determining a second playing start frame based on the target frame tracking mode; playback of the second playback start frame is performed based on the PTS 2.
As shown in fig. 5, it is assumed that the time of network disconnection or deadlock is td, the time corresponding to the start of buffering is t3, the time corresponding to the completion of buffering is t4, and the first target play timestamp PTS1= PTS _ startTime + Δ t2 cost (ii) a Wherein PTS _ startTime = t3-td, Δ t2 cost T4-t3. And if the decoding time consumption duration corresponding to the PTS1 is delta T _ decode, the PTS2= PTS1+ delta T _ decode.
Therefore, the time-consuming duration of the first frame chasing can be further subjected to secondary frame chasing, the timeliness of the frame chasing can be further improved through the secondary frame chasing, and the effect that the player end and the release end keep synchronous is achieved.
In some embodiments, for Δ T _ decode, the decoding elapsed time corresponding to the IPB frame may be calculated during the video start-up, and the elapsed time required for decoding may be predicted in advance.
In practical application, for network buffering and buffering interruption caused by network failure, the live video stream can be pulled up again, and the playing can be recalculated when the live video stream is pulled up again.
An embodiment of the present disclosure provides a playback processing apparatus, as shown in fig. 6, the playback processing apparatus may include: a first determining module 601, configured to determine a first target playing timestamp in response to detecting that a preset trigger event occurs; a second determining module 602, configured to determine a target frame tracking mode from the candidate frame tracking modes; a third determining module 603, configured to determine a play start frame based on the target frame tracking mode; the playing control module 604 is configured to perform playing of the playing start frame based on the first target playing time stamp.
In some embodiments, the third determining module 603 includes: and the first determining sub-module is used for taking an intra-frame encoding frame closest to the first target playing time stamp as a decoding start frame and determining the intra-frame encoding frame as a playing start frame under the condition that the target frame tracking mode is the first frame tracking mode. Correspondingly, the playing control module 604 includes: and the first control submodule is used for playing the analyzed image frame at a first preset speed from the beginning of playing the frame until the default speed is restored to play when the first target playing time stamp is reached.
In some embodiments, the third determining module 603 further includes: and the second determining sub-module is used for taking an intra-frame coding frame closest to the first target playing time stamp as a decoding start frame, starting decoding from the decoding start frame and determining a frame decoded at the first target playing time stamp as a playing start frame under the condition that the target frame tracking mode is the second frame tracking mode. Correspondingly, the playing control module 604 further includes: and the second control submodule is used for playing the playing start frame at the moment corresponding to the first target playing time stamp.
In some embodiments, the playback processing apparatus may include: a fourth determining module 605 (not shown in the figure) for determining the frame corresponding to the first target playing time stamp as the decoding-start frame and determining the frame corresponding to the first target playing time stamp as the playing-start frame in response to detecting that the type of the frame corresponding to the first target playing time stamp is the intra-coded frame, in case that the target frame tracking mode is the first frame tracking mode or the second frame tracking mode. The playing control module 604 includes: and the third control sub-module is used for playing the playing start frame at the moment corresponding to the first target playing time stamp.
In some embodiments, the third determining module 603 includes: and the third determining sub-module is used for taking the first intra-frame as a decoding start frame and determining the intra-frame as a playing start frame under the condition that the target frame tracking mode is the third frame tracking mode. Correspondingly, the playing control module 604 further includes: and the fourth control submodule is used for rendering and playing the analyzed image frame at the second preset multiple speed from the playing start frame until the first target playing time stamp is reached, and restoring the default multiple speed playing.
In some embodiments, the first determining module 601 includes: the fourth determining submodule is used for determining a playing scene corresponding to the preset triggering event; and the fifth determining submodule is used for determining the first target playing time stamp based on the playing scene.
In some embodiments, a fifth determination submodule to: under the condition that the playing scene is a first scene, acquiring a first time corresponding to starting playing and starting and a second time corresponding to finishing playing; determining a first time duration for starting playing based on the first time and the second time; based on the first time duration, a first target play timestamp of the first scene is determined.
In some embodiments, the fifth determining sub-module is further configured to: under the condition that the playing scene is a second scene, acquiring a third time corresponding to the starting of buffering and a fourth time corresponding to the completion of buffering; determining a second buffered time duration based on the third time and the fourth time; and determining a first target playing time stamp of the second scene based on the occurrence time point and the second time length corresponding to the preset trigger event.
In some embodiments, the playback processing apparatus may further include: an obtaining module 606 (not shown in the figure) configured to obtain a time duration from occurrence of a preset trigger event to a time corresponding to the first target playing time stamp; a fifth determining module 607 (not shown in the figure) for determining the second target playing time stamp based on the elapsed time length if the elapsed time length is greater than the preset threshold. Correspondingly, the second determining module 602 is further configured to determine a second playing start frame based on the target frame tracking mode; and the playing control module is also used for executing the playing of the second playing starting frame based on the second target playing time stamp.
It should be understood by those skilled in the art that the functions of each processing module in the playing processing apparatus according to the embodiment of the present disclosure may be understood by referring to the foregoing description of the playing processing method, and each processing module in the playing processing apparatus according to the embodiment of the present disclosure may be implemented by an analog circuit that implements the functions described in the embodiment of the present disclosure, or may be implemented by running software that implements the functions described in the embodiment of the present disclosure on an electronic device.
The playing processing device of the embodiment of the disclosure can improve the timeliness of playing.
The embodiment of the present disclosure further provides a first scene schematic diagram of playing processing, as shown in fig. 7, an electronic device, such as a cloud server, receives a preset trigger event sent by a terminal; determining a first target playing time stamp and a target frame tracking mode for the terminal; determining a playing start frame for the terminal based on the target frame tracking mode; and the terminal executes the playing of the playing start frame based on the first target playing time stamp. Therefore, the timeliness of playing can be improved; the problems of video blocking and time delay in different scenes are fundamentally solved; the fluency of playing the video is improved, and comfortable and fluent watching experience can be provided for the user.
The embodiment of the present disclosure further provides a scene schematic diagram of play processing, as shown in fig. 8, when detecting that a preset trigger event occurs, a terminal acquires a live data stream from an electronic device such as a cloud server, and the terminal determines a first target play timestamp and a target frame tracking mode; determining a playing start frame based on the target frame tracking mode; and the terminal executes the playing of the playing start frame based on the first target playing time stamp. Therefore, the timeliness of playing can be improved; the problems of video blocking and time delay in different scenes are fundamentally solved; the fluency of playing the video is improved, and comfortable and fluent watching experience can be provided for the user.
The number of the terminals and the electronic devices is not limited in the present disclosure, and a plurality of terminals and a plurality of electronic devices may be included in practical applications.
It should be understood that the scene diagrams shown in fig. 7 and fig. 8 are only illustrative and not restrictive, and those skilled in the art may make various obvious changes and/or substitutions based on the examples of fig. 7 and fig. 8, and the obtained technical solutions still belong to the disclosure scope of the embodiments of the present disclosure.
In the technical scheme of the disclosure, the acquisition, storage, application and the like of the personal information of the related user all accord with the regulations of related laws and regulations, and do not violate the good customs of the public order.
The present disclosure also provides an electronic device, a readable storage medium, and a computer program product according to embodiments of the present disclosure.
FIG. 9 illustrates a schematic block diagram of an example electronic device 900 that can be used to implement embodiments of the present disclosure. Electronic devices are intended to represent various forms of digital computers, such as laptops, desktops, workstations, personal digital assistants, servers, blade servers, mainframes, and other appropriate computers. The electronic device may also represent various forms of mobile devices, such as personal digital assistants, cellular telephones, smart phones, wearable devices, and other similar computing devices. The components shown herein, their connections and relationships, and their functions, are meant to be examples only, and are not meant to limit implementations of the disclosure described and/or claimed herein.
As shown in fig. 9, the apparatus 900 includes a computing unit 901, which can perform various appropriate actions and processes according to a computer program stored in a Read-Only Memory (ROM) 902 or a computer program loaded from a storage unit 908 into a Random Access Memory (RAM) 903. In the RAM 903, various programs and data required for the operation of the device 900 can also be stored. The calculation unit 901, ROM 902, and RAM 903 are connected to each other via a bus 904. An Input/Output (I/O) interface 905 is also connected to bus 904.
A number of components in the device 900 are connected to the I/O interface 905, including: an input unit 906 such as a keyboard, a mouse, and the like; an output unit 907 such as various types of displays, speakers, and the like; a storage unit 908 such as a magnetic disk, optical disk, or the like; and a communication unit 909 such as a network card, a modem, a wireless communication transceiver, and the like. The communication unit 909 allows the device 900 to exchange information/data with other devices through a computer network such as the internet and/or various telecommunication networks.
The computing unit 901 may be a variety of general and/or special purpose processing components having processing and computing capabilities. Some examples of the computing Unit 901 include, but are not limited to, a Central Processing Unit (CPU), a Graphics Processing Unit (GPU), various specialized Artificial Intelligence (AI) computing chips, various computing units running machine learning model algorithms, a Digital Signal Processor (DSP), and any suitable Processor, controller, microcontroller, and the like. The computing unit 901 performs the respective methods and processes described above, such as the playback processing method. For example, in some embodiments, the playback processing method may be implemented as a computer software program tangibly embodied in a machine-readable medium, such as storage unit 908. In some embodiments, part or all of the computer program may be loaded and/or installed onto device 900 via ROM 902 and/or communications unit 909. When the computer program is loaded into the RAM 903 and executed by the computing unit 901, one or more steps of the play processing method described above may be performed. Alternatively, in other embodiments, the computing unit 901 may be configured to perform the playback processing method in any other suitable manner (e.g., by means of firmware).
Various implementations of the systems and techniques described here above may be realized in digital electronic circuitry, integrated circuitry, field Programmable Gate Arrays (FPGAs), application Specific Integrated Circuits (ASICs), application-Specific Standard Products (ASSPs), system-on-Chip (SOC), complex Programmable Logic Devices (CPLDs), computer hardware, firmware, software, and/or combinations thereof. These various embodiments may include: implemented in one or more computer programs that are executable and/or interpretable on a programmable system including at least one programmable processor, which may be special or general purpose, receiving data and instructions from, and transmitting data and instructions to, a storage system, at least one input device, and at least one output device.
Program code for implementing the methods of the present disclosure may be written in any combination of one or more programming languages. These program codes may be provided to a processor or controller of a general purpose computer, special purpose computer, or other programmable data processing apparatus, such that the program codes, when executed by the processor or controller, cause the functions/operations specified in the flowchart and/or block diagram to be performed. The program code may execute entirely on the machine, partly on the machine, as a stand-alone software package, partly on the machine and partly on a remote machine or entirely on the remote machine or server.
In the context of this disclosure, a machine-readable medium may be a tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. The machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium. A machine-readable medium may include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples of a machine-readable storage medium would include an electrical connection based on one or more wires, a portable computer diskette, a hard Disk, a random access Memory, a Read-Only Memory, an Erasable Programmable Read-Only Memory (EPROM), a flash Memory, an optical fiber, a Compact disc Read-Only Memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
To provide for interaction with a user, the systems and techniques described here can be implemented on a computer having: a Display device (e.g., a Cathode Ray Tube (CRT) or Liquid Crystal Display (LCD) monitor) for displaying information to a user; and a keyboard and a pointing device (e.g., a mouse or a trackball) by which a user may provide input to the computer. Other kinds of devices may also be used to provide for interaction with a user; for example, feedback provided to the user can be any form of sensory feedback (e.g., visual feedback, auditory feedback, or tactile feedback); and input from the user can be received in any form, including acoustic, speech, or tactile input.
The systems and techniques described here can be implemented in a computing system that includes a back-end component (e.g., as a data server), or that includes a middleware component (e.g., an application server), or that includes a front-end component (e.g., a user computer having a graphical user interface or a web browser through which a user can interact with an implementation of the systems and techniques described here), or any combination of such back-end, middleware, or front-end components. The components of the system can be interconnected by any form or medium of digital data communication (e.g., a communication network). Examples of communication networks include: local Area Networks (LANs), wide Area Networks (WANs), and the internet.
The computer system may include clients and servers. A client and server are generally remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client and server relationship to each other. The server may be a cloud server, a server of a distributed system, or a server with a combined blockchain.
It should be understood that various forms of the flows shown above, reordering, adding or deleting steps, may be used. For example, the steps described in the present disclosure may be executed in parallel or sequentially or in different orders, and are not limited herein as long as the desired results of the technical solutions disclosed in the present disclosure can be achieved.
The above detailed description should not be construed as limiting the scope of the disclosure. It should be understood by those skilled in the art that various modifications, combinations, sub-combinations and substitutions may be made, depending on design requirements and other factors. Any modification, equivalent replacement, and improvement made within the principle of the present disclosure should be included in the scope of protection of the present disclosure.

Claims (21)

1. A playback processing method, comprising:
in response to detecting that a preset trigger event occurs, determining a first target playing time stamp;
determining a target frame tracking mode from the candidate frame tracking modes;
determining a playing start frame based on the target frame tracking mode;
and executing the playing of the playing start frame based on the first target playing time stamp.
2. The method of claim 1, wherein the determining a playback start frame based on the target chase frame mode comprises:
under the condition that the target frame tracking mode is a first frame tracking mode, taking an intra-frame coding frame closest to the first target playing time stamp as a decoding start frame, and determining the intra-frame coding frame as the playing start frame;
the performing of the playing start frame based on the first target playing time stamp includes:
and playing the analyzed image frame at a first preset speed from the playing start frame until the first target playing time stamp is reached, and recovering the default speed playing.
3. The method of claim 1, wherein the determining a playback start frame based on the target chase frame mode comprises:
under the condition that the target frame tracking mode is a second frame tracking mode, taking an intra-frame coding frame closest to the first target playing time stamp as a decoding start frame, starting decoding from the decoding start frame, and determining a frame decoded at the first target playing time stamp as the playing start frame;
the performing of the playing start frame based on the first target playing time stamp includes:
and playing the playing start frame at the moment corresponding to the first target playing time stamp.
4. The method of claim 2 or 3, further comprising:
in response to the fact that the type of the frame corresponding to the first target playing time stamp is detected to be an intra-frame coding frame, taking the frame corresponding to the first target playing time stamp as the decoding start frame, and determining the frame corresponding to the first target playing time stamp as the playing start frame;
the performing of the playing start frame based on the first target playing time stamp includes:
and playing the playing start frame at the moment corresponding to the first target playing time stamp.
5. The method of claim 1, wherein the determining a playback start frame based on the target chase frame mode comprises:
under the condition that the target frame tracking mode is a third frame tracking mode, taking a first intra-frame coding frame as a decoding initial frame, and determining the intra-frame coding frame as the playing initial frame;
the performing of the playing of the play start frame based on the first target play time stamp includes:
and rendering and playing the analyzed image frames according to a second preset speed from the playing start frame until the first target playing time stamp is reached, and recovering the default speed playing.
6. The method of claim 1, wherein said determining a first target play timestamp in response to detecting occurrence of a preset trigger event comprises:
determining a playing scene corresponding to the preset trigger event;
determining the first target play timestamp based on the play scene.
7. The method of claim 6, wherein said determining the first target playback timestamp based on the playback scenario comprises:
under the condition that the playing scene is a first scene, acquiring a first time corresponding to starting playing and a second time corresponding to finishing starting playing;
determining a first time duration for starting playing based on the first time and the second time;
determining the first target play timestamp for the first scene based on the first age.
8. The method of claim 6, wherein said determining the first target playback timestamp based on the playback scene comprises:
under the condition that the playing scene is a second scene, acquiring a third time corresponding to the starting of buffering and a fourth time corresponding to the completion of buffering;
determining a second buffered time duration based on the third time and the fourth time;
and determining the first target playing time stamp of the second scene based on the occurrence time point corresponding to the preset trigger event and the second time duration.
9. The method of claim 1, further comprising:
acquiring the time-consuming duration from the occurrence of the preset trigger event to the time corresponding to the first target playing time stamp;
determining a second target playing time stamp based on the time-consuming duration under the condition that the time-consuming duration is greater than a preset threshold;
determining a second playing start frame based on the target frame tracking mode;
performing playback of the second playback start frame based on the second target playback time stamp.
10. A playback processing apparatus comprising:
the first determining module is used for determining a first target playing time stamp in response to the detection of the occurrence of the preset triggering event;
the second determining module is used for determining a target frame chasing mode from the candidate frame chasing modes;
a third determining module, configured to determine a play start frame based on the target frame tracking mode;
and the playing control module is used for executing the playing of the playing start frame based on the first target playing time stamp.
11. The apparatus of claim 10, wherein the third determining means comprises:
a first determining sub-module, configured to, when the target frame tracking mode is a first frame tracking mode, use an intra-frame encoded frame closest to the first target play timestamp as a decoding start frame, and determine the intra-frame encoded frame as the play start frame;
the play control module includes:
and the first control submodule is used for playing the analyzed image frame at a first preset speed from the playing start frame until the default speed is recovered to play when the first target playing time stamp is reached.
12. The apparatus of claim 10, wherein the third determining means comprises:
a second determining sub-module, configured to, when the target frame tracking mode is a second frame tracking mode, use an intra-frame encoded frame closest to the first target play timestamp as a decoding start frame, start decoding from the decoding start frame, and determine a frame decoded at the first target play timestamp as the play start frame;
the play control module comprises:
and the second control submodule is used for playing the playing start frame at the moment corresponding to the first target playing time stamp.
13. The apparatus of claim 11 or 12, further comprising:
a fourth determining module, configured to, in response to detecting that the type of the frame corresponding to the first target play timestamp is an intra-frame coded frame, use the frame corresponding to the first target play timestamp as the decoding start frame, and determine the frame corresponding to the first target play timestamp as the play start frame;
the play control module includes:
and the third control sub-module is used for playing the playing start frame at the moment corresponding to the first target playing time stamp.
14. The apparatus of claim 10, wherein the third determining means comprises:
a third determining sub-module, configured to determine a first intra-frame as a decoding start frame and determine the intra-frame as the playing start frame when the target frame tracking mode is a third frame tracking mode;
the play control module includes:
and the fourth control submodule is used for rendering and playing the analyzed image frame at a second preset multiple speed from the playing start frame until the first target playing time stamp is reached, and restoring the default multiple speed playing.
15. The apparatus of claim 10, wherein the first determining means comprises:
the fourth determining submodule is used for determining a playing scene corresponding to the preset triggering event;
a fifth determining sub-module, configured to determine the first target playing time stamp based on the playing scene.
16. The apparatus of claim 15, wherein the fifth determination submodule is to:
under the condition that the playing scene is a first scene, acquiring a first time corresponding to starting playing and starting and a second time corresponding to finishing playing;
determining a first time duration for starting playing based on the first time and the second time;
determining the first target play timestamp for the first scene based on the first time duration.
17. The apparatus of claim 15, wherein the fifth determination submodule is to:
under the condition that the playing scene is a second scene, acquiring a third time corresponding to the start of buffering and a fourth time corresponding to the completion of buffering;
determining a second buffered time duration based on the third time and the fourth time;
and determining the first target playing time stamp of the second scene based on the occurrence time point corresponding to the preset trigger event and the second time duration.
18. The apparatus of claim 10, further comprising:
the acquisition module is used for acquiring the time-consuming duration from the occurrence of the preset trigger event to the time corresponding to the first target playing time stamp;
a fifth determining module, configured to determine a second target playing timestamp based on the time-consuming duration when the time-consuming duration is greater than a preset threshold;
the second determining module is further configured to determine a second playing start frame based on the target frame tracking mode;
the playing control module is further configured to execute playing of the second playing start frame based on the second target playing timestamp.
19. An electronic device, comprising:
at least one processor; and
a memory communicatively coupled to the at least one processor; wherein, the first and the second end of the pipe are connected with each other,
the memory stores instructions executable by the at least one processor to enable the at least one processor to perform the method of any one of claims 1-9.
20. A non-transitory computer readable storage medium having stored thereon computer instructions for causing the computer to perform the method of any one of claims 1-9.
21. A computer program product comprising a computer program which, when executed by a processor, implements the method according to any one of claims 1-9.
CN202211152210.3A 2022-09-21 2022-09-21 Playing processing method, device, equipment and storage medium Pending CN115567757A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211152210.3A CN115567757A (en) 2022-09-21 2022-09-21 Playing processing method, device, equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211152210.3A CN115567757A (en) 2022-09-21 2022-09-21 Playing processing method, device, equipment and storage medium

Publications (1)

Publication Number Publication Date
CN115567757A true CN115567757A (en) 2023-01-03

Family

ID=84741468

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211152210.3A Pending CN115567757A (en) 2022-09-21 2022-09-21 Playing processing method, device, equipment and storage medium

Country Status (1)

Country Link
CN (1) CN115567757A (en)

Similar Documents

Publication Publication Date Title
WO2020248909A1 (en) Video decoding method and apparatus, computer device, and storage medium
WO2019170073A1 (en) Media playback
CN110430445B (en) Video synchronous playing method, device, equipment and medium
EP3697088A1 (en) Video sending and receiving method, device, and terminal
CN110582012B (en) Video switching method, video processing device and storage medium
CN102883217B (en) Method and device for controlling video playing
US20160180496A1 (en) Information processor, information processing method, and program
CN109040830B (en) Live broadcast pause prediction method, switching method and device
CN113225585A (en) Video definition switching method and device, electronic equipment and storage medium
CN106470291A (en) Recover in the interruption in time synchronized from audio/video decoder
CN104053002A (en) Video decoding method and device
WO2020237466A1 (en) Video transmission method and apparatus, and aircraft, playback device, and storage medium
US8411755B2 (en) Video transmission apparatus and control method for video transmission apparatus
CN113766261A (en) Method and device for determining pre-pulling duration, electronic equipment and storage medium
CN112468876A (en) Resource playing method, device and system and readable storage medium
CN115567757A (en) Playing processing method, device, equipment and storage medium
CN115297343B (en) Video data sending method and device and electronic equipment
JP5062110B2 (en) Video transmission apparatus, video transmission method, and video transmission program
RU2662648C1 (en) Method and device for data processing
CN111467797B (en) Game data processing method and device, computer storage medium and electronic equipment
CN114339415A (en) Client video playing method and device, electronic equipment and readable medium
CN116939212A (en) Video processing method, device, computer readable storage medium and computer equipment
CN113038237A (en) Live broadcast information processing method, device, equipment and storage medium
CN113824985B (en) Live streaming method, apparatus, device, storage medium and computer program product
CN111935549B (en) Method and device for updating playing sequence

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination