CN114222165B - Video playing method, device, equipment and computer storage medium - Google Patents

Video playing method, device, equipment and computer storage medium Download PDF

Info

Publication number
CN114222165B
CN114222165B CN202111677497.7A CN202111677497A CN114222165B CN 114222165 B CN114222165 B CN 114222165B CN 202111677497 A CN202111677497 A CN 202111677497A CN 114222165 B CN114222165 B CN 114222165B
Authority
CN
China
Prior art keywords
video
action
processed
determining
playback
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111677497.7A
Other languages
Chinese (zh)
Other versions
CN114222165A (en
Inventor
李荣生
李晓耕
陶嘉伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China Mobile Communications Group Co Ltd
MIGU Video Technology Co Ltd
MIGU Culture Technology Co Ltd
Original Assignee
China Mobile Communications Group Co Ltd
MIGU Video Technology Co Ltd
MIGU Culture Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China Mobile Communications Group Co Ltd, MIGU Video Technology Co Ltd, MIGU Culture Technology Co Ltd filed Critical China Mobile Communications Group Co Ltd
Priority to CN202111677497.7A priority Critical patent/CN114222165B/en
Publication of CN114222165A publication Critical patent/CN114222165A/en
Application granted granted Critical
Publication of CN114222165B publication Critical patent/CN114222165B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234381Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by altering the temporal resolution, e.g. decreasing the frame rate by frame skipping
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/432Content retrieval operation from a local storage medium, e.g. hard-disk
    • H04N21/4325Content retrieval operation from a local storage medium, e.g. hard-disk by playing back content from the storage medium
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • H04N21/440281Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display by altering the temporal resolution, e.g. by frame skipping
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8456Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Abstract

The embodiment of the application relates to the technical field of multimedia playing and discloses a method which comprises the following steps: acquiring a video to be processed; the video to be processed comprises a plurality of actions; identifying the video to be processed to obtain the action types of all actions; respectively determining playback time of each action according to the action type; processing each action according to the playback time length to obtain a processed video; and playing the processed video. Through the mode, the video watching experience of the user is improved.

Description

Video playing method, device, equipment and computer storage medium
Technical Field
The embodiment of the application relates to the technical field of multimedia playing, in particular to a video playing method, a video playing device, video playing equipment and a computer storage medium.
Background
In video playback, particularly for video with a fast picture change rate, there is often a need for slow motion playback by users. To achieve slow motion playback of video, the video may be subject to an inter-frame process.
The inventor of the present application finds that in the process of implementing the embodiment of the present application, the current video frame inserting method has the problem that the video viewing experience of the user is poor.
Disclosure of Invention
In view of the above problems, embodiments of the present invention provide a video playing method, apparatus, device, and computer storage medium, which are used to solve the problem in the prior art that the video viewing experience of a user is poor.
According to an aspect of an embodiment of the present invention, there is provided a video playing method, including:
acquiring a video to be processed; the video to be processed comprises a plurality of actions;
identifying the video to be processed to obtain action types of all actions;
respectively determining playback time of each action according to the action type;
processing each action according to the playback time length to obtain a processed video;
and playing the processed video.
In an alternative manner, the video to be processed includes a plurality of video frames; one of the actions corresponds to at least one of the video frames; the method further comprises the steps of:
extracting features of the video to be processed to obtain image features corresponding to each video frame;
clustering the video frames according to the image characteristics to obtain a clustering result;
and determining each action and the action type corresponding to the action according to the clustering result.
In an alternative, the method further comprises:
determining type characteristic information of each action type;
determining information entropy corresponding to each action type according to all the type characteristic information;
and determining the playback time according to the information entropy.
In an optional manner, the type feature information includes at least one of action complexity, popularity, and action difficulty of the action type; the method further comprises the steps of:
respectively determining play weight values corresponding to the action types according to at least one of the action complexity, the popularity and the action difficulty of the action types;
and respectively determining the information entropy corresponding to each action type according to the play weight value.
In an alternative, the method further comprises:
determining the playback proportion of each action type according to the information entropy;
determining the total playback time length corresponding to the video to be processed;
and determining the playback time according to the playback proportion and the total playback time.
In an alternative, the method further comprises:
determining time stamp information and current playing parameters of each action;
Determining the frame inserting information corresponding to each action according to the playback time length, the time stamp information and the current playing parameters;
and carrying out frame inserting processing on each action according to the frame inserting information to obtain the processed video. .
In an alternative, the method further comprises:
marking a frame interval where the action after the frame insertion processing is located in the video to be processed to obtain a marked fragment;
and playing the marked fragments when detecting the operation on the marked fragments.
According to another aspect of an embodiment of the present invention, there is provided a video playing apparatus including:
the acquisition module is used for acquiring the video to be processed; the video to be processed comprises a plurality of actions;
the identification module is used for identifying the video to be processed to obtain action types of all actions;
the determining module is used for determining the playback time length of each action according to the action type;
the processing module is used for respectively processing each action according to the playback time length to obtain a processed video;
and the playing module is used for playing the processed video.
According to another aspect of an embodiment of the present invention, there is provided a video playback apparatus including: the device comprises a processor, a memory, a communication interface and a communication bus, wherein the processor, the memory and the communication interface complete communication with each other through the communication bus;
The memory is used for storing at least one executable instruction, and the executable instruction enables the processor to execute the operation of the video playing method.
According to yet another aspect of an embodiment of the present invention, there is provided a computer-readable storage medium having stored therein at least one executable instruction for causing a video playback device to perform the operations of the video playback method.
The embodiment of the invention obtains the video to be processed; the video to be processed comprises a plurality of actions; identifying the video to be processed to obtain the action types of all actions; respectively determining playback time of each action according to the action type; processing each action according to the playback time length to obtain a processed video; finally, playing the processed video, which is different from the problem of poor playing effect caused by global playback in the prior art, the embodiment of the invention can determine the playback time according to the action types of the actions included in the video to be processed, thereby realizing the playback with different time according to the characteristic adaptability of different actions, and enabling the user to acquire the action detail information with the quantity matched with the action types, thereby improving the video watching experience of the user.
The foregoing description is only an overview of the technical solutions of the embodiments of the present invention, and may be implemented according to the content of the specification, so that the technical means of the embodiments of the present invention can be more clearly understood, and the following specific embodiments of the present invention are given for clarity and understanding.
Drawings
The drawings are only for purposes of illustrating embodiments and are not to be construed as limiting the invention. Also, like reference numerals are used to designate like parts throughout the figures. In the drawings:
fig. 1 shows a flow chart of a video playing method according to an embodiment of the present invention;
fig. 2 shows a schematic marking diagram of a frame interval to be processed according to an embodiment of the present invention;
fig. 3 shows a schematic playback diagram of a processed video according to an embodiment of the present invention;
fig. 4 is a schematic structural diagram of a video playing device according to an embodiment of the present invention;
fig. 5 shows a schematic structural diagram of a video playing device according to an embodiment of the present invention.
Detailed Description
Exemplary embodiments of the present invention will be described in more detail below with reference to the accompanying drawings. While exemplary embodiments of the present invention are shown in the drawings, it should be understood that the present invention may be embodied in various forms and should not be limited to the embodiments set forth herein.
Fig. 1 shows a flowchart of a video playing method according to an embodiment of the present invention, where the method is performed by a computer processing device. The computer processing device may include a cell phone, a notebook computer, and the like. As shown in fig. 1, the method comprises the steps of:
step 10: acquiring a video to be processed; the video to be processed comprises a plurality of actions.
In one embodiment of the invention, the video to be processed may include one or more moving objects, and the plurality of actions may correspond to the same moving object or different moving objects. Alternatively, the moving object has a high-speed motion in the video to be processed, such as a player in racing sports or a ball in a ball game, etc.
In yet another embodiment of the present invention, the video to be processed may also be a highlight video clip obtained through pre-filtering, such as a video clip previously filtered and clipped from a video stream according to the degree of highlighting or association of actions. The action precision and the chroma can be determined according to the continuity of the action, the action difficulty, the completion degree of the action and the like, and the association degree can be the association degree of the video clip and the current playing scene information, wherein the current playing scene information comprises the information such as the current playing video information, the current playing time, the geographic position, the user characteristic information and the like.
Step 20: and identifying the video to be processed to obtain the action type of each action.
In one embodiment of the invention, the video to be processed is divided into a plurality of continuous video frames, image feature recognition is carried out on each video frame, and clustering analysis is carried out according to the recognition result to obtain the action type corresponding to each video frame. Wherein the action type may be one of selectable types pre-marked for a video type of the video to be processed, and the video type may be a type of motion included in the video or a type of a moving object. When the type of movement of the video is a ski game, alternative types may include slip assistance, take off, air flight, and landing. For video types where the moving object is a table tennis ball, alternative types include landing, table landing, racket landing, net wiping, and in-air sports, among others.
In one embodiment of the present invention, the video to be processed includes a plurality of video frames; one of the actions corresponds to at least one of the video frames; step 20 further comprises:
step 201: and identifying the video to be processed to obtain image features corresponding to the video frames.
In one embodiment of the invention, image feature extraction is performed on each video frame in the video to be processed, and the extraction mode can be through a convolutional neural network, wherein the convolutional neural network comprises a convolutional layer, a pooling layer and a connecting layer. The image features may include RGB features, gray features, etc. of the video frame.
Step 202: and clustering the video frames according to the image characteristics to obtain a clustering result.
In one embodiment of the invention, the clustering can be performed according to a K-means algorithm, c groups are preset, corresponding to c action types, and the action types corresponding to each video frame are obtained as a clustering result by clustering the video frames under the corresponding groups.
Step 203: and determining each action and the action type corresponding to the action according to the clustering result.
In one embodiment of the present invention, video frames that have the same action type and continuously appear may be determined as a frame interval corresponding to one action, so as to obtain each action, and then the action type corresponding to each action is determined according to the action type of the video frame in the frame interval.
In still another embodiment of the present invention, the video to be processed may be detected according to a preset motion detection algorithm to obtain each motion and a corresponding video frame, and then the motion type of the video frame with the largest duty ratio is determined as the motion type of the motion according to the proportion of the motion types of the corresponding video frames.
Step 30: and respectively determining the playback time of each action according to the action type.
In one embodiment of the invention, the playback time length is used for representing the degree of slow action of the action, and the longer the playback time length is, the more time is available for the user to enjoy the action, so that more detailed information of the action can be obtained. Considering that the watching interestingness of the user for different types of actions is different, aiming at actions with complex actions, greater action difficulty or more wonderful actions, the user tends to watch more details and longer time, therefore, the playing weight value of each action type can be determined according to the type characteristic information of the action type, and the playing weight value is used for representing the proportion of the playback time length corresponding to the action type in the total playback time length.
When determining the play weight value according to the type feature information, the information entropy corresponding to the action type can be determined according to the type feature information, and then the play weight value is determined according to the information entropy, wherein the information entropy is used for representing information which can be provided by the action type, and specifically, if the source symbol has n values: u1 … Ui … Un, the corresponding probability is: p1 … Pi … Pn, and the occurrence of various symbols are independent of each other. The average uncertainty of the source should be a statistical average (E) of the single symbol uncertainties-logPi, which may be referred to as information entropy. The greater the entropy of the information, the more information is provided, and the greater the uncertainty that can be eliminated, and therefore the greater the viewing weight of the action type for the user.
Thus, in yet another embodiment of the present invention, step 30 further comprises: step 301: and determining type characteristic information of each action type.
In one embodiment of the invention, a plurality of evaluation dimensions corresponding to the action type can be used as corresponding feature dimensions to obtain type feature information, wherein the type feature information comprises at least one of action complexity, popularity and action difficulty of the action type. The evaluation dimension may include an action complexity, an action attraction, and an action difficulty.
The action complexity is used for representing the number of frames required for completing the action, and the more the number of frames required for completing the action, the more the steps for representing the action, the more the details are contained, and the more enough frames are required to be allocated for slow release. The action attractiveness characterizes the user's level of interest in the action, with higher levels of interest indicating that the user wishes to see more details and playback of the action. The action difficulty represents the difficulty of realizing the action, and the greater the difficulty, the greater the visibility of the action, and enough frames should be allocated for slow release.
In still another embodiment of the present invention, the dimensional values under the multiple evaluation dimensions may be weighted and summed to obtain the play weight value corresponding to each action type.
Step 302: and determining the information entropy corresponding to each action type according to all the type characteristic information.
In one embodiment of the present invention, if the source symbol has n values, according to the following definition of the information entropy: u1 … Ui … Un, the corresponding probability is: p1 … Pi … Pn, and the occurrence of various symbols are independent of each other. The average uncertainty of the source should be a statistical average (E) of the single symbol uncertainties-logPi, which may be referred to as information entropy.
Specifically, the information entropy of each action type can be determined according to the dimension value of all action types under each type feature dimension and the number of actions under each action type. Wherein the probability of each action type is determined according to the proportion of the number of actions to the total number of all actions.
In yet another embodiment of the present invention, step 302 further comprises:
step 3021: and respectively determining play weight values corresponding to the action types according to at least one of the action complexity, the popularity and the action difficulty of the action types.
In one embodiment of the invention, the sum of at least one characteristic dimension of motion complexity, popularity and motion difficulty is performed according to all motion types, and then the play weight value is determined according to the proportion of the dimension value of each motion type in the sum of the dimension values under the characteristic dimension of at least one item.
Step 3022: and respectively determining the information entropy corresponding to each action type according to the play weight value.
In one embodiment of the present invention, according to the definition of the information entropy, an action type is determined as a value of a source symbol, and a play weight value is used as a corresponding probability of the source symbol corresponding to the action type, so as to calculate and obtain the information entropy corresponding to each action type.
Step 303: and determining the playback time according to the information entropy.
In one embodiment of the invention, the playback proportion is determined according to the information entropy, and the playback time length corresponding to each action is determined according to the playback proportion and the total playback time length corresponding to the video to be processed. Wherein, the playback proportion is determined according to the information entropy, and the playback proportion can be adjusted on the basis of the reference value of the basic playback proportion according to the information entropy. And determining the playback time length corresponding to each action respectively according to the product of the playback proportion corresponding to each action type and the total playback time length.
Thus, in yet another embodiment of the present invention, step 303 further comprises:
step 3031: and determining the playback proportion of each action type according to the information entropy.
In one embodiment of the present invention, a reference value of the playback ratio may be set, and the playback ratio of each action type may be obtained by adjusting the reference value according to the ratio of the information entropy of each action type.
Step 3032: and determining the total playback time length corresponding to the video to be processed.
In one embodiment of the invention, the total playback duration may be adaptively determined according to the length of the video to be processed.
Step 3033: and determining the playback time according to the playback proportion and the total playback time.
In one embodiment of the present invention, the playback time period may be determined by the following formula:
the information entropy H (U) is corresponding to different playback time t n And calculating the playback time corresponding to the different types of actions. Let the sortable action type be N, and for different action types, corresponding to different information entropies H (U), calculating the information entropy H (U) N for each action value, the information entropy corresponding to all kinds of actions is H (U) all,
let the total playback time of all action types be T, then the playback time T of the corresponding different actions n Is that
In still another embodiment of the present invention, the relationship between playback time length and the frame insertion information may be determined as follows:
t n =S n *(f en -f sn )/FPS;
wherein t is n For the playback duration of action n, S n For playback ratio of action n, f sn For the start frame number of action n, f en The FPS is the current frame rate for the ending frame number of action n.
Step 40: and respectively processing each action according to the playback time length to obtain a processed video.
In one embodiment of the invention, the frame inserting process can be performed on each action, and the playing time length of each action is adjusted to be the playback time length, wherein when the frame inserting process is performed, the frame inserting information corresponding to the corresponding action is determined according to the playback time length, the current playing parameters of each action and the time stamp information.
Specifically, the frame inserting information refers to parameters for performing frame inserting processing on the video to be processed, and may include, for example, a frame interval to be processed, a total inserted frame number, an inserted frame number corresponding to each action, and the like. The target frame rate is used for adjusting the number of frames of video contained in the action, and the larger the target frame rate is, the more picture frames can be received by a user in unit time, so that more details of the action can be appreciated.
Thus, in yet another embodiment of the present invention, step 40 further comprises: step 401: and determining the time stamp information and the current playing parameters of each action.
In one embodiment of the invention, the timestamp information of an action includes a sequence number of a corresponding start frame and a sequence number of an end frame of the action in the video to be processed. Specifically, the timestamp information may be determined according to the maximum value and the minimum value among all frame numbers corresponding to each action. The current play parameter includes a current frame rate for each of the actions. The current frame rate may be a current playing frame rate of the video to be processed.
Step 402: and determining the frame inserting information corresponding to each action according to the playback time length, the time stamp information and the current playing parameters.
In one embodiment of the present invention, the frame insertion information includes the number of frames inserted corresponding to each action, and is determined according to the following formula:
h n =t n *P n -(f en -f sn );
wherein h is n For the number of inserted frames corresponding to action n, P n For the target frame rate of action n, f sn For the start frame number of action n, f en An end frame number for action n.
For example, there are three actions M1, M2 and M3, where the frame number intervals are (2000,2050), (2051, 2075), (2076, 2100), respectively, the corresponding action types are C1, C2 and C3, respectively, and the corresponding playback ratios are S, respectively 1 =3,S 2 =2.5,S 3 =2, the target frame rates are P 1 =50,P 2 =30,P 3 The corresponding current frame rate is fps=25, and according to the foregoing steps, the playback durations corresponding to M1, M2, and M3 are respectively 6s, 2.5s, and 2s, and the corresponding frame insertion numbers are respectively 250, 50, and 35.
Step 403: and carrying out frame inserting processing on each action according to the frame inserting information to obtain the processed video.
In one embodiment of the invention, the video to be processed is subjected to frame inserting processing according to the number of frame inserting frames, so that the duration of the video segment corresponding to each action after frame inserting is the corresponding playback duration.
Considering that in the prior art, a mode of global slow-release processing is adopted for the video to be processed, which causes that the video is blocked when being slowly released and influences the watching experience of a user, in a further embodiment of the invention, the video to be processed can be firstly subjected to highlight frame screening, the screened slow-release processing is carried out, and the frame inserting information also comprises a frame interval to be processed; the action is located in the frame interval to be processed. The method further comprises, before step 403:
step 4031: and performing highlight motion recognition on the video to be processed to obtain the frame interval to be processed.
In one embodiment of the invention, image feature extraction is performed on the video to be processed, moving object recognition is performed according to the extracted image features, motion precision is determined according to motion feature information of the moving object, and a frame interval to be processed is determined from the video to be processed according to the motion precision of each motion. The image features may include RGB features, gray features, displacement features of pixels, and the like of each video frame, and the motion feature information may include information such as a motion type, a motion height, a motion speed, and a rotation number.
Step 50: and playing the processed video.
In one embodiment of the invention, in order to remind a user that the video to be processed includes an action which can be slowly released through frame insertion processing, the position of a frame interval to be processed for representing the occurrence of a highlight action can be marked, and a corresponding control is generated, so that the user can display the slowly released video of the corresponding action when operating the control.
Thus, in yet another embodiment of the present invention, step 50 further comprises:
step 501: and marking a frame interval in which the action after the frame insertion processing is positioned in the video to be processed to obtain a marked fragment.
In one embodiment of the present invention, the marking mode may be marking the frame interval where the action after the frame inserting process is located on the time axis, or may be generating the corresponding control to display on the frame interval where the action after the frame inserting process is located, so as to prompt the user to click and trigger the playback of the action. Wherein the generated control may be as shown with reference to fig. 2. The control may be to start the presentation when it is detected that the playing of the video to be processed starts on the user side.
Step 502: and playing the marked fragments when detecting the operation on the marked fragments.
In one embodiment of the invention, the detected operation on the markup fragment can be a user clicking on a timeline of the markup fragment or clicking on a function key in a control presented on the markup fragment that corresponds to a playback function. The playing mode of the mark segment may be in the form of a small window shown in fig. 3.
The video playing method provided by the embodiment of the invention obtains the video to be processed; the video to be processed comprises a plurality of actions; identifying the video to be processed to obtain the action types of all actions; respectively determining playback time of each action according to the action type; processing each action according to the playback time length to obtain a processed video; finally, playing the processed video, which is different from the problem of poor playing effect caused by global playback in the prior art, the video playing method provided by the embodiment of the invention can determine the playback time according to the action types of the actions included in the video to be processed, so that the playback with different time periods can be adaptively performed according to the characteristics of different actions, and a user can acquire action detail information matched with the action types, thereby improving the video watching experience of the user.
Fig. 4 is a schematic structural diagram of a video playing device according to an embodiment of the present invention. As shown in fig. 4, the apparatus 600 includes: the device comprises an acquisition module 601, an identification module 602, a determination module 603, a processing module 604 and a playing module 605.
The acquiring module 601 is configured to acquire a video to be processed; the video to be processed comprises a plurality of actions;
The identifying module 602 is configured to identify the video to be processed, and obtain an action type of each action;
a determining module 603, configured to determine playback durations of the respective actions according to the action types;
the processing module 604 is configured to process each of the actions according to the playback duration, so as to obtain a processed video;
and the playing module 605 is used for playing the processed video.
In an alternative manner, the video to be processed includes a plurality of video frames; one of the actions corresponds to at least one of the video frames; the identification module 602 is further configured to:
identifying the video to be processed to obtain image features corresponding to each video frame;
clustering the video frames according to the image characteristics to obtain a clustering result;
determining each action and the action type corresponding to the action according to the clustering result;
and determining the time stamp information of each action according to the frame sequence number of the video frame corresponding to each action.
In an alternative way, the determining module 603 is further configured to:
determining type characteristic information of each action type;
determining information entropy corresponding to each action type according to all the type characteristic information;
And determining the playback time according to the information entropy.
In an optional manner, the type feature information includes at least one of action complexity, popularity, and action difficulty of the action type; the determining module 603 is further configured to:
respectively determining play weight values corresponding to the action types according to at least one of the action complexity, the popularity and the action difficulty of the action types;
and respectively determining the information entropy corresponding to each action type according to the play weight value.
In an alternative way, the determining module 603 is further configured to:
determining the playback proportion of each action type according to the information entropy;
determining the total playback time length corresponding to the video to be processed;
and determining the playback time according to the playback proportion and the total playback time.
In an alternative approach, the identification module 602 is further configured to:
determining time stamp information and current playing parameters of each action;
determining the frame inserting information corresponding to each action according to the playback time length, the time stamp information and the current playing parameters;
and carrying out frame inserting processing on each action according to the frame inserting information to obtain the processed video.
In an alternative manner, the playing module 605 is further configured to:
marking a frame interval where the action after the frame insertion processing is located in the video to be processed to obtain a marked fragment;
and playing the marked fragments when detecting the operation on the marked fragments.
The video playing device provided by the embodiment of the invention obtains the video to be processed; the video to be processed comprises a plurality of actions; identifying the video to be processed to obtain the action types of all actions; respectively determining playback time of each action according to the action type; processing each action according to the playback time length to obtain a processed video; finally, the processed video is played, which is different from the problem of poor playing effect caused by global playback in the prior art, and the video playing device provided by the embodiment of the invention can determine the playback time according to the action types of the actions included in the video to be processed, so that the playback with different time is adaptively performed according to the characteristics of different actions, and a user can acquire action detail information matched with the action types, thereby improving the video watching experience of the user.
Fig. 5 shows a schematic structural diagram of a video playing device according to an embodiment of the present invention, and the embodiment of the present invention is not limited to the specific implementation of the video playing device.
As shown in fig. 5, the video playback device may include: a processor 702, a communication interface (Communications Interface), a memory 706, and a communication bus 708.
Wherein: processor 702, communication interface 704, and memory 706 perform communication with each other via a communication bus 708. A communication interface 704 for communicating with network elements of other devices, such as clients or other servers. The processor 702 is configured to execute the program 710, and may specifically perform the relevant steps in the embodiment of the video playing method described above.
In particular, program 710 may include program code including computer-executable instructions.
The processor 702 may be a Central Processing Unit (CPU), or a specific integrated circuit ASIC (Application Specific Integrated Circuit), or one or more integrated circuits configured to implement embodiments of the present invention. The one or more processors included in the video playback device may be the same type of processor, such as one or more CPUs; but may also be different types of processors such as one or more CPUs and one or more ASICs.
Memory 706 for storing programs 710. The memory 706 may comprise high-speed RAM memory or may further comprise non-volatile memory (non-volatile memory), such as at least one disk memory.
The program 710 may be specifically invoked by the processor 702 to cause the video playback device to:
acquiring a video to be processed; the video to be processed comprises a plurality of actions;
identifying the video to be processed to obtain action types of all actions;
respectively determining playback time of each action according to the action type;
processing each action according to the playback time length to obtain a processed video;
and playing the processed video.
In an alternative manner, the video to be processed includes a plurality of video frames; one of the actions corresponds to at least one of the video frames; the program 710 is invoked by the processor 702 to cause the video playback device to:
extracting features of the video to be processed to obtain image features corresponding to each video frame;
clustering the video frames according to the image characteristics to obtain a clustering result;
and determining each action and the action type corresponding to the action according to the clustering result. In an alternative, the program 710 is invoked by the processor 702 to cause the video playback device to:
Determining type characteristic information of each action type;
determining information entropy corresponding to each action type according to all the type characteristic information;
and determining the playback time according to the information entropy.
In an optional manner, the type feature information includes at least one of action complexity, popularity, and action difficulty of the action type; the program 710 is invoked by the processor 702 to cause the video playback device to:
respectively determining play weight values corresponding to the action types according to at least one of the action complexity, the popularity and the action difficulty of the action types;
and respectively determining the information entropy corresponding to each action type according to the play weight value.
In an alternative, the program 710 is invoked by the processor 702 to cause the video playback device to:
determining the playback proportion of each action type according to the information entropy;
determining the total playback time length corresponding to the video to be processed;
and determining the playback time according to the playback proportion and the total playback time.
In an alternative, the program 710 is invoked by the processor 702 to cause the video playback device to:
Determining time stamp information and current playing parameters of each action;
determining the frame inserting information corresponding to each action according to the playback time length, the time stamp information and the current playing parameters;
and carrying out frame inserting processing on each action according to the frame inserting information to obtain the processed video. In an alternative, the program 710 is invoked by the processor 702 to cause the video playback device to:
marking a frame interval where the action after the frame insertion processing is located in the video to be processed to obtain a marked fragment;
and playing the marked fragments when detecting the operation on the marked fragments.
The video playing device provided by the embodiment of the invention obtains the video to be processed; the video to be processed comprises a plurality of actions; identifying the video to be processed to obtain the action types of all actions; respectively determining playback time of each action according to the action type; processing each action according to the playback time length to obtain a processed video; finally, the processed video is played, which is different from the problem of poor playing effect caused by global playback in the prior art, and the video playing device provided by the embodiment of the invention can determine the playback time according to the action types of the actions included in the video to be processed, so that the playback with different time is adaptively performed according to the characteristics of different actions, and a user can acquire action detail information matched with the action types, thereby improving the video watching experience of the user.
An embodiment of the present invention provides a computer readable storage medium storing at least one executable instruction that, when executed on a video playback device, causes the video playback device to perform the video playback method in any of the method embodiments described above.
The executable instructions may be specifically configured to cause a video playback device to:
acquiring a video to be processed; the video to be processed comprises a plurality of actions;
identifying the video to be processed to obtain action types of all actions;
respectively determining playback time of each action according to the action type;
processing each action according to the playback time length to obtain a processed video;
and playing the processed video.
In an alternative manner, the video to be processed includes a plurality of video frames; one of the actions corresponds to at least one of the video frames; the executable instructions cause the video playback device to:
extracting features of the video to be processed to obtain image features corresponding to each video frame;
clustering the video frames according to the image characteristics to obtain a clustering result;
And determining each action and the action type corresponding to the action according to the clustering result.
In an alternative manner, the executable instructions cause the video playback device to:
determining type characteristic information of each action type;
determining information entropy corresponding to each action type according to all the type characteristic information;
and determining the playback time according to the information entropy.
In an optional manner, the type feature information includes at least one of action complexity, popularity, and action difficulty of the action type; the executable instructions cause the video playback device to:
respectively determining play weight values corresponding to the action types according to at least one of the action complexity, the popularity and the action difficulty of the action types;
and respectively determining the information entropy corresponding to each action type according to the play weight value.
In an alternative manner, the executable instructions cause the video playback device to:
determining the playback proportion of each action type according to the information entropy;
Determining the total playback time length corresponding to the video to be processed;
and determining the playback time according to the playback proportion and the total playback time.
In an alternative manner, the executable instructions cause the video playback device to:
determining time stamp information and current playing parameters of each action;
determining the frame inserting information corresponding to each action according to the playback time length, the time stamp information and the current playing parameters;
and carrying out frame inserting processing on each action according to the frame inserting information to obtain the processed video.
In an alternative manner, the executable instructions cause the video playback device to:
marking a frame interval where the action after the frame insertion processing is located in the video to be processed to obtain a marked fragment;
and playing the marked fragments when detecting the operation on the marked fragments.
The computer storage medium provided by the embodiment of the invention acquires the video to be processed; the video to be processed comprises a plurality of actions; identifying the video to be processed to obtain the action types of all actions; respectively determining playback time of each action according to the action type; processing each action according to the playback time length to obtain a processed video; finally, the processed video is played, which is different from the problem of poor playing effect caused by global playback in the prior art, and the computer storage medium provided by the embodiment of the invention can determine the playback time according to the action types of the actions included in the video to be processed, so that the playback with different time periods can be adaptively performed according to the characteristics of different actions, and a user can acquire action detail information matched with the action types, thereby improving the video watching experience of the user.
The embodiment of the invention provides a video playing device for executing the video playing method.
An embodiment of the present invention provides a computer program that can be invoked by a processor to cause a video playback device to perform the video playback method of any of the method embodiments described above.
An embodiment of the present invention provides a computer program product, including a computer program stored on a computer readable storage medium, the computer program including program instructions which, when run on a computer, cause the computer to perform the video playback method of any of the method embodiments described above.
The algorithms or displays presented herein are not inherently related to any particular computer, virtual system, or other apparatus. Various general-purpose systems may also be used with the teachings herein. The required structure for a construction of such a system is apparent from the description above. In addition, embodiments of the present invention are not directed to any particular programming language. It will be appreciated that the teachings of the present invention described herein may be implemented in a variety of programming languages, and the above description of specific languages is provided for disclosure of enablement and best mode of the present invention.
In the description provided herein, numerous specific details are set forth. However, it is understood that embodiments of the invention may be practiced without these specific details. In some instances, well-known methods, structures and techniques have not been shown in detail in order not to obscure an understanding of this description.
Similarly, it should be appreciated that in the above description of exemplary embodiments of the invention, various features of the embodiments of the invention are sometimes grouped together in a single embodiment, figure, or description thereof for the purpose of streamlining the disclosure and aiding in the understanding of one or more of the various inventive aspects. However, the disclosed method should not be construed as reflecting the intention that: i.e., the claimed invention requires more features than are expressly recited in each claim.
Those skilled in the art will appreciate that the modules in the apparatus of the embodiments may be adaptively changed and disposed in one or more apparatuses different from the embodiments. The modules or units or components of the embodiments may be combined into one module or unit or component, and they may be divided into a plurality of sub-modules or sub-units or sub-components. Any combination of all features disclosed in this specification (including any accompanying claims, abstract and drawings), and all of the processes or units of any method or apparatus so disclosed, may be used in combination, except insofar as at least some of such features and/or processes or units are mutually exclusive. Each feature disclosed in this specification (including any accompanying claims, abstract and drawings), may be replaced by alternative features serving the same, equivalent or similar purpose, unless expressly stated otherwise.
It should be noted that the above-mentioned embodiments illustrate rather than limit the invention, and that those skilled in the art will be able to design alternative embodiments without departing from the scope of the appended claims. In the claims, any reference signs placed between parentheses shall not be construed as limiting the claim. The word "comprising" does not exclude the presence of elements or steps not listed in a claim. The word "a" or "an" preceding an element does not exclude the presence of a plurality of such elements. The invention may be implemented by means of hardware comprising several distinct elements, and by means of a suitably programmed computer. In the unit claims enumerating several means, several of these means may be embodied by one and the same item of hardware. The use of the words first, second, third, etc. do not denote any order. These words may be interpreted as names. The steps in the above embodiments should not be construed as limiting the order of execution unless specifically stated.

Claims (9)

1. A video playing method, the method comprising:
acquiring a video to be processed; the video to be processed comprises a plurality of actions;
Identifying the video to be processed to obtain action types of the actions;
respectively determining playback time of each action according to the action type; the determining the playback time of the action according to the action type comprises the following steps:
determining type characteristic information of each action type;
determining information entropy corresponding to each action type according to all the type characteristic information;
determining the playback time length according to the information entropy;
processing each action according to the playback time length to obtain a processed video;
and playing the processed video.
2. The method of claim 1, wherein the video to be processed comprises a plurality of video frames; one of the actions corresponds to at least one of the video frames; the identifying the video to be processed to obtain the action types of the actions comprises the following steps:
extracting features of the video to be processed to obtain image features corresponding to each video frame;
clustering the video frames according to the image characteristics to obtain a clustering result;
and determining each action and the action type corresponding to the action according to the clustering result.
3. The method of claim 1, wherein the type feature information includes at least one of a motion complexity, a popularity, and a motion difficulty of the motion type; the determining the information entropy corresponding to each action type according to all the type characteristic information comprises the following steps:
respectively determining play weight values corresponding to the action types according to at least one of the action complexity, the popularity and the action difficulty of the action types;
and respectively determining the information entropy corresponding to each action type according to the play weight value.
4. The method of claim 1, wherein said determining said playback duration from said information entropy comprises:
determining the playback proportion of each action type according to the information entropy;
determining the total playback time length corresponding to the video to be processed;
and determining the playback time according to the playback proportion and the total playback time.
5. The method according to claim 1, wherein the processing the video to be processed according to the playback time length to obtain a processed video includes:
determining time stamp information and current playing parameters of each action;
Determining the frame inserting information corresponding to each action according to the playback time length, the time stamp information and the current playing parameters;
and carrying out frame inserting processing on each action according to the frame inserting information to obtain the processed video.
6. The method of claim 1, wherein playing the processed video comprises:
marking a frame interval where the action after the frame insertion processing is located in the video to be processed to obtain a marked fragment;
and playing the marked fragments when detecting the operation on the marked fragments.
7. A video playback device, the device comprising:
the acquisition module is used for acquiring the video to be processed; the video to be processed comprises a plurality of actions;
the identification module is used for identifying the video to be processed to obtain action types of the actions;
the determining module is used for determining the playback time length of each action according to the action type; the determining the playback time of the action according to the action type comprises the following steps:
determining type characteristic information of each action type;
determining information entropy corresponding to each action type according to all the type characteristic information;
Determining the playback time length according to the information entropy;
the processing module is used for respectively processing each action according to the playback time length to obtain a processed video;
and the playing module is used for playing the processed video.
8. A video playback device, comprising: the device comprises a processor, a memory, a communication interface and a communication bus, wherein the processor, the memory and the communication interface complete communication with each other through the communication bus;
the memory is configured to store at least one executable instruction that causes the processor to perform the operations of the video playing method according to any one of claims 1-6.
9. A computer readable storage medium having stored therein at least one executable instruction that, when executed on a video playback device, causes the video playback device to perform the operations of the video playback method of any one of claims 1-6.
CN202111677497.7A 2021-12-31 2021-12-31 Video playing method, device, equipment and computer storage medium Active CN114222165B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111677497.7A CN114222165B (en) 2021-12-31 2021-12-31 Video playing method, device, equipment and computer storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111677497.7A CN114222165B (en) 2021-12-31 2021-12-31 Video playing method, device, equipment and computer storage medium

Publications (2)

Publication Number Publication Date
CN114222165A CN114222165A (en) 2022-03-22
CN114222165B true CN114222165B (en) 2023-11-10

Family

ID=80707494

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111677497.7A Active CN114222165B (en) 2021-12-31 2021-12-31 Video playing method, device, equipment and computer storage medium

Country Status (1)

Country Link
CN (1) CN114222165B (en)

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2004091195A1 (en) * 2003-04-11 2004-10-21 Id (Event Marketing) Monaco Sam Method of and apparatus for providing a visual presentation
US9661384B1 (en) * 2016-04-05 2017-05-23 Arris Enterprises Llc Trick play user activity reconstruction
CN109964275A (en) * 2016-10-14 2019-07-02 乐威指南公司 For providing the system and method for slow motion video stream simultaneously with normal speed video flowing when detecting event
CN111147955A (en) * 2019-12-31 2020-05-12 咪咕视讯科技有限公司 Video playing method, server and computer readable storage medium
CN112153462A (en) * 2019-06-26 2020-12-29 腾讯科技(深圳)有限公司 Video processing method, device, terminal and storage medium
CN112422863A (en) * 2019-08-22 2021-02-26 华为技术有限公司 Intelligent video recording method and device
CN113014995A (en) * 2021-02-26 2021-06-22 北京蜂巢世纪科技有限公司 Video playback method, medium, glasses, and program product
CN113515997A (en) * 2020-12-28 2021-10-19 腾讯科技(深圳)有限公司 Video data processing method and device and readable storage medium

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104735385B (en) * 2015-03-31 2018-10-12 小米科技有限责任公司 Control method for playing back and device, electronic equipment

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2004091195A1 (en) * 2003-04-11 2004-10-21 Id (Event Marketing) Monaco Sam Method of and apparatus for providing a visual presentation
US9661384B1 (en) * 2016-04-05 2017-05-23 Arris Enterprises Llc Trick play user activity reconstruction
CN109964275A (en) * 2016-10-14 2019-07-02 乐威指南公司 For providing the system and method for slow motion video stream simultaneously with normal speed video flowing when detecting event
CN112153462A (en) * 2019-06-26 2020-12-29 腾讯科技(深圳)有限公司 Video processing method, device, terminal and storage medium
CN112422863A (en) * 2019-08-22 2021-02-26 华为技术有限公司 Intelligent video recording method and device
CN111147955A (en) * 2019-12-31 2020-05-12 咪咕视讯科技有限公司 Video playing method, server and computer readable storage medium
CN113515997A (en) * 2020-12-28 2021-10-19 腾讯科技(深圳)有限公司 Video data processing method and device and readable storage medium
CN113014995A (en) * 2021-02-26 2021-06-22 北京蜂巢世纪科技有限公司 Video playback method, medium, glasses, and program product

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
"Detection of slow-motion replay segments in sports video for highlights generation";H. Pan等;《2001 IEEE International Conference on Acoustics, Speech, and Signal Processing. Proceedings》;全文 *
"基于关键帧提取的视频智能快进技术";董乐乐;《中国优秀硕士学位论文全文数据库》;全文 *
MOOC在线学习行为建模与分析;周国昌;《中国硕士学位论文全文库》;全文 *

Also Published As

Publication number Publication date
CN114222165A (en) 2022-03-22

Similar Documents

Publication Publication Date Title
US9600717B1 (en) Real-time single-view action recognition based on key pose analysis for sports videos
CN110166827B (en) Video clip determination method and device, storage medium and electronic device
EP3473016B1 (en) Method and system for automatically producing video highlights
WO2021109952A1 (en) Video editing method, apparatus and server, and computer readable storage medium
CN111988638B (en) Method and device for acquiring spliced video, electronic equipment and storage medium
CN110505519A (en) A kind of video clipping method, electronic equipment and storage medium
CN107295362B (en) Live broadcast content screening method, device and equipment based on image and storage medium
EP2800057B1 (en) Mobile determination of properties of a trajectory of a ball
CN111491173A (en) Live broadcast cover determining method and device, computer equipment and storage medium
CN110309784A (en) Action recognition processing method, device, equipment and storage medium
CN112183588A (en) Video processing method and device, electronic equipment and storage medium
CN107454437A (en) A kind of video labeling method and its device, server
CN115396705B (en) Screen operation verification method, platform and system
CN114359343A (en) Motion trail management method, device and equipment and computer readable storage medium
CN110727629B (en) Playing method of audio electronic book, electronic equipment and computer storage medium
CN114222165B (en) Video playing method, device, equipment and computer storage medium
CN110287934B (en) Object detection method and device, client and server
US11917264B2 (en) Systems and methods for providing video enhancement for sporting events
CN113992976B (en) Video playing method, device, equipment and computer storage medium
CN111726649B (en) Video stream processing method, device, computer equipment and medium
US11749309B2 (en) Information processor, information processing method, and program
CN108121963B (en) Video data processing method and device and computing equipment
CN113971693A (en) Live broadcast picture generation method, system and device and electronic equipment
CN114584680A (en) Motion data display method and device, computer equipment and storage medium
CN113992975B (en) Video playing method, device, equipment and computer storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant