CN112492382A - Video frame extraction method and device, electronic equipment and storage medium - Google Patents

Video frame extraction method and device, electronic equipment and storage medium Download PDF

Info

Publication number
CN112492382A
CN112492382A CN202011308180.1A CN202011308180A CN112492382A CN 112492382 A CN112492382 A CN 112492382A CN 202011308180 A CN202011308180 A CN 202011308180A CN 112492382 A CN112492382 A CN 112492382A
Authority
CN
China
Prior art keywords
video
target video
frames
extracting
target
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202011308180.1A
Other languages
Chinese (zh)
Other versions
CN112492382B (en
Inventor
邢波
张磊磊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chuangsheng Shilian Digital Technology Beijing Co Ltd
Original Assignee
Chuangsheng Shilian Digital Technology Beijing Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chuangsheng Shilian Digital Technology Beijing Co Ltd filed Critical Chuangsheng Shilian Digital Technology Beijing Co Ltd
Priority to CN202011308180.1A priority Critical patent/CN112492382B/en
Publication of CN112492382A publication Critical patent/CN112492382A/en
Application granted granted Critical
Publication of CN112492382B publication Critical patent/CN112492382B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/44008Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream

Abstract

The embodiment of the application provides a video frame extraction method, which comprises the following steps: acquiring an application mode of a video frame, wherein the application mode of the video frame comprises a cover, auditing and dragging preview; when the application mode of the video frame is a cover, extracting one video frame in the target video; when the application mode of the video frames is auditing, extracting a first number of video frames according to first information of a target video, wherein the first information comprises a scene type of the target video, the scene type of the target video at least comprises comprehensive entertainment, education and live game broadcasting, and the first number is determined by the scene type of the target video; and when the application mode of the video frames is dragging preview, extracting a second number of video frames according to the duration of the target video, wherein the second number is determined by the duration of the target video. According to the video frame extraction method, the video frame extraction system and the storage medium, different methods are adopted for extraction according to different application modes of the video frame, so that the extracted video frame is more practical and closer to the user requirements.

Description

Video frame extraction method and device, electronic equipment and storage medium
Technical Field
The embodiment of the application relates to the field of multimedia, in particular to a video frame extraction method and device, electronic equipment and a storage medium.
Background
With the continuous development of multimedia technology, more and more users choose to watch videos on the network. Video frames have important roles in video, such as: a video playing provider generally needs to check the content of a video before uploading the video to a network, so as to prevent the video related to yellow storm from being uploaded, and the checking of the content of the video is generally realized by extracting and checking a plurality of video frames in the video; for another example: a video provider extracts a certain video frame in the video to be used as a cover so as to improve the click intention of a user; for another example: a video playing provider can extract a plurality of video frames in a video and provide a dragging preview function for a user so as to improve the watching experience of the user. Generally, a video or a video playing provider extracts video frames according to experience, and the extraction mode is single.
Disclosure of Invention
In view of the above, in a first aspect, an embodiment of the present application provides a method for extracting a video frame, including:
acquiring an application mode of a video frame, wherein the application mode of the video frame comprises a cover, auditing and dragging preview;
when the application mode of the video frame is a cover, extracting one video frame in the target video;
when the application mode of the video frames is auditing, extracting a first number of video frames according to first information of a target video, wherein the first information comprises a scene type of the target video, the scene type of the target video at least comprises comprehensive entertainment, education and live game broadcasting, and the first number is determined by the scene type of the target video;
and when the application mode of the video frames is dragging preview, extracting a second number of video frames according to the duration of the target video, wherein the second number is determined by the duration of the target video.
Optionally, in a specific embodiment, when the application mode of the video frame is a cover page, extracting one video frame in the target video includes: and extracting all elements in a first element class and video frames with highest image quality in the target video, wherein the first element class is a set of elements with the occurrence frequency larger than a first threshold value in the target video, and the first threshold value is a value which enables the first element class to at least comprise one element.
Optionally, in a specific embodiment, the first information is determined according to a change frequency of the target video.
Optionally, in a specific embodiment, the determining the first information according to the change frequency of the target video includes:
extracting video frames in a target video at equal intervals;
taking the proportion of key frames in the video frames extracted at equal intervals as the change frequency of the target video;
the first information is determined according to the change frequency of the target video.
Optionally, in a specific embodiment, the first number is determined by a scene type of the target video, and includes:
when the scene type of the target video is the entertainment, the first number is at least 12 frames;
when the scene type of the target video is education, the first number is at least 8 frames;
when the scene type of the target video is game live, the first number is at least 24 frames.
Optionally, in a specific embodiment, the second number is determined by a duration of the target video, and includes:
when the duration of the target video is greater than 10 seconds, taking the result of rounding up the quotient of the duration of the target video and 10 seconds as a second quantity;
when the duration of the target video is less than 10s, the second number is greater than or equal to 2 and less than or equal to the number of video frames in the target video.
Optionally, in a specific embodiment, when the application mode of the video frames is a drag preview, extracting a second number of video frames according to the duration of the target video includes:
determining a second video frame extraction interval according to the duration and the second number of the target video;
a second number of video frames is extracted according to a second video frame extraction interval.
Optionally, in a specific embodiment, when the application mode of the video frames is a drag preview, extracting a second number of video frames according to the duration of the target video includes:
dividing the target video into a second number of sub-videos;
extracting key frames in each section of sub-video and respectively forming dynamic pictures to obtain a second number of video frames;
in a second aspect, an embodiment of the present application provides a video frame extraction apparatus, including:
the acquisition module is used for acquiring the application mode of the video frame, wherein the application mode of the video frame comprises a cover, audit and dragging preview;
the cover extraction module is used for extracting one video frame in the target video when the application mode of the video frame is the cover;
the auditing and extracting module is used for extracting a first number of video frames according to first information of the target video when the application mode of the video frames is auditing, wherein the first information comprises the scene type of the target video, the scene type of the target video at least comprises comprehensive entertainment, education and live game broadcast, and the first number is determined by the scene type of the target video;
and the preview extraction module is used for extracting a second number of video frames according to the duration of the target video when the application mode of the video frames is dragging preview, wherein the second number is determined by the duration of the target video.
Optionally, in a specific embodiment, the audit extraction module is further configured to determine the first information according to a change frequency of the target video.
Optionally, in a specific embodiment, the audit extraction module is further configured to determine the first information according to a change frequency of the target video, including:
extracting video frames in a target video at equal intervals;
taking the proportion of key frames in the video frames extracted at equal intervals as the change frequency of the target video;
the first information is determined according to the change frequency of the target video.
In a third aspect, an embodiment of the present application provides an electronic device, including a memory and a processor, where the memory stores an executable program, and the processor executes the executable program to perform the following steps:
acquiring an application mode of a video frame, wherein the application mode of the video frame comprises a cover, auditing and dragging preview;
when the application mode of the video frame is a cover, extracting one video frame in the target video;
when the application mode of the video frames is auditing, extracting a first number of video frames according to first information of a target video, wherein the first information comprises a scene type of the target video, the scene type of the target video at least comprises comprehensive entertainment, education and live game broadcasting, and the first number is determined by the scene type of the target video;
and when the application mode of the video frames is dragging preview, extracting a second number of video frames according to the duration of the target video, wherein the second number is determined by the duration of the target video.
Optionally, in a specific embodiment, the method further includes:
the first information is determined according to the change frequency of the target video.
Optionally, in a specific embodiment, the determining the first information according to the change frequency of the target video includes:
extracting video frames in a target video at equal intervals;
taking the proportion of key frames in the video frames extracted at equal intervals as the change frequency of the target video;
the first information is determined according to the change frequency of the target video.
In a fourth aspect, the present application provides a storage medium having a computer program stored thereon, where the computer program is executed by a processor to implement the method according to any one of the first aspect.
6. The method of claim 1, wherein the second number is determined by a duration of the target video, and comprises:
when the duration of the target video is greater than 10 seconds, taking the result of rounding up the quotient of the duration of the target video and 10 seconds as a second number;
when the duration of the target video is less than 10s, the second number is greater than or equal to 2 and less than or equal to the number of video frames in the target video.
10. The video frame extraction apparatus of claim 9, wherein the audit extraction module is further configured to determine the first information according to a change frequency of the target video.
11. The apparatus according to claim 10, wherein the audit extraction module is further configured to determine first information according to a change frequency of the target video, including:
extracting video frames in the target video at equal intervals;
taking the proportion of key frames in the video frames extracted at equal intervals as the change frequency of the target video;
and determining first information according to the change frequency of the target video.
13. The electronic device of claim 12, further comprising:
and determining first information according to the change frequency of the target video.
14. The electronic device of claim 12, wherein determining the first information according to the change frequency of the target video comprises:
extracting video frames in the target video at equal intervals;
taking the proportion of key frames in the video frames extracted at equal intervals as the change frequency of the target video;
and determining first information according to the change frequency of the target video.
According to the video frame extraction method and device, the electronic equipment and the storage medium, the video frame is extracted by adopting different methods according to different application modes of the video frame, so that the extracted video frame is more practical and closer to the user requirements.
Drawings
Some specific embodiments of the present application will be described in detail hereinafter by way of illustration and not limitation with reference to the accompanying drawings. The same reference numbers in the drawings identify the same or similar elements or components. Those skilled in the art will appreciate that the drawings are not necessarily drawn to scale. In the drawings:
fig. 1 is a flowchart of a video frame extraction method according to an embodiment of the present disclosure;
fig. 2 is a flowchart of a method for determining first information according to an embodiment of the present disclosure;
fig. 3 is a schematic diagram of a video frame extracting apparatus according to an embodiment of the present application;
fig. 4 is a schematic view of an electronic device according to an embodiment of the present application.
Detailed Description
The following further describes a specific implementation of the embodiments of the present application with reference to the drawings of the embodiments of the present application.
Example one
Referring to fig. 1, an embodiment of the present application provides a method for extracting a video frame, including:
s101: acquiring an application mode of a video frame, wherein the application mode of the video frame comprises a cover, auditing and dragging preview;
specifically, the application mode of the video frame is a cover page, which means that the extracted video frame is used as a cover page of the target video; the application mode of the video frames is auditing, namely whether the corresponding target video is illegal or illegal is judged by auditing whether the extracted video frames are illegal or illegal; the application mode of the video frames is that dragging preview refers to extracting representative video frames in the target video, so that a user can preview the representative video frames in advance to know the video content approximately.
S102 a: when the application mode of the video frame is a cover, extracting one video frame in the target video;
generally, selecting a video frame, which is the most capable of expressing the content of the target video, from the videos as a cover page can improve the click intention of the target user. Optionally, in a specific embodiment, when the application mode of the video frame is a cover page, extracting one video frame in the target video includes: and extracting all elements in a first element class and video frames with highest image quality in the target video, wherein the first element class is a set of elements with the occurrence frequency larger than a first threshold value in the target video, and the first threshold value is a value which enables the first element class to at least comprise one element.
In detail, the target video may be a person, the frequency of occurrence of each person in the target video may be obtained through AI image recognition, in a preferred embodiment, the person with the highest frequency of occurrence is used as a main person, the person with the second highest frequency of occurrence is used as a secondary person, the first threshold is between the frequency of occurrence of the secondary person and the third highest frequency of occurrence, and a video frame with the highest image quality, i.e., the highest resolution, which includes both the main person and the secondary person in the target video is extracted as a cover of the target video. Therefore, the method is beneficial to obtaining the video frame which can represent the content of the target video most in the target video, and is beneficial to improving the click intention of the target user. Of course, the elements in the target video may also be scenes or objects, which is not limited in this application.
S102 b: when the application mode of the video frames is auditing, extracting a first number of video frames according to first information of a target video, wherein the first information comprises a scene type of the target video, the scene type of the target video at least comprises comprehensive entertainment, education and live game broadcasting, and the first number is determined by the scene type of the target video;
specifically, the scene type of the target video may be determined in various ways, for example, based on a speech recognition technology, speech in the target video may be recognized, and keywords may be extracted to obtain speech content, so as to determine the scene type of the video according to the speech content of the target video; for another example, the image information in the target video may be identified based on an image processing technique, and the target video may be classified according to the image information, so as to determine the scene type of the target video; for another example, the scene type of the target video may be determined by recognizing the character information in the target video based on a face recognition technology.
Optionally, in a specific embodiment, the first information is determined according to a change frequency of the target video.
Referring to fig. 2, optionally, in a specific embodiment, the determining the first information according to the variation frequency of the target video includes:
extracting video frames in a target video at equal intervals;
taking the proportion of key frames in the video frames extracted at equal intervals as the change frequency of the target video;
the first information is determined according to the change frequency of the target video.
In particular, a key frame of a target video refers to the frame in which a key action in a character or object motion or change in the target video is located.
Specifically, a mapping relationship between the change frequency and the first information may be pre-established, for example, a mapping relationship between a scene type of a target video with the change frequency located in a [ a, b ] interval and the entertainment, a mapping relationship between a scene type of a target video with the change frequency located in a [ c, d ] interval and the education, a mapping relationship between a scene type of a target video with the change frequency located in a [ e, f ] interval and the live game broadcast may be established, and a, b, c, d, e, f are numerical values representing frequencies, which may be preset by a person as needed, and the present application does not limit this. Therefore, the scene type of the target video is associated with the change frequency of the target video, and different scene types are determined according to different change frequencies, so that the accuracy of the auditing result is improved.
Optionally, in a specific embodiment, the first number is determined by a scene type of the target video, and includes:
when the scene type of the target video is the entertainment, the first number is at least 12 frames;
when the scene type of the target video is education, the first number is at least 8 frames;
when the scene type of the target video is game live, the first number is at least 24 frames.
Therefore, the video frame representativeness of the selected target video is higher, the video content can be reflected more comprehensively, and the accuracy of the auditing result can be improved.
When the application mode of the video frame is the auditing mode, other types of modes for extracting the video frame can be adopted, for example, a main character and a secondary character are extracted based on AI image recognition, the number of pictures respectively required by the main character and the secondary character is calculated according to the weight preset by the user, and the video frames including the main character and the secondary character are respectively extracted. For example, if the user sets the primary character weight to 3, the secondary character weight to 1, and the first number to 8, 6 pictures of the primary character and 2 pictures of the secondary character are obtained. The risk of the content of the video frame can be scored through machine learning based on AI image recognition, for example, the frame score related to yellow blast can be set to be the highest, and finally, the video frames ten or twenty before the score are extracted for auditing according to the score from high to low, and of course, other numbers can be extracted, which is only an exemplary illustration and does not represent that the application is limited thereto.
S102 c: and when the application mode of the video frames is dragging preview, extracting a second number of video frames according to the duration of the target video, wherein the second number is determined by the duration of the target video.
Specifically, a mapping relation between the duration of the target video and the second number may be established, and the mapping relation satisfies that the duration of the target video is positively correlated with the second number, and the longer the duration of the target video is, the larger the second number is, and preferably, the second number is in a direct proportional functional relationship or a linear functional relationship with the duration of the target video. Therefore, the number of the video frames to be extracted is related to the duration of the target video, and the video frames with different numbers are extracted according to the videos with different durations and serve as the video frames for dragging preview, so that the content of the videos can be more completely and truly displayed to the user.
Optionally, in a specific embodiment, the second number is determined by a duration of the target video, and includes:
when the duration of the target video is greater than 10 seconds, taking the result of rounding up the quotient of the duration of the target video and 10 seconds as a second quantity;
when the duration of the target video is less than 10s, the second number is greater than or equal to 2 and less than or equal to the number of video frames in the target video.
In this manner, the content of the video is advantageously more completely and realistically presented to the user.
Optionally, in a specific embodiment, when the application mode of the video frames is a drag preview, extracting a second number of video frames according to the duration of the target video includes:
determining a second video frame extraction interval according to the duration and the second number of the target video;
a second number of video frames is extracted according to a second video frame extraction interval.
Specifically, the quotient of the duration of the target video and the second number can be used as a second video frame extraction interval, and the video frames are extracted at equal intervals according to the second video frame extraction interval, so that the obtained second number of video frames can more completely and truly show the content of the video to the user.
It should be noted that the second number of video frames can be combined into a large map to save transmission resources.
Optionally, in a specific embodiment, when the application mode of the video frames is a drag preview, extracting a second number of video frames according to the duration of the target video includes:
dividing the target video into a second number of sub-videos;
and extracting key frames in each section of the sub-video and respectively forming dynamic pictures to obtain a second number of video frames.
Preferably, the target video may be divided into a second number of segments of sub-videos at equal intervals, at least one key frame in each segment of sub-video is extracted, and each key frame constitutes a dynamic picture, where the dynamic picture may be in gif format or in other formats, which is not limited in this application. Therefore, when the user performs dragging preview, the preview effect is dynamic, the method is more friendly to the user, and the method is favorable for improving the click willingness of the user.
In a specific embodiment, the target video may be further divided into multiple segments of sub-video at intervals of 10 seconds, each 10-second segment of sub-video, all key frames of the first two seconds are intercepted, and the key frames are extracted and kept as gif images, so as to obtain a second number of video frames. Therefore, when the user performs dragging preview, the preview effect is dynamic, the method is more friendly to the user, and the method is favorable for improving the click willingness of the user.
Example two
Referring to fig. 3, an embodiment of the present invention provides a video frame extracting apparatus 20, including:
the acquisition module 201 is configured to acquire an application mode of a video frame, where the application mode of the video frame includes a cover, review, and drag preview;
specifically, the application mode of the video frame is a cover page, which means that the extracted video frame is used as a cover page of the target video; the application mode of the video frames is auditing, namely whether the corresponding target video is illegal or illegal is judged by auditing whether the extracted video frames are illegal or illegal; the application mode of the video frames is that dragging preview refers to extracting representative video frames in the target video, so that a user can preview the representative video frames in advance to know the video content approximately.
The cover extraction module 202, the cover extraction module 202 is configured to extract one video frame in the target video when the application mode of the video frame is a cover;
generally, selecting a video frame, which is the most capable of expressing the content of the target video, from the videos as a cover page can improve the click intention of the target user. Optionally, in a specific embodiment, when the application mode of the video frame is a cover page, extracting one video frame in the target video includes: and extracting all elements in a first element class and video frames with highest image quality in the target video, wherein the first element class is a set of elements with the occurrence frequency larger than a first threshold value in the target video, and the first threshold value is a value which enables the first element class to at least comprise one element.
In detail, the target video may be a person, the frequency of occurrence of each person in the target video may be obtained through AI image recognition, in a preferred embodiment, the person with the highest frequency of occurrence is used as a main person, the person with the second highest frequency of occurrence is used as a secondary person, the first threshold is between the frequency of occurrence of the secondary person and the third highest frequency of occurrence, and a video frame with the highest image quality, i.e., the highest resolution, which includes both the main person and the secondary person in the target video is extracted as a cover of the target video. Therefore, the method is beneficial to obtaining the video frame which can represent the content of the target video most in the target video, and is beneficial to improving the click intention of the target user. Of course, the elements in the target video may also be scenes or objects, which is not limited in this application.
The auditing and extracting module 203 is used for extracting a first number of video frames according to first information of the target video when the application mode of the video frames is auditing, wherein the first information comprises a scene type of the target video, the scene type of the target video at least comprises comprehensive entertainment, education and live game broadcast, and the first number is determined by the scene type of the target video;
specifically, the scene type of the target video may be determined in various ways, for example, based on a speech recognition technology, speech in the target video may be recognized, and keywords may be extracted to obtain speech content, so as to determine the scene type of the video according to the speech content of the target video; for another example, the image information in the target video may be identified based on an image processing technique, and the target video may be classified according to the image information, so as to determine the scene type of the target video; for another example, the scene type of the target video may be determined by recognizing the character information in the target video based on a face recognition technology.
Optionally, in a specific embodiment, the audit extraction module 203 is further configured to determine the first information according to a change frequency of the target video.
Optionally, in a specific embodiment, the audit extraction module 203 is further configured to determine first information according to a change frequency of the target video, including:
extracting video frames in a target video at equal intervals;
taking the proportion of key frames in the video frames extracted at equal intervals as the change frequency of the target video;
the first information is determined according to the change frequency of the target video.
Specifically, the key frame of the target video specifically refers to the frame in which the key action in the character or object motion or change in the target video is located.
Specifically, a mapping relationship between the change frequency and the first information may be pre-established, for example, a mapping relationship between a scene type of a target video with the change frequency located in a [ a, b ] interval and the entertainment, a mapping relationship between a scene type of a target video with the change frequency located in a [ c, d ] interval and the education, a mapping relationship between a scene type of a target video with the change frequency located in a [ e, f ] interval and the live game broadcast may be established, and a, b, c, d, e, f are numerical values representing frequencies, which may be preset by a person as needed, and the present application does not limit this. Therefore, the scene type of the target video is associated with the change frequency of the target video, and different scene types are determined according to different change frequencies, so that the accuracy of the auditing result is improved.
Optionally, in a specific embodiment, the first number is determined by a scene type of the target video, and includes:
when the scene type of the target video is the entertainment, the first number is at least 12 frames;
when the scene type of the target video is education, the first number is at least 8 frames;
when the scene type of the target video is game live, the first number is at least 24 frames.
Therefore, the video frame representativeness of the selected target video is higher, the video content can be reflected more comprehensively, and the accuracy of the auditing result can be improved.
And the preview extracting module 204, where the preview extracting module 204 is configured to extract a second number of video frames according to the duration of the target video when the application mode of the video frames is dragging preview, where the second number is determined by the duration of the target video.
Specifically, a mapping relation between the duration of the target video and the second number may be established, and the mapping relation satisfies that the duration of the target video is positively correlated with the second number, and the longer the duration of the target video is, the larger the second number is, and preferably, the second number is in a direct proportional functional relationship or a linear functional relationship with the duration of the target video. Therefore, the number of the video frames to be extracted is related to the duration of the target video, and the video frames with different numbers are extracted according to the videos with different durations and serve as the video frames for dragging preview, so that the content of the videos can be more completely and truly displayed to the user.
Optionally, in a specific embodiment, the second number is determined by a duration of the target video, and includes:
when the duration of the target video is greater than 10 seconds, taking the result of rounding up the quotient of the duration of the target video and 10 seconds as a second quantity;
when the duration of the target video is less than 10s, the second number is greater than or equal to 2 and less than or equal to the number of video frames in the target video.
In this manner, the content of the video is advantageously more completely and realistically presented to the user.
Optionally, in a specific embodiment, when the application mode of the video frames is a drag preview, extracting a second number of video frames according to the duration of the target video includes:
determining a second video frame extraction interval according to the duration and the second number of the target video;
a second number of video frames is extracted according to a second video frame extraction interval.
Specifically, the quotient of the duration of the target video and the second number can be used as a second video frame extraction interval, and the video frames are extracted at equal intervals according to the second video frame extraction interval, so that the obtained second number of video frames can more completely and truly show the content of the video to the user.
It should be noted that the second number of video frames can be combined into a large map to save transmission resources.
Optionally, in a specific embodiment, when the application mode of the video frames is a drag preview, extracting a second number of video frames according to the duration of the target video includes:
dividing the target video into a second number of sub-videos;
and extracting key frames in each section of the sub-video and respectively forming dynamic pictures to obtain a second number of video frames.
Preferably, the target video may be divided into a second number of segments of sub-videos at equal intervals, at least one key frame in each segment of sub-video is extracted, and each key frame constitutes a dynamic picture, where the dynamic picture may be in gif format or in other formats, which is not limited in this application. Therefore, when the user performs dragging preview, the preview effect is dynamic, the method is more friendly to the user, and the method is favorable for improving the click willingness of the user.
In a specific embodiment, the target video may be further divided into multiple segments of sub-video at intervals of 10 seconds, each 10-second segment of sub-video, all key frames of the first two seconds are intercepted, and the key frames are extracted and kept as gif images, so as to obtain a second number of video frames. Therefore, when the user performs dragging preview, the preview effect is dynamic, the method is more friendly to the user, and the method is favorable for improving the click willingness of the user.
EXAMPLE III
Referring to fig. 4, an electronic device 30 according to an embodiment of the present application includes a memory 301 and a processor 302, where the memory 301 stores an executable program, and the processor 302 executes the executable program to perform the following steps:
acquiring an application mode of a video frame, wherein the application mode of the video frame comprises a cover, auditing and dragging preview;
specifically, the application mode of the video frame is a cover page, which means that the extracted video frame is used as a cover page of the target video; the application mode of the video frames is auditing, namely whether the corresponding target video is illegal or illegal is judged by auditing whether the extracted video frames are illegal or illegal; the application mode of the video frames is that dragging preview refers to extracting representative video frames in the target video, so that a user can preview the representative video frames in advance to know the video content approximately.
When the application mode of the video frame is a cover, extracting one video frame in the target video;
generally, selecting a video frame, which is the most capable of expressing the content of the target video, from the videos as a cover page can improve the click intention of the target user. Optionally, in a specific embodiment, when the application mode of the video frame is a cover page, extracting one video frame in the target video includes: and extracting all elements in a first element class and video frames with highest image quality in the target video, wherein the first element class is a set of elements with the occurrence frequency larger than a first threshold value in the target video, and the first threshold value is a value which enables the first element class to at least comprise one element.
In detail, the target video may be a person, the frequency of occurrence of each person in the target video may be obtained through AI image recognition, in a preferred embodiment, the person with the highest frequency of occurrence is used as a main person, the person with the second highest frequency of occurrence is used as a secondary person, the first threshold is between the frequency of occurrence of the secondary person and the third highest frequency of occurrence, and a video frame with the highest image quality, i.e., the highest resolution, which includes both the main person and the secondary person in the target video is extracted as a cover of the target video. Therefore, the method is beneficial to obtaining the video frame which can represent the content of the target video most in the target video, and is beneficial to improving the click intention of the target user. Of course, the elements in the target video may also be scenes or objects, which is not limited in this application.
When the application mode of the video frames is auditing, extracting a first number of video frames according to first information of a target video, wherein the first information comprises a scene type of the target video, the scene type of the target video at least comprises comprehensive entertainment, education and live game broadcasting, and the first number is determined by the scene type of the target video;
specifically, the scene type of the target video may be determined in various ways, for example, based on a speech recognition technology, speech in the target video may be recognized, and keywords may be extracted to obtain speech content, so as to determine the scene type of the video according to the speech content of the target video; for another example, the image information in the target video may be identified based on an image processing technique, and the target video may be classified according to the image information, so as to determine the scene type of the target video; for another example, the scene type of the target video may be determined by recognizing the character information in the target video based on a face recognition technology.
Optionally, in a specific embodiment, the first information is determined according to a change frequency of the target video.
Optionally, in a specific embodiment, the determining the first information according to the change frequency of the target video includes:
extracting video frames in a target video at equal intervals;
taking the proportion of key frames in the video frames extracted at equal intervals as the change frequency of the target video;
the first information is determined according to the change frequency of the target video.
Specifically, the key frame of the target video specifically refers to the frame in which the key action in the character or object motion or change in the target video is located.
Specifically, a mapping relationship between the change frequency and the first information may be pre-established, for example, a mapping relationship between a scene type of a target video with the change frequency located in a [ a, b ] interval and the entertainment, a mapping relationship between a scene type of a target video with the change frequency located in a [ c, d ] interval and the education, a mapping relationship between a scene type of a target video with the change frequency located in a [ e, f ] interval and the live game broadcast may be established, and a, b, c, d, e, f are numerical values representing frequencies, which may be preset by a person as needed, and the present application does not limit this. Therefore, the scene type of the target video is associated with the change frequency of the target video, and different scene types are determined according to different change frequencies, so that the accuracy of the auditing result is improved.
Optionally, in a specific embodiment, the first number is determined by a scene type of the target video, and includes:
when the scene type of the target video is the entertainment, the first number is at least 12 frames;
when the scene type of the target video is education, the first number is at least 8 frames;
when the scene type of the target video is game live, the first number is at least 24 frames.
Therefore, the video frame representativeness of the selected target video is higher, the video content can be reflected more comprehensively, and the accuracy of the auditing result can be improved.
And when the application mode of the video frames is dragging preview, extracting a second number of video frames according to the duration of the target video, wherein the second number is determined by the duration of the target video.
Specifically, a mapping relation between the duration of the target video and the second number may be established, and the mapping relation satisfies that the duration of the target video is positively correlated with the second number, and the longer the duration of the target video is, the larger the second number is, and preferably, the second number is in a direct proportional functional relationship or a linear functional relationship with the duration of the target video. Therefore, the number of the video frames to be extracted is related to the duration of the target video, and the video frames with different numbers are extracted according to the videos with different durations and serve as the video frames for dragging preview, so that the content of the videos can be more completely and truly displayed to the user.
Optionally, in a specific embodiment, the second number is determined by a duration of the target video, and includes:
when the duration of the target video is greater than 10 seconds, taking the result of rounding up the quotient of the duration of the target video and 10 seconds as a second quantity;
when the duration of the target video is less than 10s, the second number is greater than or equal to 2 and less than or equal to the number of video frames in the target video.
In this manner, the content of the video is advantageously more completely and realistically presented to the user.
Optionally, in a specific embodiment, when the application mode of the video frames is a drag preview, extracting a second number of video frames according to the duration of the target video includes:
determining a second video frame extraction interval according to the duration and the second number of the target video;
a second number of video frames is extracted according to a second video frame extraction interval.
Specifically, the quotient of the duration of the target video and the second number can be used as a second video frame extraction interval, and the video frames are extracted at equal intervals according to the second video frame extraction interval, so that the obtained second number of video frames can more completely and truly show the content of the video to the user.
It should be noted that the second number of video frames can be combined into a large map to save transmission resources.
Optionally, in a specific embodiment, when the application mode of the video frames is a drag preview, extracting a second number of video frames according to the duration of the target video includes:
dividing the target video into a second number of sub-videos;
and extracting key frames in each section of the sub-video and respectively forming dynamic pictures to obtain a second number of video frames.
Preferably, the target video may be divided into a second number of segments of sub-videos at equal intervals, at least one key frame in each segment of sub-video is extracted, and each key frame constitutes a dynamic picture, where the dynamic picture may be in gif format or in other formats, which is not limited in this application. Therefore, when the user performs dragging preview, the preview effect is dynamic, the method is more friendly to the user, and the method is favorable for improving the click willingness of the user.
In a specific embodiment, the target video may be further divided into multiple segments of sub-video at intervals of 10 seconds, each 10-second segment of sub-video, all key frames of the first two seconds are intercepted, and the key frames are extracted and kept as gif images, so as to obtain a second number of video frames. Therefore, when the user performs dragging preview, the preview effect is dynamic, the method is more friendly to the user, and the method is favorable for improving the click willingness of the user.
Example four
The embodiment of the application provides a storage medium, wherein a computer program is stored on the storage medium, and when a processor executes the computer program, the method of any one of the embodiments is realized.
According to the video frame extraction method and device, the electronic equipment and the storage medium, the video frame is extracted by adopting different methods according to different application modes of the video frame, so that the extracted video frame is more practical and closer to the user requirements.
The storage medium of the embodiments of the present application exists in various forms, including but not limited to:
(1) a mobile communication device: such devices are characterized by mobile communications capabilities and are primarily targeted at providing voice, data communications. Such terminals include: smart phones (e.g., iphones), multimedia phones, functional phones, and low-end phones, among others.
(2) Ultra mobile personal computer device: the equipment belongs to the category of personal computers, has calculation and processing functions and generally has the characteristic of mobile internet access. Such terminals include: PDA, MID, and UMPC devices, etc., such as ipads.
(3) A portable entertainment device: such devices can display and play multimedia content. This type of device comprises: audio, video players (e.g., ipods), handheld game consoles, electronic books, and smart toys and portable car navigation devices.
(4) And other electronic equipment with data interaction function.
Thus, particular embodiments of the present subject matter have been described. Other embodiments are within the scope of the following claims. In some cases, the actions recited in the claims can be performed in a different order and still achieve desirable results. In addition, the processes depicted in the accompanying figures do not necessarily require the particular order shown, or sequential order, to achieve desirable results. In some embodiments, multitasking and parallel processing may be advantageous.
In the 90 s of the 20 th century, improvements in a technology could clearly distinguish between improvements in hardware (e.g., improvements in circuit structures such as diodes, transistors, switches, etc.) and improvements in software (improvements in process flow). However, as technology advances, many of today's process flow improvements have been seen as direct improvements in hardware circuit architecture. Designers almost always obtain the corresponding hardware circuit structure by programming an improved method flow into the hardware circuit. Thus, it cannot be said that an improvement in the process flow cannot be realized by hardware physical modules. For example, a Programmable Logic Device (PLD), such as a Field Programmable Gate Array (FPGA), is an integrated circuit whose Logic functions are determined by programming the Device by a user. A digital system is "integrated" on a PLD by the designer's own programming without requiring the chip manufacturer to design and fabricate application-specific integrated circuit chips. Furthermore, nowadays, instead of manually making an Integrated Circuit chip, such Programming is often implemented by "logic compiler" software, which is similar to a software compiler used in program development and writing, but the original code before compiling is also written by a specific Programming Language, which is called Hardware Description Language (HDL), and HDL is not only one but many, such as abel (advanced Boolean Expression Language), ahdl (alternate Hardware Description Language), traffic, pl (core universal Programming Language), HDCal (jhdware Description Language), lang, Lola, HDL, laspam, hardward Description Language (vhr Description Language), vhal (Hardware Description Language), and vhigh-Language, which are currently used in most common. It will also be apparent to those skilled in the art that hardware circuitry that implements the logical method flows can be readily obtained by merely slightly programming the method flows into an integrated circuit using the hardware description languages described above.
The controller may be implemented in any suitable manner, for example, the controller may take the form of, for example, a microprocessor or processor and a computer-readable medium storing computer-readable program code (e.g., software or firmware) executable by the (micro) processor, logic gates, switches, an Application Specific Integrated Circuit (ASIC), a programmable logic controller, and an embedded microcontroller, examples of which include, but are not limited to, the following microcontrollers: ARC 625D, Atmel AT91SAM, Microchip PIC18F26K20, and Silicone Labs C8051F320, the memory controller may also be implemented as part of the control logic for the memory. Those skilled in the art will also appreciate that, in addition to implementing the controller as pure computer readable program code, the same functionality can be implemented by logically programming method steps such that the controller is in the form of logic gates, switches, application specific integrated circuits, programmable logic controllers, embedded microcontrollers and the like. Such a controller may thus be considered a hardware component, and the means included therein for performing the various functions may also be considered as a structure within the hardware component. Or even means for performing the functions may be regarded as being both a software module for performing the method and a structure within a hardware component.
The systems, devices, modules or units illustrated in the above embodiments may be implemented by a computer chip or an entity, or by a product with certain functions. One typical implementation device is a computer. In particular, the computer may be, for example, a personal computer, a laptop computer, a cellular telephone, a camera phone, a smartphone, a personal digital assistant, a media player, a navigation device, an email device, a game console, a tablet computer, a wearable device, or a combination of any of these devices.
For convenience of description, the above devices are described as being divided into various units by function, and are described separately. Of course, the functionality of the units may be implemented in one or more software and/or hardware when implementing the present application.
As will be appreciated by one skilled in the art, embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the application. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
In a typical configuration, a computing device includes one or more processors (CPUs), input/output interfaces, network interfaces, and memory.
The memory may include forms of volatile memory in a computer readable medium, Random Access Memory (RAM) and/or non-volatile memory, such as Read Only Memory (ROM) or flash memory (flash RAM). Memory is an example of a computer-readable medium.
Computer-readable media, including both non-transitory and non-transitory, removable and non-removable media, may implement information storage by any method or technology. The information may be computer readable instructions, data structures, modules of a program, or other data. Examples of computer storage media include, but are not limited to, phase change memory (PRAM), Static Random Access Memory (SRAM), Dynamic Random Access Memory (DRAM), other types of Random Access Memory (RAM), Read Only Memory (ROM), Electrically Erasable Programmable Read Only Memory (EEPROM), flash memory or other memory technology, compact disc read only memory (CD-ROM), Digital Versatile Discs (DVD) or other optical storage, magnetic cassettes, magnetic tape magnetic disk storage or other magnetic storage devices, or any other non-transmission medium that can be used to store information that can be accessed by a computing device. As defined herein, a computer readable medium does not include a transitory computer readable medium such as a modulated data signal and a carrier wave.
It should also be noted that the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.
As will be appreciated by one skilled in the art, embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The application may be described in the general context of computer-executable instructions, such as program modules, being executed by a computer. Generally, program modules include routines, programs, objects, components, data structures, etc. that perform particular transactions or implement particular abstract data types. The application may also be practiced in distributed computing environments where transactions are performed by remote processing devices that are linked through a communications network. In a distributed computing environment, program modules may be located in both local and remote computer storage media including memory storage devices.
The embodiments in the present specification are described in a progressive manner, and the same and similar parts among the embodiments are referred to each other, and each embodiment focuses on the differences from the other embodiments. In particular, for the system embodiment, since it is substantially similar to the method embodiment, the description is simple, and for the relevant points, reference may be made to the partial description of the method embodiment.
The above description is only an example of the present application and is not intended to limit the present application. Various modifications and changes may occur to those skilled in the art. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present application should be included in the scope of the claims of the present application.

Claims (10)

1. A method for extracting a video frame, comprising:
acquiring an application mode of a video frame, wherein the application mode of the video frame comprises a cover, auditing and dragging preview;
when the application mode of the video frame is a cover, extracting one video frame in the target video;
when the application mode of the video frames is auditing, extracting a first number of video frames according to first information of a target video, wherein the first information comprises a scene type of the target video, the scene type of the target video at least comprises integrated entertainment, education and live game, and the first number is determined by the scene type of the target video;
and when the application mode of the video frames is dragging preview, extracting a second number of video frames according to the duration of the target video, wherein the second number is determined by the duration of the target video.
2. The method of claim 1, wherein when the video frame is applied in a cover page, extracting a video frame from the target video comprises: extracting all elements in a first element class which is a set of elements with the occurrence frequency larger than a first threshold value in a target video and has the highest image quality from the target video, wherein the first element class is a value which enables the first element class to at least comprise one element.
3. The method of claim 1, further comprising:
and determining first information according to the change frequency of the target video.
4. The method of claim 3, wherein determining the first information according to the change frequency of the target video comprises:
extracting video frames in the target video at equal intervals;
taking the proportion of key frames in the video frames extracted at equal intervals as the change frequency of the target video;
and determining first information according to the change frequency of the target video.
5. The method of claim 1, wherein the first number is determined by a scene type of the target video, and comprises:
when the scene type of the target video is entertainment, the first number is at least 12 frames;
when the scene type of the target video is education, the first number is at least 8 frames;
when the scene type of the target video is game live, the first number is at least 24 frames.
6. The method of claim 1, wherein when the application mode of the video frame is dragging preview, extracting a second number of video frames according to the duration of the target video comprises:
determining a second video frame extraction interval according to the duration and the second number of the target video;
extracting a second number of video frames according to the second video frame extraction interval.
7. The method of claim 1, wherein when the application mode of the video frame is dragging preview, extracting a second number of video frames according to the duration of the target video comprises:
dividing the target video into a second number of segments of sub-video;
and extracting key frames in each section of the sub-video and respectively forming dynamic pictures to obtain a second number of video frames.
8. A video frame extraction apparatus, comprising:
the system comprises an acquisition module, a display module and a display module, wherein the acquisition module is used for acquiring the application mode of a video frame, and the application mode of the video frame comprises a cover, audit and dragging preview;
the cover extraction module is used for extracting one video frame in the target video when the application mode of the video frame is the cover;
the auditing and extracting module is used for extracting a first number of video frames according to first information of a target video when the application mode of the video frames is auditing, wherein the first information comprises a scene type of the target video, the scene type of the target video at least comprises comprehensive entertainment, education and live game, and the first number is determined by the scene type of the target video;
and the preview extraction module is used for extracting a second number of video frames according to the duration of the target video when the application mode of the video frames is dragging preview, and the second number is determined by the duration of the target video.
9. An electronic device comprising a memory and a processor, the memory having an executable program stored thereon, the processor executing the executable program to perform the steps of:
acquiring an application mode of a video frame, wherein the application mode of the video frame comprises a cover, auditing and dragging preview;
when the application mode of the video frame is a cover, extracting one video frame in the target video;
when the application mode of the video frames is auditing, extracting a first number of video frames according to first information of a target video, wherein the first information comprises a scene type of the target video, the scene type of the target video at least comprises integrated entertainment, education and live game, and the first number is determined by the scene type of the target video;
and when the application mode of the video frames is dragging preview, extracting a second number of video frames according to the duration of the target video, wherein the second number is determined by the duration of the target video.
10. A storage medium, characterized in that the storage medium has stored thereon a computer program which, when executed by a processor, carries out the method according to any one of claims 1-8.
CN202011308180.1A 2020-11-19 2020-11-19 Video frame extraction method and device, electronic equipment and storage medium Active CN112492382B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011308180.1A CN112492382B (en) 2020-11-19 2020-11-19 Video frame extraction method and device, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011308180.1A CN112492382B (en) 2020-11-19 2020-11-19 Video frame extraction method and device, electronic equipment and storage medium

Publications (2)

Publication Number Publication Date
CN112492382A true CN112492382A (en) 2021-03-12
CN112492382B CN112492382B (en) 2022-01-21

Family

ID=74932243

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011308180.1A Active CN112492382B (en) 2020-11-19 2020-11-19 Video frame extraction method and device, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN112492382B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114528923A (en) * 2022-01-25 2022-05-24 山东浪潮科学研究院有限公司 Video target detection method, device, equipment and medium based on time domain context
CN115174812A (en) * 2022-07-01 2022-10-11 维沃移动通信有限公司 Video generation method, video generation device and electronic equipment

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070098267A1 (en) * 2005-10-27 2007-05-03 Electronics And Telecommunications Research Institute Method and apparatus for distinguishing obscene video using visual feature
CN105323634A (en) * 2014-06-27 2016-02-10 Tcl集团股份有限公司 Method and system for generating thumbnail of video
CN105898345A (en) * 2016-04-14 2016-08-24 张愚 Previewing video service system
CN106911943A (en) * 2017-02-21 2017-06-30 腾讯科技(深圳)有限公司 A kind of video display method and its device
CN108040262A (en) * 2018-01-25 2018-05-15 湖南机友科技有限公司 Live audio and video are reflected yellow method and device in real time
CN111182314A (en) * 2018-11-12 2020-05-19 阿里巴巴集团控股有限公司 Live stream processing method and device and data processing method

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070098267A1 (en) * 2005-10-27 2007-05-03 Electronics And Telecommunications Research Institute Method and apparatus for distinguishing obscene video using visual feature
CN105323634A (en) * 2014-06-27 2016-02-10 Tcl集团股份有限公司 Method and system for generating thumbnail of video
CN105898345A (en) * 2016-04-14 2016-08-24 张愚 Previewing video service system
CN106911943A (en) * 2017-02-21 2017-06-30 腾讯科技(深圳)有限公司 A kind of video display method and its device
CN108040262A (en) * 2018-01-25 2018-05-15 湖南机友科技有限公司 Live audio and video are reflected yellow method and device in real time
CN111182314A (en) * 2018-11-12 2020-05-19 阿里巴巴集团控股有限公司 Live stream processing method and device and data processing method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
ZHAO SHIWEI等: "RESEARCH ON KEY THECHNOLOGIES OF PORNOGRAPHIC IMAGE/VIDEO RECOGNITION IN COMPRESSED DOMAIN", 《JOURNAL OF ELECTRONICS (CHINA)》 *
张维刚等: "融合视觉内容分析的网络视频缩略图推荐", 《智能计算机与应用》 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114528923A (en) * 2022-01-25 2022-05-24 山东浪潮科学研究院有限公司 Video target detection method, device, equipment and medium based on time domain context
CN114528923B (en) * 2022-01-25 2023-09-26 山东浪潮科学研究院有限公司 Video target detection method, device, equipment and medium based on time domain context
CN115174812A (en) * 2022-07-01 2022-10-11 维沃移动通信有限公司 Video generation method, video generation device and electronic equipment

Also Published As

Publication number Publication date
CN112492382B (en) 2022-01-21

Similar Documents

Publication Publication Date Title
CN109618222B (en) A kind of splicing video generation method, device, terminal device and storage medium
CN110012302B (en) Live network monitoring method and device and data processing method
US8966372B2 (en) Systems and methods for performing geotagging during video playback
CN112492382B (en) Video frame extraction method and device, electronic equipment and storage medium
CN111580921B (en) Content creation method and device
CN111131876B (en) Control method, device and terminal for live video and computer readable storage medium
CN110177295B (en) Subtitle out-of-range processing method and device and electronic equipment
CN112672184A (en) Video auditing and publishing method
CN103997687A (en) Techniques for adding interactive features to videos
CN112399249A (en) Multimedia file generation method and device, electronic equipment and storage medium
CN111279709A (en) Providing video recommendations
WO2012111226A1 (en) Time-series document summarization device, time-series document summarization method and computer-readable recording medium
CN112291614A (en) Video generation method and device
KR20210091082A (en) Image processing apparatus, control method thereof and computer readable medium having computer program recorded therefor
CN114943006A (en) Singing bill display information generation method and device, electronic equipment and storage medium
US20150111189A1 (en) System and method for browsing multimedia file
CN106936830B (en) Multimedia data playing method and device
CN110197459B (en) Image stylization generation method and device and electronic equipment
CN112672202B (en) Bullet screen processing method, equipment and storage medium
CN114925285A (en) Book information processing method, device, equipment and storage medium
CN108197113B (en) Article information conversion method, device, equipment and computer readable medium
KR20220000459A (en) Method of evaluating language difficulty in video and computer program for the same
CN111046232A (en) Video classification method, device and system
CN109218771A (en) A kind of recommended method of video program, device, electronic equipment and storage medium
CN110909204A (en) Video publishing method and device and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant