CN112235613B - Video processing method and device, electronic equipment and storage medium - Google Patents

Video processing method and device, electronic equipment and storage medium Download PDF

Info

Publication number
CN112235613B
CN112235613B CN202010983900.8A CN202010983900A CN112235613B CN 112235613 B CN112235613 B CN 112235613B CN 202010983900 A CN202010983900 A CN 202010983900A CN 112235613 B CN112235613 B CN 112235613B
Authority
CN
China
Prior art keywords
video
information
user
video segment
segment
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010983900.8A
Other languages
Chinese (zh)
Other versions
CN112235613A (en
Inventor
史南胜
谢马林
曹姣
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Baidu Netcom Science and Technology Co Ltd
Original Assignee
Beijing Baidu Netcom Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Baidu Netcom Science and Technology Co Ltd filed Critical Beijing Baidu Netcom Science and Technology Co Ltd
Priority to CN202010983900.8A priority Critical patent/CN112235613B/en
Publication of CN112235613A publication Critical patent/CN112235613A/en
Application granted granted Critical
Publication of CN112235613B publication Critical patent/CN112235613B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/258Client or end-user data management, e.g. managing client capabilities, user preferences or demographics, processing of multiple end-users preferences to derive collaborative data
    • H04N21/25866Management of end-user data
    • H04N21/25891Management of end-user data being end-user preferences

Abstract

The application discloses a video processing method and device, electronic equipment and a storage medium, and relates to the field of video processing. The specific implementation scheme is as follows: acquiring identification information of a target video clip interested by a user based on a video account concerned by the user and a label subscribed by the user in advance; and sending a video notification message to the user based on the identification information of the target video segment. According to the technical scheme, the video clip can be accurately pushed, and the accuracy and the pushing efficiency of video pushing are effectively improved.

Description

Video processing method and device, electronic equipment and storage medium
Technical Field
The present application relates to the field of computer technologies, and in particular, to a video processing method and apparatus, an electronic device, and a storage medium.
Background
In recent years, with the burning of videos, users can view various videos in which they are interested through various network platforms capable of providing video services.
In practical application, a plurality of video accounts can be maintained in each network platform, and countless videos of each video account can be managed. The user can browse videos of all video accounts on each network platform through the network platforms. In the prior art, in order to improve the service quality, each network platform may also adopt a certain policy to push links of videos in which the user is interested to the user in a notification message manner. Through this link, the user can open and view the entire video.
However, in the video push method in the prior art, a user sees all videos, and in practical applications, the user usually focuses on only a certain segment in the videos, so that the accuracy of the existing video push is low.
Disclosure of Invention
In order to solve the above problem, the present application provides a video processing method, an apparatus, an electronic device, and a storage medium.
According to an aspect of the present application, there is provided a video processing method, wherein the method includes the steps of:
acquiring identification information of a target video clip interested by a user based on a video account concerned by the user and a label subscribed by the user in advance;
and sending a video notification message to the user based on the identification information of the target video segment.
According to another aspect of the present application, there is provided a video processing apparatus, wherein the apparatus comprises:
the acquisition module is used for acquiring identification information of a target video clip which is interested by a user based on a video account concerned by the user and a label subscribed by the user in advance;
and the sending module is used for sending a video notification message to the user based on the identification information of the target video clip.
According to still another aspect of the present application, there is provided an electronic apparatus including:
at least one processor; and
a memory communicatively coupled to the at least one processor; wherein the content of the first and second substances,
the memory stores instructions executable by the at least one processor to enable the at least one processor to perform the method as described above.
According to yet another aspect of the present application, there is provided a non-transitory computer readable storage medium having stored thereon computer instructions for causing the computer to perform the method as described above.
According to yet another aspect of the application, a computer program product is provided, comprising a computer program which, when executed by a processor, implements the method as described above.
According to the technology of the application, identification information of a target video clip which is interested by a user is obtained through a video account concerned by the user and a label subscribed by the user in advance; based on the identification information of the target video clip, the video notification message is sent to the user, so that the video clip can be accurately pushed, and the accuracy and the pushing efficiency of the video pushing are effectively improved.
It should be understood that the statements in this section do not necessarily identify key or critical features of the embodiments of the present disclosure, nor do they limit the scope of the present disclosure. Other features of the present disclosure will become apparent from the following description.
Drawings
The drawings are included to provide a better understanding of the present solution and are not to be considered limiting of the present application. Wherein:
FIG. 1 is a schematic diagram according to a first embodiment of the present application;
FIG. 2 is a schematic diagram according to a second embodiment of the present application;
FIG. 3 is a schematic illustration according to a third embodiment of the present application;
FIG. 4 is a schematic illustration of a fourth embodiment according to the present application;
fig. 5 is a block diagram of an electronic device for implementing a video processing method according to an embodiment of the present application.
Detailed description of the preferred embodiments
The following description of the exemplary embodiments of the present application, taken in conjunction with the accompanying drawings, includes various details of the embodiments of the application for the understanding of the same, which are to be considered exemplary only. Accordingly, those of ordinary skill in the art will recognize that various changes and modifications of the embodiments described herein can be made without departing from the scope and spirit of the present application. Also, descriptions of well-known functions and constructions are omitted in the following description for clarity and conciseness.
FIG. 1 is a schematic illustration according to a first embodiment of the present application; as shown in fig. 1, this embodiment provides a video processing method, which may specifically include the following steps:
s101, acquiring identification information of a target video clip interested by a user based on a video account concerned by the user and a label subscribed by the user in advance;
and S102, sending a video notification message to the user based on the identification information of the target video clip.
The execution main body of the video processing method of the embodiment is a video processing device, and the video processing device may be disposed on a video server, or may also be independent of the video server, but may communicate with the video server, and may acquire a video resource on the video server, so as to achieve more accurate transmission of a video notification message to a user.
The video server of the embodiment can manage and maintain a plurality of video accounts, each video account can provide video service for a user, and the video content and the type provided by different video accounts can be different. For example, some video accounts are used for providing videos for selling cosmetics in a live mode for users, some video accounts are used for providing videos for selling clothes in a live mode for users, some video accounts are used for providing entertainment bagua information for users, some video accounts are used for providing game comment information for users, and other video accounts are used for providing training information of various professional examinations for users. In short, the relationship between the video account and the user in the embodiment may be a relationship between a main broadcasting and a fan, a relationship between a teacher and a student, and other various relationships.
On the video server, each video account which normally runs can update the video content regularly, and a user can browse all videos of the concerned video account regularly; the video notification service may also be pushed to the user by the video processing device based on a video account that the user is interested in.
In this embodiment, before the video notification service is pushed to the user, it is further necessary to determine a tag subscribed by the user in advance, where the tag is a tag of a video, and identify attribute information to which video content belongs, for example, in a video of live broadcast sales, a brand of a sold good may be used as a tag of the video, or names of the sold goods, such as facial cleanser, eye cream, and face cream, may also be used as a tag, and the like. The video account of the eight diagrams for entertainment can adopt the star of the eight diagrams as a video label, or can also adopt the area of the star of the eight diagrams as a video label, such as a harbor platform, a inland, a Japanese-Korean game and the like, or can also adopt the entertainment company of the star of the eight diagrams as a video label and the like. In short, a video account can be preset with a plurality of labels based on the coverage area of the operation service. The user can subscribe the interested labels based on a plurality of labels preset by the user.
The video processing apparatus in this embodiment may obtain, based on the account concerned by the user and the tag subscribed in advance by the user, the identification information of the target video segment related to the tag subscribed by the user from the video published by the video account concerned by the user, as the identification information of the target video segment interested by the user. A video notification message may then be sent to the user based on the identification information of the target video segment. In this way, after the client where the user is located receives the information of the target video segment, the target video segment can be opened through the video player based on the identification information of the target video segment, and the target video is just the segment in which the user is most interested in the source video, so that accurate video push can be realized. For example, especially in a shopping video, the advertisement can be accurately pushed to interested commodities to the user, and the effect of advertisement is effectively achieved.
In the video processing method of the embodiment, identification information of a target video clip which is interested by a user is acquired based on a video account concerned by the user and a tag subscribed by the user in advance; based on the identification information of the target video clip, the video notification message is sent to the user, so that the video clip can be accurately pushed, and the accuracy and the pushing efficiency of the video pushing are effectively improved.
FIG. 2 is a schematic diagram according to a second embodiment of the present application; as shown in fig. 2, the video processing method of this embodiment further introduces the technical solution of the present application in more detail on the basis of the technical solution of the embodiment shown in fig. 1. As shown in fig. 2, the video processing method of this embodiment may specifically include the following steps:
s201, displaying all labels of a video account concerned by a user;
s202, receiving a subscription request carrying an identification and a label of a user;
s203, establishing a corresponding relation between the user identification, the video account and the label;
the above steps S201 to S203 of this embodiment are processes of user subscription. For example, a user may click to focus on a video account while browsing various video accounts on a video server through a web browser. And further under the video account, all tags of the video account can be clicked to be requested to be viewed, and at the moment, the video processing device can respond to the request of the user and display all tags of the video account. The user may select one, two, or more of the tags of interest and click on the subscription. At the video processing device side, a subscription request of the user carrying the user identifier and the selected tag can be received, and the corresponding relation among the user identifier, the video account and the tag is further established and stored, so that when video clips are subsequently pushed to the user, the video account concerned by the user and the tag subscribed by the user can be referred to, and the information of the target video clip interested by the user can be screened.
S204, analyzing a source video issued by the video account to obtain a plurality of video segments and summary information of each video segment;
for example, in a specific implementation, the step S204 may specifically include the following steps:
(a1) Segmenting a source video to obtain a plurality of video segments;
for example, the present embodiment may be evenly divided according to a preset length, and the length of the present embodiment may be 5 frames, 10 frames, or other number of consecutive frames. That is, for each video segment, the start time point and the end time point of the video segment can be known at the time of slicing. Or alternatively, the cutting may not be performed uniformly according to preset lengths, which is not limited herein.
(b1) Extracting at least one of each frame image in each video segment in the plurality of video segments, character information in each frame image and voice information of each video segment;
in this embodiment, the text information in each frame image may include subtitle information in each frame image and text information of a part other than a subtitle in an identified image, where text identification is performed on each frame image.
(c1) And generating summary information of each video segment based on at least one of each frame image in each video segment, text information in each frame image and voice information of each video segment.
In this embodiment, the process of generating the summary information of each video segment may be implemented by means of some tools. Or may be implemented using a pre-trained video analysis model. For example, the video analysis model of the present embodiment is implemented by using a pre-trained neural network model. For each video segment, when generating the summary information of the corresponding video segment in this embodiment, at least one of each frame image, text information in each frame image, and voice information of each video segment in the video segment needs to be input into the video analysis model, and the video analysis model can predict and output the summary information of the video segment based on the input information. In practical application, preferably, the frame images in the video segment, the text information in the frame images, and the voice information of the video segment are input simultaneously, that is, the three types of information of the image, the text, and the voice of the video segment are input simultaneously, and the obtained summary information of the video segment is more accurate. I.e. the richer the type of information input, the more accurate the summary information of the video segment obtained. However, in practical applications, some videos do not have subtitles and/or voice information, and in this case, only all types of information of the video segments that can be acquired may be included.
In practical application, when the video analysis model is trained, which type of information is adopted for training, and the summary information of the video segments is predicted, the corresponding type of information also needs to be input, so that accurate prediction can be realized. The training process of the video analysis model of this embodiment is similar to the principle of the use process, and is not described herein again.
By the method, the summary information of each video segment can be accurately acquired, so that the accuracy of acquiring the subsequent target video segments is improved.
S205, acquiring information of a target video segment based on the summary information and the label of each video segment; the information of the target video clip comprises address information of a source video to which the target video clip belongs, and a starting time point and an ending time point of the target video clip;
for example, when the step is implemented specifically, the step may specifically include the following steps:
(a2) Calculating the similarity between the abstract information of each video segment and the label;
specifically, for each video segment, the summary information and the tag of the video segment may be respectively embedded and expressed to obtain a corresponding vector representation, and then the vector similarity between the summary information and the tag of the video segment is calculated as the similarity between the summary information and the tag of the video segment.
(b2) Combining at least two video segments which have similarity greater than or equal to a preset similarity threshold and are continuously adjacent in a source video to serve as a target video segment;
the preset similarity threshold of the present embodiment may be set to 80%,85% or other values according to actual requirements. Because the units of the video segments are small, if the similarity between at least two consecutive adjacent video segments in the source video and the label is greater than or equal to the preset similarity threshold, at least two consecutive adjacent video segments can be merged to be used as the target video segment.
Optionally, if multiple target video segments exist in the same source video according to the above processing, the similarity between each target video segment and the tag may be sorted, and the target video segment corresponding to the maximum similarity is obtained as the target video segment to be finally obtained.
The similarity between each target video segment and the label may be an average value of the similarity between the summary information of each video segment in the target video segment and the label, or the similarity between the target video segment and the label may be calculated based on the similarity between the summary information of each video segment in the target video segment and the label by using summation or other mathematical operations.
In addition, optionally, in a specific implementation, the video segments may be sequentially analyzed according to the sequence of the video segments in the source video. For example, first, a first video segment is analyzed, the first video segment is taken as a current video segment, the similarity between the summary information of the current video segment and the label is calculated by adopting the manner of the above embodiment, and whether the similarity is greater than or equal to a preset similarity threshold value is judged; if not, updating the second video segment to be the current video segment and restarting the analysis; if so, continuing to analyze the similarity between the summary information of the next nearest neighbor video segment of the current video segment and the label, and if not, merging the next nearest neighbor video segment and the current video segment to form a video segment, continuing to analyze the next nearest neighbor video segment according to a similar mode, and repeating the steps until the similarity between the summary information of the next nearest neighbor video segment and the label is less than the preset similarity threshold, and taking the previously merged video segment as a target video segment.
In a similar manner, a plurality of target video segments may also be acquired in one source video, and in the same manner as in the above embodiment, a target video segment corresponding to the maximum similarity value of the tag in the plurality of target video segments is acquired as a target video segment to be finally acquired.
(c2) And acquiring the address information of the source video and the starting time point and the ending time point of the target video segment.
Since the start time point and the end time point of each video segment are already determined, and the target video segment is formed by splicing at least two continuous video segments, the start time point and the end time point of the target video segment can be obtained based on the start time point and the end time point of the at least two continuous video segments. The address information of the source video refers to the address of the source video on the video server, the address information of the source video is the address information of the target video segment, and the target video segment can be accurately positioned based on the address information of the source video and by combining the starting time point and the ending time point of the target video segment in the source video.
By adopting the method, the accuracy of the acquired target video clip information can be effectively ensured, and the accuracy and the pushing efficiency of subsequent video pushing are improved.
S206, sending a video notification message carrying the address information of the source video and the starting time point and the ending time point of the target video clip to the user.
After the video client used by the user receives the video notification message, the video client can open the source video through the video player based on the address information of the source video, and can accurately position the target video segment based on the starting time point and the ending time point of the target video segment.
The number of video accounts that the user pays attention to and the number of tags subscribed by the user are not limited in the above implementation manner of the embodiment. In practical application, the technical solution of the present embodiment may be specifically adapted and adjusted in combination with a specific usage scenario and a specific operation policy.
For example, in case of a first specific use, the user pays attention to multiple video accounts, but only subscribes to the same tag of the multiple video accounts, and the information of N target video clips is required to be pushed to the user every preset time period.
In this scenario, each source video issued by each video account within a preset time period may be analyzed in the manner of the above embodiment, and for each source video issued by each video account, information of one target video clip may be acquired according to the above embodiment; then, the relevance of each target video segment and the label can be respectively calculated, and specifically, the text similarity between the abstract information of each target video segment and the label can be calculated. And then selecting the information of the N target video segments with the highest correlation as the information of the finally screened N target video segments, and sending a video notification message to the user based on the information of the N target video segments. Wherein N in this case may be a positive integer greater than or equal to 1.
And in the second case, in specific use, the user pays attention to a plurality of video accounts, subscribes to a plurality of labels in total, and requires to push information of N target video clips to the user within a preset time period.
In this scenario, each tag may be analyzed separately, for each tag, each source video published by the video account to which the user subscribes in the preset time period needs to be analyzed according to the method of the above-mentioned case one, and for each source video published by each video account to which the user subscribes, information of one target video clip may be obtained according to the above-mentioned embodiment, and a correlation between the tag and each target video clip may also be recorded. And finally, sequencing all the obtained target video clips according to the correlation degree between the target video clips and the labels adopted by the target video clips, and finally screening the information of the N target video clips, and sending a video notification message to the user based on the information of the N target video clips. Likewise, N in this case may be a positive integer greater than or equal to 1.
By adopting the above scheme, the video processing method of the embodiment can establish the corresponding relationship among the user identifier, the video account and the tag based on the subscription of the user, so as to accurately send the video notification message to the user based on the corresponding relationship. Further, in this embodiment, by analyzing the summary information of each video segment in the source video and the similarity of the tag, and merging at least two consecutive adjacent video segments in the source video, each of which has a similarity greater than or equal to a preset similarity threshold, as the target video segment, the accuracy of the obtained target video segment can be effectively ensured, and further, the address information of the source video, the start time point and the end time point of the target video segment are obtained as the information of the target video segment, so as to send a corresponding video notification message to the user, thereby being capable of accurately pushing the target video segment interested by the user, and effectively improving the accuracy and the pushing efficiency of video pushing.
FIG. 3 is a schematic illustration according to a third embodiment of the present application; as shown in fig. 3, the present embodiment provides a video processing apparatus 300, including:
the obtaining module 301 is configured to obtain identification information of a target video clip that a user is interested in based on a video account that the user pays attention to and a tag that the user subscribes in advance;
a sending module 302, configured to send a video notification message to the user based on the identification information of the target video segment.
The video processing apparatus 300 of this embodiment implements the video processing principle and technical effect by using the modules, which are the same as the implementation of the related method embodiment described above, and reference may be made to the description of the related method embodiment in detail, which is not repeated herein.
FIG. 4 is a schematic illustration according to a fourth embodiment of the present application; as shown in fig. 4, the video processing apparatus 300 of the present embodiment further describes the technical solution of the present application in more detail on the basis of the technical solution of the embodiment shown in fig. 3.
As shown in fig. 4, in the video processing apparatus 300 of the present embodiment, the obtaining module 301 includes:
the summary obtaining unit 3011 is configured to analyze a source video published by a video account, and obtain a plurality of video segments and summary information of each video segment;
an information obtaining unit 3012, configured to obtain information of the target video segment based on the summary information and the tag of each video segment.
Further optionally, the summary obtaining unit 3011 is configured to:
segmenting a source video to obtain a plurality of video segments;
extracting at least one of each frame image in each video segment in the plurality of video segments, character information in each frame image and voice information of each video segment;
and generating summary information of each video segment based on at least one of each frame image in each video segment, character information in each frame image and voice information of each video segment.
Further optionally, the summary obtaining unit 3011 is configured to generate summary information of each video segment based on at least one of each frame image in each video segment, text information in each frame image, and voice information of each video segment by using a video analysis model trained in advance.
Further optionally, the information obtaining unit 3012 includes:
calculating the similarity between the abstract information of each video segment and the label;
combining at least two video segments which have similarity greater than or equal to a preset similarity threshold and are continuously adjacent in a source video to serve as a target video segment;
and acquiring the address information of the source video and the starting time point and the ending time point of the target video segment.
Further optionally, in the video processing apparatus 300 of this embodiment, the sending module 302 is configured to:
and sending a video notification message carrying the address information of the source video and the starting time point and the ending time point of the target video segment to the user.
Further optionally, as shown in fig. 4, the video processing apparatus 300 of this embodiment further includes:
a display module 303, configured to display all tags of the video account;
a receiving module 304, configured to receive a subscription request carrying an identifier and a tag of a user;
the establishing module 305 is configured to establish a correspondence between the user identifier, the video account, and the tag.
The video processing apparatus 300 of this embodiment implements the video processing principle and technical effect by using the modules, which are the same as the implementation of the related method embodiment described above, and reference may be made to the description of the related method embodiment in detail, which is not repeated herein.
According to an embodiment of the present application, an electronic device and a readable storage medium are also provided.
Fig. 5 is a block diagram of an electronic device implementing a video processing method according to an embodiment of the present application. Electronic devices are intended to represent various forms of digital computers, such as laptops, desktops, workstations, personal digital assistants, servers, blade servers, mainframes, and other appropriate computers. The electronic device may also represent various forms of mobile devices, such as personal digital processing, cellular phones, smart phones, wearable devices, and other similar computing devices. The components shown herein, their connections and relationships, and their functions, are meant to be examples only, and are not meant to limit implementations of the present application that are described and/or claimed herein.
As shown in fig. 5, the electronic apparatus includes: one or more processors 501, memory 502, and interfaces for connecting the various components, including high-speed interfaces and low-speed interfaces. The various components are interconnected using different buses and may be mounted on a common motherboard or in other manners as desired. The processor may process instructions for execution within the electronic device, including instructions stored in or on the memory to display graphical information of a GUI on an external input/output apparatus (such as a display device coupled to the interface). In other embodiments, multiple processors and/or multiple buses may be used, along with multiple memories and multiple memories, as desired. Also, multiple electronic devices may be connected, with each device providing portions of the necessary operations (e.g., as a server array, a group of blade servers, or a multi-processor system). In fig. 5, one processor 501 is taken as an example.
Memory 502 is a non-transitory computer readable storage medium as provided herein. Wherein the memory stores instructions executable by at least one processor to cause the at least one processor to perform the video processing method provided herein. The non-transitory computer-readable storage medium of the present application stores computer instructions for causing a computer to perform the video processing method provided by the present application.
The memory 502, which is a non-transitory computer readable storage medium, may be used to store non-transitory software programs, non-transitory computer executable programs, and modules, such as program instructions/modules (e.g., related modules shown in fig. 3 and 4) corresponding to the video processing method in the embodiments of the present application. The processor 501 executes various functional applications of the server and data processing, i.e., implements the video processing method in the above-described method embodiments, by running non-transitory software programs, instructions, and modules stored in the memory 502.
The memory 502 may include a storage program area and a storage data area, wherein the storage program area may store an operating system, an application program required for at least one function; the storage data area may store data created according to use of an electronic device implementing the video processing method, and the like. Further, the memory 502 may include high speed random access memory, and may also include non-transitory memory, such as at least one magnetic disk storage device, flash memory device, or other non-transitory solid state storage device. In some embodiments, memory 502 optionally includes memory located remotely from processor 501, which may be connected via a network to an electronic device implementing the video processing method. Examples of such networks include, but are not limited to, the internet, intranets, local area networks, mobile communication networks, and combinations thereof.
The electronic device implementing the video processing method may further include: an input device 503 and an output device 504. The processor 501, the memory 502, the input device 503 and the output device 504 may be connected by a bus or other means, and fig. 5 illustrates the connection by a bus as an example.
The input device 503 may receive input numeric or character information and generate key signal inputs related to user settings and function control of an electronic apparatus implementing the video processing method, such as an input device such as a touch screen, a keypad, a mouse, a track pad, a touch pad, a pointing stick, one or more mouse buttons, a track ball, a joystick, or the like. The output devices 504 may include a display device, auxiliary lighting devices (e.g., LEDs), and haptic feedback devices (e.g., vibrating motors), among others. The display device may include, but is not limited to, a Liquid Crystal Display (LCD), a Light Emitting Diode (LED) display, and a plasma display. In some implementations, the display device can be a touch screen.
Various implementations of the systems and techniques described here can be realized in digital electronic circuitry, integrated circuitry, application specific ASICs (application specific integrated circuits), computer hardware, firmware, software, and/or combinations thereof. These various embodiments may include: implemented in one or more computer programs that are executable and/or interpretable on a programmable system including at least one programmable processor, which may be special or general purpose, receiving data and instructions from, and transmitting data and instructions to, a storage system, at least one input device, and at least one output device.
These computer programs (also known as programs, software applications, or code) include machine instructions for a programmable processor, and may be implemented using high-level procedural and/or object-oriented programming languages, and/or assembly/machine languages. As used herein, the terms "machine-readable medium" and "computer-readable medium" refer to any computer program product, apparatus, and/or device (e.g., magnetic discs, optical disks, memory, programmable Logic Devices (PLDs)) used to provide machine instructions and/or data to a programmable processor, including a machine-readable medium that receives machine instructions as a machine-readable signal. The term "machine-readable signal" refers to any signal used to provide machine instructions and/or data to a programmable processor.
To provide for interaction with a user, the systems and techniques described here can be implemented on a computer having: a display device (e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor) for displaying information to a user; and a keyboard and a pointing device (e.g., a mouse or a trackball) by which a user can provide input to the computer. Other kinds of devices may also be used to provide for interaction with a user; for example, feedback provided to the user can be any form of sensory feedback (e.g., visual feedback, auditory feedback, or tactile feedback); and input from the user may be received in any form, including acoustic, speech, or tactile input.
The systems and techniques described here can be implemented in a computing system that includes a back-end component (e.g., as a data server), or that includes a middleware component (e.g., an application server), or that includes a front-end component (e.g., a user computer having a graphical user interface or a web browser through which a user can interact with an implementation of the systems and techniques described here), or any combination of such back-end, middleware, or front-end components. The components of the system can be interconnected by any form or medium of digital data communication (e.g., a communication network). Examples of communication networks include: local Area Networks (LANs), wide Area Networks (WANs), the Internet, and blockchain networks.
The computer system may include clients and servers. A client and server are generally remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
According to the technical scheme of the embodiment of the application, the information of the target video clip which is interested by the user is obtained through the video account concerned by the user and the label subscribed by the user in advance; based on the information of the target video clip, the video notification message is sent to the user, so that the video clip can be accurately pushed, and the accuracy and the pushing efficiency of the video pushing are effectively improved.
According to the technical scheme of the embodiment of the application, the corresponding relation among the user identification, the video account and the label can be established based on the subscription of the user, so that the video notification message can be accurately sent to the user based on the corresponding relation. Further, in this embodiment, by analyzing the summary information of each video segment in the source video and the similarity of the tag, and merging at least two consecutive adjacent video segments in the source video, each of which has a similarity greater than or equal to a preset similarity threshold, as the target video segment, the accuracy of the obtained target video segment can be effectively ensured, and further, the address information of the source video, the start time point and the end time point of the target video segment are obtained as the information of the target video segment, so as to send a corresponding video notification message to the user, thereby being capable of accurately pushing the target video segment interested by the user, and effectively improving the accuracy and the pushing efficiency of video pushing.
It should be understood that various forms of the flows shown above may be used, with steps reordered, added, or deleted. For example, the steps described in the present application may be executed in parallel, sequentially, or in different orders, and the present invention is not limited thereto as long as the desired results of the technical solutions disclosed in the present application can be achieved.
The above-described embodiments should not be construed as limiting the scope of the present application. It should be understood by those skilled in the art that various modifications, combinations, sub-combinations and substitutions may be made in accordance with design requirements and other factors. Any modification, equivalent replacement, and improvement made within the spirit and principle of the present application shall be included in the protection scope of the present application.

Claims (12)

1. A video processing method, wherein the method comprises the steps of:
acquiring identification information of a target video clip interested by a user based on a video account concerned by the user and a label subscribed by the user in advance; the label is preset based on the range covered by the operation service of the video account;
sending a video notification message to the user based on the identification information of the target video clip;
acquiring identification information of a target video clip interested by a user based on a video account concerned by the user and a label subscribed by the user in advance, wherein the identification information comprises the following steps:
analyzing a source video issued by the video account to acquire a plurality of video segments and summary information of each video segment;
acquiring identification information of the target video segment based on the summary information of each video segment and the label;
the method further comprises the following steps: if a plurality of target video clips exist in the source video clip, acquiring identification information of the target video clip corresponding to the maximum similarity of the tags, and taking the identification information as the identification information of the target video clip to be finally acquired; the similarity is an average value or a sum of the similarity of the summary information of each video segment included in the target video segment and the label;
sending a video notification message to the user based on the information of the target video clip, including:
and sending a video notification message carrying the address information of the source video and the starting time point and the ending time point of the target video segment to the user.
2. The method of claim 1, wherein analyzing the source video published by the video account to obtain a plurality of video segments and summary information of each of the video segments comprises:
segmenting the source video to obtain a plurality of video segments;
extracting at least one of each frame image in each video segment in the plurality of video segments, text information in each frame image and voice information of each video segment;
and generating summary information of each video segment based on at least one of each frame image in each video segment, text information in each frame image and voice information of each video segment.
3. The method of claim 2, wherein generating summary information for each of the video segments based on at least one of each frame image within each of the video segments, text information in each frame image, and speech information for each of the video segments comprises:
and generating summary information of each video segment based on at least one of each frame image in each video segment, text information in each frame image and voice information of each video segment by adopting a pre-trained video analysis model.
4. The method of claim 2, wherein obtaining identification information of the target video segment based on the summary information of each video segment and the tag comprises:
calculating the similarity between the summary information of each video segment and the label;
combining at least two video segments which have similarity greater than or equal to a preset similarity threshold and are continuously adjacent in the source video to serve as the target video segment;
and acquiring the address information of the source video and the starting time point and the ending time point of the target video clip.
5. The method of any of claims 1-4, wherein the method further comprises:
displaying all tags of the video account;
receiving a subscription request carrying the user identification and the label;
and establishing a corresponding relation among the user identification, the video account and the label.
6. A video processing apparatus, wherein the apparatus comprises:
the acquisition module is used for acquiring identification information of a target video clip which is interested by a user based on a video account concerned by the user and a label subscribed by the user in advance; the label is preset based on the range covered by the operation service of the video account;
a sending module, configured to send a video notification message to the user based on the identification information of the target video segment;
the acquisition module includes:
the abstract acquiring unit is used for analyzing a source video issued by the video account and acquiring a plurality of video segments and abstract information of each video segment;
an information obtaining unit, configured to obtain identification information of the target video segment based on the summary information of each video segment and the tag;
the information obtaining unit is further configured to: if a plurality of target video clips exist in the source video clip, acquiring identification information of the target video clip corresponding to the maximum similarity of the tags, and taking the identification information as the identification information of the target video clip to be finally acquired; the similarity is the average value or the sum of the similarity of the abstract information of each video segment included in the target video segment and the label;
the sending module is configured to:
and sending a video notification message carrying the address information of the source video and the starting time point and the ending time point of the target video segment to the user.
7. The apparatus of claim 6, wherein the digest acquisition unit is configured to:
segmenting the source video to obtain a plurality of video segments;
extracting at least one of each frame image in each video segment in the plurality of video segments, text information in each frame image and voice information of each video segment;
and generating summary information of each video segment based on at least one of each frame image in each video segment, text information in each frame image and voice information of each video segment.
8. The apparatus of claim 7, wherein the digest acquisition unit is configured to:
and generating summary information of each video segment based on at least one of each frame image in each video segment, text information in each frame image and voice information of each video segment by adopting a pre-trained video analysis model.
9. The apparatus of claim 7, wherein the information obtaining unit comprises:
calculating the similarity between the summary information of each video segment and the label;
combining at least two video segments which have similarity greater than or equal to a preset similarity threshold and are continuously adjacent in the source video to serve as the target video segment;
and acquiring the address information of the source video and the starting time point and the ending time point of the target video segment.
10. The apparatus of any of claims 6-9, wherein the apparatus further comprises:
the display module is used for displaying all labels of the video account;
a receiving module, configured to receive a subscription request carrying the identifier of the user and the tag;
and the establishing module is used for establishing the corresponding relation among the user identification, the video account and the label.
11. An electronic device, comprising:
at least one processor; and
a memory communicatively coupled to the at least one processor; wherein the content of the first and second substances,
the memory stores instructions executable by the at least one processor to enable the at least one processor to perform the method of any one of claims 1-5.
12. A non-transitory computer readable storage medium having stored thereon computer instructions for causing the computer to perform the method of any one of claims 1-5.
CN202010983900.8A 2020-09-17 2020-09-17 Video processing method and device, electronic equipment and storage medium Active CN112235613B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010983900.8A CN112235613B (en) 2020-09-17 2020-09-17 Video processing method and device, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010983900.8A CN112235613B (en) 2020-09-17 2020-09-17 Video processing method and device, electronic equipment and storage medium

Publications (2)

Publication Number Publication Date
CN112235613A CN112235613A (en) 2021-01-15
CN112235613B true CN112235613B (en) 2023-03-21

Family

ID=74108438

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010983900.8A Active CN112235613B (en) 2020-09-17 2020-09-17 Video processing method and device, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN112235613B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116208821A (en) * 2021-06-02 2023-06-02 宁波星巡智能科技有限公司 Target video capturing method, device, equipment and medium based on image screening
CN113438395A (en) * 2021-06-16 2021-09-24 北京京东乾石科技有限公司 Video transmitting method, video transmitting device, electronic equipment and computer readable medium
CN114245171B (en) * 2021-12-15 2023-08-29 百度在线网络技术(北京)有限公司 Video editing method and device, electronic equipment and medium
CN114449327B (en) * 2021-12-31 2024-03-26 北京百度网讯科技有限公司 Video clip sharing method and device, electronic equipment and readable storage medium

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8117283B2 (en) * 2008-02-04 2012-02-14 Echostar Technologies L.L.C. Providing remote access to segments of a transmitted program
CN109120954B (en) * 2018-09-30 2021-09-07 武汉斗鱼网络科技有限公司 Video message pushing method and device, computer equipment and storage medium
CN111343483B (en) * 2020-02-18 2022-07-19 北京奇艺世纪科技有限公司 Method and device for prompting media content segment, storage medium and electronic device
CN111556326A (en) * 2020-03-27 2020-08-18 威比网络科技(上海)有限公司 Public class video clip pushing method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN112235613A (en) 2021-01-15

Similar Documents

Publication Publication Date Title
CN112235613B (en) Video processing method and device, electronic equipment and storage medium
CN111107392B (en) Video processing method and device and electronic equipment
CN112261423B (en) Method, device, equipment and storage medium for pushing information
CN111626202A (en) Method and device for identifying video
CN111246257B (en) Video recommendation method, device, equipment and storage medium
CN112102448B (en) Virtual object image display method, device, electronic equipment and storage medium
CN111680517A (en) Method, apparatus, device and storage medium for training a model
CN111984825A (en) Method and apparatus for searching video
CN112818224A (en) Information recommendation method and device, electronic equipment and readable storage medium
CN111694983A (en) Information display method, information display device, electronic equipment and storage medium
CN111177462B (en) Video distribution timeliness determination method and device
CN111444819A (en) Cutting frame determining method, network training method, device, equipment and storage medium
CN111949820B (en) Video associated interest point processing method and device and electronic equipment
CN112702619A (en) Anchor interface display method, device, equipment and storage medium
CN111797801A (en) Method and apparatus for video scene analysis
CN111783013A (en) Comment information publishing method, device, equipment and computer-readable storage medium
CN111770182B (en) Data pushing method and device
CN113365090B (en) Object recommendation method, object recommendation device, electronic equipment and readable storage medium
CN111683140B (en) Method and apparatus for distributing messages
CN115278326A (en) Video display method and device, computer readable medium and electronic equipment
CN113965798A (en) Video information generating and displaying method, device, equipment and storage medium
CN112783543A (en) Generation method, device, equipment and medium for small program distribution materials
CN111352685A (en) Input method keyboard display method, device, equipment and storage medium
CN112541145A (en) Page display method, device, equipment and storage medium
CN111914120A (en) Video classification method and device, electronic equipment and computer-readable storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant