CN112672208A - Video playing method, device, electronic equipment, server and system - Google Patents

Video playing method, device, electronic equipment, server and system Download PDF

Info

Publication number
CN112672208A
CN112672208A CN202011381477.0A CN202011381477A CN112672208A CN 112672208 A CN112672208 A CN 112672208A CN 202011381477 A CN202011381477 A CN 202011381477A CN 112672208 A CN112672208 A CN 112672208A
Authority
CN
China
Prior art keywords
video
information
playing
content
videos
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202011381477.0A
Other languages
Chinese (zh)
Other versions
CN112672208B (en
Inventor
杨立
张曼
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Dajia Internet Information Technology Co Ltd
Original Assignee
Beijing Dajia Internet Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Dajia Internet Information Technology Co Ltd filed Critical Beijing Dajia Internet Information Technology Co Ltd
Priority to CN202011381477.0A priority Critical patent/CN112672208B/en
Publication of CN112672208A publication Critical patent/CN112672208A/en
Application granted granted Critical
Publication of CN112672208B publication Critical patent/CN112672208B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The embodiment of the disclosure provides a video playing method, a video playing device, electronic equipment, a server and a video playing system. The method comprises the following steps: responding to a first playing triggering operation, acquiring a first video to be played and associated video information, wherein the associated video information comprises video information of at least one second video which is associated with the first video on video content; playing the first video; and when the playing state of the first video is detected to meet a preset condition, displaying the playing entry of the second video according to the associated video information. Therefore, the embodiment of the disclosure can provide the playing entrance of the associated video in due time according to the state of the user watching the video, so that the user can continuously watch the associated video without complicated operations, and the video watching experience of the user is improved.

Description

Video playing method, device, electronic equipment, server and system
Technical Field
The present disclosure relates to the field of computer technologies, and in particular, to a video playing method, an apparatus, an electronic device, a server, and a system.
Background
With the rapid development of mobile communication technology, various short video APPs emerge endlessly. And the user watches the short videos published by other users in the video APP.
At present, users usually find the interested short videos in the short video APP by means of sliding up and down and browsing at will. After watching an interested short video, a user also wants to watch the short video related to the short video (for example, other short videos in a short video set with a topic published by an author of the short video), and often needs to find a homepage of the author of the short video to search or input a topic of the short video to search, which results in that the user cannot watch the related short video continuously, and short video watching experience of the user is affected.
Disclosure of Invention
The embodiment of the disclosure provides a video playing method, a video playing device, an electronic device, a server and a video playing system, and the technical scheme of the disclosure is as follows:
according to a first aspect of the embodiments of the present disclosure, there is provided a video playing method, the method including:
responding to a first playing triggering operation, acquiring a first video to be played and associated video information, wherein the associated video information comprises video information of at least one second video which is associated with the first video on video content;
playing the first video;
and when the playing state of the first video is detected to meet a preset condition, displaying the playing entry of the second video according to the associated video information.
Optionally, the process of acquiring the associated video information includes:
acquiring videos with authors the same as the author of the first video from pre-stored videos to serve as candidate videos;
and under the condition that the content information of the first video is identified, at least one second video which is associated with the first video on the video content is selected from the candidate videos according to the content information of the first video, and the video information of the selected at least one second video is used as the associated video information.
Optionally, the process of identifying content information of the first video includes:
acquiring characteristic information of the first video;
and inputting the characteristic information of the first video into a pre-established content identification model for identifying video content, and outputting the content information of the first video.
Optionally, the obtaining the feature information of the first video includes:
extracting key frames of the first video;
acquiring color features of each key frame of the first video to serve as feature information of a first dimension of the first video;
acquiring color features of the first video to serve as feature information of a second dimension of the first video;
identifying feature information of a moving object in each key frame of the first video as feature information of a third dimension of the first video;
and acquiring audio characteristic information of the first video to serve as the characteristic information of the fourth dimension of the first video.
Optionally, the obtaining the color feature of the first video includes:
acquiring a first target type of a subject object in each key frame of the first video;
determining a weight value corresponding to the first target type according to a predetermined corresponding relation between the object type and the weight value, wherein the weight value is used as the weight value of a key frame where a subject object to which the first target type belongs is located;
obtaining a product of a color feature and a weight value of each key frame of the first video, wherein the product is used as a first parameter of each key frame of the first video;
and acquiring the sum of the first parameters of each key frame of the first video to serve as the color feature of the first video.
Optionally, the process of establishing the content recognition model includes:
acquiring characteristic information of a plurality of sample videos;
and taking the characteristic information of the sample video and the predetermined content information of the sample video as a training sample, and training the training sample by adopting a recurrent neural network to obtain the content recognition model.
Optionally, the obtaining feature information of a plurality of sample videos includes:
for each of the sample videos, performing the following process:
extracting key frames of the sample video;
acquiring color features of each key frame of the sample video to serve as feature information of a first dimension of the sample video;
acquiring color features of the sample video to serve as feature information of a second dimension of the sample video;
identifying feature information of a moving object in each key frame of the sample video as feature information of a third dimension of the sample video;
and acquiring audio characteristic information of the sample video to serve as the characteristic information of the fourth dimension of the sample video.
Optionally, the acquiring the color feature of the sample video includes:
acquiring a second target type of the subject object in each key frame of the sample video;
determining a weight value corresponding to the second target type according to a predetermined corresponding relation between the object type and the weight value, wherein the weight value is used as the weight value of a key frame where a subject object to which the second target type belongs is located;
obtaining a product of the color feature and the weight value of each key frame of the sample video, and using the product as a second parameter of each key frame of the sample video;
and acquiring the sum of the second parameters of each key frame of the sample video to serve as the color feature of the sample video.
Optionally, the content information includes: episode names of episodes to which the video belongs;
the selecting at least one second video associated with the first video on the video content from the candidate videos according to the content information of the first video comprises:
selecting at least one candidate video with the episode name matched with the episode name of the first video from the candidate videos as a second video related to the first video on the video content.
Optionally, the process of acquiring the associated video information further includes:
under the condition that the content information of the first video is not identified, selecting a candidate video with the similarity between the video cover and the video cover of the first video larger than a preset threshold value from the candidate videos, and taking the video information of the selected candidate video as the associated video information.
Optionally, the selecting, from the candidate videos, a candidate video whose similarity between the video cover and the video cover of the first video is greater than a preset threshold includes:
and selecting a candidate video with the similarity between the video cover and the video cover of the first video larger than a preset threshold value from the candidate videos by adopting a similar image retrieval technology based on a perceptual hash algorithm.
Optionally, the selecting, from the candidate videos, a candidate video with a similarity between a video cover and a video cover of the first video being greater than a preset threshold by using a similar image retrieval technology based on a perceptual hash algorithm includes:
acquiring a first hash value of a video cover of each candidate video and a second hash value of the video cover of the first video;
acquiring the Hamming distance between the second hash value and each first hash value;
and taking the candidate video to which the video cover to which the first Hash value corresponding to the Hamming distance smaller than the preset distance belongs as the candidate video with the similarity between the video cover and the video cover of the first video larger than the preset threshold value.
Optionally, the video information of the second video includes identification information of the second video;
the playing entry for displaying the second video according to the associated video information includes:
acquiring a video cover of a second video to which the target identification information belongs according to the target identification information to serve as a target video cover, wherein the target identification information comprises identification information of the second video related to the associated video information;
and displaying the target video cover as a playing inlet of the second video.
Optionally, the video information of the second video further includes an episode number of the second video;
the displaying the target video cover as a playing entrance of the second video comprises:
displaying the target video cover as a playing entrance of the second video according to the sequence of the target episode numbers;
wherein the target episode number comprises an episode number of a second video to which the associated video information relates.
Optionally, the method further includes:
receiving a second play trigger operation on one of the video covers of the displayed second video;
and responding to the second playing triggering operation, and playing a second video to which the video cover aimed by the second playing triggering operation belongs.
Optionally, the video information of the second video includes an episode number of the second video; the method further comprises the following steps:
displaying a preset icon;
receiving a first preset operation on the preset icon;
and responding to the first preset operation, acquiring the episode serial number of the first video, and playing a second video to which the next serial number of the episode serial number of the first video belongs.
Optionally, the process of acquiring the associated video information includes:
and receiving the associated video information sent by the server.
According to a second aspect of the embodiments of the present disclosure, there is provided a video playing method, the method including:
sending a first video and associated video information to client equipment so as to play the first video on the client equipment, and displaying a play entry of a second video on the client equipment according to the associated video information when the play state of the first video meets a preset condition;
wherein the associated video information comprises video information of at least one of the second videos associated with the first video on video content.
According to a third aspect of the embodiments of the present disclosure, there is provided a video playback apparatus, the apparatus including:
the device comprises an associated information acquisition module, a first display triggering operation and a second display triggering operation, wherein the associated information acquisition module is configured to respond to the first display triggering operation and acquire a first video to be displayed and associated video information, and the associated video information comprises video information of at least one second video which is associated with the first video on video content;
a first playing module configured to play the first video;
and the entrance display module is configured to display the playing entrance of the second video according to the associated video information when detecting that the playing state of the first video meets a preset condition.
According to a fourth aspect of the embodiments of the present disclosure, there is provided a video playback apparatus, the apparatus including:
the information sending module is configured to send a first video and associated video information to client equipment so as to play the first video on the client equipment, and when the playing state of the first video meets a preset condition, a playing entrance of a second video is displayed on the client equipment according to the associated video information;
wherein the associated video information comprises video information of at least one of the second videos associated with the first video on video content.
According to a fifth aspect of embodiments of the present disclosure, there is provided an electronic apparatus including:
a processor;
a memory for storing the processor-executable instructions;
wherein the processor is configured to execute the instructions to implement the video playing method according to the first aspect.
According to a sixth aspect of embodiments of the present disclosure, there is provided a server including:
a processor;
a memory for storing the processor-executable instructions;
wherein the processor is configured to execute the instructions to implement the video playing method according to the second aspect.
According to a seventh aspect of the embodiments of the present disclosure, there is provided a video playing system, including the electronic device of the fifth aspect and the server of the sixth aspect.
According to an eighth aspect of the embodiments of the present disclosure, there is provided a storage medium, wherein instructions of the storage medium, when executed by a processor of an electronic device, enable the electronic device to execute any one of the video playing methods described above.
According to a ninth aspect of the embodiments of the present disclosure, there is provided a computer program product containing instructions which, when run on a computer, cause the computer to implement the video playback method of any one of the above.
According to the technical scheme provided by the embodiment of the disclosure, the first video to be played and the associated video information can be acquired in response to the first playing triggering operation, so that the first video is played, and when the playing state of the first video is detected to meet the preset condition, the playing entry of the second video is displayed according to the associated video information, wherein the associated video information comprises video information of at least one second video managed on the video content with the first video.
Therefore, in the embodiment of the disclosure, when a play trigger operation for a first video is received, video information of the first video and at least one second video associated with the first video on video content is acquired, and when a play state of the first video meets a preset condition, a play entry of the second video can be displayed according to the video information of the second videos, so that the play entry of the associated video can be timely provided according to a state of watching the video by a user, the user can continuously watch the associated video without complicated operations, and the video watching experience of the user is improved.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the disclosure.
Drawings
FIG. 1 is a flow diagram illustrating a video playback method in accordance with an exemplary embodiment;
FIG. 2 is a flow diagram illustrating another method of video playback in accordance with an exemplary embodiment;
FIG. 3 is a flow diagram illustrating a process for building a content recognition model in accordance with an exemplary embodiment;
FIG. 4 is a display diagram of an object video cover, shown in accordance with an exemplary embodiment;
FIG. 5 is a flow diagram illustrating another method of video playback in accordance with an exemplary embodiment;
FIG. 6 is a flow diagram illustrating another method of video playback in accordance with an exemplary embodiment;
FIG. 7 is a block diagram illustrating a video playback device in accordance with an exemplary embodiment;
FIG. 8 is a block diagram illustrating another video playback device in accordance with an exemplary embodiment;
FIG. 9 is a block diagram illustrating an electronic device in accordance with an exemplary embodiment;
FIG. 10 is a block diagram illustrating another electronic device in accordance with an exemplary embodiment;
FIG. 11 is a block diagram illustrating a server in accordance with an example embodiment.
Detailed Description
In order to make the technical solutions of the present disclosure better understood by those of ordinary skill in the art, the technical solutions in the embodiments of the present disclosure will be clearly and completely described below with reference to the accompanying drawings.
It should be noted that the terms "first," "second," and the like in the description and claims of the present disclosure and in the above-described drawings are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used is interchangeable under appropriate circumstances such that the embodiments of the disclosure described herein are capable of operation in sequences other than those illustrated or otherwise described herein. The implementations described in the exemplary embodiments below are not intended to represent all implementations consistent with the present disclosure. Rather, they are merely examples of apparatus and methods consistent with certain aspects of the present disclosure, as detailed in the appended claims.
In order to solve the problem that a user cannot continuously watch associated short videos in the prior art, so that the short video watching experience of the user is affected, embodiments of the present disclosure provide a video playing method, an apparatus, an electronic device, a server, and a system.
In one aspect of the disclosed embodiments, a video playing method is provided, which can be applied to an electronic device. As shown in fig. 1, the video playing method may include the following steps:
step 101: and responding to the first playing triggering operation, and acquiring the first video to be played and the associated video information.
The first playing trigger operation may be a click operation on a playing entry of the first video (e.g., a video cover of the first video).
In addition, the associated video information includes video information of at least one second video associated with the first video on the video content. Related videos on video content belong to the same episode, where an episode is typically a video in the form of a series or a story. Optionally, the video information of the second video may include identification information of the second video. The identification information may be a video ID.
As can be seen from the above, in the embodiment of the present disclosure, when the play trigger operation for the first video is received, the video information of the first video and at least one second video associated with the first video on the video content are acquired together.
Step 102: and playing the first video.
Step 103: and when the playing state of the first video is detected to meet a preset condition, displaying the playing entry of the second video according to the associated video information.
The playing entry of the second video is an entry into the playing interface of the second video, and for example, if the playing entry of the second video is clicked, the second video can be played.
In a possible implementation manner, the preset condition includes that the remaining playing time of the first video reaches a preset time. That is, the client device may display the play entry of the second video when the remaining play time of the first video reaches a preset time (e.g., 3 seconds) after the first video is played.
As can be seen from the foregoing steps 101 to 103, the video playing method provided in the embodiment of the present disclosure can obtain, in response to the first play trigger operation, a first video to be played and associated video information, so as to play the first video, and when it is detected that a playing state of the first video meets a preset condition, show a playing entry of a second video according to the associated video information, where the associated video information includes video information of at least one second video managed on video content with the first video.
Therefore, according to the video playing method provided by the embodiment of the disclosure, when a play trigger operation for a first video is received, video information of the first video and at least one second video associated with the first video on video content can be acquired, and when a play state of the first video meets a preset condition, a play entry of the second video can be displayed according to the video information of the second videos, so that the play entry of the associated video can be timely provided according to a state of watching the video by a user, the user can continuously watch the associated video without complicated operations, and the video watching experience of the user is improved.
In another aspect of the present disclosure, a video playing method is also provided, and the video playing method may be applied to an electronic device. As shown in fig. 2, the video playing method may include the following steps:
step 201: and responding to the first playing triggering operation, and acquiring the first video to be played and the associated video information.
The first playing trigger operation may be a click operation on a playing entry of the first video (e.g., a video cover of the first video).
In addition, the associated video information includes video information of at least one second video associated with the first video on the video content. Related videos on video content belong to the same episode, where an episode is typically a video in the form of a series or a story. The video information of the second video includes identification information of the second video. The identification information may be a video ID.
As can be seen from the above, in the embodiment of the present disclosure, when the play trigger operation for the first video is received, the video information of the first video and at least one second video associated with the first video on the video content are acquired together.
Step 202: and playing the first video.
Step 203: and when the playing state of the first video is detected to meet the preset condition, acquiring a video cover of a second video to which the target identification information belongs according to the target identification information to serve as a target video cover.
The playing entry of the second video is an entry into the playing interface of the second video, and for example, if the playing entry of the second video is clicked, the second video can be played.
In a possible implementation manner, the preset condition includes that the remaining playing time of the first video reaches a preset time. That is, the client device may execute the step 203 when the remaining playing time of the first video reaches a preset time (e.g., 3 seconds) after the first video is played.
In addition, the target identification information includes identification information of a second video to which the associated video information relates. That is, in the embodiment of the present disclosure, when it is detected that the playing state of the first video meets the preset condition, a video cover of the second video is obtained according to the identification information of the second video related to the associated video information, so as to be displayed as a playing entry of the second video. For example, the identification information of the second video related to the associated video information includes identification information of A, B, C three videos, and video covers of the three videos need to be acquired A, B, C for displaying.
In addition, after the client device receives the first play trigger operation, the client device itself may obtain a video cover of the second video according to the identification information of the second video, or the client device may request the server for the video cover of the second video according to the identification information of the second video.
Step 204: and displaying the target video cover as a playing inlet of the second video.
As can be seen from the foregoing steps 201 to 204, the video playing method provided by the embodiment of the present disclosure can obtain, in response to the first play trigger operation, the first video to be played and the associated video information, so as to play the first video, and when it is detected that the play state of the first video meets the preset condition, obtain, according to the identification information of the second video related to the associated video information, a video cover of the second video, so as to serve as a play entry of the second video for displaying, where the associated video information includes video information of at least one second video that is managed on the video content with the first video.
Therefore, according to the video playing method provided by the embodiment of the disclosure, when a play trigger operation for a first video is received, identification information of the first video and at least one second video associated with the first video on video content is acquired, and when a play state of the first video meets a preset condition, a video cover of the second video can be acquired according to the identification information of the second videos to be displayed as a play entry of the second video, so that the play entry of the associated video can be timely provided according to a state of watching the video by a user, the user can continuously watch the associated video without complicated operations, and the video watching experience of the user is improved.
In one possible implementation, the process of obtaining the associated video information includes:
acquiring videos with authors the same as the author of the first video from pre-stored videos to serve as candidate videos;
and under the condition that the content information of the first video is identified, at least one second video which is associated with the first video on the video content is selected from the candidate videos according to the content information of the first video, and the video information of the selected at least one second video is used as the associated video information.
Optionally, the content information includes an episode name, an episode number, and an episode type.
As can be seen from the above, in the embodiment of the present disclosure, after receiving the first play trigger operation, the client device may automatically identify content information of the first video, so that at least one second video associated with the first video on video content may be selected from videos that have the same author as the author of the first video and are stored in advance according to the content information of the first video.
In addition, when a user watches a video in a short video APP by using a client device, after watching a video, if other videos related to the video on the video content need to be searched, the videos of the same author are often focused on, but in the embodiment of the present disclosure, videos that belong to the same author as a first video can be screened out from pre-stored videos to be used as candidate videos, and then a second video related to the first video on the video content is selected from the candidate videos, that is, in the embodiment of the present disclosure, a second video that belongs to the same author as the first video and is related to the video content can be automatically selected out, so that a video cover of the second video that belongs to the same author as the first video and is related to the video content can be displayed on the client, so as to be closer to the actual use needs of the user, and further, the user does not need to screen the video for the author by himself.
In one possible implementation, the process of identifying content information of the first video includes:
acquiring characteristic information of the first video;
and inputting the characteristic information of the first video into a pre-established content identification model for identifying video content, and outputting the content information of the first video.
Therefore, in the embodiment of the disclosure, a content identification model for identifying video content can be established in advance, and then when the content information of the first video needs to be identified, the feature information of the first video can be extracted, so that the feature information of the first video is input into the content identification model, and then the content information of the first video can be output, and the whole process is simple and convenient.
Based on the prior art, a mode of determining at least one second video associated with a first video on video content is easily conceivable, and often, a setting option is provided in a short video APP, and in a process of uploading videos in the short video APP by an author, content information of the uploaded short videos is marked in the setting option, so that a server counts which videos are associated on the video content according to the content information of each video marked by the author. However, the development difficulty of the short video APP is high, and the author uploading the video needs to be guided to label the video content information, which increases the burden on the author uploading the video.
In the embodiment of the disclosure, the content information of the video is automatically identified through the pre-established content identification model, setting options do not need to be provided in the short video APP, and an author does not need to manually label the content information of the video, so that the process of identifying the content information of the video is greatly simplified, and the interactive operation experience of a user is further improved.
In a possible implementation manner, the obtaining feature information of the first video includes:
extracting key frames of the first video;
acquiring color features of each key frame of the first video to serve as feature information of a first dimension of the first video;
acquiring color features of the first video to serve as feature information of a second dimension of the first video;
identifying feature information of a moving object in each key frame of the first video as feature information of a third dimension of the first video;
and acquiring audio characteristic information of the first video to serve as the characteristic information of the fourth dimension of the first video.
The frame is a single image frame of the minimum unit in the motion picture, and is equivalent to each shot on the motion picture film. The frames appear as a grid or a marker on the time axis of the animation software. The key frame, which is equivalent to the original picture in the two-dimensional animation, refers to the frame where the key action in the movement or change of the character or object is located.
In addition, the color feature may be an RGB feature.
Further, the above-mentioned audio feature information may include at least one of an instrument adopted by the background music in the video, a pitch, a frequency, a decibel of a sound other than the background music present in the video, and voice information of a person present in the video.
As can be seen from this, the video feature information of the first video includes feature information of a plurality of dimensions, for example, the above-mentioned feature information of the first dimension of the first video (may be referred to as a frame-level picture color feature), the feature information of the second dimension (may be referred to as a video-level color feature), the feature information of the third dimension, and the feature information of the fourth dimension. Compared with the method for identifying the content information of the first video according to the single-dimensional feature information of the first video, in the embodiment of the disclosure, the content information of the first video is identified by extracting the feature information of multiple dimensions of the first video, so that the content information of the first video can be identified more accurately, that is, the accuracy of identifying the content information of the video is further improved.
In a possible implementation, the obtaining the color feature of the first video includes:
acquiring a first target type of a subject object in each key frame of the first video;
determining a weight value corresponding to the first target type according to a predetermined corresponding relation between the object type and the weight value, wherein the weight value is used as the weight value of a key frame where a subject object to which the first target type belongs is located;
obtaining a product of a color feature and a weight value of each key frame of the first video, wherein the product is used as a first parameter of each key frame of the first video;
and acquiring the sum of the first parameters of each key frame of the first video to serve as the color feature of the first video.
For example, 5 key frames are extracted from the first video, and the color features of each key frame are X1, X2, X3, X4, and X5, respectively, and the weight values of each key frame are k1, k2, k3, k4, and k5, respectively, then the feature information of the second dimension of the first video is X1 × k1+ X2 × k2+ X3 × k3+ X4 × 4+ X5 × k 5.
In a possible implementation, the establishing process of the content recognition model includes:
acquiring characteristic information of a plurality of sample videos;
and taking the characteristic information of the sample video and the predetermined content information of the sample video as a training sample, and training the training sample by adopting a recurrent neural network to obtain the content recognition model.
Therefore, in the embodiment of the disclosure, the content information of a plurality of sample videos can be predetermined, then the feature information of the plurality of sample videos is extracted, further, the feature information and the content information of one sample video are used as one training sample, the recurrent neural network is adopted to train the training sample, and finally, the content identification model for identifying the video content can be obtained.
In a possible implementation, the obtaining feature information of a plurality of sample videos includes:
for each of the sample videos, performing the following process:
extracting key frames of the sample video;
acquiring color features of each key frame of the sample video to serve as feature information of a first dimension of the sample video;
acquiring color features of the sample video to serve as feature information of a second dimension of the sample video;
identifying feature information of a moving object in each key frame of the sample video as feature information of a third dimension of the sample video;
and acquiring audio characteristic information of the sample video to serve as the characteristic information of the fourth dimension of the sample video.
Therefore, the process of obtaining the feature information of each sample video is the same as the process of obtaining the feature information of the first video, and is not repeated here.
The method comprises the steps of extracting characteristic information of multiple dimensions of a sample video, and establishing a content identification model according to the characteristic information, so that the established content identification model can identify the content information of the video more accurately.
In a possible implementation, the obtaining the color feature of the sample video includes:
acquiring a second target type of the subject object in each key frame of the sample video;
determining a weight value corresponding to the second target type according to a predetermined corresponding relation between the object type and the weight value, wherein the weight value is used as the weight value of a key frame where a subject object to which the second target type belongs is located;
obtaining a product of the color feature and the weight value of each key frame of the sample video, and using the product as a second parameter of each key frame of the sample video;
and acquiring the sum of the second parameters of each key frame of the sample video to serve as the color feature of the sample video.
Therefore, the process of obtaining the color feature of each sample video is the same as the process of obtaining the color feature of the first video, and is not repeated here.
As can be seen from the above, in the embodiment of the present disclosure, a specific implementation of establishing a content identification module may be as shown in fig. 3. The details are as follows:
firstly, collecting a plurality of video samples;
secondly, video features are extracted, namely the following processes are executed for each sample video:
the first step is as follows: extracting key frames of the sample video (namely, a video frame with a complete picture, such as a shaking head video, and a video frame with a complete face as a key frame), and extracting audio information;
the second step is that: aiming at the key frames, acquiring the RGB characteristics of each key frame;
the third step: obtaining a subject object included in each key frame, determining the weight of each key frame according to the corresponding relation between the predetermined object type and the weight, and then calculating the product of the RGB characteristics of the key frame and the weight thereof, so as to add all the products to obtain a weighted average value, wherein the weighted average value is the video-level RGB characteristics;
the fourth step: identifying which objects in the keyframes are moving objects;
the fifth step: extracting audio features from the audio information;
and finally: and establishing a multi-modal fused classification/identification model (namely establishing a content identification model) according to the different dimensional features extracted in the steps.
In one possible implementation, the content information includes: episode names of episodes to which the video belongs;
the selecting at least one second video associated with the first video on the video content from the candidate videos according to the content information of the first video comprises:
selecting at least one candidate video with the episode name matched with the episode name of the first video from the candidate videos as a second video related to the first video on the video content.
As can be seen, in the embodiment of the present disclosure, when the episode name of the first video is identified, the second video associated with the first video on the video content may be directly searched for in the candidate videos according to the episode name.
In one possible embodiment, the method further comprises:
under the condition that the content information of the first video is not identified, selecting a candidate video with the similarity between the video cover and the video cover of the first video larger than a preset threshold value from the candidate videos, and taking the video information of the selected candidate video as the associated video information.
Therefore, in the embodiment of the disclosure, if the content information of the first video is not identified, the second video related to the first video on the video content can be selected from the candidate videos according to the video cover of the first video, so that the deficiency that the content information of the video cannot be identified is made up by using the way that the video cover identifies the related video.
In a possible implementation manner, the selecting, from the candidate videos, a candidate video whose video cover has a similarity with a video cover of the first video greater than a preset threshold includes:
and selecting a candidate video with the similarity between the video cover and the video cover of the first video larger than a preset threshold value from the candidate videos by adopting a similar image retrieval technology based on a perceptual hash algorithm.
Therefore, the video cover with the similarity larger than the preset value can be selected by adopting the image retrieval technology based on the perceptual hash algorithm. It is to be understood that the method for selecting a video cover with a similarity greater than a preset value is not limited thereto.
In a possible implementation manner, the selecting, from the candidate videos, a candidate video with a similarity between a video cover and a video cover of the first video greater than a preset threshold value by using a similar image retrieval technology based on a perceptual hash algorithm includes:
acquiring a first hash value of a video cover of each candidate video and a second hash value of the video cover of the first video;
acquiring the Hamming distance between the second hash value and each first hash value;
and taking the candidate video to which the video cover to which the first Hash value corresponding to the Hamming distance smaller than the preset distance belongs as the candidate video with the similarity between the video cover and the video cover of the first video larger than the preset threshold value.
The perceptual hash algorithm is to generate a 'fingerprint' character string for each picture by using the hash algorithm, then compare fingerprints of different pictures, and the smaller the hamming distance between the fingerprints is, the more similar the pictures are. For example, if the hamming distance is 0, it means that the two pictures are very similar; for example, a hamming distance less than 5 indicates a difference, but a close proximity; for example, a hamming distance greater than 10 indicates a completely different picture.
In addition, the process of calculating the hash value of a video cover can be as follows:
firstly, reducing the size of a video cover; among them, the fastest method to remove high frequency and detail is to downscale the picture to 8 × 8 size for a total of 64 pixels, wherein the aspect ratio is not maintained but only changed to 8 × 8 squares, so that the picture difference caused by different sizes and scales can be eliminated.
Secondly, simplifying colors, namely converting 8-by-8 pictures into gray level images;
thirdly, calculating an average value, namely calculating the gray level average value of all 64 pixels;
thirdly, comparing the gray scale of the pixels, namely comparing the gray scale of each pixel with the average value, wherein the gray scale is greater than or equal to the average value and is marked as 1; less than the average, noted as 0.
And finally, calculating a hash value, namely combining the comparison results of the previous step together to form a 64-bit integer, namely the hash value of the video cover.
In one possible embodiment, the content information further includes an episode number of the video in an episode to which the video belongs, and the method further includes:
acquiring an episode serial number of the second video;
and storing the second video and the first video in a target container according to the sequence of the episode numbers of the second video and the identified episode number of the first video.
That is, in the embodiments of the present disclosure, videos associated on video content are stored in one target container and sorted according to their episode number.
If the second video related to the first video on the video content is not found in the candidate videos, a target container is newly established, so that the first video is stored in the newly established target container, and after other videos are subsequently received, the first video is matched with the newly established target container on the video content.
In one possible embodiment, the content information further includes an episode type of an episode to which the video belongs, and the method further includes:
and classifying and storing the target container according to the episode type of the video stored in the target container.
The videos stored in the same target container are related on the video content, so that the episode types of the videos stored in the same target container are the same or similar. And further according to the episode types of the videos stored in the target container, the target container is classified and stored, so that the videos with the same or similar episode types can be further stored together, and further the classification management of the videos is facilitated.
In one possible embodiment, the video information of the second video further includes an episode number of the second video;
the displaying the target video cover as a playing entrance of the second video comprises:
displaying the target video cover as a playing entrance of the second video according to the sequence of the target episode numbers;
wherein the target episode number comprises an episode number of a second video to which the associated video information relates.
The target video covers are displayed according to the sequence of the episode numbers, so that a user can be more clear of the episode sequence of the second videos to which the target video covers belong, and the user can conveniently select which second video to watch.
In addition, the episode number of the second video belonging to the target video cover can be displayed on the target video cover, so that the user can know the episode number of the second video belonging to the target video cover more clearly, and the user can select the second video to be watched conveniently.
Specifically, for example, the target video covers include a first cover 401, a second cover 402, and a third cover 403, and the order of the episode numbers is the first cover 401, the second cover 402, and the third cover 403, so that the three video covers can be displayed according to the order of the episode numbers and according to the position distribution from top to bottom, as shown in fig. 4.
In one possible embodiment, the method further comprises:
receiving a second play trigger operation on one of the video covers of the displayed second video;
and responding to the second playing triggering operation, and playing a second video to which the video cover aimed by the second playing triggering operation belongs.
The second playing triggering operation may be a click operation on a video cover of the second video.
Therefore, in the embodiment of the disclosure, a user can select one video cover from target video covers displayed on the client device, so that the client device can be triggered to play the second video to which the selected video cover belongs, that is, the user can select which second video associated with the first video on the video content to watch according to the own requirement, and the interactive operation experience of the user is further improved.
In one possible embodiment, the video information of the second video includes an episode number of the second video; the method further comprises the following steps:
displaying a preset icon;
receiving a first preset operation on the preset icon;
and responding to the first preset operation, acquiring the episode serial number of the first video, and playing a second video to which the next serial number of the episode serial number of the first video belongs.
The preset icon can be displayed when the playing state of the first video meets the preset condition. In addition. The first preset operation may be a click operation on a preset icon.
As can be seen from the above, in the embodiment of the present disclosure, the client device may further obtain the episode number of the first video and the episode number of the second video associated with the first video on the video content, so that after the user performs the first preset operation on the preset icon displayed on the client device, the client device may automatically play the second video to which the next sequence number of the episode number of the first video belongs, thereby avoiding the user from selecting the second video to which the next sequence number of the episode number of the first video belongs by himself, and further improving the interactive operation experience of the user.
In one possible implementation, the process of obtaining the associated video information includes:
and receiving the associated video information sent by the server.
As can be seen, in embodiments of the present disclosure, the associated video information (i.e., the video information of the at least one second video associated with the first video on the video content) may also be sent by the server to the client device. In this case, the server needs to acquire the associated video information first and then send the associated video information to the client device. The process of acquiring the management video information by the server may be the same as the process of acquiring the associated video information by the client device, which is described in the foregoing, and is not repeated here.
Therefore, in the embodiment of the present disclosure, when the server issues the first video to the client device, the server may issue the first video to the client device together with the video information of the at least one second video associated with the first video on the video content.
Specifically, the server may send the first video together with video information of at least one second video associated with the first video on the video to the client device when pushing the first video to the client device; or the server may further send the first video and video information of at least one second video associated with the first video on the video content to the client device after receiving a request message indicating that the client device requests the first video.
After receiving the video information of the first video pushed by the server and the at least one second video associated with the first video, the client device may play the first video when receiving a first play trigger operation of a user for the first video, so as to send a request message for requesting a video cover of the second video to the server when a play state of the first video meets a preset condition.
For example, after a certain short video APP is started, a server pushes a first video to a client device according to a browsing record of a user, and then the server can push the first video to the client device together with identification information of at least one second video associated with the first video on video content when pushing the first video, so that the client device starts to play the first video after receiving a first play trigger operation of the user for the first video, and further sends first request information to the server by the client device when the remaining play duration of the first video reaches a preset duration, that is, a video cover of the second video to which the identification information belongs is requested from the server according to the identification information of the at least one second video, and the video covers are displayed after receiving the video cover of the second video sent by the server, so that the client device does not need to switch a play interface of the first video to other interfaces, the user can watch the video cover of at least one second video related to the first video on the video content, and the tedious operation that the user enters an author brief introduction interface to search the second videos related to the first video on the video content one by one is avoided.
In summary, for more and more short video APP producers can upload videos related to a plurality of sets of contents, the embodiment of the present disclosure can employ a Multimedia intelligent identification (MMU) technology to identify video contents, associate the video contents with episodes related to an author, and then issue identification information of the associated episodes to a client device, so that a user can quickly and smoothly locate a next set of currently browsed videos, viewing experience is improved, and content that is laid by a producer and is elaborately acquired by viewers quickly, thereby establishing a relationship between the producer and the consumers.
In another aspect of the present disclosure, a video playing method is also provided, and the video playing method may be applied to a server. As shown in fig. 5, the video playing method may include the following steps:
step 501: sending a first video and associated video information to client equipment so as to play the first video on the client equipment, and displaying a playing entry of a second video on the client equipment according to the associated video information when the playing state of the first video meets a preset condition.
Wherein the associated video information comprises video information of at least one of the second videos associated with the first video on video content. Related videos on video content belong to the same episode, where an episode is typically a video in the form of a series or a story. Optionally, the video information of the second video may include identification information of the second video. The identification information may be a video ID.
As can be seen from the above, in the embodiments of the present disclosure, when pushing a first video to a client device, a server may push to the client device together with video information of at least one second video associated with the first video on video content.
In addition, the playing entry of the second video is an entry into the playing interface of the second video, and for example, if the playing entry of the second video is clicked, the second video can be played.
In a possible implementation manner, the preset condition includes that the remaining playing time of the first video reaches a preset time. That is, the client device may display the play entry of the second video when the remaining play time of the first video reaches a preset time (e.g., 3 seconds) after the first video is played.
To sum up, the video playing method provided by the embodiment of the present disclosure, when sending the first video to the client device, sends the first video to the client device together with the video information of at least one second video associated with the first video on the video content, so that when the state of playing the first video by the client device meets the preset condition, the playing entry of the second video is displayed according to the associated video information, thereby realizing that the playing entry of the associated video is timely provided according to the state of watching the video by the user, so that the user can continuously watch the associated video without complicated operations, and improving the video watching experience of the user.
In another aspect of the present disclosure, a video playing method is also provided, and the video playing method may be applied to a server. As shown in fig. 6, the video playing method may include the following steps:
step 601: in the pre-stored video, a video with the same author as that of the first video is acquired as a candidate video.
Wherein, when a user watches a video in the short video APP, after finishing watching a video, if other videos related to the video on the video content need to be searched, the video of the same author is often focused on, in the embodiment of the disclosure, a video belonging to the same author as the first video may be screened out from the pre-stored videos as a candidate video, a second video associated with the first video on the video content is then selected from the candidate videos, that is, in the embodiment of the present disclosure, a second video belonging to the same author as the first video and associated with the video content can be automatically selected, so that a video cover of a second video belonging to the same author as the first video and associated on the video content can be displayed on the client, therefore, the method is closer to the actual use needs of the users, and the users do not need to screen the videos aiming at the authors.
Step 602: and under the condition that the content information of the first video is identified, at least one second video which is associated with the first video on the video content is selected from the candidate videos according to the content information of the first video, and the video information of the selected at least one second video is used as the associated video information.
In one possible embodiment, the content information includes an episode name, an episode number, and an episode type.
In an embodiment of the present disclosure, content information of a first video may be automatically identified, so that at least one second video associated with the first video on video content may be selected from among pre-stored videos having the same author as the author of the first video, according to the content information of the first video.
Step 603: and sending the first video and the associated video information to client equipment so as to play the first video on the client equipment, and displaying a playing entry of a second video on the client equipment according to the associated video information when the playing state of the first video meets a preset condition.
As can be seen from the above, the associated video information includes video information of at least one of the second videos associated with the first video on the video content. Related videos on video content belong to the same episode, where an episode is typically a video in the form of a series or a story. Optionally, the video information of the second video may include identification information of the second video. The identification information may be a video ID.
As can be seen, in embodiments of the present disclosure, a server may push a first video to a client device along with video information of at least one second video associated with the first video on video content.
The playing entry of the second video is an entry into the playing interface of the second video, and for example, if the playing entry of the second video is clicked, the second video can be played.
In a possible implementation manner, the preset condition includes that the remaining playing time of the first video reaches a preset time. That is, the client device may display the play entry of the second video when the remaining play time of the first video reaches a preset time (e.g., 3 seconds) after the first video is played.
In summary, the video playing method provided by the embodiment of the present disclosure can automatically identify the content information of the first video, thereby screening a second video associated with the first video on the video content among candidate videos identical to the author of the first video according to the content information of the first video, and further sending the first video to the client device together with video information of at least one second video associated with the first video on the video content when the first video is sent to the client device, so that when the state of playing the first video by the client device meets the preset condition, the playing entrance of the second video is displayed according to the associated video information, thereby realizing that the playing entrance of the associated video is provided at the right time according to the state of the user watching the video, therefore, the user can continuously watch the associated video without complicated operation, and the video watching experience of the user is improved.
In one possible implementation, the process of identifying content information of the first video includes:
acquiring characteristic information of the first video;
and inputting the characteristic information of the first video into a pre-established content identification model for identifying video content, and outputting the content information of the first video.
Therefore, in the embodiment of the disclosure, a content identification model for identifying video content can be established in advance, and then when the content information of the first video needs to be identified, the feature information of the first video can be extracted, so that the feature information of the first video is input into the content identification model, and then the content information of the first video can be output, and the whole process is simple and convenient.
Based on the prior art, a mode of determining at least one second video associated with a first video on video content is easily conceivable, and often, a setting option is provided in a short video APP, and in a process of uploading videos in the short video APP by an author, content information of the uploaded short videos is marked in the setting option, so that a server counts which videos are associated on the video content according to the content information of each video marked by the author. However, the development difficulty of the short video APP is high, and the author uploading the video needs to be guided to label the video content information, which increases the burden on the author uploading the video.
In the embodiment of the disclosure, the content information of the video is automatically identified through the pre-established content identification model, setting options do not need to be provided in the short video APP, and an author does not need to manually label the content information of the video, so that the process of identifying the content information of the video is greatly simplified, and the interactive operation experience of a user is further improved.
In a possible implementation manner, the obtaining feature information of the first video includes:
extracting key frames of the first video;
acquiring color features of each key frame of the first video to serve as feature information of a first dimension of the first video;
acquiring color features of the first video to serve as feature information of a second dimension of the first video;
identifying feature information of a moving object in each key frame of the first video as feature information of a third dimension of the first video;
and acquiring audio characteristic information of the first video to serve as the characteristic information of the fourth dimension of the first video.
The frame is a single image frame of the minimum unit in the motion picture, and is equivalent to each shot on the motion picture film. The frames appear as a grid or a marker on the time axis of the animation software. The key frame, which is equivalent to the original picture in the two-dimensional animation, refers to the frame where the key action in the movement or change of the character or object is located.
In addition, the color feature may be an RGB feature.
Further, the above-mentioned audio feature information may include at least one of an instrument adopted by the background music in the video, a pitch, a frequency, a decibel of a sound other than the background music present in the video, and voice information of a person present in the video.
As can be seen from this, the video feature information of the first video includes feature information of a plurality of dimensions, for example, the above-mentioned feature information of the first dimension of the first video (may be referred to as a frame-level picture color feature), the feature information of the second dimension (may be referred to as a video-level color feature), the feature information of the third dimension, and the feature information of the fourth dimension. Compared with the method for identifying the content information of the first video according to the single-dimensional feature information of the first video, in the embodiment of the disclosure, the content information of the first video is identified by extracting the feature information of multiple dimensions of the first video, so that the content information of the first video can be identified more accurately, that is, the accuracy of identifying the content information of the video is further improved.
In a possible implementation, the obtaining the color feature of the first video includes:
acquiring a first target type of a subject object in each key frame of the first video;
determining a weight value corresponding to the first target type according to a predetermined corresponding relation between the object type and the weight value, wherein the weight value is used as the weight value of a key frame where a subject object to which the first target type belongs is located;
obtaining a product of a color feature and a weight value of each key frame of the first video, wherein the product is used as a first parameter of each key frame of the first video;
and acquiring the sum of the first parameters of each key frame of the first video to serve as the color feature of the first video.
For example, 5 key frames are extracted from the first video, and the color features of each key frame are X1, X2, X3, X4, and X5, respectively, and the weight values of each key frame are k1, k2, k3, k4, and k5, respectively, then the feature information of the second dimension of the first video is X1 × k1+ X2 × k2+ X3 × k3+ X4 × 4+ X5 × k 5.
In a possible implementation, the establishing process of the content recognition model includes:
acquiring characteristic information of a plurality of sample videos;
and taking the characteristic information of the sample video and the predetermined content information of the sample video as a training sample, and training the training sample by adopting a recurrent neural network to obtain the content recognition model.
Therefore, in the embodiment of the disclosure, the content information of a plurality of sample videos can be predetermined, then the feature information of the plurality of sample videos is extracted, further, the feature information and the content information of one sample video are used as one training sample, the recurrent neural network is adopted to train the training sample, and finally, the content identification model for identifying the video content can be obtained.
In a possible implementation, the obtaining feature information of a plurality of sample videos includes:
for each of the sample videos, performing the following process:
extracting key frames of the sample video;
acquiring color features of each key frame of the sample video to serve as feature information of a first dimension of the sample video;
acquiring color features of the sample video to serve as feature information of a second dimension of the sample video;
identifying feature information of a moving object in each key frame of the sample video as feature information of a third dimension of the sample video;
and acquiring audio characteristic information of the sample video to serve as the characteristic information of the fourth dimension of the sample video.
Therefore, the process of obtaining the feature information of each sample video is the same as the process of obtaining the feature information of the first video, and is not repeated here.
The method comprises the steps of extracting characteristic information of multiple dimensions of a sample video, and establishing a content identification model according to the characteristic information, so that the established content identification model can identify the content information of the video more accurately.
In a possible implementation, the obtaining the color feature of the sample video includes:
acquiring a second target type of the subject object in each key frame of the sample video;
determining a weight value corresponding to the second target type according to a predetermined corresponding relation between the object type and the weight value, wherein the weight value is used as the weight value of a key frame where a subject object to which the second target type belongs is located;
obtaining a product of the color feature and the weight value of each key frame of the sample video, and using the product as a second parameter of each key frame of the sample video;
and acquiring the sum of the second parameters of each key frame of the sample video to serve as the color feature of the sample video.
Therefore, the process of obtaining the color feature of each sample video is the same as the process of obtaining the color feature of the first video, and is not repeated here.
As can be seen from the above, in the embodiment of the present disclosure, a specific implementation of establishing a content identification module may be as shown in fig. 3. The details are as follows:
firstly, collecting a plurality of video samples;
secondly, video features are extracted, namely the following processes are executed for each sample video:
the first step is as follows: extracting key frames of the sample video (namely, a video frame with a complete picture, such as a shaking head video, and a video frame with a complete face as a key frame), and extracting audio information;
the second step is that: aiming at the key frames, acquiring the RGB characteristics of each key frame;
the third step: obtaining a subject object included in each key frame, determining the weight of each key frame according to the corresponding relation between the predetermined object type and the weight, and then calculating the product of the RGB characteristics of the key frame and the weight thereof, so as to add all the products to obtain a weighted average value, wherein the weighted average value is the video-level RGB characteristics;
the fourth step: identifying which objects in the keyframes are moving objects;
the fifth step: extracting audio features from the audio information;
and finally: and establishing a multi-modal fused classification/identification model (namely establishing a content identification model) according to the different dimensional features extracted in the steps.
In one possible implementation, the content information includes: episode names of episodes to which the video belongs;
the selecting at least one second video associated with the first video on the video content from the candidate videos according to the content information of the first video comprises:
selecting at least one candidate video with the episode name matched with the episode name of the first video from the candidate videos as a second video related to the first video on the video content.
As can be seen, in the embodiment of the present disclosure, when the episode name of the first video is identified, the second video associated with the first video on the video content may be directly searched for in the candidate videos according to the episode name.
In one possible embodiment, the method further comprises:
under the condition that the content information of the first video is not identified, selecting a candidate video with the similarity between the video cover and the video cover of the first video larger than a preset threshold value from the candidate videos, and taking the video information of the selected candidate video as the associated video information.
Therefore, in the embodiment of the disclosure, if the content information of the first video is not identified, the second video related to the first video on the video content can be selected from the candidate videos according to the video cover of the first video, so that the deficiency that the content information of the video cannot be identified is made up by using the way that the video cover identifies the related video.
In a possible implementation manner, the selecting, from the candidate videos, a candidate video whose video cover has a similarity with a video cover of the first video greater than a preset threshold includes:
and selecting a candidate video with the similarity between the video cover and the video cover of the first video larger than a preset threshold value from the candidate videos by adopting a similar image retrieval technology based on a perceptual hash algorithm.
Therefore, the video cover with the similarity larger than the preset value can be selected by adopting the image retrieval technology based on the perceptual hash algorithm. It is to be understood that the method for selecting a video cover with a similarity greater than a preset value is not limited thereto.
In a possible implementation manner, the selecting, from the candidate videos, a candidate video with a similarity between a video cover and a video cover of the first video greater than a preset threshold value by using a similar image retrieval technology based on a perceptual hash algorithm includes:
acquiring a first hash value of a video cover of each candidate video and a second hash value of the video cover of the first video;
acquiring the Hamming distance between the second hash value and each first hash value;
and taking the candidate video to which the video cover to which the first Hash value corresponding to the Hamming distance smaller than the preset distance belongs as the candidate video with the similarity between the video cover and the video cover of the first video larger than the preset threshold value.
Therefore, the video cover with the similarity larger than the preset value can be selected by adopting the image retrieval technology based on the perceptual hash algorithm. It is to be understood that the method for selecting a video cover with a similarity greater than a preset value is not limited thereto.
The perceptual hash algorithm is to generate a 'fingerprint' character string for each picture by using the hash algorithm, then compare fingerprints of different pictures, and the smaller the hamming distance between the fingerprints is, the more similar the pictures are. For example, if the hamming distance is 0, it means that the two pictures are very similar; for example, a hamming distance less than 5 indicates a difference, but a close proximity; for example, a hamming distance greater than 10 indicates a completely different picture.
In addition, the process of calculating the hash value of a video cover can be as follows:
firstly, reducing the size of a video cover; among them, the fastest method to remove high frequency and detail is to downscale the picture to 8 × 8 size for a total of 64 pixels, wherein the aspect ratio is not maintained but only changed to 8 × 8 squares, so that the picture difference caused by different sizes and scales can be eliminated.
Secondly, simplifying colors, namely converting 8-by-8 pictures into gray level images;
thirdly, calculating an average value, namely calculating the gray level average value of all 64 pixels;
thirdly, comparing the gray scale of the pixels, namely comparing the gray scale of each pixel with the average value, wherein the gray scale is greater than or equal to the average value and is marked as 1; less than the average, noted as 0.
And finally, calculating a hash value, namely combining the comparison results of the previous step together to form a 64-bit integer, namely the hash value of the video cover.
In one possible embodiment, the content information further includes an episode number of the video in an episode to which the video belongs, and the method further includes:
acquiring an episode serial number of the second video;
and storing the second video and the first video in a target container according to the sequence of the episode numbers of the second video and the identified episode number of the first video.
That is, in the embodiments of the present disclosure, videos associated on video content are stored in one target container and sorted according to their episode number.
If the second video related to the first video on the video content is not found in the candidate videos, a target container is newly established, so that the first video is stored in the newly established target container, and after other videos are subsequently received, the first video is matched with the newly established target container on the video content.
In one possible embodiment, the content information further includes an episode type of an episode to which the video belongs, and the method further includes:
and classifying and storing the target container according to the episode type of the episode to which the video stored in the target container belongs.
The videos stored in the same target container are related on the video content, so that the episode types of the videos stored in the same target container are the same or similar. And further according to the episode types of the videos stored in the target container, the target container is classified and stored, so that the videos with the same or similar episode types can be further stored together, and further the classification management of the videos is facilitated.
In another aspect of the present disclosure, there is also provided a video playing apparatus, as shown in fig. 7, the video playing apparatus 700 includes:
an associated information obtaining module 701, configured to, in response to a first play trigger operation, obtain a first video to be played and associated video information, where the associated video information includes video information of at least one second video associated with the first video on video content;
a first playing module 702 configured to play the first video;
an entry showing module 703, configured to show the play entry of the second video according to the associated video information when detecting that the play state of the first video meets a preset condition.
In a possible implementation manner, the association information obtaining module 701 includes:
the first screening submodule 7011 is configured to obtain, from videos stored in advance, a video with an author that is the same as that of the first video, as a candidate video;
the second filtering sub-module 7012 is configured to, if the content information of the first video is identified, select, from the candidate videos, at least one second video that is associated with the first video on the video content according to the content information of the first video, and use the video information of the at least one selected second video as the associated video information.
In a possible implementation, the apparatus further includes a content identification module 704 configured to identify content information of the first video;
wherein the content identification module 704 comprises:
a first feature information acquisition sub-module 7041 configured to acquire feature information of the first video;
a content information output sub-module 7042 configured to input the feature information of the first video to a pre-established content recognition model for recognizing the content of the video, and output the content information of the first video.
In a possible implementation, the first feature information obtaining sub-module 7041 is specifically configured to:
extracting key frames of the first video;
acquiring color features of each key frame of the first video to serve as feature information of a first dimension of the first video;
acquiring color features of the first video to serve as feature information of a second dimension of the first video;
identifying feature information of a moving object in each key frame of the first video as feature information of a third dimension of the first video;
and acquiring audio characteristic information of the first video to serve as the characteristic information of the fourth dimension of the first video.
In a possible implementation, when acquiring the color feature of the first video, the first feature information acquiring sub-module 7041 is specifically configured to:
acquiring a first target type of a subject object in each key frame of the first video;
determining a weight value corresponding to the first target type according to a predetermined corresponding relation between the object type and the weight value, wherein the weight value is used as the weight value of a key frame where a subject object to which the first target type belongs is located;
obtaining a product of a color feature and a weight value of each key frame of the first video, wherein the product is used as a first parameter of each key frame of the first video;
and acquiring the sum of the first parameters of each key frame of the first video to serve as the color feature of the first video.
In a possible implementation, the apparatus further comprises a model building module 705 configured to build the content recognition model;
wherein the model building module 705 comprises:
a second feature information obtaining sub-module 7051 configured to obtain feature information of a plurality of sample videos;
the model training sub-module 7052 is configured to use the feature information of one sample video and the predetermined content information of the sample video as a training sample, and train the training sample by using a recurrent neural network to obtain the content recognition model.
In one possible implementation, the second feature information obtaining sub-module 7051 is specifically configured to:
for each of the sample videos, performing the following process:
extracting key frames of the sample video;
acquiring color features of each key frame of the sample video to serve as feature information of a first dimension of the sample video;
acquiring color features of the sample video to serve as feature information of a second dimension of the sample video;
identifying feature information of a moving object in each key frame of the sample video as feature information of a third dimension of the sample video;
and acquiring audio characteristic information of the sample video to serve as the characteristic information of the fourth dimension of the sample video.
In a possible implementation, when acquiring the color feature of the sample video, the second feature information acquiring sub-module 7051 is specifically configured to:
acquiring a second target type of the subject object in each key frame of the sample video;
determining a weight value corresponding to the second target type according to a predetermined corresponding relation between the object type and the weight value, wherein the weight value is used as the weight value of a key frame where a subject object to which the second target type belongs is located;
obtaining a product of the color feature and the weight value of each key frame of the sample video, and using the product as a second parameter of each key frame of the sample video;
and acquiring the sum of the second parameters of each key frame of the sample video to serve as the color feature of the sample video.
In one possible implementation, the content information includes: episode names of episodes to which the video belongs; the second screening submodule 7012 is specifically configured to:
selecting at least one candidate video with the episode name matched with the episode name of the first video from the candidate videos as a second video related to the first video on the video content.
In a possible implementation manner, the association information obtaining module 701 further includes:
a third filtering sub-module 7013 configured to, in a case where the content information of the first video is not identified, select, from the candidate videos, a candidate video in which a similarity between a video cover and a video cover of the first video is greater than a preset threshold, and use video information of the selected candidate video as the associated video information.
In one possible implementation, the third screening submodule 7013 is specifically configured to:
and selecting a candidate video with the similarity between the video cover and the video cover of the first video larger than a preset threshold value from the candidate videos by adopting a similar image retrieval technology based on a perceptual hash algorithm.
In a possible implementation manner, when the similar image retrieval technology based on the perceptual hash algorithm is adopted to select, from the candidate videos, a candidate video whose similarity between the video cover and the video cover of the first video is greater than a preset threshold, the third filtering sub-module 7013 is specifically configured to:
acquiring a first hash value of a video cover of each candidate video and a second hash value of the video cover of the first video;
acquiring the Hamming distance between the second hash value and each first hash value;
and taking the candidate video to which the video cover to which the first Hash value corresponding to the Hamming distance smaller than the preset distance belongs as the candidate video with the similarity between the video cover and the video cover of the first video larger than the preset threshold value.
In one possible implementation, the video information of the second video includes identification information of the second video; the portal presentation module 703 is specifically configured to:
acquiring a video cover of a second video to which the target identification information belongs according to the target identification information to serve as a target video cover, wherein the target identification information comprises identification information of the second video related to the associated video information;
and displaying the target video cover as a playing inlet of the second video.
In one possible embodiment, the video information of the second video further includes an episode number of the second video; the entry display module 703, when displaying the target video cover as the play entry of the second video, is specifically configured to:
displaying the target video cover as a playing entrance of the second video according to the sequence of the target episode numbers;
wherein the target episode number comprises an episode number of a second video to which the associated video information relates.
In a possible embodiment, the apparatus further comprises:
a first receiving module 706 configured to receive a second play trigger operation on one of the video covers of the second video being displayed;
a second playing module 707, configured to, in response to the second playing trigger operation, play a second video to which a video cover to which the second playing trigger operation is directed belongs.
In one possible embodiment, the video information of the second video includes an episode number of the second video; the device further comprises:
an icon display module 708 configured to display a preset icon;
a second receiving module 709 configured to receive a first preset operation on the preset icon;
the third playing module 710 is configured to, in response to the first preset operation, acquire an episode number of the first video, and play a second video to which a next sequence number of the episode number of the first video belongs.
In a possible implementation manner, the association information obtaining module 701 is specifically configured to:
and receiving the associated video information sent by the server.
As can be seen from the above, the video playing apparatus provided in the embodiment of the present disclosure can respond to the first play trigger operation, acquire the first video to be played and the associated video information, thereby playing the first video, and when it is detected that the playing state of the first video meets the preset condition, show the playing entry of the second video according to the associated video information, where the associated video information includes video information of at least one second video that is managed on the video content with the first video.
Therefore, when the video playing device provided by the embodiment of the disclosure receives the play trigger operation for the first video, the video information of the first video and at least one second video associated with the first video on the video content can be acquired, and when the play state of the first video meets the preset condition, the play entry of the second video can be displayed according to the video information of the second videos, so that the play entry of the associated video can be timely provided according to the state of the user watching the video, the user can continuously watch the associated video without complicated operations, and the video watching experience of the user is improved.
In another aspect of the present disclosure, there is also provided a video playing apparatus, as shown in fig. 8, the video playing apparatus 800 includes:
an information sending module 801 configured to send a first video and associated video information to a client device, so as to play the first video on the client device, and when a playing state of the first video meets a preset condition, show a playing entry of a second video on the client device according to the associated video information;
wherein the associated video information comprises video information of at least one of the second videos associated with the first video on video content.
In a possible implementation, the apparatus further includes an information obtaining module 802 configured to obtain the associated video information;
the information obtaining module 802 includes:
a first filtering sub-module 8021 configured to obtain, from a pre-stored video, a video whose author is the same as that of the first video, as a candidate video;
a second filtering submodule 8022, configured to, in a case where the content information of the first video is identified, select, from the candidate videos, at least one second video associated with the first video on the video content according to the content information of the first video, and use the video information of the selected at least one second video as the associated video information.
In a possible implementation, the apparatus further comprises a content identification module 803 configured to identify content information of the first video;
wherein the content identification module 803 comprises:
a first feature information acquisition sub-module 8031 configured to acquire feature information of the first video;
a content information output sub-module 8032 configured to input the feature information of the first video to a pre-established content identification model for identifying video content, and output the content information of the first video.
In a possible implementation, the first feature information obtaining sub-module 8031 is specifically configured to:
extracting key frames of the first video;
acquiring color features of each key frame of the first video to serve as feature information of a first dimension of the first video;
acquiring color features of the first video to serve as feature information of a second dimension of the first video;
identifying feature information of a moving object in each key frame of the first video as feature information of a third dimension of the first video;
and acquiring audio characteristic information of the first video to serve as the characteristic information of the fourth dimension of the first video.
In a possible implementation, the first feature information obtaining sub-module 8031, when obtaining the color feature of the first video, is specifically configured to:
acquiring a first target type of a subject object in each key frame of the first video;
determining a weight value corresponding to the first target type according to a predetermined corresponding relation between the object type and the weight value, wherein the weight value is used as the weight value of a key frame where a subject object to which the first target type belongs is located;
obtaining a product of a color feature and a weight value of each key frame of the first video, wherein the product is used as a first parameter of each key frame of the first video;
and acquiring the sum of the first parameters of each key frame of the first video to serve as the color feature of the first video.
In a possible implementation, the apparatus further comprises a model building module 804 configured to build the content recognition model;
wherein the model building module 804 includes:
a second feature information acquisition sub-module 8041 configured to acquire feature information of the plurality of sample videos;
the model training sub-module 8042 is configured to use feature information of one sample video and predetermined content information of the sample video as a training sample, and train the training sample by using a recurrent neural network to obtain the content identification model.
In a possible implementation, the second feature information obtaining sub-module 8041 is specifically configured to:
for each of the sample videos, performing the following process:
extracting key frames of the sample video;
acquiring color features of each key frame of the sample video to serve as feature information of a first dimension of the sample video;
acquiring color features of the sample video to serve as feature information of a second dimension of the sample video;
identifying feature information of a moving object in each key frame of the sample video as feature information of a third dimension of the sample video;
and acquiring audio characteristic information of the sample video to serve as the characteristic information of the fourth dimension of the sample video.
In a possible implementation, the second feature information obtaining sub-module 8041, when obtaining the color feature of the sample video, is specifically configured to:
acquiring a second target type of the subject object in each key frame of the sample video;
determining a weight value corresponding to the second target type according to a predetermined corresponding relation between the object type and the weight value, wherein the weight value is used as the weight value of a key frame where a subject object to which the second target type belongs is located;
obtaining a product of the color feature and the weight value of each key frame of the sample video, and using the product as a second parameter of each key frame of the sample video;
and acquiring the sum of the second parameters of each key frame of the sample video to serve as the color feature of the sample video.
In one possible implementation, the content information includes: episode names of episodes to which the video belongs; the second screening submodule 8022 is specifically configured to:
selecting at least one candidate video with the episode name matched with the episode name of the first video from the candidate videos as a second video related to the first video on the video content.
In a possible implementation manner, the information obtaining module 802 further includes:
a third filtering submodule 8023, configured to, in a case where the content information of the first video is not identified, select, from the candidate videos, a candidate video in which a similarity between a video cover and a video cover of the first video is greater than a preset threshold, and use video information of the selected candidate video as the associated video information.
In one possible implementation, the third screening submodule 8023 is specifically configured to:
and selecting a candidate video with the similarity between the video cover and the video cover of the first video larger than a preset threshold value from the candidate videos by adopting a similar image retrieval technology based on a perceptual hash algorithm.
In a possible implementation, when the similar image retrieval technology based on the perceptual hash algorithm is adopted to select, from the candidate videos, a candidate video whose similarity between the video cover and the video cover of the first video is greater than a preset threshold, the third filtering sub-module 8023 is specifically configured to:
acquiring a first hash value of a video cover of each candidate video and a second hash value of the video cover of the first video;
acquiring the Hamming distance between the second hash value and each first hash value;
and taking the candidate video to which the video cover to which the first Hash value corresponding to the Hamming distance smaller than the preset distance belongs as the candidate video with the similarity between the video cover and the video cover of the first video larger than the preset threshold value.
In a possible implementation, the content information further includes an episode number of the video in an episode to which the video belongs, and the apparatus further includes:
a serial number obtaining module 805 configured to obtain an episode serial number of the second video;
a storage module 806 configured to store the second video and the first video in a target container in an order of the episode number of the second video and the identified episode number of the first video.
In one possible embodiment, the content information further includes an episode type of an episode to which the video belongs, and the apparatus further includes:
a classifying module 807 configured to classify and store the target container according to the episode type of the episode to which the video stored in the target container belongs.
Therefore, the video playing device provided by the embodiment of the present disclosure can acquire the first video and the video information of at least one second video associated with the first video on the video content when receiving the play trigger operation for the first video, and can show the play entry of the second video according to the video information of the second videos when the play state of the first video meets the preset condition, so that the play entry of the associated video can be timely provided according to the state of the user watching the video, the user can continuously watch the associated video without complicated operations, and the video watching experience of the user is improved.
In yet another aspect of the disclosed implementation, an electronic device is also provided. Referring to fig. 9, the electronic device includes:
a processor 910;
a memory 920 for storing instructions executable by the processor 910;
wherein the processor 910 is configured to execute the instructions to implement the video playing method applied to the electronic device.
In yet another aspect of the disclosed implementation, an electronic device is also provided. As shown in fig. 10, the electronic device 1000 may be a mobile phone, a computer, a digital broadcast terminal, a messaging device, a game console, a tablet device, a medical device, an exercise device, a personal digital assistant, or the like.
Referring to fig. 10, electronic device 1000 may include one or more of the following components: processing component 1002, memory 1004, power component 1006, multimedia component 1008, audio component 1010, input/output (I/O) interface 1012, sensor component 1014, and communications component 1016.
The processing component 1002 generally controls overall operation of the electronic device 1000, such as operations associated with display, telephone calls, data communications, camera operations, and recording operations. The processing components 1002 may include one or more processors 1020 to execute instructions to perform all or a portion of the steps of the methods described above. Further, processing component 1002 may include one or more modules that facilitate interaction between processing component 1002 and other components. For example, the processing component 1002 may include a multimedia module to facilitate interaction between the multimedia component 1008 and the processing component 1002.
The memory 1004 is configured to store various types of data to support operations at the electronic device 1000. Examples of such data include instructions for any application or method operating on the electronic device 1000, contact data, phonebook data, messages, pictures, videos, and so forth. The memory 1004 may be implemented by any type or combination of volatile or non-volatile memory devices such as Static Random Access Memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable programmable read-only memory (EPROM), programmable read-only memory (PROM), read-only memory (ROM), magnetic memory, flash memory, magnetic or optical disks.
The power supply component 1006 provides power to the various components of the electronic device 1000. The power components 1006 may include a power management system, one or more power sources, and other components associated with generating, managing, and distributing power for the electronic device 1000.
The multimedia component 1008 includes a screen that provides an output interface between the electronic device 1000 and a user. In some embodiments, the screen may include a Liquid Crystal Display (LCD) and a Touch Panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive an input signal from a user. The touch panel includes one or more touch sensors to sense touch, slide, and gestures on the touch panel. The touch sensor may not only sense the boundary of a touch or slide action, but also detect the duration and pressure associated with the touch or slide operation. In some embodiments, the multimedia component 1008 includes a front facing camera and/or a rear facing camera. The front camera and/or the rear camera may receive external multimedia data when the electronic device 1000 is in an operating mode, such as a shooting mode or a video mode. Each front camera and rear camera may be a fixed optical lens system or have a focal length and optical zoom capability.
The audio component 1010 is configured to output and/or input audio signals. For example, the audio component 1010 may include a Microphone (MIC) configured to receive external audio signals when the electronic device 1000 is in an operational mode, such as a call mode, a recording mode, and a voice recognition mode. The received audio signal may further be stored in the memory 1004 or transmitted via the communication component 1016. In some embodiments, audio component 1010 also includes a speaker for outputting audio signals.
I/O interface 1012 provides an interface between processing component 1002 and peripheral interface modules, which may be keyboards, click wheels, buttons, etc. These buttons may include, but are not limited to: a home button, a volume button, a start button, and a lock button.
The sensor assembly 1014 includes one or more sensors for providing various aspects of status assessment for the electronic device 1000. For example, the sensor assembly 1014 may detect an open/closed state of the electronic device 1000, the relative positioning of components, such as a display and keypad of the electronic device 1000, the sensor assembly 1014 may also detect a change in position of the electronic device 1000 or a component of the electronic device 1000, the presence or absence of user contact with the electronic device 1000, orientation or acceleration/deceleration of the electronic device 1000, and a change in temperature of the electronic device 1000. The sensor assembly 1014 may include a proximity sensor configured to detect the presence of a nearby object without any physical contact. The sensor assembly 1014 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications. In some embodiments, the sensor assembly 1014 may also include an acceleration sensor, a gyroscope sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
The communication component 1016 is configured to facilitate wired or wireless communication between the electronic device 1000 and other devices. The electronic device 1000 may access a wireless network based on a communication standard, such as WiFi, a carrier network (such as 2G, 3G, 8G, or 5G), or a combination thereof. In an exemplary embodiment, the communication component 1016 receives a broadcast signal or broadcast related information from an external broadcast management system via a broadcast channel. In an exemplary embodiment, the communications component 1016 further includes a Near Field Communication (NFC) module to facilitate short-range communications. For example, the NFC module may be implemented based on Radio Frequency Identification (RFID) technology, infrared data association (IrDA) technology, Ultra Wideband (UWB) technology, Bluetooth (BT) technology, and other technologies.
In an exemplary embodiment, the electronic device 1000 may be implemented by one or more Application Specific Integrated Circuits (ASICs), Digital Signal Processors (DSPs), Digital Signal Processing Devices (DSPDs), Programmable Logic Devices (PLDs), Field Programmable Gate Arrays (FPGAs), controllers, micro-controllers, microprocessors or other electronic components for performing the video playback method described above.
In an exemplary embodiment, a non-transitory computer-readable storage medium comprising instructions, such as the memory 1004 comprising instructions, executable by the processor 1020 of the electronic device 1000 to perform the above-described method is also provided. Alternatively, for example, the storage medium may be a non-transitory computer-readable storage medium, which may be, for example, a ROM, a Random Access Memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, and the like.
In yet another aspect of the present disclosure, there is also provided a server, as shown in fig. 11, including:
a processor 1110;
a memory 1120 for storing instructions executable by the processor 1110;
wherein the processor 1110 is configured to execute the instructions to implement the video playing method applied to the server.
In another aspect of the present disclosure, a video playing system is further provided, which includes the electronic device and the server.
Wherein, in the video playing system, the electronic equipment is used as the client equipment.
That is, in the video playing system provided by the embodiment of the present disclosure, a first video and associated video information are sent to the electronic device through the server, where the associated video information includes video information of at least one second video associated with the first video in video content;
playing the first video through the electronic equipment;
and when the electronic equipment detects that the playing state of the first video meets a preset condition, displaying a playing inlet of the second video through the electronic equipment according to the associated video information.
Therefore, the video playing system provided by the embodiment of the disclosure sends the first video and the video information (namely, the associated video information) of the at least one second video associated with the first video on the video content to the electronic device through the server, so that when the state of playing the first video of the electronic device meets the preset condition, the electronic device displays the playing entry of the second video according to the associated video information, the playing entry of the associated video is timely provided according to the state of watching the video by the user, the user can continuously watch the associated video without complicated operations, and the video watching experience of the user is improved.
In yet another aspect of the present disclosure, there is also provided a storage medium, wherein instructions of the storage medium, when executed by a processor of an electronic device, enable the electronic device to execute the video playing method described above.
In yet another aspect of the embodiments of the present disclosure, there is also provided a computer program product containing instructions, which when run on a computer, causes the computer to implement the video playing method described above.
The video playback schemes provided herein are not inherently related to any particular computer, virtual system, or other apparatus. Various general purpose systems may also be used with the teachings herein. The structure required to construct a system incorporating aspects of the present invention will be apparent from the description above. Moreover, the present invention is not directed to any particular programming language. It is appreciated that a variety of programming languages may be used to implement the teachings of the present invention as described herein, and any descriptions of specific languages are provided above to disclose the best mode of the invention.
In the description provided herein, numerous specific details are set forth. It is understood, however, that embodiments of the invention may be practiced without these specific details. In some instances, well-known methods, structures and techniques have not been shown in detail in order not to obscure an understanding of this description.
Similarly, it should be appreciated that in the foregoing description of exemplary embodiments of the invention, various features of the invention are sometimes grouped together in a single embodiment, figure, or description thereof for the purpose of streamlining the disclosure and aiding in the understanding of one or more of the various inventive aspects. However, the disclosed method should not be interpreted as reflecting an intention that: that the invention as claimed requires more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive aspects lie in less than all features of a single foregoing disclosed embodiment. Thus, the claims following the detailed description are hereby expressly incorporated into this detailed description, with each claim standing on its own as a separate embodiment of this invention.
Those skilled in the art will appreciate that the modules in the device in an embodiment may be adaptively changed and disposed in one or more devices different from the embodiment. The modules or units or components of the embodiments may be combined into one module or unit or component, and furthermore they may be divided into a plurality of sub-modules or sub-units or sub-components. All of the features disclosed in this specification (including any accompanying claims, abstract and drawings), and all of the processes or elements of any method or apparatus so disclosed, may be combined in any combination, except combinations where at least some of such features and/or processes or elements are mutually exclusive. Each feature disclosed in this specification (including any accompanying claims, abstract and drawings) may be replaced by alternative features serving the same, equivalent or similar purpose, unless expressly stated otherwise.
Furthermore, those skilled in the art will appreciate that while some embodiments described herein include some features included in other embodiments, rather than other features, combinations of features of different embodiments are meant to be within the scope of the invention and form different embodiments. For example, in the claims, any of the claimed embodiments may be used in any combination.
The various component embodiments of the invention may be implemented in hardware, or in software modules running on one or more processors, or in a combination thereof. Those skilled in the art will appreciate that a microprocessor or Digital Signal Processor (DSP) may be used in practice to implement some or all of the functions of some or all of the components in an information extraction scheme according to embodiments of the present disclosure. The present invention may also be embodied as apparatus or device programs (e.g., computer programs and computer program products) for performing a portion or all of the methods described herein. Such programs implementing the present invention may be stored on computer-readable media or may be in the form of one or more signals. Such a signal may be downloaded from an internet website or provided on a carrier signal or in any other form.
It should be noted that the above-mentioned embodiments illustrate rather than limit the invention, and that those skilled in the art will be able to design alternative embodiments without departing from the scope of the appended claims. In the claims, any reference signs placed between parentheses shall not be construed as limiting the claim. The word "comprising" does not exclude the presence of elements or steps not listed in a claim. The word "a" or "an" preceding an element does not exclude the presence of a plurality of such elements. The invention may be implemented by means of hardware comprising several distinct elements, and by means of a suitably programmed computer. In the unit claims enumerating several means, several of these means may be embodied by one and the same item of hardware. The usage of the words first, second and third, etcetera do not indicate any ordering. These words may be interpreted as names.

Claims (10)

1. A video playing method, characterized by:
responding to a first playing triggering operation, acquiring a first video to be played and associated video information, wherein the associated video information comprises video information of at least one second video which is associated with the first video on video content;
playing the first video;
and when the playing state of the first video is detected to meet a preset condition, displaying the playing entry of the second video according to the associated video information.
2. The video playing method according to claim 1, wherein the process of obtaining the associated video information comprises:
acquiring videos with authors the same as the author of the first video from pre-stored videos to serve as candidate videos;
and under the condition that the content information of the first video is identified, at least one second video which is associated with the first video on the video content is selected from the candidate videos according to the content information of the first video, and the video information of the selected at least one second video is used as the associated video information.
3. The video playing method according to claim 2, wherein the process of identifying the content information of the first video comprises:
acquiring characteristic information of the first video;
and inputting the characteristic information of the first video into a pre-established content identification model for identifying video content, and outputting the content information of the first video.
And receiving the associated video information sent by the server.
4. A video playback method, the method comprising:
sending a first video and associated video information to client equipment so as to play the first video on the client equipment, and displaying a play entry of a second video on the client equipment according to the associated video information when the play state of the first video meets a preset condition;
wherein the associated video information comprises video information of at least one of the second videos associated with the first video on video content.
5. A video playback apparatus, comprising:
the device comprises an associated information acquisition module, a first display triggering operation and a second display triggering operation, wherein the associated information acquisition module is configured to respond to the first display triggering operation and acquire a first video to be displayed and associated video information, and the associated video information comprises video information of at least one second video which is associated with the first video on video content;
a first playing module configured to play the first video;
and the entrance display module is configured to display the playing entrance of the second video according to the associated video information when detecting that the playing state of the first video meets a preset condition.
6. A video playback apparatus, comprising:
the information sending module is configured to send a first video and associated video information to client equipment so as to play the first video on the client equipment, and when the playing state of the first video meets a preset condition, a playing entrance of a second video is displayed on the client equipment according to the associated video information;
wherein the associated video information comprises video information of at least one of the second videos associated with the first video on video content.
7. An electronic device, comprising:
a processor;
a memory for storing the processor-executable instructions;
wherein the processor is configured to execute the instructions to implement the video playback method of any of claims 1 to 3.
8. A server, comprising:
a processor;
a memory for storing the processor-executable instructions;
wherein the processor is configured to execute the instructions to implement the video playback method of claim 4.
9. A video playback system comprising an electronic device according to claim 7 and a server according to claim 8.
10. A storage medium, wherein instructions in the storage medium, when executed by a processor of an electronic device, enable the electronic device to perform the video playback method of any one of claims 1 to 3, or perform the video playback method of claim 4.
CN202011381477.0A 2020-11-30 2020-11-30 Video playing method, device, electronic equipment, server and system Active CN112672208B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011381477.0A CN112672208B (en) 2020-11-30 2020-11-30 Video playing method, device, electronic equipment, server and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011381477.0A CN112672208B (en) 2020-11-30 2020-11-30 Video playing method, device, electronic equipment, server and system

Publications (2)

Publication Number Publication Date
CN112672208A true CN112672208A (en) 2021-04-16
CN112672208B CN112672208B (en) 2023-06-20

Family

ID=75403930

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011381477.0A Active CN112672208B (en) 2020-11-30 2020-11-30 Video playing method, device, electronic equipment, server and system

Country Status (1)

Country Link
CN (1) CN112672208B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113613053A (en) * 2021-07-26 2021-11-05 北京达佳互联信息技术有限公司 Video recommendation method and device, electronic equipment and storage medium
CN113721807A (en) * 2021-08-30 2021-11-30 北京字跳网络技术有限公司 Information display method and device, electronic equipment and storage medium
CN114125545A (en) * 2021-11-29 2022-03-01 北京达佳互联信息技术有限公司 Video information processing method and device, electronic equipment and storage medium
CN114443897A (en) * 2022-02-10 2022-05-06 北京字跳网络技术有限公司 Video recommendation method and device, electronic equipment and storage medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102868929A (en) * 2012-08-31 2013-01-09 深圳市龙视传媒有限公司 Video display method and system, and digital television terminal
US20140068013A1 (en) * 2012-09-04 2014-03-06 Wistron Corporation Method of playing internet video and related electronic device
CN109640129A (en) * 2018-12-12 2019-04-16 北京字节跳动网络技术有限公司 Video recommendation method, device, client device, server and storage medium
CN110519621A (en) * 2019-09-20 2019-11-29 北京字节跳动网络技术有限公司 Video recommendation method, device, electronic equipment and computer-readable medium
CN110866563A (en) * 2019-11-20 2020-03-06 咪咕文化科技有限公司 Similar video detection and recommendation method, electronic device and storage medium
CN111770376A (en) * 2020-06-29 2020-10-13 百度在线网络技术(北京)有限公司 Information display method, device, system, electronic equipment and storage medium

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102868929A (en) * 2012-08-31 2013-01-09 深圳市龙视传媒有限公司 Video display method and system, and digital television terminal
US20140068013A1 (en) * 2012-09-04 2014-03-06 Wistron Corporation Method of playing internet video and related electronic device
CN103686342A (en) * 2012-09-04 2014-03-26 纬创资通股份有限公司 Method for playing network film and electronic device thereof
CN109640129A (en) * 2018-12-12 2019-04-16 北京字节跳动网络技术有限公司 Video recommendation method, device, client device, server and storage medium
CN110519621A (en) * 2019-09-20 2019-11-29 北京字节跳动网络技术有限公司 Video recommendation method, device, electronic equipment and computer-readable medium
CN110866563A (en) * 2019-11-20 2020-03-06 咪咕文化科技有限公司 Similar video detection and recommendation method, electronic device and storage medium
CN111770376A (en) * 2020-06-29 2020-10-13 百度在线网络技术(北京)有限公司 Information display method, device, system, electronic equipment and storage medium

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113613053A (en) * 2021-07-26 2021-11-05 北京达佳互联信息技术有限公司 Video recommendation method and device, electronic equipment and storage medium
CN113721807A (en) * 2021-08-30 2021-11-30 北京字跳网络技术有限公司 Information display method and device, electronic equipment and storage medium
CN113721807B (en) * 2021-08-30 2023-08-22 北京字跳网络技术有限公司 Information display method and device, electronic equipment and storage medium
CN114125545A (en) * 2021-11-29 2022-03-01 北京达佳互联信息技术有限公司 Video information processing method and device, electronic equipment and storage medium
CN114125545B (en) * 2021-11-29 2024-03-12 北京达佳互联信息技术有限公司 Video information processing method, device, electronic equipment and storage medium
CN114443897A (en) * 2022-02-10 2022-05-06 北京字跳网络技术有限公司 Video recommendation method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN112672208B (en) 2023-06-20

Similar Documents

Publication Publication Date Title
US11622141B2 (en) Method and apparatus for recommending live streaming room
CN107454465B (en) Video playing progress display method and device and electronic equipment
US11520824B2 (en) Method for displaying information, electronic device and system
CN107818180B (en) Video association method, video display device and storage medium
CN112672208B (en) Video playing method, device, electronic equipment, server and system
EP3327590B1 (en) Method and device for adjusting video playback position
RU2597232C1 (en) Method for providing a video in real time and device for its implementation, as well as a server and a terminal device
US11394675B2 (en) Method and device for commenting on multimedia resource
CN109189987A (en) Video searching method and device
US9953221B2 (en) Multimedia presentation method and apparatus
CN111556366A (en) Multimedia resource display method, device, terminal, server and system
WO2022037307A1 (en) Information recommendation method and apparatus, and electronic device
CN108227950B (en) Input method and device
CN106331761A (en) Live broadcast list display method and apparatuses
CN111553372B (en) Training image recognition network, image recognition searching method and related device
CN105809174A (en) Method and device for identifying image
WO2020119254A1 (en) Method and device for filter recommendation, electronic equipment, and storage medium
CN107229403B (en) Information content selection method and device
CN106550252A (en) The method for pushing of information, device and equipment
CN109766473B (en) Information interaction method and device, electronic equipment and storage medium
CN112464031A (en) Interaction method, interaction device, electronic equipment and storage medium
CN106331328B (en) Information prompting method and device
CN106547850A (en) Expression annotation method and device
CN111629270A (en) Candidate item determination method and device and machine-readable medium
CN112115341B (en) Content display method, device, terminal, server, system and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant