CN112866762A - Processing method and device for acquiring video associated information, electronic equipment and server - Google Patents

Processing method and device for acquiring video associated information, electronic equipment and server Download PDF

Info

Publication number
CN112866762A
CN112866762A CN202011641038.9A CN202011641038A CN112866762A CN 112866762 A CN112866762 A CN 112866762A CN 202011641038 A CN202011641038 A CN 202011641038A CN 112866762 A CN112866762 A CN 112866762A
Authority
CN
China
Prior art keywords
video
image
data
server
client
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011641038.9A
Other languages
Chinese (zh)
Inventor
平淑文
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Dajia Internet Information Technology Co Ltd
Original Assignee
Beijing Dajia Internet Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Dajia Internet Information Technology Co Ltd filed Critical Beijing Dajia Internet Information Technology Co Ltd
Priority to CN202011641038.9A priority Critical patent/CN112866762A/en
Publication of CN112866762A publication Critical patent/CN112866762A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/258Client or end-user data management, e.g. managing client capabilities, user preferences or demographics, processing of multiple end-users preferences to derive collaborative data
    • H04N21/25866Management of end-user data
    • H04N21/25875Management of end-user data involving end-user authentication
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/485End-user interface for client configuration

Abstract

The disclosure relates to a processing method and device for acquiring video associated information, electronic equipment and a server. In one embodiment of the method, the characteristic data of the picture image uploaded by the client can be searched in a characteristic data set of a pre-stored key frame image, information associated with the picture image is obtained, and the information is returned to the client. The client can display the video related information searched according to the picture images. By the scheme of the embodiment of the disclosure, a new video interaction mode is provided, video associated information can be searched in an image mode, more accurate search results can be obtained, searching can be performed without inputting text information, and the response speed of displaying the search results by a client is improved.

Description

Processing method and device for acquiring video associated information, electronic equipment and server
Technical Field
The present disclosure relates to the field of computer data processing technologies, and in particular, to a processing method and apparatus for acquiring video associated information, an electronic device, and a server.
Background
Short video applications of terminals are gradually popularized at present. The user can summarize the keywords describing the video content according to the watched video content, and then search the related content of the video. However, many video contents are difficult to accurately describe by characters, so that the input keywords are not accurate enough, and the accuracy of the search result is influenced.
Disclosure of Invention
The present disclosure provides a processing method, an apparatus, an electronic device, and a server for obtaining video associated information, which provide a new video interaction manner to at least solve the problem in the related art that searching for video associated information by inputting characters is inaccurate. The technical scheme of the disclosure is as follows:
according to an aspect of the embodiments of the present disclosure, a processing method for acquiring video related information is provided, including:
determining a picture image of a playing video of a client;
searching target characteristic data matched with the characteristic data of the picture image in a characteristic data set, wherein the characteristic data set comprises characteristic data extracted from a key frame image, and the key frame image is determined based on video data uploaded by a video publishing account;
when the target characteristic data is found, video associated information corresponding to the target characteristic data is obtained;
and sending the video associated information to the client, wherein the video associated information is used for displaying at the client.
According to another aspect of the embodiments of the present disclosure, in the method, the determining a picture image of a playing video of a client includes: receiving the picture images uploaded by the client side,
alternatively, the first and second electrodes may be,
and receiving a playing time label uploaded by the client, and acquiring a picture image of the video data according to the time label.
According to another aspect of the disclosed embodiments, in the method, the key frame image of the video data comprises:
a frame image at a specified location in the video data;
the picture images comprise waiting interface images displayed after the video is played by the client.
According to another aspect of the embodiments of the present disclosure, in the method, the video-related information includes one of:
account information of a target object in the picture image, a name of a provenance product of the video data corresponding to the target characteristic data, and image association information of which the matching degree with the characteristic data of the picture image meets a first requirement.
According to another aspect of the embodiments of the present disclosure, in the method, after the target feature data is found, the method further includes:
determining a name of a person as a starting product when the picture image is determined to contain the person based on the feature data of the picture image;
searching according to the name of the provenance work to obtain work association information associated with the name of the provenance work;
correspondingly, the video associated information comprises the work associated information.
According to another aspect of the embodiments of the present disclosure, in the method, after the target feature data is found, the method further includes:
determining a person name of a person when it is determined that the screen image contains the person based on the feature data of the screen image;
searching according to the person name to acquire person association information associated with the person name;
correspondingly, the video related information comprises the person related information.
According to another aspect of the embodiments of the present disclosure, in the method, after the target feature data is found, the method further includes:
searching for first video data containing the picture image and/or second video data having a matching degree with the feature data of the picture image according to a second requirement when it is determined that the picture image does not contain a person object based on the feature data of the picture image;
correspondingly, the video associated information includes the first video data and/or the second video data.
According to another aspect of the embodiments of the present disclosure, a processing method for acquiring video related information is provided, which includes:
determining a picture image of a playing video;
uploading the picture image to a server;
receiving video associated information corresponding to target feature data returned by a server, wherein the target feature data comprise feature data matched with the feature data of the picture image and searched by the server in a feature data set, and the feature data are obtained by performing feature extraction on a key frame image determined by video data uploaded to a video publishing account;
and at least displaying part of the video related information.
According to another aspect of the embodiments of the present disclosure, in the method, the determining a picture image of the playing video includes:
intercepting a picture of a playing video to obtain a picture image;
alternatively, the first and second electrodes may be,
the method comprises the following steps of intercepting a picture image of a playing video, determining a playing time tag of the picture in the playing video, and uploading the picture to a server, wherein the uploading step comprises the following steps: and uploading the playing time tag to a server.
According to another aspect of the embodiments of the present disclosure, in the method, the video-related information includes one of:
account information contained in the picture image, a name of a place of sale of the video data corresponding to the target characteristic data, and image association information of which the matching degree with the characteristic data of the picture image meets a first requirement.
According to another aspect of the embodiments of the present disclosure, in the method, after the server finds the target feature data, the method further includes:
when the picture image is determined to contain a person based on the feature data of the picture image, the returned video related information includes one of the following information:
searching for the obtained work associated information according to the name of the person;
and searching the acquired person associated information according to the person name of the person.
According to another aspect of the embodiments of the present disclosure, in the method, after the server finds the target feature data, the method further includes:
when the picture image is determined to contain a person based on the characteristic data of the picture image, the returned video related information comprises one of the following information searched by the server:
first video data containing the picture image;
and second video data having a degree of matching with the feature data of the picture image in accordance with a second requirement.
According to another aspect of the embodiments of the present disclosure, in the method, the acquiring a picture image of a playing video includes:
and intercepting the picture image based on the image search interface element provided by the trigger.
According to another aspect of the disclosed embodiments, in the method, the screen image includes:
and displaying the waiting interface image after the video is played.
According to another aspect of the embodiments of the present disclosure, there is also provided a processing apparatus for acquiring video related information, including: the system comprises an image determining module, a feature matching module, an information searching module and a sending module;
the image determination module is configured to determine a picture image of a playing video of a client;
the feature matching module is configured to search for target feature data matched with feature data of the picture image in a feature data set, wherein the feature data set comprises feature data extracted from a key frame image, and the key frame image is determined based on video data uploaded by a video publishing account;
the information searching module is configured to acquire video associated information corresponding to the target characteristic data when the target characteristic data is searched;
the sending module is configured to send the video associated information to the client, the video associated information being for presentation at the client.
According to another aspect of the embodiments of the present disclosure, in the apparatus, the determining, by the image determining module, a picture image of the playing video of the client includes: receiving the picture images uploaded by the client side,
alternatively, the first and second electrodes may be,
and receiving a playing time label uploaded by the client, and acquiring a picture image of the video data according to the time label.
According to another aspect of the disclosed embodiment, in the apparatus, the key frame image of the video data in the feature matching module includes:
a frame image at a specified location in the video data;
the picture images comprise waiting interface images displayed after the video is played by the client.
According to another aspect of the embodiment of the present disclosure, in the apparatus, the video related information obtained by the information search module includes one of:
account information of a target object in the picture image, a name of a provenance product of the video data corresponding to the target characteristic data, and image association information of which the matching degree with the characteristic data of the picture image meets a first requirement.
According to another aspect of the embodiments of the present disclosure, the apparatus further includes:
a character work determination module configured to determine a name of a work of origin of a character when the picture image is determined to contain the character based on the feature data of the picture image after the target feature data is found;
the work searching module is configured to search according to the name of the provenance work and acquire work association information associated with the name of the provenance work;
correspondingly, the video associated information comprises the work associated information.
According to another aspect of the disclosed embodiments, the apparatus further comprises:
a person name determining module configured to determine a person name of a person after finding the target feature data and when determining that the screen image contains the person based on the feature data of the screen image;
a name searching module configured to search according to the person name and acquire person related information associated with the person name;
correspondingly, the video related information comprises the person related information.
According to another aspect of the disclosed embodiments, the apparatus further comprises:
a non-human search module configured to search first video data including the picture image and/or second video data matching the feature data of the picture image to a second degree that meets a second requirement, when it is determined that the picture image does not include a human object based on the feature data of the picture image;
correspondingly, the video associated information includes the first video data and/or the second video data.
According to another aspect of the embodiments of the present disclosure, there is also provided a processing apparatus for acquiring video related information, including:
an image acquisition module configured to determine a picture image of a playing video;
an upload module configured to upload the screen image to a server;
the relevant information receiving module is configured to receive video relevant information corresponding to target characteristic data returned by the server, wherein the target characteristic data comprise characteristic data which are searched by the server in a characteristic data set and are matched with the characteristic data of the picture image, and the characteristic data are obtained by performing characteristic extraction on a key frame image determined by video data uploaded to a video publishing account;
a presentation module configured to present at least a portion of the video-associated information.
According to another aspect of the embodiments of the present disclosure, in the apparatus, the deciding, by the image obtaining module, to play the picture image of the video includes:
intercepting a picture of a playing video to obtain a picture image;
alternatively, the first and second electrodes may be,
intercepting a picture image of a playing video, and determining a playing time label of the picture in the playing video; the uploading module uploads the picture to a server and comprises: and uploading the playing time tag to a server.
According to another aspect of the disclosed embodiment, in the apparatus, the video associated information includes one of:
account information contained in the picture image, a name of a place of sale of the video data corresponding to the target characteristic data, and image association information of which the matching degree with the characteristic data of the picture image meets a first requirement.
According to another aspect of the embodiments of the present disclosure, in the apparatus, after the server finds the target feature data, the apparatus further includes:
when the picture image is determined to contain a person based on the feature data of the picture image, the returned video related information includes one of the following information:
searching for the obtained work associated information according to the name of the person;
and searching the acquired person associated information according to the person name of the person.
According to another aspect of the embodiments of the present disclosure, in the apparatus, after the server finds the target feature data, the apparatus further includes:
when the picture image is determined to contain a person based on the characteristic data of the picture image, the returned video related information comprises one of the following information searched by the server:
first video data containing the picture image;
and second video data having a degree of matching with the feature data of the picture image in accordance with a second requirement.
According to another aspect of the disclosed embodiment, in the apparatus, the acquiring the picture image of the playing video includes:
and intercepting the picture image based on the image search interface element provided by the trigger.
According to another aspect of the disclosed embodiments, in the apparatus, the screen image includes:
and displaying the waiting interface image after the video is played.
According to another aspect of the embodiments of the present disclosure, there is also provided an electronic device, including:
a processor;
a memory for storing the processor-executable instructions;
wherein the processor is configured to execute the instructions to implement the method of any embodiment of the present disclosure implemented on a client.
According to another aspect of the embodiments of the present disclosure, there is also provided a storage medium, wherein instructions, when executed by a processor of an electronic device, enable the electronic device to perform the method according to any one of the embodiments of the present disclosure implemented on a client side.
According to another aspect of the embodiments of the present disclosure, there is also provided a server, including:
at least one processor;
a memory for storing the processor-executable instructions;
wherein the processor is configured to execute the instructions to implement the method of any embodiment of the present disclosure implemented in a server.
According to another aspect of the embodiments of the present disclosure, there is also provided a storage medium, wherein instructions in the storage medium, when executed by a processor of a server, enable the server to perform the method of any one of the present disclosure implemented in the server.
The technical scheme provided by the embodiment of the disclosure at least brings the following beneficial effects:
the server can process the video uploaded by the video publishing account in advance, intercept the key frame image in the video, extract the feature data of the key frame image and store the key frame image. When the video watching account at the client wants to acquire the relevant information in the watching video, the client can intercept the picture image of the video being played, and the picture image or the corresponding playing time tag can be uploaded to the server. The server can determine the picture image of the video played by the client, search in the feature data set of the key frame image stored in advance according to the feature data of the picture image, acquire the information associated with the picture image intercepted by the user, and return the information to the client. And the client can display the video related information searched according to the picture images. According to the scheme of the embodiment of the disclosure, the picture content of the played video can be intercepted through the video watching account, the video related information is searched in an image mode, and as the server stores the characteristic data of the related video in advance, a more accurate search result can be obtained. According to the scheme of the embodiment of the disclosure, a new video interaction mode is provided, the server can automatically identify the picture image and return the search result, the video watching account can be searched without inputting text information, and the response speed of the client for displaying the search result is improved. And moreover, video interaction modes are enriched, and the experience of a video watching account is improved.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the disclosure.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the present disclosure and, together with the description, serve to explain the principles of the disclosure and are not to be construed as limiting the disclosure.
Fig. 1 is an application environment diagram illustrating a processing method for acquiring video associated information according to an exemplary embodiment.
Fig. 2 is a flowchart illustrating a processing method for acquiring video associated information according to an exemplary embodiment.
Fig. 3 is a scene diagram illustrating a processing method for acquiring video-related information according to an exemplary embodiment.
Fig. 4 is a flowchart illustrating a processing method for acquiring video associated information according to an exemplary embodiment.
Fig. 5 is a diagram illustrating an application scenario of a processing method for acquiring video related information according to an exemplary embodiment.
Fig. 6 is a flowchart illustrating a processing method for acquiring video associated information according to an exemplary embodiment.
Fig. 7 is a flowchart illustrating a processing method for acquiring video associated information according to an exemplary embodiment.
Fig. 8 is a diagram illustrating an application scenario of a processing method for acquiring video related information according to an exemplary embodiment.
Fig. 9 is a flowchart illustrating a processing method for acquiring video associated information according to an exemplary embodiment.
Fig. 10 is a block diagram illustrating a processing device for acquiring video associated information according to an exemplary embodiment.
Fig. 11 is a block diagram illustrating a processing device for acquiring video associated information according to an example embodiment.
Fig. 12 is a block diagram illustrating a processing device for acquiring video associated information according to an exemplary embodiment.
Fig. 13 is a block diagram illustrating a processing device for acquiring video associated information according to an example embodiment.
Fig. 14 is a block diagram illustrating a processing device for acquiring video associated information according to an exemplary embodiment.
Fig. 15 is an internal block diagram of an electronic device shown in accordance with an example embodiment.
Fig. 16 is an internal block diagram of a server according to an example embodiment.
Detailed Description
In order to make the technical solutions of the present disclosure better understood by those of ordinary skill in the art, the technical solutions in the embodiments of the present disclosure will be clearly and completely described below with reference to the accompanying drawings.
It should be noted that the terms "first," "second," and the like in the description and claims of the present disclosure and in the above-described drawings are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used is interchangeable under appropriate circumstances such that the embodiments of the disclosure described herein are capable of operation in sequences other than those illustrated or otherwise described herein. The implementations described in the exemplary embodiments below are not intended to represent all implementations consistent with the present disclosure. Rather, they are merely examples of apparatus and methods consistent with certain aspects of the present disclosure, as detailed in the appended claims. The terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, the presence of additional identical or equivalent elements in a process, method, article, or apparatus that comprises the recited elements is not excluded. For example, if the terms first, second, etc. are used to denote names, they do not denote any particular order.
The processing method for acquiring the video related information provided by the present disclosure can be applied to the application environment shown in fig. 1. Wherein the terminal 110 may interact with the server 120 through a network connection. The server 120 may transmit video data to the terminal 110, and the terminal 110 may generate and play a video locally after receiving the video data. The video may be played when a video playing Application (APP) is opened on the terminal 110 side, or may be played when the video playing window is activated by switching from another APP to the video playing application. The account may intercept a screen image of the video playing content when the terminal 110 plays the video (including a preview interface after the video playing), and send the screen image to the server 120. The server 120 may find corresponding video related information according to the feature data of the picture image, and return the video related information to the terminal 110. The terminal 110 may present the video associated information in whole or in part. Therefore, the account can be searched without inputting the text description, and the relevant information of the playing video can be searched by carrying out video screenshot on the terminal 110. The terminal 110 described in the embodiments of the present disclosure may include, but is not limited to, various personal computers, notebook computers, smart phones, tablet computers, wearable devices, vehicle-mounted devices, medical devices, and the like. The servers 120 described in the embodiments of the present disclosure may include, but are not limited to, independent servers, server clusters, distributed processing servers, blockchain servers, cloud computing platforms, and the like, and combinations thereof. The server 120 may be understood as one or more servers on the side opposite the client 110, such as a server that may include image feature extraction and/or matching, and a server that may include an understanding of the particular content in the image, such as the identification of people and sources in the image.
Generally, an operation object for playing a video is usually an account for logging in the video playing application, and in other application scenarios, an account for viewing a video that is not logged in the video playing application may also be used. The manipulation behavior may include a manipulation of the terminal 110 itself, such as pressing a physical button, tilting a screen, shaking, inserting an earphone, and the like, and may also include a manipulation of an interface of the terminal 110, such as clicking a virtual button, sliding on a touch screen, dragging an interface element, and the like.
Fig. 2 is a flowchart illustrating a processing method for acquiring video-related information according to an exemplary embodiment, which may be used in the server 120 as shown in fig. 2 and may include the following steps.
In step S202, a screen image of the play video of the client is determined.
The determining of the picture image of the video played by the client may include processing capable of determining an image captured by the video viewing account at the client according to the picture image uploaded by the client or information related to the picture image. For example, the server may obtain a picture image of a playing video captured by the account at the client, and determine the picture image according to the playing time tag uploaded by the receiving client. Thus, in some embodiments, the determining the picture image of the playing video of the client may include: and receiving the picture images uploaded by the client. In general, video can be viewed as a collection of sequences of successive frame images. The video playing may be displaying frame images included in the video data at a certain frame rate according to a time sequence. For example, frame images in video data are played at a frame rate of 24 frames/second or more, and a human eye sees a continuous screen at this time. The played video may include a video being played at the client, or may include a waiting interface after the video is played. For example, after the video is played, the waiting interface includes the last frame image of the video data, and the picture image captured by the video watching account at this time still belongs to the picture image of the played video described in this embodiment. The video watching account can be controlled by physical keys, virtual interface elements, gestures, sensory operations (such as blinking and the like) and the like during video playing, so that the client can capture picture images of the played video. The picture image of the video playing captured at the client may be an image in a captured video playing window, or an image of a display interface on the client, including, for example, information of electric quantity and signal strength at the top of the display interface, and an image of a menu bar and a tool button in the display interface.
The client can automatically upload the captured picture images to the server, and can also send an uploading control instruction based on the video viewing account to the server. For example, the video viewing account may upload the screen image to the server after clicking a "start recognition" virtual button of the client interface. The client can upload one picture image at a time, and also can upload a plurality of picture images to the server at a time after capturing a plurality of picture images of the played video.
In step S204, target feature data matched with the feature data of the picture image is searched for in a feature data set, where the feature data set includes feature data extracted from a key frame image, and the key frame image is determined based on video data uploaded by a video publishing account.
The video publishing account may upload video data via a client or other device. The server can process the video data uploaded by the video publishing account, and intercept a plurality of frame images in the video data as key frame images. The video key frames can be generally used for extracting feature data in a preset mode. For example, a frame image obtained by cutting out video data every 250 milliseconds may be set as a key frame image. Other embodiments may also be used to determine key frame images from video data uploaded from a video distribution account. For example, the first frame image (or cover image) of the video data may be taken as the key frame image.
After the key frame image is obtained, feature extraction may be performed on the key frame image to determine feature data of the key frame image. The feature data may typically include data information characterizing the nature of the image itself and/or what the image characterizes. The feature data of the image may be extracted based on one or more features of color features, texture features, shape features, spatial relationship features, and the like of the image. Algorithms for extracting feature data may include, but are not limited to, Scale Invariant Feature Transform (SIFT), Histogram of Oriented Gradient (HOG), Local Binary Pattern (LBP), etc., and machine learning algorithms based on Convolutional Neural Networks (CNN), etc.
It is understood that the characteristic data may include data obtained by the foregoing algorithm, and may also include data obtained by further processing the data obtained by the algorithm. For example, after feature data Date _1 of the key frame image Pic _1 is extracted by SIFT, feature data Date _11 obtained by encoding the feature data Date _1 is stored in the image change data look-up table Sheet _ 1. Or adding the name Work _1 of the video data to which the key frame image Pic _1 belongs to the feature data Date _1 to identify that the feature data Date _1 comes from the name Work _1 of the provenance Work.
The feature data of the one or more video data stored by the server may be referred to as a feature data set. The feature data set may be stored in a storage unit of the server, or may be stored in a dedicated storage unit independent of the server or a storage unit of a third party. The server may perform query/search in a local storage unit, or may obtain query/search results from the dedicated storage unit or a third-party storage unit. For example, the server may store the aforementioned image encoding data look-up table Sheet _1 in a storage unit local to the server.
Similarly, the server may perform feature extraction on the picture image of the playing video by using the same algorithm as that used for obtaining the feature data of the key frame image, so as to obtain the feature data of the picture image. The server may search the feature data set for feature data that matches the feature data of the picture image (for convenience of distinction and description, the found matching feature data is referred to herein as target feature data). The searching mode may include comparing the feature data of the picture image with the feature data in the data set one by one, or may adopt a database index constructed in advance for searching.
The matching may include that the matching degree meets a preset requirement. For example, when the similarity between the feature data of the screen image and a certain feature data in the feature data set reaches 90% or more, it may be determined that the feature data of the screen image matches the feature data, and the feature data may be set as target feature data. If more than two feature data meeting the preset requirement of matching degree are found, the feature data with the highest matching degree can be selected as the target feature data in one embodiment. Of course, this disclosure does not exclude embodiments that may include multiple target feature data, and accordingly, video associated information of each target feature data may be acquired and returned.
It should be understood that, in the above description, the "provenance work name" is that, in general, when (after) video data is uploaded to a video publishing account, the video data may be named. In some application scenarios, the name named by the video publishing account for the video data may not be the same as the name of the work from which the video data originates. For example, if the short video uploaded by the video publishing account is a segment of the television series XXXXA, the video publishing account may name the short video as "domestic drama that has to be watched". The provenance work name in some embodiments of the disclosure may represent a source of a work from which the video data comes, and when it is determined that there is a need, video related information of the provenance work name may be returned, so as to provide a more accurate search result. If the short video uploaded by the account is a segment of the television series xxxxxa, the provenance title may be "xxxxxa". The provenance product name can be set to be filled in when (after) the video data are uploaded from the video publishing account, and can also be determined after the video data are subjected to video processing by the server or a third-party processing server.
In step S206, when the target feature data is found, video related information corresponding to the target feature data is obtained.
The video associated information may comprise information content related to video data of the target video data object. Specifically, the information related to the video data itself may be included, for example, a name of a place of business, a director, a lead actor, and a showing time corresponding to the video data, or the information related to the content in the video data may be included, for example, other movie works of the director of the video data, or a chinese knot is included in the video data, and the video related information may include information related to making of the chinese knot, meaning of the chinese knot, and the like. Some of the video related information may be determined based on information filled in (after) the video publishing account uploads the video data, for example, names of actors and names of departures filled in the video publishing account uploads the video.
If the server finds the target characteristic data matched with the characteristic data of the picture image in the characteristic data set, the server can acquire the video associated information corresponding to the target characteristic data.
In step S208, the video-related information is sent to the client, and the video-related information is used for displaying on the client.
The server may return all or part of the obtained video related information to the client. After receiving the video associated information returned by the server, the client can be used as a video watching account to display the search result of the picture image intercepted by the client.
According to the processing method for searching the video associated information, the video uploaded by the video publishing account can be processed in advance, the key frame image in the video is intercepted, and the feature data of the key frame image is extracted and then stored. When the video watching account at the client wants to acquire the relevant information in the watching video, the picture image of the video being played can be intercepted by the client and then uploaded to the server. Therefore, the server can search in the feature data set of the key frame image stored in advance according to the feature data of the picture image uploaded by the client, acquire the information associated with the picture image intercepted by the user and return the information to the client. And the client can display the video related information searched according to the picture images. According to the scheme of the embodiment of the disclosure, the picture content of the played video can be intercepted through the video watching account, the video related information is searched in an image mode, and as the server stores the characteristic data of the related video in advance, a more accurate search result can be obtained. According to the scheme of the embodiment of the disclosure, a new video interaction mode is provided, the server can automatically identify the picture image and return the search result, the video watching account can be searched without inputting text information, and the response speed of the client for displaying the search result is improved. And moreover, video interaction modes are enriched, and the experience of a video watching account is improved.
In other embodiments, the determining the picture image of the playing video of the client may include: and receiving a playing time label uploaded by the client, and acquiring a picture image of the video data according to the time label.
Because the data volume of the playing time tag is usually much smaller than that of the image data, the client can send the playing time tag without sending the picture image to the server, so that the interactive data flow can be reduced, the sending speed of the client is improved, and the response time is reduced.
As described in the previous embodiments, in some implementations, the video viewing account may intercept images of the video being played and then upload to the server. In other implementation scenarios, after the video is finished, the video may stay on the video playing page for a certain time, and the last frame of the video just played is continuously displayed. At this time, the video viewing account may capture a screen of the waiting interface after the video is played, and upload the screen to the server. Correspondingly, when processing the video data uploaded by the video publishing account, the server side may also use the last frame image corresponding to the video data as the key frame image. Of course, in other embodiments of the present disclosure, the first frame image corresponding to the video data may be used as a key frame image, or another designated frame image may be used as a key frame image. Therefore, in another exemplary embodiment of the present disclosure, in step S204, the key frame image of the video data may include:
a frame image at a specified location in the video data.
Accordingly, in another embodiment of the present disclosure, in step S202, the picture images uploaded by the client may also include a waiting interface image displayed after the client has played the post-video.
The designated position may include the last frame or the first frame or other designated frame image position, etc. In the embodiment of the present disclosure, the waiting interface image displayed after the video data is played by the client generally includes an image included in the video that is just played by the client. As shown in fig. 3, after the video V1 is played, a specific implementation example may count down for 5 seconds in the video playing interface, and within the 5 seconds of waiting time, the waiting interface continues to display the last frame image Pic _ last of the video just played, and may simultaneously display a manipulation button of the "enter next video" information. At this time, the video viewing account may capture a screen of the waiting interface after the video is played, obtain the picture image Scr _1, and upload the picture image Scr _1 to the server. The intercepted picture image Scr _1 may include the last frame image Pic _ last, and may also include other information. That is, the screen image may include a waiting interface image displayed after the video is played. For example, the image of the whole display interface of the client, captured by the video viewing account by pressing the entity key or the virtual screenshot button, may include power information, signal strength information, tool buttons in the display interface, and the like at the top of the display interface. In this way, if the video viewing account fails to capture the picture image during the video playing process, the waiting interface image displayed after the video playing process can be captured, so that the control steps and the overall search processing time (the time for playing back the video performed for capturing the picture image) caused by playing back the video under the conditions that the video viewing account forgets to capture the picture image or the video duration is unknown and the like are reduced.
In another exemplary embodiment of the present disclosure, in step S206, the video associated information may include one of:
account information of a target object in the picture image, a name of a provenance product of the video data corresponding to the target characteristic data, and image association information of which the matching degree with the characteristic data of the picture image meets a first requirement.
The video types played by the client can comprise various types, and various objects can also be included in the video, such as videos related to people, videos related to dynamic states, videos related to automobiles and the like. The captured screen image may also include one or more than one target object (in some embodiments of the present disclosure, the object included in the captured screen image may be referred to as a target object). The target object may be defined, set or filtered according to the type label of the video data or the identified image content in advance. For example, a television play segment (which may be determined according to remark information filled in when video data is uploaded by a video publishing account or a selected tag) is played, and a picture image captured by a user includes multiple objects such as characters, lakes, landmarks and the like, so that it may be predicted that the video viewing account is more likely to be the name of the television play or the real name of the characters (including the art name) according to the played video type tag and the like. The target object may be set as a person, and video related information related to the person included in the picture image may be preferentially acquired when the video related information is acquired.
Generally, the images of the playing video captured by the video viewing account usually contain video content of major interest or interest to the video viewing account. The video related information returned by the embodiment can include account information of the target object in the picture image, and can better accord with an expected search result of a video watching account, so that the search result is more accurate and reliable. The image-related information described in some embodiments of the present disclosure may refer to information associated with the screen image. The matching degree can be determined by the similarity between the feature data based on the picture image and certain feature data in the feature data set, or the similarity between the feature data based on the picture image and the features of information such as characters, pictures, videos and the like contained in the feature data set or other data information. The first requirement can be set in a self-defined mode, for example, the matching degree reaches 90%. If the image associated information with the matching degree meeting the first requirement conflicts with or coincides with other video associated information, a corresponding processing strategy can be preset. For example, a place where information content other than the target object and the name of the work of origin, such as a scene shown in the picture image, is viewed, may be set.
Similarly, the name of the video watched by the video watching account or the image associated information of which the matching degree with the feature data of the picture image meets the first requirement is returned, and the expected search result of the video watching account is also better met, so that the search result is more accurate and reliable, and further description is omitted.
Other embodiments of the disclosure provide technical solutions for searching related information content containing people in videos. Fig. 4 is a flowchart illustrating a processing method for acquiring video related information according to an exemplary embodiment, as shown in fig. 4, in another exemplary embodiment of the present disclosure, after finding the target feature data, the method may further include:
s402: when it is determined that the screen image contains a person based on the feature data of the screen image, the name of the person as a source is determined.
The characters may generally include natural people, such as Zhang three, Li four, etc. However, in some embodiments of the present disclosure, the character may be an unnatural person such as a robot or a virtual character, or may be a modified person or a modified person. These characters may be the roles of some film and television works. Such as the robot Mxx _007 from the future in the science fiction movie, the virtual singing Ji LuoTYxx, etc. The server may identify whether the picture image includes a person according to the feature data of the picture image, for example, may identify whether the image includes a face by using a feature face algorithm (Eigenface), and then determine whether the image includes a person image. If the picture image is determined to contain the person, the work name of the place can be obtained based on the video data of the target feature data corresponding to the picture image.
S404: and searching according to the name of the work of the place, and acquiring the work related information related to the name of the work of the place.
Correspondingly, the video associated information may include the composition associated information.
After obtaining the work provenance name, the server may perform further searching using the work provenance name to obtain information associated with the work provenance name (which may be referred to herein as work associated information). For example, the server recognizes the picture Pic _2, determines that the person Peo _2 is included, and determines that the video data of the picture Pic _2 is from the drama XXXXA. The server may search the library of works for "xxxxxa" for information associated with other video segments of the television series "xxxxxxxa", actor names, etc. The composition associated information may be provided as video associated information or as part of video associated information.
In a related information content searching scene of a video containing a character, if a video watching object is interested in the watched video, the name of a starting work of the character can be further searched, other information content related to the watched video is deeply mined, and then the information content is returned to a client side as video related information to be displayed. Therefore, the video associated information is further searched by taking the provenance product name as a keyword, more and more accurate search results can be returned to the video watching account, the expected search requirement of the video watching account can be met, and the search times and the search time are saved.
Fig. 5 is a flowchart illustrating a processing method for acquiring video related information according to an exemplary embodiment, where as shown in fig. 5, in another exemplary embodiment of the present disclosure, after the target feature data is found, the method further includes:
s502: when it is determined that the screen image contains a person based on the feature data of the screen image, the person name of the person is determined.
The name of the person may generally include the real name of the person, such as zhang san and lie si, or may be the art name of the person, or a name agreed or widely known in the art. For example, the name of a person from a robot in the future in the famous science fiction movie "xxxxxm" may be "M _ 007", and the name of a person of a virtual singing Ji, which is introduced by a music production company, may be "Luoty", and the like. The server may identify whether the picture image includes a person according to the feature data of the picture image, for example, may identify whether the image includes a face by using a feature face algorithm (Eigenface), and then determine whether the image includes a person image. If it is determined that the screen image contains a person, the name of the person may be acquired based on the video data of the target feature data corresponding to the screen image.
S504: and searching according to the person name to acquire the person related information related to the person name.
Correspondingly, the video related information comprises the person related information.
After obtaining the person name, the server may perform further search using the person name to obtain information associated with the person name (which may be referred to herein as person-associated information). For example, the server may identify the picture Pic _2 and determine that the person Peo _2 is included, or may determine that the video data of the picture Pic _2 is from the drama XXXXA. The server may search for "Peo _ 2" to obtain information related to Peo _2, such as account information of Peo _2 in a video playing application, live broadcast information, video clips of other participating movie and television works, and the like. The personal related information may be the video related information or a part of the video related information.
In a related information content search scene of a video containing people, if a person in the watched video is interested by a video watching object, the name of the person can be further searched, other information content related to the name of the person is deeply mined, and then the information content is returned to a client side as video related information to be displayed. Therefore, the person name is used as the keyword to further search the video related information, more and more accurate search results can be returned to the video watching account, the expected search requirement of the video watching account can be met, and the search times and the search time are saved.
Of course, in some embodiments, the video related information may include various information of the image related information, the work related information, and the person related information. When the captured picture image of the broadcast video contains a person, the result displayed at the client may include person-related information such as "short video with three pictures" and "live broadcast with Liqu" as shown in fig. 6, or work-related information such as "catkin of XXXXXA".
Fig. 7 is a flowchart illustrating a processing method for acquiring video related information according to an exemplary embodiment, as shown in fig. 7, in another exemplary embodiment of the present disclosure, after finding the target feature data, the method may further include:
702: when it is determined that the picture image does not contain the person object based on the feature data of the picture image, first video data containing the picture image and/or second video data matching the feature data of the picture image to a degree that meets a second requirement are searched.
Accordingly, the video associated information may include the first video data and/or the second video data.
For videos without people, the server may perform a search using feature data of the picture images. The searched content may include first video data including the picture image, and may include second video data matching the feature data of the picture image to a second degree in accordance with a second requirement. The second requirement may be specifically referred to the description of the aforementioned first requirement. The second requirement may be the same as or different from the first requirement. The background image understanding system utilizes the key frame data characteristic data coding lookup table of the server to identify the video containing the image or highly matched with the image
For example, in an exemplary scenario, the client plays a video clip about the chinese knot, and the captured picture image Pic _3 is a chinese knot image containing no character. The server identifies that the picture image Pic _3 does not contain any person, and then finds the first video data containing the picture image Pic _3 in the feature data set. If the first video data is not found, three second video data which have the highest matching degree with the feature data of the picture image Pic _3 and have the matching degree larger than 50% can be found in the feature data set or other data units. Of course, the second video data may also be searched and obtained under the condition that the first video data is searched. The video associated data may include both the first video data and the second video data.
Of course, in a scene of a video without a character, if the screen capture does not include a character, the returned video related information may include other pictures or text information in addition to the first video data or the second video data. As shown in fig. 8, in the application scenario schematic diagram of an embodiment, when the captured image does not include a person, the interface content of the video related information displayed by the client may include text information content related to a chinese knot, such as a text introduction page link of "meaning of the chinese knot," or a peripheral video related to the chinese knot, such as "chinese knot making", "video 1", and the like.
In a related information content searching scene of a video without characters, if a video watching object is interested in a target object in the watched video, the intercepted image content can be further searched, video related information related to the picture image is deeply mined, more and more accurate searching results can be returned to a video watching account, meanwhile, the expected searching requirement of the video watching account can be met, and the searching times and the searching time are saved.
The present disclosure also provides a processing method for obtaining video related information, which can be used on the side of the client 110. In a specific example embodiment, as shown in fig. 9, the method may include:
s902: determining a picture image of a playing video;
s904: uploading the picture image to a server;
s906: receiving video associated information corresponding to target feature data returned by a server, wherein the target feature data comprises feature data matched with the feature data of the picture image and searched by the server in a feature data set, and the feature data is obtained by extracting features from a key frame image of the video data uploaded to a video publishing account;
s908: and at least displaying part of the video related information.
The video watching account can intercept the picture of the playing video and acquire a picture image. One or more of the picture images may then be uploaded to a server. The server can extract the characteristics of the picture image, and match the characteristics with the characteristics in the stored characteristic data set to obtain the target characteristic data. Therefore, the server can acquire the video related information according to the target data and return the video related information to the client. The client can display the video associated information completely, and can also display the video associated information after screening processing or adding other information. Therefore, when the client watches the video, the video watching account can search the related information of the video by intercepting the picture image, the server side can automatically identify the picture image and return the search result, and the faster and more accurate search result can be obtained. According to the scheme of the embodiment of the disclosure, the client provides a new interaction mode in the interaction with the video watching account, the video watching account can be searched without inputting text information, the response speed of the client for displaying the search result is improved, the operation steps of the video watching account through text information search are reduced, and the quicker and more convenient video related information search is realized.
The present disclosure also provides another processing method for obtaining video related information, which can be used at the client 110 side. In a specific exemplary embodiment, the determining the picture image of the played video includes:
intercepting a picture of a playing video to obtain a picture image;
alternatively, the first and second electrodes may be,
the method comprises the following steps of intercepting a picture image of a playing video, determining a playing time tag of the picture in the playing video, and uploading the picture to a server, wherein the uploading step comprises the following steps: and uploading the playing time tag to a server.
The present disclosure also provides another processing method for obtaining video related information, which can be used at the client 110 side. In a specific exemplary embodiment, the video related information includes one of:
account information of a target object contained in the picture image, a source product name of video data corresponding to the target characteristic data, and image association information of which the matching degree with the characteristic data of the picture image meets a first requirement.
The present disclosure also provides another processing method for obtaining video related information, which can be used at the client 110 side. In a specific exemplary embodiment, after the server finds the target feature data, the method may further include:
when the picture image is determined to contain a person based on the feature data of the picture image, the returned video related information includes one of the following information:
searching for the obtained work associated information according to the name of the person;
and searching the acquired person associated information according to the person name of the person.
The present disclosure also provides another processing method for obtaining video related information, which can be used at the client 110 side. In a specific exemplary embodiment, after the server finds the target feature data, the method further includes:
when the picture image is determined to contain a person based on the characteristic data of the picture image, the returned video related information at least comprises one of the following information searched by the server:
first video data containing the picture image;
and second video data having a degree of matching with the feature data of the picture image in accordance with a second requirement.
The present disclosure also provides another processing method for obtaining video related information, which can be used at the client 110 side. In a specific exemplary embodiment, the acquiring the picture image of the playing video may include:
and intercepting the picture image based on the image search interface element provided by the trigger.
In this embodiment, an image search interface element may be set at the client. The image search interface element may be a page element displayed on the display interface, such as a plug-in, a control, a switch, and the like, and may be used to control a behavior of a video playing application or a client. In this embodiment, when the image search interface element may be triggered, an action of capturing a picture image of a playing video may be executed to obtain the picture image, thereby implementing subsequent video related information search through the image. The image search interface element may be temporary, such as displaying the image search interface element while playing a video. Or the video watching account can be resident in a display interface of the client, and the video watching account can intercept the picture image at any time. In an exemplary embodiment, the image search interface element may be an image search button, and in playing a video, the video viewing account may click on the virtual image search button to capture a frame image of the playing video.
The above method embodiments that can be used on the client side may refer to the description of the server side related embodiments. The image search interface specially used for intercepting and playing the video image can be provided in the embodiment, so that the video watching account can conveniently and quickly intercept the image, further, the video related information search based on the image can be quickly carried out, the interaction mode is simpler and more convenient, the acquisition time of the video related information is integrally reduced, and the acquisition efficiency of the data information is improved.
It is understood that the embodiments of the client-side or server-side method described above are described in a progressive manner, and the same/similar parts of the embodiments may be referred to each other, and each embodiment focuses on differences from other embodiments. Reference may be made to the description of other method embodiments for relevant points.
It should be understood that, although the steps in the flowcharts in fig. 2 to 9 are sequentially shown as indicated by arrows, the steps are not necessarily performed sequentially as indicated by the arrows. The steps are not performed in the exact order shown and described, and may be performed in other orders, unless explicitly stated otherwise. Moreover, at least some of the steps in fig. 2-9 may include multiple steps or multiple stages, which are not necessarily performed at the same time, but may be performed at different times, which are not necessarily performed in sequence, but may be performed in turn or alternately with other steps or at least some of the steps or stages of other steps.
Based on the above description of the page data processing embodiment of the video medium, the present disclosure further provides a processing device for acquiring video related information. The apparatus may include systems (including distributed systems), software (applications), modules, components, servers, clients, etc. that use the methods described in embodiments of the present specification in conjunction with any necessary apparatus to implement the hardware. Based on the same innovative concept, the embodiments of the present disclosure provide an apparatus in one or more embodiments as described in the following embodiments. Since the implementation scheme of the apparatus for solving the problem is similar to that of the method, the specific implementation of the apparatus in the embodiment of the present specification may refer to the implementation of the foregoing method, and repeated details are not repeated. As used hereinafter, the term "unit" or "module" may be a combination of software and/or hardware that implements a predetermined function. Although the means described in the embodiments below are preferably implemented in software, an implementation in hardware, or a combination of software and hardware is also possible and contemplated.
Fig. 10 is a block diagram illustrating a processing device for acquiring video associated information according to an exemplary embodiment. The device may be a server, such as the server 120 described above. Specifically, referring to fig. 10, the apparatus 100 may include an image determining module 102, a feature matching module 104, an information searching module 106, and a sending module 108.
The image determination module 102 is configured to determine a picture image of a playing video of a client;
the feature matching module 104 is configured to search for target feature data matching feature data of the picture image in a feature data set, where the feature data set includes feature data extracted from a key frame image, and the key frame image is determined based on video data uploaded by a video publishing account;
the information search module 106 is configured to, when the target feature data is found, obtain video associated information corresponding to the target feature data;
the sending module 108 is configured to send the video associated information to the client, where the video associated information is used for presentation at the client.
In an exemplary embodiment, the image determination module determining a picture image of a playing video of a client includes:
receiving the picture images uploaded by the client side,
alternatively, the first and second electrodes may be,
and receiving a playing time label uploaded by the client, and acquiring a picture image of the video data according to the time label.
In an exemplary embodiment, the key frame images of the video data in the feature matching module 104 include:
a frame image at a specified location in the video data;
the picture images comprise waiting interface images displayed after the video is played by the client.
In an exemplary embodiment, the video related information obtained in the information search module 106 includes one of the following:
account information of a target object in the picture image, a name of a provenance product of the video data corresponding to the target characteristic data, and image association information of which the matching degree with the characteristic data of the picture image meets a first requirement.
An exemplary embodiment is shown in fig. 11, and fig. 11 is a block diagram of a processing device for acquiring video associated information according to an exemplary embodiment. Referring to fig. 11, the apparatus may further include:
a character work determination module 1102 configured to determine a name of a work of origin of a person when the picture image is determined to contain the person based on the feature data of the picture image after the target feature data is found;
a work search module 1104 configured to search according to the name of the offer work and obtain work association information associated with the name of the offer work;
correspondingly, the video associated information comprises the work associated information.
An exemplary embodiment is shown in fig. 12, and fig. 12 is a block diagram of a processing device for acquiring video associated information according to an exemplary embodiment. Referring to fig. 12, the apparatus may further include:
a person name determining module 1202 configured to determine a person name of a person after finding the target feature data and when determining that the screen image contains the person based on the feature data of the screen image;
a name search module 1204 configured to perform a search according to the person name, and acquire person related information associated with the person name;
correspondingly, the video related information comprises the person related information.
An exemplary embodiment is shown in fig. 13, and fig. 13 is a block diagram of a processing device for acquiring video associated information according to an exemplary embodiment. Referring to fig. 13, the apparatus may further include:
a non-person searching module 1302 configured to search, when it is determined that the screen image does not contain a person object based on the feature data of the screen image, first video data containing the screen image and/or second video data matching the feature data of the screen image to a degree that meets a second requirement;
correspondingly, the video associated information includes the first video data and/or the second video data.
Fig. 14 is a block diagram illustrating another processing device for obtaining video associated information according to an example embodiment. The apparatus may be a client 120, and specifically, referring to fig. 14, the apparatus 200 may include:
an image acquisition module 1402 configured to determine a picture image of a playing video;
an upload module 1404 configured to upload the screen image to a server;
the related information receiving module 1406 is configured to receive video related information corresponding to target feature data returned by the server, where the target feature data includes feature data matched with the feature data of the picture image and found in a feature data set by the server, and the feature data is obtained by performing feature extraction on a key frame image determined by video data uploaded to a video publishing account;
a presentation module 1408 configured to present at least a portion of the video-related information.
In another processing apparatus for acquiring video related information shown in an exemplary embodiment, the deciding of the picture image of the video to be played by the image acquiring module includes:
intercepting a picture of a playing video to obtain a picture image;
alternatively, the first and second electrodes may be,
intercepting a picture image of a playing video, and determining a playing time label of the picture in the playing video; the uploading module uploads the picture to a server and comprises: and uploading the playing time tag to a server.
Another processing apparatus for acquiring video related information according to an exemplary embodiment of the present invention is a processing apparatus for acquiring video related information, where the video related information includes one of:
account information contained in the picture image, a name of a place of sale of the video data corresponding to the target characteristic data, and image association information of which the matching degree with the characteristic data of the picture image meets a first requirement.
Another processing apparatus for acquiring video related information shown in an exemplary embodiment, after the server finds the target feature data, the processing apparatus further includes:
when the picture image is determined to contain a person based on the feature data of the picture image, the returned video related information includes one of the following information:
searching for the obtained work associated information according to the name of the person;
and searching the acquired person associated information according to the person name of the person.
Another processing apparatus for acquiring video related information shown in an exemplary embodiment, after the server finds the target feature data, the processing apparatus further includes:
when the picture image is determined to contain a person based on the characteristic data of the picture image, the returned video related information comprises one of the following information searched by the server:
first video data containing the picture image;
and second video data having a degree of matching with the feature data of the picture image in accordance with a second requirement.
Another processing apparatus for acquiring video related information according to an exemplary embodiment of the present invention, wherein the acquiring a picture image of a playing video includes:
and intercepting the picture image based on the image search interface element provided by the trigger.
Another processing apparatus for acquiring video-related information according to an exemplary embodiment is one in which the screen image includes:
and displaying the waiting interface image after the video is played.
With regard to the apparatus in the above-described embodiment, the specific manner in which each module performs the operation has been described in detail in the embodiment related to the method, and will not be elaborated here.
Fig. 15 is a block diagram illustrating a processing apparatus Z00 for acquiring video associated information according to an exemplary embodiment. The device Z00 may be an electronic device on the video viewing account side, for example, device Z00 may be a mobile phone, computer, digital broadcast terminal, messaging device, game console, tablet device, medical device, fitness device, personal digital assistant, or the like.
Referring to fig. 15, device Z00 may include one or more of the following components: a processing component Z02, a memory Z04, a power component Z06, a multimedia component Z08, an audio component Z10, an interface for input/output (I/O) Z12, a sensor component Z14 and a communication component Z16.
The processing component Z02 generally controls the overall operation of the device Z00, such as operations associated with display, telephone calls, data communications, camera operations, and recording operations. The processing component Z02 may include one or more processors Z20 to execute instructions to perform all or part of the steps of the method described above. Further, the processing component Z02 may include one or more modules that facilitate interaction between the processing component Z02 and other components. For example, the processing component Z02 may include a multimedia module to facilitate interaction between the multimedia component Z08 and the processing component Z02.
The memory Z04 is configured to store various types of data to support operations at device Z00. Examples of such data include instructions for any application or method operating on device Z00, contact data, phonebook data, messages, pictures, videos, etc. The memory Z04 may be implemented by any type or combination of volatile or non-volatile storage devices, such as static random access memory (12RAM), electrically erasable programmable read-only memory (EEPROM), erasable programmable read-only memory (EPROM), programmable read-only memory (PROM), read-only memory (ROM), magnetic memory, flash memory, magnetic or optical disks.
The power supply component Z06 provides power to the various components of the device Z00. The power component Z06 may include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power for the device Z00.
The multimedia component Z08 comprises a screen between the device Z00 and the object providing an output interface. In some embodiments, the screen may include a Liquid Crystal Display (LCD) and a Touch Panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive an input signal from an object. The touch panel includes one or more touch sensors to sense touch, slide, and gestures on the touch panel. The touch sensor may not only sense the boundary of a touch or slide action, but also detect the duration and pressure associated with the touch or slide operation. In some embodiments, the multimedia component Z08 includes a front facing camera and/or a rear facing camera. When device Z00 is in an operating mode, such as a capture mode or a video mode, the front-facing camera and/or the rear-facing camera may receive external multimedia data. Each front camera and rear camera may be a fixed optical lens system or have a focal length and optical zoom capability.
The audio component Z10 is configured to output and/or input an audio signal. For example, the audio component Z10 includes a Microphone (MIC) configured to receive external audio signals when the device Z00 is in an operational mode, such as a call mode, a recording mode, and a voice recognition mode. The received audio signal may further be stored in the memory Z04 or transmitted via the communication component Z16. In some embodiments, the audio component Z10 further includes a speaker for outputting audio signals.
The I/O interface Z12 provides an interface between the processing component Z02 and peripheral interface modules, which may be keyboards, click wheels, buttons, etc. These buttons may include, but are not limited to: a home button, a volume button, a start button, and a lock button.
The sensor assembly Z14 includes one or more sensors for providing status assessment of various aspects to the device Z00. For example, sensor assembly Z14 may detect the open/closed state of device Z00, the relative positioning of the components, such as the display and keypad of device Z00, sensor assembly Z14 may also detect a change in the position of one component of device Z00 or device Z00, the presence or absence of an object in contact with device Z00, the orientation or acceleration/deceleration of device Z00, and a change in the temperature of device Z00. The sensor assembly Z14 may include a proximity sensor configured to detect the presence of a nearby object without any physical contact. The sensor assembly Z14 may also include a light sensor, such as a CMO12 or CCD image sensor, for use in imaging applications. In some embodiments, the sensor assembly Z14 may also include an acceleration sensor, a gyroscope sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
The communication component Z16 is configured to facilitate wired or wireless communication between device Z00 and other devices. Device Z00 may access a wireless network based on a communication standard, such as WiFi, a carrier network (such as 2G, 3G, 4G, or 5G), or a combination thereof. In an exemplary embodiment, the communication component Z16 receives a broadcast signal or broadcast related information from an external broadcast management system via a broadcast channel. In an exemplary embodiment, the communication component Z16 further includes a Near Field Communication (NFC) module to facilitate short-range communications. For example, the NFC module may be implemented based on Radio Frequency Identification (RFID) technology, infrared data association (IrDA) technology, Ultra Wideband (UWB) technology, Bluetooth (BT) technology, and other technologies.
In an exemplary embodiment, the device Z00 may be implemented by one or more application specific integrated circuits (a12IC), a digital signal processor (D12P), a digital signal processing device (D12PD), a Programmable Logic Device (PLD), a Field Programmable Gate Array (FPGA), a controller, a microcontroller, a microprocessor or other electronic components for performing the above processing method for obtaining video related information, which may be implemented on the client side.
It should be noted that the device Z00 may be an exemplary description of an electronic device on the side of a video viewing object, and in some end products, it may not be necessary to include all of the above components or all of the functional units under a certain component.
In an exemplary embodiment, there is also provided a non-transitory computer readable storage medium, such as the memory Z04, comprising instructions executable by the processor Z20 of the device Z00 to perform the above processing method for obtaining video association information, which may be implemented on the client side. For example, the non-transitory computer readable storage medium may be a ROM, a Random Access Memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, and the like.
Fig. 16 is a block diagram illustrating a processing device S00 for acquiring video associated information according to an exemplary embodiment. For example, the apparatus S00 may be a combination of one or more servers, such as a server for extracting and/or matching features of a picture image or a key frame image, a server for recognizing and understanding contents in a picture image, such as recognizing a person and a place in an image, or a server for performing a further search on the recognized contents in a picture image, such as a server for performing a further search by using a name of a recognized place product or a name of a person as a keyword. Referring to FIG. 16, device S00 includes a processing component S20 that further includes one or more processors and memory resources represented by memory S22 for storing instructions, e.g., applications, that are executable by processing component S20. The application program stored in the memory S22 may include one or more modules each corresponding to a set of instructions. Further, the processing component S20 is configured to execute instructions to perform the above-described processing method of obtaining video related information that may be implemented at a server.
The device S00 may also include a power supply component S24 configured to perform power management of the device S00, a wired or wireless network interface S26 configured to connect the device S00 to a network, and an input-output (I/O) interface S28. The device S00 may operate based on an operating system stored in memory S22, such as Window 1212 erver, Mac O12X, Unix, Linux, FreeB12D, or the like.
In an exemplary embodiment, there is also provided a storage medium, such as the memory S22, including instructions executable by the processor of the device S00 to perform the above processing method for obtaining video associated information, which may be implemented in a server. The storage medium may be a non-transitory computer readable storage medium, which may be, for example, a ROM, a Random Access Memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, and the like. And other implementations of a readable storage medium, such as quantum storage, graphene storage, and so forth.
The embodiments in the present specification are described in a progressive manner, and the same and similar parts among the embodiments are referred to each other, and each embodiment focuses on the differences from the other embodiments. In particular, for the hardware + program class embodiment, since it is substantially similar to the method embodiment, the description is simple, and the relevant points can be referred to the partial description of the method embodiment.
It should be noted that, the descriptions of the above-mentioned apparatuses, devices, servers, and the like according to the method embodiments may also include other embodiments, and specific implementations may refer to the descriptions of the related method embodiments. Meanwhile, the new embodiment formed by the mutual combination of the features of the methods, the devices, the equipment and the server embodiments still belongs to the implementation range covered by the present disclosure, and the details are not repeated herein.
For convenience of description, the above devices are described as being divided into various modules by functions, and are described separately. Of course, when implementing one or more of the present description, the functions of each module may be implemented in one or more software and/or hardware, or a module implementing the same function may be implemented by a combination of multiple sub-modules or sub-units, etc. The above-described embodiments of the apparatus are merely illustrative, and for example, a module or a unit may be divided into only one logical function, and may be implemented in other ways, for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted or not executed. In addition, the coupling, communication connection, etc. between the devices or units shown or described may be realized by direct and/or indirect coupling/connection, and may be realized by some standard or customized interfaces, protocols, etc., in an electrical, mechanical or other form.
Other embodiments of the disclosure will be apparent to those skilled in the art from consideration of the specification and practice of the disclosure disclosed herein. This disclosure is intended to cover any variations, uses, or adaptations of the disclosure following, in general, the principles of the disclosure and including such departures from the present disclosure as come within known or customary practice within the art to which the disclosure pertains. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the disclosure being indicated by the following claims.
It will be understood that the present disclosure is not limited to the precise arrangements described above and shown in the drawings and that various modifications and changes may be made without departing from the scope thereof.

Claims (10)

1. A processing method for acquiring video associated information is characterized by comprising the following steps:
determining a picture image of a playing video of a client;
searching target characteristic data matched with the characteristic data of the picture image in a characteristic data set, wherein the characteristic data set comprises characteristic data extracted from a key frame image, and the key frame image is determined based on video data uploaded by a video publishing account;
when the target characteristic data is found, video associated information corresponding to the target characteristic data is obtained;
and sending the video associated information to the client, wherein the video associated information is used for displaying at the client.
2. The method of claim 1, wherein determining the picture image of the playing video of the client comprises: receiving the picture images uploaded by the client side,
alternatively, the first and second electrodes may be,
and receiving a playing time label uploaded by the client, and acquiring a picture image of the video data according to the time label.
3. The method of claim 1, wherein the key frame images of the video data comprise:
a frame image at a specified location in the video data;
the picture images comprise waiting interface images displayed after the video is played by the client.
4. A processing method for acquiring video associated information is characterized by comprising the following steps:
determining a picture image of a playing video;
uploading the picture image to a server;
receiving video associated information corresponding to target feature data returned by a server, wherein the target feature data comprise feature data matched with the feature data of the picture image and searched by the server in a feature data set, and the feature data are obtained by performing feature extraction on a key frame image determined by video data uploaded to a video publishing account;
and at least displaying part of the video related information.
5. A processing apparatus for obtaining video-related information, comprising: the system comprises an image determining module, a feature matching module, an information searching module and a sending module;
the image determination module is configured to determine a picture image of a playing video of a client;
the feature matching module is configured to search for target feature data matched with feature data of the picture image in a feature data set, wherein the feature data set comprises feature data extracted from a key frame image, and the key frame image is determined based on video data uploaded by a video publishing account;
the information searching module is configured to acquire video associated information corresponding to the target characteristic data when the target characteristic data is searched;
the sending module is configured to send the video associated information to the client, the video associated information being for presentation at the client.
6. A processing apparatus for obtaining video associated information, the apparatus comprising:
an image acquisition module configured to determine a picture image of a playing video;
an upload module configured to upload the screen image to a server;
the relevant information receiving module is configured to receive video relevant information corresponding to target characteristic data returned by the server, wherein the target characteristic data comprise characteristic data which are searched by the server in a characteristic data set and are matched with the characteristic data of the picture image, and the characteristic data are obtained by performing characteristic extraction on a key frame image determined by video data uploaded to a video publishing account;
a presentation module configured to present at least a portion of the video-associated information.
7. An electronic device, comprising:
a processor;
a memory for storing the processor-executable instructions;
wherein the processor is configured to execute the instructions to implement the method of claim 4.
8. A storage medium, wherein instructions in the storage medium, when executed by a processor of an electronic device, enable the electronic device to perform the method of claim 4.
9. A server, comprising:
at least one processor;
a memory for storing the processor-executable instructions;
wherein the processor is configured to execute the instructions to implement the method of any one of claims 1 to 3.
10. A storage medium, wherein instructions in the storage medium, when executed by a processor of a server, enable the server to perform the method of any one of claims 1 to 3.
CN202011641038.9A 2020-12-31 2020-12-31 Processing method and device for acquiring video associated information, electronic equipment and server Pending CN112866762A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011641038.9A CN112866762A (en) 2020-12-31 2020-12-31 Processing method and device for acquiring video associated information, electronic equipment and server

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011641038.9A CN112866762A (en) 2020-12-31 2020-12-31 Processing method and device for acquiring video associated information, electronic equipment and server

Publications (1)

Publication Number Publication Date
CN112866762A true CN112866762A (en) 2021-05-28

Family

ID=76000840

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011641038.9A Pending CN112866762A (en) 2020-12-31 2020-12-31 Processing method and device for acquiring video associated information, electronic equipment and server

Country Status (1)

Country Link
CN (1) CN112866762A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113407850A (en) * 2021-07-15 2021-09-17 北京百度网讯科技有限公司 Method and device for determining and acquiring virtual image and electronic equipment

Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104125492A (en) * 2013-04-23 2014-10-29 深圳市快播科技有限公司 Video playing method and device
CN104754413A (en) * 2013-12-30 2015-07-01 北京三星通信技术研究有限公司 Image search based television signal identification and information recommendation method and device
US20150341457A1 (en) * 2013-06-28 2015-11-26 Tencent Technology (Shenzhen) Company Limited Method and system for pushing information to end users adaptively
CN105320706A (en) * 2014-08-05 2016-02-10 阿里巴巴集团控股有限公司 Processing method and device of search result
CN105791906A (en) * 2014-12-15 2016-07-20 深圳Tcl数字技术有限公司 Information pushing method and system
CN105992041A (en) * 2014-12-22 2016-10-05 青岛海信电器股份有限公司 Method and device for encoding a captured screenshot and controlling program content switching based on the captured screenshot
US20170339287A1 (en) * 2016-05-20 2017-11-23 Beijing Xiaomi Mobile Software Co., Ltd. Image transmission method and apparatus
CN107404661A (en) * 2017-08-23 2017-11-28 韩城黄河指尖影院有限公司 A kind of micro- film video acquisition methods and device
US20180025405A1 (en) * 2016-07-25 2018-01-25 Facebook, Inc. Presentation of content items synchronized with media display
CN107977145A (en) * 2017-12-15 2018-05-01 广东欧珀移动通信有限公司 Electronic device, screenshot method and Related product
CN108255970A (en) * 2017-12-26 2018-07-06 努比亚技术有限公司 A kind of video retrieval method, terminal and computer readable storage medium
CN108462889A (en) * 2017-02-17 2018-08-28 阿里巴巴集团控股有限公司 Information recommendation method during live streaming and device
CN109068185A (en) * 2018-09-25 2018-12-21 湖南快乐阳光互动娱乐传媒有限公司 A kind of video screening technique and system
CN110019933A (en) * 2018-01-02 2019-07-16 阿里巴巴集团控股有限公司 Video data handling procedure, device, electronic equipment and storage medium
CN110362714A (en) * 2019-07-25 2019-10-22 腾讯科技(深圳)有限公司 The searching method and device of video content
CN110909209A (en) * 2019-11-26 2020-03-24 北京达佳互联信息技术有限公司 Live video searching method and device, equipment, server and storage medium
CN111163367A (en) * 2020-01-08 2020-05-15 百度在线网络技术(北京)有限公司 Information search method, device, equipment and medium based on playing video
US20200275048A1 (en) * 2017-12-20 2020-08-27 Hisense Visual Technology Co., Ltd. Smart television and method for displaying graphical user interface of television screen shot

Patent Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104125492A (en) * 2013-04-23 2014-10-29 深圳市快播科技有限公司 Video playing method and device
US20150341457A1 (en) * 2013-06-28 2015-11-26 Tencent Technology (Shenzhen) Company Limited Method and system for pushing information to end users adaptively
CN104754413A (en) * 2013-12-30 2015-07-01 北京三星通信技术研究有限公司 Image search based television signal identification and information recommendation method and device
CN105320706A (en) * 2014-08-05 2016-02-10 阿里巴巴集团控股有限公司 Processing method and device of search result
CN105791906A (en) * 2014-12-15 2016-07-20 深圳Tcl数字技术有限公司 Information pushing method and system
CN105992041A (en) * 2014-12-22 2016-10-05 青岛海信电器股份有限公司 Method and device for encoding a captured screenshot and controlling program content switching based on the captured screenshot
US20170339287A1 (en) * 2016-05-20 2017-11-23 Beijing Xiaomi Mobile Software Co., Ltd. Image transmission method and apparatus
US20180025405A1 (en) * 2016-07-25 2018-01-25 Facebook, Inc. Presentation of content items synchronized with media display
CN108462889A (en) * 2017-02-17 2018-08-28 阿里巴巴集团控股有限公司 Information recommendation method during live streaming and device
CN107404661A (en) * 2017-08-23 2017-11-28 韩城黄河指尖影院有限公司 A kind of micro- film video acquisition methods and device
CN107977145A (en) * 2017-12-15 2018-05-01 广东欧珀移动通信有限公司 Electronic device, screenshot method and Related product
US20200275048A1 (en) * 2017-12-20 2020-08-27 Hisense Visual Technology Co., Ltd. Smart television and method for displaying graphical user interface of television screen shot
CN108255970A (en) * 2017-12-26 2018-07-06 努比亚技术有限公司 A kind of video retrieval method, terminal and computer readable storage medium
CN110019933A (en) * 2018-01-02 2019-07-16 阿里巴巴集团控股有限公司 Video data handling procedure, device, electronic equipment and storage medium
CN109068185A (en) * 2018-09-25 2018-12-21 湖南快乐阳光互动娱乐传媒有限公司 A kind of video screening technique and system
CN110362714A (en) * 2019-07-25 2019-10-22 腾讯科技(深圳)有限公司 The searching method and device of video content
CN110909209A (en) * 2019-11-26 2020-03-24 北京达佳互联信息技术有限公司 Live video searching method and device, equipment, server and storage medium
CN111163367A (en) * 2020-01-08 2020-05-15 百度在线网络技术(北京)有限公司 Information search method, device, equipment and medium based on playing video

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
深响(界面官方财经号): "揭秘《青春有你2》火爆背后,新技术带来的内容变革" *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113407850A (en) * 2021-07-15 2021-09-17 北京百度网讯科技有限公司 Method and device for determining and acquiring virtual image and electronic equipment
CN113407850B (en) * 2021-07-15 2022-08-26 北京百度网讯科技有限公司 Method and device for determining and acquiring virtual image and electronic equipment

Similar Documents

Publication Publication Date Title
CN110662083B (en) Data processing method and device, electronic equipment and storage medium
CN110517185B (en) Image processing method, device, electronic equipment and storage medium
KR101910346B1 (en) Picture processing method and apparatus
CN104615769B (en) Picture classification method and device
KR101680714B1 (en) Method for providing real-time video and device thereof as well as server, terminal device, program, and recording medium
WO2018000585A1 (en) Interface theme recommendation method, apparatus, terminal and server
CN112153400B (en) Live broadcast interaction method and device, electronic equipment and storage medium
CN105631403A (en) Method and device for human face recognition
KR101985955B1 (en) Face photo album based music playing method, apparatus and terminal device and storage medium
CN106331761A (en) Live broadcast list display method and apparatuses
CN105654039A (en) Image processing method and device
CN110572706B (en) Video screenshot method, terminal and computer-readable storage medium
CN112672208B (en) Video playing method, device, electronic equipment, server and system
CN105678266A (en) Method and device for combining photo albums of human faces
CN111526287A (en) Image shooting method, image shooting device, electronic equipment, server, image shooting system and storage medium
CN107895041B (en) Shooting mode setting method and device and storage medium
CN112464031A (en) Interaction method, interaction device, electronic equipment and storage medium
CN109145878B (en) Image extraction method and device
CN110019907B (en) Image retrieval method and device
CN110019897B (en) Method and device for displaying picture
CN112004020B (en) Image processing method, image processing device, electronic equipment and storage medium
CN112866762A (en) Processing method and device for acquiring video associated information, electronic equipment and server
CN113888543A (en) Skin color segmentation method and device, electronic equipment and storage medium
CN113596574A (en) Video processing method, video processing apparatus, electronic device, and readable storage medium
CN113032627A (en) Video classification method and device, storage medium and terminal equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20210528