CN113315691A - Video processing method and device and electronic equipment - Google Patents

Video processing method and device and electronic equipment Download PDF

Info

Publication number
CN113315691A
CN113315691A CN202110553846.8A CN202110553846A CN113315691A CN 113315691 A CN113315691 A CN 113315691A CN 202110553846 A CN202110553846 A CN 202110553846A CN 113315691 A CN113315691 A CN 113315691A
Authority
CN
China
Prior art keywords
target
target video
video
user
input
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110553846.8A
Other languages
Chinese (zh)
Other versions
CN113315691B (en
Inventor
杭宇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Vivo Mobile Communication Co Ltd
Original Assignee
Vivo Mobile Communication Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Vivo Mobile Communication Co Ltd filed Critical Vivo Mobile Communication Co Ltd
Priority to CN202110553846.8A priority Critical patent/CN113315691B/en
Publication of CN113315691A publication Critical patent/CN113315691A/en
Priority to PCT/CN2022/092837 priority patent/WO2022242577A1/en
Application granted granted Critical
Publication of CN113315691B publication Critical patent/CN113315691B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/04Real-time or near real-time messaging, e.g. instant messaging [IM]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/06Message adaptation to terminal or network requirements
    • H04L51/063Content adaptation, e.g. replacement of unsuitable content
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/07User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail characterised by the inclusion of specific contents
    • H04L51/10Multimedia information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/44008Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting

Abstract

The application discloses a video processing method and device and electronic equipment, and belongs to the technical field of video processing. The video processing method comprises the following steps: receiving a target video sent by a target chat object; and displaying a first identifier under the condition that the first object in the target video is matched with a preset second object, wherein the first identifier is used for indicating that the target video comprises the second object.

Description

Video processing method and device and electronic equipment
Technical Field
The application belongs to the technical field of videos, and particularly relates to a video processing method and device.
Background
With the development of communication technology, users can acquire massive videos through electronic equipment. However, in the prior art, in the case that the electronic device receives a video, a user may need to click to watch the video to determine whether content of interest of the user exists in the video, and particularly, in the case that a plurality of videos are received, the user needs to play the watching videos one by one to determine which video is of interest of the user, and the determination process is tedious and time-consuming.
Disclosure of Invention
The embodiment of the application aims to provide a video processing method and device, and the problems that in the prior art, the process of determining videos in which users are interested is complicated and time-consuming can be solved.
In a first aspect, an embodiment of the present application provides a video processing method, where the method includes:
receiving a target video sent by a target chat object;
and displaying a first identifier under the condition that the first object in the target video is matched with a preset second object, wherein the first identifier is used for indicating that the target video comprises the second object.
In a second aspect, an embodiment of the present application provides a video processing apparatus, including:
the receiving module is used for receiving a target video sent by a target chat object;
the first display module is used for displaying a first identifier under the condition that a first object in the target video is matched with a preset second object, and the first identifier is used for indicating that the target video comprises the second object.
In a third aspect, embodiments of the present application provide an electronic device, which includes a processor, a memory, and a program or instructions stored on the memory and executable on the processor, where the program or instructions, when executed by the processor, implement the steps of the method according to the first aspect.
In a fourth aspect, embodiments of the present application provide a readable storage medium on which a program or instructions are stored, which when executed by a processor, implement the steps of the method according to the first aspect.
In a fifth aspect, embodiments of the present application provide a chip, where the chip includes a processor and a communication interface, where the communication interface is coupled to the processor, and the processor is configured to execute a program or instructions to implement the method according to the first aspect.
In the embodiment of the application, a target video sent by a target chat object is received, and a first identifier is displayed when a first object in the target video is matched with a preset second object, wherein the first identifier is used for indicating that the target video comprises the second object. By matching the first object with the second object, the first identifier can be displayed on the target video comprising the second object, and the method and the device for displaying the first identifier are beneficial to a user to efficiently acquire the video with the content of interest.
Drawings
Fig. 1 is a schematic flowchart of a video processing method provided in an embodiment of the present application;
FIG. 2 is one of exemplary diagrams of a session interface in an embodiment of the application;
FIG. 3 is a second exemplary diagram of a session interface in an embodiment of the present application;
FIG. 4 is a third exemplary diagram of a session interface in an embodiment of the present application;
FIG. 5 is one of exemplary diagrams of a target window in an embodiment of the present application;
FIG. 6 is a second exemplary diagram of a target window in an embodiment of the present application;
FIG. 7 is a third exemplary diagram of a target window in an embodiment of the present application;
FIG. 8 is one of exemplary diagrams of a setup interface in an embodiment of the present application;
FIG. 9 is a second exemplary illustration of a setup interface in an embodiment of the present application;
FIG. 10 is a third exemplary view of a setup interface in an embodiment of the present application;
FIG. 11 is a fourth illustration of an example of a setup interface in an embodiment of the present application;
FIG. 12 is a fifth illustration of an example of a setup interface in an embodiment of the present application;
fig. 13 is a schematic flowchart of a video processing method provided in an embodiment of the present application in a specific application example;
fig. 14 is a schematic structural diagram of a video processing apparatus according to an embodiment of the present application;
fig. 15 is a schematic structural diagram of an electronic device provided in an embodiment of the present application;
fig. 16 is a schematic hardware structure diagram of an electronic device according to an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be described clearly below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are some, but not all, embodiments of the present application. All other embodiments that can be derived by one of ordinary skill in the art from the embodiments given herein are intended to be within the scope of the present disclosure.
The terms first, second and the like in the description and in the claims of the present application are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It will be appreciated that the data so used may be interchanged under appropriate circumstances such that embodiments of the application may be practiced in sequences other than those illustrated or described herein, and that the terms "first," "second," and the like are generally used herein in a generic sense and do not limit the number of terms, e.g., the first term can be one or more than one. In addition, "and/or" in the specification and claims means at least one of connected objects, a character "/" generally means that a preceding and succeeding related objects are in an "or" relationship.
The following describes in detail a video processing method and apparatus provided in the embodiments of the present application with reference to the accompanying drawings.
As shown in fig. 1, a video processing method provided in an embodiment of the present application includes:
step 101, receiving a target video sent by a target chat object;
and 102, displaying a first identifier under the condition that a first object in the target video is matched with a preset second object, wherein the first identifier is used for indicating that the target video comprises the second object. The mark in the present application is used for indicating words, symbols, images and the like of information, and a control or other container can be used as a carrier for displaying information, including but not limited to a word mark, a symbol mark, an image mark.
The video processing method provided by the embodiment of the present application may be applied to an electronic device, which may be a personal computer or a mobile terminal, and is not limited specifically here.
In combination with some application scenarios, when a video application or a social application is running in the electronic device, videos sent by a server or other types of terminal devices may be received through the applications. The target video received in step 101 may be a video received by the electronic device from another terminal device.
For example, in a social application, there may be chat groups. The video sent by a user a in a chat group in the chat group is received by other users in the chat group. In this case, the user a can consider the target chat object as described above.
The user B may be another user in the chat group, and after receiving the video sent by the user a through the electronic device, the user B may have difficulty in directly knowing whether the video has content of interest, such as people or articles of interest. Generally, the user B needs to click to play the video when entering the corresponding session interface, so as to be able to confirm whether the video includes the content of interest.
In some scenarios, user a may have transmitted multiple videos, only some of which may include content of interest to user B. From the perspective of the user B, after the multiple videos are received through the electronic device, the multiple videos may be clicked one by one to be played after entering the session interface, so as to determine the video including the content of interest of the user B.
Therefore, the user B has a problem of low acquisition efficiency in the process of acquiring the video of the content of interest of the user B.
Of course, in practical applications, the target chat object for sending the target video may be a robot service, or the like, besides the user. For example, the robot service may send a video to the electronic device according to a keyword sent by a user of the electronic device. Alternatively, the robot customer service may periodically send a video or the like to the electronic device according to a history viewing record of a user of the electronic device or a preset interest tag.
Similarly, in the case where the robot service sends multiple videos, the user may need to click and play the multiple videos one by one, and determine the video including the content of interest.
In conjunction with the above example of the application scenario, in step 102, the preset second object may correspond to content in which a user of the electronic device is interested.
Specifically, in this embodiment, the second object may be an object preset by a user, or may also be an object automatically determined by the electronic device according to historical operation data of the user, and the like. In terms of content, the second object may be a person, an animal, an article, or the like, and is not particularly limited herein.
In connection with an example, the second object may be a target person, and when the second object is set, the user may select an image or a video obtained by shooting the target person as a material to be added to the material library. And the electronic device may identify a second object in the material based on a machine learning algorithm.
Of course, in practical applications, the second object may also be an animal, such as a dog, a cat, or an item, such as a house, a vehicle, etc. When the second objects are set, images or videos obtained by shooting the second objects may be selected as the identification material.
Similarly to the second object, the first object, such as a person or an object, is also generally included in the target video. The first object in the target video can also be obtained by identifying the target video through a machine learning algorithm.
For simplicity of description, the video processing method provided by the embodiment of the present application is mainly described below by taking the first object and the second object as characters as examples.
In combination with the above example of the application scenario of the chat group, the user B may set itself as the second object. From the perspective of the electronic device, after receiving a selection input of the user B for an image including the face of the user B while displaying the related setting interface, the image selected by the user B may be recognized, and the second object may be determined.
After receiving the target video, the electronic device used by the user B may identify the image frames in the target video to obtain the first object. The first object may also be a human face. When it is determined that the first object and the second object indicate the same person according to the face feature comparison, the first object in the target video may be considered to be matched with a preset second object.
Of course, in practical applications, the features selected for determining whether the first object and the second object are matched may be not only human face features, but also body features, and the like, and are not limited herein.
The identification process of the image frames in the target video by the electronic device can be performed in the background. For example, when a session interface is displayed in the electronic device, the target video received at the session interface may be cached, and the target video may be played in the background to identify the image frames in the target video, so as to obtain the first object.
In some possible embodiments, the electronic device may also buffer the target video and identify the image frame when the target video is received through the social application. That is, in the case where the conversation interface of the chat group is not displayed, the first object in the target video may still be acquired.
Of course, the electronic device may also acquire the first object in the target video when the user clicks to play the target video.
In other words, the electronic device does not specifically limit the timing of acquiring the first object in the target video.
In this embodiment, in the case that the first object in the target video matches a preset second object, a first identifier may be displayed, where the first identifier is used to indicate that the target video is a video matching the second object.
For example, in the case where a conversation interface is displayed in the electronic device, there may typically be thumbnails of the received videos in the conversation interface. When one of the videos, i.e., the first object in the target video, is matched with the second object, a first identifier may be displayed at a position corresponding to the thumbnail of the target video.
The first identification may be a text identification, for example, the content of the first identification may be "@ user B", "video that you may be interested in", or the like. Of course, the first identifier may also be a picture identifier or an icon identifier, and the like, which is not limited herein.
In other examples, the first identifier may be a preset hover control, such as a hover ball or a hover icon. That is, in the case that a first object in the target video matches a preset second object, the electronic device may display a hover control in a preset region to prompt that a video including the second object has been received.
In general, in step 102, the user may be prompted by displaying the first identifier that a video including the second object has been received. Or, further, the specific display position of the thumbnail of the target video in the current display interface can be indicated through the first identifier.
The video processing method provided by the embodiment of the application receives a target video sent by a target chat object, and displays a first identifier under the condition that a first object in the target video is matched with a preset second object, wherein the first identifier is used for indicating that the target video comprises the second object. By matching the first object with the second object, the first identifier can be displayed on the target video comprising the second object, and the method and the device for displaying the first identifier are beneficial to a user to efficiently acquire the video with the content of interest.
Optionally, the target chat object is an object in the target chat group;
correspondingly, in step 101, after receiving the target video sent by the target chat object, the video processing method may further include:
acquiring a preset second object associated with the target chat group;
comparing the similarity of the first object and the second object;
in the case that the similarity of the first object to the second object is greater than the similarity threshold, it is determined that the first object matches the second object.
As shown by way of example in the above embodiment, the target chat object may be a video sent in a chat group. In other words, the target chat object may be an object in a chat group. To simplify the description, the target chat object may be considered an object in the target chat group, and the target chat object is a target video sent in the target chat group.
Again, in the above example of the application scenario of the chat group, there may be a plurality of contents of interest to the user B, and the user B may be interested in the person X and the person Y (both of which may correspond to the second object) at the same time. However, the video including person X and the video including person Y may be generally received in different chat groups.
In order to avoid matching a first object in a target video by using a person X and a person Y in the case of receiving the target video in one chat group, in this embodiment, an association relationship between the chat group and a second object may be established in advance.
In particular, the target video may be a video sent to the target chat group, and the target chat group may have a corresponding second object associated therewith.
For example, in the related setting interface, the user can associate person X with chat group HX and person Y with chat group HY by input on the electronic device.
When the chat group HX receives the target video as the target chat group described above, the electronic device may determine character X as the second object associated with the target chat group.
In the above example, the association relationship between the chat group and the second object can be regarded as the association relationship between the different chat group and the second object in one social application. In practical applications, there may also be an association established between the application and the second object.
In other words, in the case that the target video is received through the target chat group, the application program to which the target chat group belongs may be determined, and the second object associated with the target chat group may be determined according to the preset association relationship between the application program and the second object.
For example, in the setting process, a character X may be associated with the application AX, and a character Y may be associated with the application AY. When the target video is received through any chat group (corresponding to the target chat group) in the application AX, the character X may be determined as the second object associated with the target chat group.
In the case that the second object associated with the target chat group is obtained, the second object associated with the target chat group can be used for carrying out similarity comparison on the first object in the target video.
In combination with an example, the first object and the second object may both be characters, and the similarity between the first object and the second object may be obtained by comparing contents such as facial features of the first object and the second object.
Of course, the similarity may also be considered as a probability of identifying the first object as the second object. For example, the second object may be a preset animal "dog", and the comparison result of the similarity between the first object and the second object may be a probability that the first object is identified as "dog", which corresponds to the above-mentioned similarity to some extent.
When the similarity of the first object to the second object is greater than the similarity threshold, it may be determined that the first object matches the second object.
For example, the similarity threshold may be preset, and the specific value is 90%. When the similarity between the first object and the second object is greater than 90%, it may be determined that the first object matches the second object.
In some examples, the process of the electronic device performing similarity comparison on the first object in the target video using the second object not associated with the target chat group may be omitted.
Therefore, in this embodiment, in the case of receiving the target video sent to the target chat group, the second object associated with the target chat group may be obtained, and the similarity between the second object associated with the target chat group and the first object in the target video may be compared, so that consumption of computing resources caused by the similarity comparison process may be effectively reduced.
In some examples, the association relationship between the chat group and the second object may be specifically associated with the second object through group information of the chat group.
For example, the group information of the chat group may be a name of the chat group or a number corresponding to the chat group. During the setup process, a chat group name may be associated with the second object. In the case of receiving the transmission to the target chat group, the associated second object may be determined according to the chat group name of the target chat group.
The following exemplifies a display mode of the first marker.
As shown in fig. 2, fig. 2 is a diagram illustrating an example of a session interface of a target chat group. In the target chat group, thumbnails of multiple videos (labeled VP1 in the figure) may be displayed, and these videos may be sent by user a (corresponding target chat object) to the target chat group.
In general, based on the identification of a first object in these videos, a target video including a second object may be obtained.
As shown in fig. 3, fig. 3 is a diagram showing an example of displaying a first identifier for a target video including a second object in a conversation interface of a target chat group. In this example diagram, the first logo may be displayed on the thumbnail (labeled VP2 in the figure) side of the target video, and specifically displayed as the text of "@ user B" (labeled T1 in the figure).
Specifically, the first identifier may be displayed in the form of a pop-up window or a floating window, and preset prompt text may be displayed in the pop-up window or the floating window.
Based on the displayed first identification, the user can quickly know specific videos which include the content interested by the user, and does not need to click and play each video one by one, so that the operation time of the user is effectively saved, and the use experience of the user is improved.
Of course, in some embodiments, the indicia displayed in the target chat group for the target video including the second object may not be limited to the first indicia described above.
Fig. 4 is a diagram showing an example of displaying a second mark on a target video including a second object in a conversation interface of a target chat group, as shown in fig. 4. In this example figure, the second logo may be displayed on the thumbnail side of the target video, and specifically displayed as "90%", "50%" (labeled T2 in the figure). The words may be used to indicate the matching degree of the first object and the second object in the target video.
Specifically, in these embodiments, in step 101, after receiving the target video sent by the target chat object, the video processing method may further include:
acquiring the matching degree of a first object and a preset second object in a target video; the matching degree is used for indicating the similarity between the first object and the second object, or the matching degree is used for indicating the proportion of a target image frame in the target video, the target image frame is an image frame comprising the first object, and the similarity between the first object and the second object is greater than a similarity threshold value;
and displaying a second identifier, wherein the second identifier is used for indicating the matching degree of the first object and the second object in the target video.
In combination with the application scenario in which the first object and the second object are both characters, the similarity between the first object and the second object may be the similarity of human face features, and the like. Accordingly, the similarity may be obtained through a preset face recognition model.
Of course, in other application scenarios, the second object may also be a general-purpose animal or object. For example, the second object may be an animal dog, the similarity between the first object and the second object may also be a probability that the first object is an animal dog obtained by recognizing the image frames of the target video by the corresponding recognition model.
In some examples, the similarity may be directly used as a matching degree of the target video and the second object.
The matching degree between the first object in the target video and the preset second object can also be determined by the ratio of the target image frames in the target video. The target image frame may be an image frame including a first object, a similarity between the first object and a second object being greater than a similarity threshold.
For example, the second object may be person X. In general, if a person is included in an image frame and the similarity between the person and the person X is greater than the similarity threshold, the person X may be considered to be included in the image frame. On this basis, if the target video has 100 image frames, and 60 image frames of the target video include the person X, the 60 image frames correspond to the target image frame.
Of the 100 image frames of the target video, there are 60 target image frames, i.e., the ratio of the target image frames in the target video is 60%. At this time, the matching degree of the target video and the second object may be considered to be 60%.
Of course, in practical applications, the matching degree between the first object and the second object can be obtained by combining the above similarity and the ratio of the target image frame. For example, the similarity and the ratio of the image frames may be weighted and averaged to obtain the matching degree between the first object and the second object.
Referring to fig. 4, for a target video, a second logo may be displayed at a position matching its thumbnail. The second identification may include a percentage, which may indicate a degree of match of the first object with the second object.
Of course, in practical applications, the second identifier may also include a decimal value or a score value, and may indicate a matching degree between the first object and the second object.
In this embodiment, the matching degree between the first object and the second object may be determined according to the similarity between the first object and the second object, or according to the ratio of the target image frames in the target video, and the matching degree is prompted to the user by displaying a second identifier that may be used to indicate the matching degree, which is helpful for the user to quickly locate the video of interest, and provides a selection basis for the video that can be preferentially viewed for the user, thereby improving the user experience.
In some embodiments, after receiving the target video sent by the target chat object, there may be a process of displaying the first identifier and displaying the second identifier. The specific display timings of the two marks may be the same or different.
For example, in an example, in a case that a currently displayed interface of the electronic device is a main interface of a social application, if a target video is received and a first object in the target video matches a preset second object, a first identifier may be displayed on the main interface, and the first identifier may be located near an image identifier of a target chat group. In the case where a conversation interface of the target chat group is displayed in response to the user's input, the second identifier may be displayed in the conversation interface.
In another example, the first identifier and the second identifier may be displayed simultaneously for the target video in a conversation interface of the target chat group.
In summary, in the embodiment of the application, when the electronic device matches the target video including the content of interest of the user, the electronic device may display the first identifier for the target video, and/or display the second identifier for indicating the matching degree, so that the user can be effectively reminded, and the retrieval time of the user on the target video is reduced.
Of course, after the first mark or the second mark is displayed, the marks may not affect the user's view of other unmarked videos.
In some application scenarios, the electronic device may receive the target video without displaying the target chat group. Alternatively, when the target chat group is displayed on the electronic device, if the number of received target videos is large or the intervals between the received target videos are large, it is difficult to simultaneously display thumbnails of all the target videos.
In order to enable the electronic device to intuitively prompt the user with the target video matching the second object in the above application scenario, in an embodiment, in step 101, after receiving the target video sent by the target chat object, the video processing method may further include:
and in the case that the first object in the target video is matched with the preset second object, displaying a target window, wherein the target window comprises a thumbnail of the target video.
As shown in fig. 5, fig. 5 is an exemplary diagram of displaying a target window (marked as W in the figure) in an electronic device. Wherein the target window may be a floating window.
In one example, the electronic device may, upon receiving the target video, cache the target video, play the target video in the background, identify a first object in the target video, and match the first object with a second object. The above processes can be performed in the background, and in the case that the first object in the target video matches with the second object, the target window can be displayed.
In other words, the target window may be displayed on a desktop, may be displayed on an operation interface of an application program, may be displayed on a conversation interface of the target chat group, or the like.
The target window may include a thumbnail of the target video (labeled WP in the figure). Therefore, the electronic equipment can prompt the user to receive the video which is possibly interested in the electronic equipment in different display states, and the timeliness of the prompting process is improved.
In connection with the above embodiment, there may be a degree of match between the first object and the second object in the target video. Accordingly, in one embodiment, the thumbnail of the target video includes a third identifier indicating a degree of matching of the first object and the second object in the target video.
In other words, in the target window, a degree of matching between the first object and the second object in the target video may be indicated by the third identification. The determination method of the matching degree between the first object and the second object in the target video is described in the above embodiments, and details are not repeated here.
As shown in fig. 6, in one example, the third indicator (labeled WT in the figure) described above may be displayed in the upper left corner of the thumbnail of the target video, where the degree of matching, e.g., "90%", "50%" etc., may be specifically displayed.
Of course, the third identifier may also be located at other positions in the thumbnail, and may be set according to actual needs, so that the matching degree between the first object and the second object in the target video may be embodied.
In this embodiment, the thumbnail of the target video includes the third identifier, so as to help the user quickly determine the target video of interest from the target window and provide a reference for the user to select the target video to be preferentially played.
In practical applications, the number of received videos may be plural, and the number of target videos including the first object matching the second object may also be plural.
Accordingly, in one embodiment, in the case that there are a plurality of target videos, thumbnails of the respective target videos may be displayed in a target window in an ordered manner according to matching degrees of the first object and the second object in the respective target videos.
For example, thumbnails of the respective target videos may be sequentially displayed in order of increasing degree of matching between the first object and the second object in the respective target videos. Thus, the user can preferentially view the target videos including more interesting contents according to the order of the thumbnails of the respective target videos.
Of course, in order to enable the user to more intuitively obtain the matching degree between the first object and the second object in each target video, the thumbnails of each target video may respectively include corresponding third identifiers.
In the case where the target window includes a thumbnail of the target video, the electronic device may receive user input of the thumbnail to operate on the corresponding target video.
Specifically, in some embodiments, after displaying the target window, the video processing method further comprises at least one of:
under the condition that first input of a user on a thumbnail of a target video in a target window is received, responding to the first input, and playing the target video;
in the case that a second input of the thumbnail of the target video in the target window by the user is received, responding to the second input, and storing the target video;
under the condition that a third input of the thumbnail of the target video in the target window by the user is received, responding to the third input, and forwarding the target video;
in a case where a fourth input of the thumbnail of the target video in the target window by the user is received, the target video is deleted in response to the fourth input.
The above embodiments are illustrated below with reference to some specific application scenarios.
In some examples, the first input may correspond to a user click input of a thumbnail of the target video, and the electronic device may play the target video in response to the click input if the user click input of the thumbnail of the target video is received.
The second input may correspond to a double-click input by the user on a thumbnail of the target video, and accordingly, the electronic device may store the target video in a case where the double-click input is received.
The third input may correspond to a long press input of a thumbnail of the target video by the user, and accordingly, the electronic device may forward the target video in a case where the long press input is received. In order to facilitate the user to select which application to forward the target video specifically, after receiving the long press input, icons of applications available for forwarding the target video may be displayed.
The fourth input may correspond to a drag input of the target video along the preset path by the user. For example, the user may drag the thumbnail of the target video to an outer region of the target window, and the electronic device may delete the target video from the memory for the cache file in response to the drag output of the user, and clarify the thumbnail of the target video in the target window. As a possible implementation, after all thumbnails of the target videos in the target window are deleted, the target window may be hidden.
Of course, the above are only some examples of the processing of the target video by the electronic device in response to the user's input. In practical application, the association relationship between different input operations and the processing mode of the target video can be adjusted according to actual needs.
For example, when a long press input for a thumbnail of a target video is received, two operation controls with names of "store" and "forward" may be displayed, and when a click input for the operation control by a user is received, the target video is further processed correspondingly.
Therefore, by displaying the target window, the user does not need to go deep into a specific session interface of the application program to manage the target video, the operation path is shortened and simplified, and the operation efficiency is improved.
As an optional implementation manner, in the case that the first object in the target video matches a preset second object, displaying the target window may specifically include:
under the condition that a first object in the target video is matched with a preset second object, displaying a target operation control;
in a case where an input to the target operation control is received, a target window is displayed in response to the input.
In this embodiment, a target operation control may be displayed when a target video in which a user is interested is received. The target operation control can be a floating button, and the floating button can occupy a small display area to avoid excessive shielding of the current display content of the electronic device.
And in case of receiving an input to the target operation control, for example, in case of receiving a single click, double click or long press input to the target operation control, the target window may be further loaded and a thumbnail of the target video may be displayed in the target window.
Optionally, the first object in the target video is obtained by identifying the target video by using a preset identification model;
accordingly, after displaying the first identifier, the video processing method may further include:
under the condition that a fifth input to the target video is received, training the recognition model by taking the target video as a training sample;
wherein the fifth input is used to store or forward the target video.
As shown above, in the case that the first object in the target video is a person, the first object may be obtained by recognizing an image frame in the target video through a face recognition model. Similarly, in the case that the first object is an animal or an article, the first object may be identified from the target video by using an identification model trained by the relevant sample.
It is easy to understand that, in general, for a recognition model based on a machine learning algorithm, it is usually required to use a corresponding training sample to train the recognition model sufficiently, so that the recognition model has a better recognition effect. The quantity and quality of the training samples have great influence on the recognition effect of the recognition model obtained by training.
Specifically, in the recognition model used in the embodiment of the present application, the corresponding training sample may be an image or a video that is labeled. The quality of the training samples is generally affected by the accuracy of the labeling.
In this embodiment, the target video may be regarded as an unlabelled video to a certain extent, and after the target video is input into the recognition model, a corresponding recognition result may be obtained. The recognition result may indicate whether the first object in the target video matches the second object.
If the recognition result indicates that the first object and the second object in the target video are matched and the recognition result is a correct recognition result, the target video can be used as a positive sample to perform optimization training on the recognition model.
And if the identification result is correct, the operation condition of the target video can be reflected by the user. For example, in a case that a first identifier is displayed for a certain target video, if a forwarding or storing operation of the user on the target video is received, the user is considered to be interested in the target video indeed, that is, the recognition result of the recognition model on the target video can be considered to be correct.
In this embodiment, the fifth input to the target video may correspond to the input for storing or forwarding the target video, and the like, and this may not be specifically limited here.
In general, the fifth input may be considered a user's forward operation on the target video. In the case that the fifth input to the target video is received, the target video can be used as a positive sample to train the recognition model, so that the recognition accuracy of the recognition model can be improved.
Some arrangements of the second object will be exemplified below.
In an embodiment, in step 101, before receiving the target video, the video processing method may further include:
displaying a setting interface, wherein the setting interface comprises P file identifications, one file identification indicates one image file, and P is a positive integer;
and under the condition of receiving selection input of Q file identifications, identifying Q image files indicated by the Q file identifications to obtain a second object, wherein Q is a positive integer less than or equal to P.
Generally, a picture file or a video file of a type such as a picture or a video file may be stored in advance in an electronic device. For example, the image files may be obtained by shooting the second object with a camera, or may be pictures or videos including the second object sent by other terminals.
In the setting interface, P file identifiers may be displayed, and the P file identifiers may correspond to P video files.
Specifically, in the setting interface, thumbnails of a plurality of image files may be displayed in a matrix manner, and a user may drag the interface up and down to switch and display the thumbnails of the image files.
Alternatively, in the setting interface, the image identifier of one video file may be displayed separately, for example, a video screenshot, and the user may drag the interface left and right to switch and display the image identifier of the video file.
The thumbnail images or image identifiers may be regarded as file identifiers of video files. For these file identifications, the user may make a selection input.
For example, a user may click on a file identifier to select the file identifier; or, the user may drag the file identifier to a preset display area to select the file identifier.
Of course, in practical applications, the selection input of the file identifier may not be limited to the above-mentioned exemplary manner, and a specific input manner may be set according to actual needs, which is not described herein.
It is easy to understand that the selection operation of the file identifier may be considered as a selection operation of a video file associated with the file identifier to some extent. Accordingly, the electronic device can identify the selected image files, that is, the Q image files indicated by the Q file identifiers, to obtain the second object.
Therefore, in the embodiment, by displaying the setting interface including the P file identifiers, the user can conveniently select the image file for obtaining the second object, and the operation efficiency is improved.
In one example, the number of the second objects obtained by identifying the Q video files may be one or more.
For example, when the user selects a picture including the person X and a picture including the person Y in the process of selecting and inputting the file identifier, the second object obtained by recognition may include both the person X and the person Y.
The following describes the setting process of the second object in conjunction with some application scenarios.
As shown in fig. 7, in the case of displaying a desktop or displaying an operation interface of an application program in the electronic device, a hover ball (labeled as F in the figure) named "material library" may be displayed in the interface. Upon receiving input to the hover, a material library interface may be displayed in which existing material, such as pictures or videos, for obtaining the second object may be included. In the material library interface, the user can slide left and right to switch to view existing material.
As shown in fig. 8, fig. 8 is an exemplary diagram of the interface of the material library. In this example diagram, the materials library interface may also include a button control named "upload" (labeled B1 in the figure). Upon receiving an input for an "upload" button control, thumbnails of the pictures or videos in the album may be displayed in a preset display area (labeled R in the figure). The material library interface at this time may be regarded as the above-described setting interface.
As shown in fig. 8, a button control named "bind" (labeled B2 in the figure) may also be included in the materials library interface. After the user selects the thumbnails of the pictures or videos in the album, the user can further click the 'binding' button control, and the pictures or videos corresponding to the selected thumbnails are added into the material library.
The materials in the material library can also be used for identification to obtain the second object.
As shown in the above embodiments, the chat group and the second object may have a preset association relationship. Accordingly, the preset association relationship may be obtained by setting in a related setting interface.
Accordingly, as a variation of the previous embodiment, in this embodiment, before receiving the target video sent by the target chat object in step 101, the video processing method may further include:
displaying a setting interface, wherein the setting interface comprises P file identifications, one file identification indicates one image file, and P is a positive integer;
under the condition that selection input of a user for Q file identifiers is received, displaying at least one candidate chat group option, wherein Q is a positive integer less than or equal to P;
under the condition that selection input of a user on a target chat group option in at least one candidate chat group option is received, identifying second objects in Q image files indicated by Q file identifications;
and storing the second objects in the Q image files in association with the target chat group corresponding to the target chat group option.
In this embodiment, the specific content in the setting interface and the display mode thereof may be the same as those in the previous embodiment, and are not described herein again for simplifying the description.
After receiving the selection input of the user for the Q file identifiers, the electronic device may display at least one candidate chat group option, where the candidate chat group option may include group information of the candidate chat group.
For example, the candidate chat group may be an existing chat group in the social application. The candidate chat group may generally have group information such as a chat group name, a chat group image identifier, or a number corresponding to the chat group. The candidate chat group option may be specifically displayed as a chat group name, a chat group image identifier or number, and the like of the candidate chat group, which is not limited in this respect.
In case of receiving selection input of Q file identifiers, a popup window may be displayed, or a selection page may be switched to be displayed.
For example, as shown in fig. 9, when a pop-up window is displayed, chat group names of "X school class discussion group", "Y company conference group", and "Z family life group" waiting for selection of a chat group may be displayed on the pop-up window, or a chat group image identifier formed by a group member image thumbnail combination may be displayed in front of each chat group name. These chat group names or chat group image identifications may correspond to the candidate chat group options described above.
If the user needs to identify the video including the second object in the discussion group of class X school, the user can click and input the chat group name of the discussion group of class X school. At this time, the click input of the chat group name of the user "discussion group by class X" may be regarded as the selection input of the target chat group option.
Under the condition that selection input of a target chat group option is received, the electronic equipment can identify Q image files indicated by Q file identifications to obtain a second object in the Q image files. The specific manner of identifying the image file to obtain the second object has been described in detail in the above embodiments, and is not described herein again.
The electronic device may store the second object in the Q image files in association with the target chat group corresponding to the target chat group option.
For example, the electronic device identifies the Q image files selected by the user to obtain a second object, and associates and stores the second object with "discussion group of school X". In the subsequent use process, if the electronic equipment receives a target video in a chat group sent to the discussion group corresponding to the school X, the electronic equipment can identify a first object in the target video; according to the content stored in the association mode, a second object associated with the discussion group of the school X is determined; and displaying the first identifier in case the first object matches the second object.
In the embodiment, a setting interface is displayed, and at least one candidate chat group option is displayed under the condition that selection input of a user on Q file identifiers is received; and after receiving selection input of a target chat group option in the at least one candidate chat group option, identifying second objects in the Q image files indicated by the Q file identifications, and storing the second objects in association with the target chat group corresponding to the target chat group option. The embodiment is beneficial to setting the association relationship between the second object and the target chat group by the user relatively quickly.
In one example, the user may need to perform more input processes in switching from the interface shown in fig. 8 to the interface shown in fig. 9.
For example, upon receiving user input to the "bind" button control shown in fig. 8, the electronic device may toggle to display an application selection interface as shown in fig. 10.
In the application selection interface, icons of a plurality of application programs and application names thereof may be displayed. For example, the application names may be "application a", "application B", and "application C", etc.
When a user click input to "application B" is received, a function selection interface as shown in fig. 11 may be further displayed. For example, the function selection interface may include types of options such as "select group", "share to circle of friends", and "favorite".
And after receiving the user's option to "select group", may switch to the interface shown in fig. 9.
Based on the above operation processes, on one hand, the user can associate corresponding second objects to different groups in different application programs; on the other hand, different use requirements of the user on the selected image file can be met.
Defining a setting interface for displaying the P file identifiers as a first setting interface, and in order to facilitate a user to manage a set second object, in an embodiment, the first setting interface further includes a first operation control;
correspondingly, after the first setting interface is displayed, the video processing method may further include:
under the condition that a sixth input to the first operation control is received, responding to the sixth input, and displaying a second setting interface, wherein the second setting interface comprises group information of the target chat group, a file identifier of an image file associated with the target chat group and at least one second operation control;
and under the condition that a seventh input to any one second operation control is received, responding to the seventh input, and updating the association relation between the group information of the target chat group and the image file.
The following describes a specific implementation procedure of the present embodiment with reference to an example.
As shown in fig. 8, in the interface shown in fig. 8, the first operation control may correspond to a button control named "association" (marked as B3 in the figure). Upon receiving a click input to the "associate" button control, the interface shown in FIG. 12 may be toggled.
As shown in fig. 12, fig. 12 may be considered as an example diagram of the second setting interface. Group information of the target chat group can be displayed on the second setting interface.
In general, the target chat group may refer to a chat group associated with the second object. In conjunction with the above description of the embodiment, the association relationship between the target chat group and the second object may be embodied as the association relationship between the group information of the target chat group and the second object.
As shown in fig. 12, in the second setting interface, the group information of the target chat group may include a chat group name (denoted as TN in the figure) and a chat group image identifier (denoted as TP in the figure). Wherein, the chat group name can be "group 1", "group 2", and "group 3", etc.; the chat group image identifier may be a group icon of the chat group, or the like.
Generally, the second object can be identified from the selected image file. The target chat group is associated with the second object, and therefore, the target chat group is also associated with the image file. As shown in fig. 12, a file identifier (marked as RP in the drawing) of a video file associated with a target chat group can be displayed below the group information of the target chat group. For example, the file identifier may be a thumbnail of a video file.
In addition, at least one second operation control (marked as K in the figure) can be displayed in association with each file identifier. For example, as shown in fig. 12, the second operation control includes an operation control named "modify" and an operation control named "delete".
In connection with some examples of application scenarios, group 1 is associated with image file a. When receiving an input to the "modify" operation control displayed behind the file identifier of the image file a, thumbnails of P image files, or thumbnails of the materials in the above-mentioned material library, may be displayed for the user to select. When the user selects the thumbnail of the image file B, the association relationship between the image file a and the "group 1" may be cancelled, and the association relationship between the image file B and the "group 1" may be established.
When an input to the "delete" operation control displayed behind the file identifier of the video file a is received, the association relationship between the video file a and the "group 1" may be cancelled.
Therefore, by displaying the second setting interface, the modification requirement of the user on the association relationship between the group information of the existing target chat group and the image file can be met, and the operation convenience is improved.
The following describes a video processing method provided in the embodiments of the present application with reference to a specific application example.
As shown in fig. 13, in this specific application example, the video processing method may include:
step 1301, establishing an association relation between the materials used for matching the video and the social software group;
the user may upload material for video matching and specify the material to a specific social software group.
The video herein may refer to a video received through social software. The material for video matching may be a video or a picture selected by the user on the setting interface, and the selected video or picture may be uploaded to a preset material library.
In addition, these stories may establish specified relationships with specific social software groups. For example, in an actual application, by setting the specified relationship, it is possible to match person a to a video transmitted to one group, match person B to a video transmitted to another group, and the like.
Step 1302, the group receives the video resources and checks whether there is a material library resource match; if yes, go to step 1303; if not, ending;
generally, the video assets may include first objects such as people or objects, and similarly, the story bank assets may include second objects such as people or objects.
The checking whether there is a match of the material library resources may refer to whether the first object in the video resources matches the second object in the material library resources.
Step 1303, carrying out intelligent matching according to the uploaded materials, wherein the matched contents can remind a user;
that is, in the case that the first object in the target video matches with the preset second object, a first identifier is displayed, and the first identifier may be in the form of "@ user" or a matching degree value.
Step 1304, automatically caching the extracted video into a floating ball, and sequencing according to a certain rule;
in other words, in the case that the target video is determined to be the video matched with the second object, the target video may be cached and the thumbnail of the target video may be displayed in the hover ball, or a window may be loaded to display the thumbnail of the target video after receiving the input to the hover ball.
In addition, as shown above, there may be a matching degree relationship between the target video and the second object, and the thumbnails of the target videos may be displayed in order according to the matching degree relationship, so that the user may find the target video with a high interest degree more quickly.
Of course, in practical applications, the thumbnails of the target videos may be displayed in order according to a rule such as a time sequence of receiving the target videos.
Step 1305, judging whether forward operations such as saving and forwarding are performed on the video in the floating ball by the user, if so, executing step 1306; if not, go to step 1307;
the user inputs the thumbnail of the video in the floating ball, so that the electronic equipment can perform corresponding processing on the corresponding video resource.
For example, a user may long press a thumbnail of a video, and in response to the user's long press input, a control for saving the video and a control for forwarding the video may be displayed. After the input of the user to the controls is received, the video can be saved and forwarded correspondingly.
The video is stored and forwarded, so that the forward operation of the user can be considered, and the user can be considered to be interested in the video in the floating ball to a certain extent. Conversely, the user may also perform operations to delete the video, and these operations may correspond to the reverse operations of the user.
And 1306, responding to the forward operation, storing or forwarding the video and the like, and releasing the cache resources occupied by the video in the floating ball.
And 1307, discarding the video in response to the deletion operation, and releasing the cache resources occupied by the video in the floating ball.
In combination with the application scenes, based on the video processing method provided by the embodiment of the application, under the condition that the user uploads the material resources, the electronic equipment can automatically perform a video resource matching algorithm task in the background, and after the matching is completed, the matched video is designated in the group to remind the user, so that the user can select reference. In addition, the video obtained by matching can be cached in the floating ball queue, so that the user can conveniently and quickly operate, the user can quickly perform operations such as storage, forwarding and the like, the purpose of saving the time of the user is achieved, and the user can focus on the content concerned by the user. The presentation mode can not only inform the users with the highest matching relevance, but also inform the users of a matching priority level through the matching degree proportion, more finely give reference opinions to the users, and have more accurate and reference significance for the decisions of the users.
It should be noted that, in the video processing method provided in the embodiment of the present application, the execution subject may be a video processing apparatus, or a control module in the video processing apparatus for executing the video processing method. In the embodiment of the present application, a video processing apparatus executing a video processing method is taken as an example, and the video processing apparatus provided in the embodiment of the present application is described.
As shown in fig. 14, a video processing apparatus 1400 provided in an embodiment of the present application includes:
a receiving module 1401, configured to receive a target video sent by a target chat object;
the first display module 1402 is configured to display a first identifier if a first object in the target video matches a preset second object, where the first identifier is used to indicate that the target video includes the second object.
Optionally, the target chat object is an object in the target chat group;
accordingly, the video processing apparatus 1400 may further include:
the first acquisition module is used for acquiring a preset second object associated with the target chat group;
the comparison module is used for comparing the similarity of the first object with the similarity of the second object;
and the determining module is used for determining that the first object is matched with the second object under the condition that the similarity between the first object and the second object is greater than the similarity threshold value.
Optionally, the video processing apparatus 1400 may further include:
the second acquisition module is used for acquiring the matching degree of the first object in the target video and a preset second object; the matching degree is used for indicating the similarity between the first object and the second object, or the matching degree is used for indicating the proportion of a target image frame in the target video, the target image frame is an image frame comprising the first object, and the similarity between the first object and the second object is greater than a similarity threshold value;
and the second display module is used for displaying a second identifier, and the second identifier is used for indicating the matching degree of the first object and the second object in the target video.
Optionally, the video processing apparatus 1400 may further include:
and the third display module is used for displaying a target window under the condition that the first object in the target video is matched with a preset second object, wherein the target window comprises a thumbnail of the target video.
Optionally, the thumbnail of the target video includes a third identifier, and the third identifier is used to indicate a matching degree of the first object and the second object in the target video.
Optionally, the first object in the target video is obtained by identifying the target video by using a preset identification model;
accordingly, the video processing apparatus 1400 may further include at least one of:
the playing module is used for responding to a first input to play the target video under the condition that the first input of the thumbnail of the target video in the target window is received by a user;
the first storage module is used for responding to a second input to store the target video under the condition that the second input of the thumbnail of the target video in the target window by the user is received;
the forwarding module is used for responding to a third input and forwarding the target video under the condition that the third input of the user on the thumbnail of the target video in the target window is received;
and the deleting module is used for responding to the fourth input to delete the target video under the condition that the fourth input of the user to the thumbnail of the target video in the target window is received.
Optionally, the first object in the target video is obtained by identifying the target video by using a preset identification model;
accordingly, the video processing apparatus 1400 may further include:
the training module is used for responding to a fifth input under the condition that the fifth input of the user to the target video is received, taking the target video as a training sample and training the recognition model;
wherein the fifth input is used to store or forward the target video.
Optionally, the video processing apparatus 1400 may further include:
the fourth display module is used for displaying a setting interface, the setting interface comprises P file identifiers, one file identifier indicates one image file, and P is a positive integer;
the fifth display module is used for displaying at least one candidate chat group option under the condition that selection input of Q file identifiers by a user is received, wherein Q is a positive integer less than or equal to P;
the identification module is used for identifying second objects in the Q image files indicated by the Q file identifications under the condition that selection input of a user on a target chat group option in the at least one candidate chat group option is received;
and the second storage module is used for storing the second objects in the Q image files in association with the target chat group corresponding to the target chat group option.
The video processing device provided by the embodiment of the application receives a target video sent by a target chat object, and displays a first identifier for indicating that the target video comprises a second object under the condition of a first object and a preset second object in the target video, so that a user can quickly find an interested video. In addition, in the case of the first object in the target video and the preset second object, a second identifier indicating the matching degree between the first object and the second object in the target video may be further displayed, so as to provide a basis for the user to select the target video to be preferentially viewed. Through the target window, the thumbnail of the target video is displayed, so that the user can conveniently and quickly operate the target video, and the use experience of the user is effectively improved.
The video processing apparatus in the embodiment of the present application may be an apparatus, or may be a component, an integrated circuit, or a chip in a terminal. The device can be mobile electronic equipment or non-mobile electronic equipment. By way of example, the mobile electronic device may be a mobile phone, a tablet computer, a notebook computer, a palm top computer, a vehicle-mounted electronic device, a wearable device, an ultra-mobile personal computer (UMPC), a netbook or a Personal Digital Assistant (PDA), and the like, and the non-mobile electronic device may be a server, a Network Attached Storage (NAS), a Personal Computer (PC), a Television (TV), a teller machine or a self-service machine, and the like, and the embodiments of the present application are not particularly limited.
The video processing apparatus in the embodiment of the present application may be an apparatus having an operating system. The operating system may be an Android operating system (Android), an iOS operating system, or other possible operating systems, which is not specifically limited in the embodiments of the present application.
The video processing apparatus provided in the embodiment of the present application can implement each process implemented by the method embodiments in fig. 1 to fig. 13, and is not described herein again to avoid repetition.
Optionally, as shown in fig. 15, an electronic device 1500 according to an embodiment of the present application is further provided, and includes a processor 1501, a memory 1502, and a program or an instruction stored in the memory 1502 and executable on the processor 1501, where the program or the instruction is executed by the processor 1501 to implement the processes of the video processing method embodiment, and can achieve the same technical effects, and details are not repeated here to avoid repetition.
It should be noted that the electronic devices in the embodiments of the present application include the mobile electronic device and the non-mobile electronic device described above.
Fig. 16 is a schematic hardware structure diagram of an electronic device implementing an embodiment of the present application.
The electronic device 1600 includes, but is not limited to: radio frequency unit 1601, network module 1602, audio output unit 1603, input unit 1604, sensor 1605, display unit 1606, user input unit 1607, interface unit 1608, memory 1609, and processor 1610.
Those skilled in the art will appreciate that the electronic device 1600 may further include a power supply (e.g., a battery) for supplying power to various components, which may be logically coupled to the processor 1610 via a power management system, so as to manage charging, discharging, and power consumption management functions via the power management system. The electronic device structure shown in fig. 16 does not constitute a limitation of the electronic device, and the electronic device may include more or less components than those shown, or combine some components, or arrange different components, and thus, the description thereof is omitted.
The radio frequency unit 1601 is configured to receive a target video sent by a target chat object;
and a display unit 1606, configured to display a first identifier when the first object in the target video matches a preset second object, where the first identifier is used to indicate that the target video includes the second object.
The electronic device provided by the embodiment of the application receives a target video sent by a target chat object, and displays a first identifier which can be used for indicating that the target video comprises a second object under the condition that a first object in the target video is matched with a preset second object. According to the embodiment of the application, the second object can be interesting content preset by a user, and the first identification is displayed on the target video matched with the second object, so that the user can be helped to efficiently acquire the video with the interesting content.
Optionally, the target chat object is an object in the target chat group;
accordingly, the processor 1610 is configured to obtain a preset second object associated with the target chat group;
comparing the similarity of the first object and the second object;
in the case that the similarity of the first object to the second object is greater than the similarity threshold, it is determined that the first object matches the second object.
Optionally, the processor 1610 may be configured to obtain a matching degree between a first object in the target video and a preset second object; the matching degree is used for indicating the similarity between the first object and the second object, or the matching degree is used for indicating the proportion of a target image frame in the target video, the target image frame is an image frame comprising the first object, and the similarity between the first object and the second object is greater than a similarity threshold value;
the display unit 1606 may be configured to display a second identifier, where the second identifier is used to indicate a matching degree between the first object and the second object in the target video.
Optionally, the display unit 1606 may be further configured to display a target window in a case where the first object in the target video matches a preset second object, where the target window includes a thumbnail of the target video. .
Optionally, the thumbnail of the target video includes a third identifier, and the third identifier is used to indicate a matching degree of the first object and the second object in the target video.
Optionally, processor 1610 may be configured to at least one of:
under the condition that first input of a user on a thumbnail of a target video in a target window is received, responding to the first input, and playing the target video;
in the case that a second input of the thumbnail of the target video in the target window by the user is received, responding to the second input, and storing the target video;
under the condition that a third input of the thumbnail of the target video in the target window by the user is received, responding to the third input, and forwarding the target video;
in a case where a fourth input of the thumbnail of the target video in the target window by the user is received, the target video is deleted in response to the fourth input.
Optionally, the first object in the target video is obtained by identifying the target video by using a preset identification model;
accordingly, the processor 1610 is configured to, in a case where a fifth input of the target video by the user is received, train the recognition model by using the target video as a training sample in response to the fifth input;
wherein the fifth input is used to store or forward the target video.
Optionally, the display unit 1606 is further configured to display a setting interface, where the setting interface includes P file identifiers, where one file identifier indicates one image file, and P is a positive integer;
under the condition that selection input of a user for Q file identifiers is received, displaying at least one candidate chat group option, wherein Q is a positive integer less than or equal to P;
accordingly, processor 1610 is configured to, in a case where a user selection input of a target chat group option of the at least one candidate chat group option is received, identify a second object of the Q video files indicated by the Q file identifications;
and storing the second objects in the Q image files in association with the target chat group corresponding to the target chat group option.
It should be understood that in the embodiment of the present application, the input Unit 1604 may include a Graphics Processing Unit (GPU) 16041 and a microphone 16042, and the Graphics processor 16041 processes image data of still pictures or videos obtained by an image capturing device (such as a camera) in a video capturing mode or an image capturing mode. The display unit 1606 may include a display panel 16061, and the display panel 16061 may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like. The user input unit 1607 includes a touch panel 16071 and other input devices 16072. Touch panel 16071, also referred to as a touch screen. The touch panel 16071 may include two parts of a touch detection device and a touch controller. Other input devices 16072 may include, but are not limited to, a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, a mouse, and a joystick, which are not described in detail herein. The memory 1609 may be used to store software programs as well as various data including, but not limited to, application programs and an operating system. Processor 1610 may integrate an application processor, which primarily handles operating systems, user interfaces, applications, etc., and a modem processor, which primarily handles wireless communications. It is to be appreciated that the modem processor described above may not be integrated into processor 1610.
The embodiments of the present application further provide a readable storage medium, where a program or an instruction is stored on the readable storage medium, and when the program or the instruction is executed by a processor, the program or the instruction implements each process of the video processing method embodiment, and can achieve the same technical effect, and in order to avoid repetition, details are not repeated here.
The processor is the processor in the electronic device in the above embodiment. Readable storage media, including computer-readable storage media, such as Read-Only Memory (ROM), Random Access Memory (RAM), magnetic or optical disks, etc.
The embodiment of the present application further provides a chip, where the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is configured to execute a program or an instruction to implement each process of the video processing method embodiment, and the same technical effect can be achieved.
It should be understood that the chips mentioned in the embodiments of the present application may also be referred to as system-on-chip, system-on-chip or system-on-chip, etc.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element. Further, it should be noted that the scope of the methods and apparatus of the embodiments of the present application is not limited to performing the functions in the order illustrated or discussed, but may include performing the functions in a substantially simultaneous manner or in a reverse order based on the functions involved, e.g., the methods described may be performed in an order different than that described, and various steps may be added, omitted, or combined. In addition, features described with reference to certain examples may be combined in other examples.
Through the above description of the embodiments, those skilled in the art will clearly understand that the method of the above embodiments can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware, but in many cases, the former is a better implementation manner. Based on such understanding, the technical solutions of the present application may be embodied in the form of a computer software product, which is stored in a storage medium (such as ROM/RAM, magnetic disk, optical disk) and includes instructions for enabling a terminal (such as a mobile phone, a computer, a server, or a network device) to execute the method of the embodiments of the present application.
While the present embodiments have been described with reference to the accompanying drawings, it is to be understood that the invention is not limited to the precise embodiments described above, which are meant to be illustrative and not restrictive, and that various changes may be made therein by those skilled in the art without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (11)

1. A video processing method, comprising:
receiving a target video sent by a target chat object;
and displaying a first identifier under the condition that a first object in the target video is matched with a preset second object, wherein the first identifier is used for indicating that the target video comprises the second object.
2. The method of claim 1, wherein the target chat object is an object in a target chat group;
after receiving the target video sent by the target chat object, the method further comprises:
acquiring a preset second object associated with the target chat group;
comparing the similarity of the first object with the similarity of the second object;
determining that the first object matches the second object if the similarity of the first object to the second object is greater than a similarity threshold.
3. The method of claim 1, wherein after receiving the target video sent by the target chat object, the method further comprises:
acquiring the matching degree of a first object and a preset second object in the target video; wherein the matching degree is used for indicating the similarity between the first object and the second object, or the matching degree is used for indicating the proportion of a target image frame in the target video, the target image frame is an image frame comprising a first object, and the similarity between the first object and the second object is greater than a similarity threshold value;
displaying a second identifier, wherein the second identifier is used for indicating the matching degree of the first object and the second object in the target video.
4. The method of claim 1, wherein after receiving the target video sent by the target chat object, the method further comprises:
and displaying a target window under the condition that a first object in the target video is matched with a preset second object, wherein the target window comprises a thumbnail of the target video.
5. The method of claim 4, wherein the thumbnail of the target video comprises a third identifier indicating a matching degree of the first object and the second object in the target video.
6. The method of claim 4, wherein after the displaying the target window, the method further comprises at least one of:
in the case that a first input of a thumbnail of the target video in the target window by a user is received, responding to the first input, and playing the target video;
in the case that a second input of the thumbnail of the target video in the target window by the user is received, responding to the second input, and storing the target video;
in the case that a third input of the thumbnail of the target video in the target window by the user is received, responding to the third input, and forwarding the target video;
in a case where a fourth input of the thumbnail of the target video in the target window by the user is received, the target video is deleted in response to the fourth input.
7. The method according to claim 1, wherein the first object in the target video is obtained by using a preset recognition model to recognize the target video;
after the displaying the first identifier, the method further comprises:
under the condition that a fifth input of the user to the target video is received, responding to the fifth input, taking the target video as a training sample, and training the recognition model;
wherein the fifth input is used to store or forward the target video.
8. The method of claim 1, wherein prior to receiving the target video sent by the target chat object, the method further comprises:
displaying a setting interface, wherein the setting interface comprises P file identifications, one file identification indicates one image file, and P is a positive integer;
under the condition that selection input of a user on Q file identifiers is received, displaying at least one candidate chat group option, wherein Q is a positive integer less than or equal to P;
under the condition that selection input of a user on a target chat group option in the at least one candidate chat group option is received, identifying Q second objects in the Q image files indicated by the file identification;
and storing the second objects in the Q image files in association with the target chat group corresponding to the target chat group option.
9. A video processing apparatus, comprising:
the receiving module is used for receiving a target video sent by a target chat object;
the first display module is used for displaying a first identifier under the condition that a first object in the target video is matched with a preset second object, wherein the first identifier is used for indicating that the target video comprises the second object.
10. An electronic device comprising a processor, a memory, and a program or instructions stored on the memory and executable on the processor, the program or instructions when executed by the processor implementing the steps of the video processing method according to any one of claims 1-8.
11. A readable storage medium, on which a program or instructions are stored, which, when executed by a processor, carry out the steps of the video processing method according to any one of claims 1 to 8.
CN202110553846.8A 2021-05-20 2021-05-20 Video processing method and device and electronic equipment Active CN113315691B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202110553846.8A CN113315691B (en) 2021-05-20 2021-05-20 Video processing method and device and electronic equipment
PCT/CN2022/092837 WO2022242577A1 (en) 2021-05-20 2022-05-13 Video processing method and apparatus, and electronic device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110553846.8A CN113315691B (en) 2021-05-20 2021-05-20 Video processing method and device and electronic equipment

Publications (2)

Publication Number Publication Date
CN113315691A true CN113315691A (en) 2021-08-27
CN113315691B CN113315691B (en) 2023-02-24

Family

ID=77373804

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110553846.8A Active CN113315691B (en) 2021-05-20 2021-05-20 Video processing method and device and electronic equipment

Country Status (2)

Country Link
CN (1) CN113315691B (en)
WO (1) WO2022242577A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022242577A1 (en) * 2021-05-20 2022-11-24 维沃移动通信有限公司 Video processing method and apparatus, and electronic device

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103020173A (en) * 2012-11-27 2013-04-03 北京百度网讯科技有限公司 Video image information searching method and system for mobile terminal and mobile terminal
US20130101209A1 (en) * 2010-10-29 2013-04-25 Peking University Method and system for extraction and association of object of interest in video
CN105488489A (en) * 2015-12-17 2016-04-13 掌赢信息科技(上海)有限公司 Short video message transmitting method, electronic device and system
CN110378247A (en) * 2019-06-26 2019-10-25 腾讯科技(深圳)有限公司 Virtual objects recognition methods and device, storage medium and electronic device
CN110650367A (en) * 2019-08-30 2020-01-03 维沃移动通信有限公司 Video processing method, electronic device, and medium
CN111770386A (en) * 2020-05-29 2020-10-13 维沃移动通信有限公司 Video processing method, video processing device and electronic equipment
CN112084370A (en) * 2020-09-10 2020-12-15 维沃移动通信有限公司 Video processing method and device and electronic equipment
CN112203142A (en) * 2020-12-03 2021-01-08 浙江岩华文化科技有限公司 Video processing method and device, electronic device and storage medium

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10897490B2 (en) * 2015-08-17 2021-01-19 E-Plan, Inc. Systems and methods for augmenting electronic content
CN111061912A (en) * 2018-10-16 2020-04-24 华为技术有限公司 Method for processing video file and electronic equipment
CN111314759B (en) * 2020-03-02 2021-08-10 腾讯科技(深圳)有限公司 Video processing method and device, electronic equipment and storage medium
CN113315691B (en) * 2021-05-20 2023-02-24 维沃移动通信有限公司 Video processing method and device and electronic equipment

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130101209A1 (en) * 2010-10-29 2013-04-25 Peking University Method and system for extraction and association of object of interest in video
CN103020173A (en) * 2012-11-27 2013-04-03 北京百度网讯科技有限公司 Video image information searching method and system for mobile terminal and mobile terminal
CN105488489A (en) * 2015-12-17 2016-04-13 掌赢信息科技(上海)有限公司 Short video message transmitting method, electronic device and system
CN110378247A (en) * 2019-06-26 2019-10-25 腾讯科技(深圳)有限公司 Virtual objects recognition methods and device, storage medium and electronic device
CN110650367A (en) * 2019-08-30 2020-01-03 维沃移动通信有限公司 Video processing method, electronic device, and medium
CN111770386A (en) * 2020-05-29 2020-10-13 维沃移动通信有限公司 Video processing method, video processing device and electronic equipment
CN112084370A (en) * 2020-09-10 2020-12-15 维沃移动通信有限公司 Video processing method and device and electronic equipment
CN112203142A (en) * 2020-12-03 2021-01-08 浙江岩华文化科技有限公司 Video processing method and device, electronic device and storage medium

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022242577A1 (en) * 2021-05-20 2022-11-24 维沃移动通信有限公司 Video processing method and apparatus, and electronic device

Also Published As

Publication number Publication date
CN113315691B (en) 2023-02-24
WO2022242577A1 (en) 2022-11-24

Similar Documents

Publication Publication Date Title
CN110020411B (en) Image-text content generation method and equipment
US20110016150A1 (en) System and method for tagging multiple digital images
US20150339348A1 (en) Search method and device
CN109218750B (en) Video content retrieval method, device, storage medium and terminal equipment
CN105009113A (en) Queryless search based on context
US11734370B2 (en) Method for searching and device thereof
WO2016000536A1 (en) Method for activating application program, user terminal and server
CN110691028B (en) Message processing method, device, terminal and storage medium
CN112954046B (en) Information transmission method, information transmission device and electronic equipment
WO2021254251A1 (en) Input display method and apparatus, and electronic device
CN113518026A (en) Message processing method and device and electronic equipment
CN111857460A (en) Split screen processing method, split screen processing device, electronic equipment and readable storage medium
CN112181253A (en) Information display method and device and electronic equipment
EP3486796A1 (en) Search method and device
CN113315691B (en) Video processing method and device and electronic equipment
CN113037925B (en) Information processing method, information processing apparatus, electronic device, and readable storage medium
CN113869063A (en) Data recommendation method and device, electronic equipment and storage medium
CN113590008A (en) Chat message display method and device and electronic equipment
CN113099033A (en) Information sending method, information sending device and electronic equipment
WO2023131092A1 (en) Information display method and apparatus
CN111324262A (en) Application interface control method, device, terminal and medium
CN114416664A (en) Information display method, information display device, electronic apparatus, and readable storage medium
CN113779293A (en) Image downloading method, device, electronic equipment and medium
CN113010072A (en) Searching method and device, electronic equipment and readable storage medium
CN113536037A (en) Video-based information query method, device, equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant