CN113760158A - Target object display method, object association method, device, medium and equipment - Google Patents

Target object display method, object association method, device, medium and equipment Download PDF

Info

Publication number
CN113760158A
CN113760158A CN202110484569.XA CN202110484569A CN113760158A CN 113760158 A CN113760158 A CN 113760158A CN 202110484569 A CN202110484569 A CN 202110484569A CN 113760158 A CN113760158 A CN 113760158A
Authority
CN
China
Prior art keywords
target
user
video
target object
playing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110484569.XA
Other languages
Chinese (zh)
Inventor
陈姿
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tencent Technology Shenzhen Co Ltd
Original Assignee
Tencent Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tencent Technology Shenzhen Co Ltd filed Critical Tencent Technology Shenzhen Co Ltd
Priority to CN202110484569.XA priority Critical patent/CN113760158A/en
Publication of CN113760158A publication Critical patent/CN113760158A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/451Execution arrangements for user interfaces
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/02Marketing; Price estimation or determination; Fundraising
    • G06Q30/0241Advertisements

Abstract

The application provides a target object display method, an object association method, a target object display device, an object association device, a computer-readable storage medium and an electronic device; relates to the technical field of information interaction; the method comprises the following steps: displaying a first interaction control in a playing interface for playing a target video, wherein the first interaction control displays relevant information of a first target object, and the first target object corresponds to a first target entity in the target video; and in response to the first operation of the first interaction control, displaying an object display interface of the first target object. The method and the device can provide richer and more effective display contents for the user, and improve the user experience.

Description

Target object display method, object association method, device, medium and equipment
Technical Field
The present application relates to the field of information interaction, and in particular, to a target object display method, an object association method, a target object display apparatus, an object association apparatus, a computer-readable storage medium, and an electronic device.
Background
In order to promote goods, services, and the like, an advertiser, a video service provider, or the like inserts an advertisement at the beginning, middle, or end of a video, and when a user views the video, the advertisement is played from the inserted position, thereby achieving the effect of advertising. However, the user experience brought by this method is not good, for example, the inserted advertisement content cannot meet the user requirement, the video playing is easily interrupted many times when the advertisement is inserted much, and the like.
It is to be noted that the information disclosed in the above background section is only for enhancement of understanding of the background of the present application and therefore may include information that does not constitute prior art known to a person of ordinary skill in the art.
Disclosure of Invention
The application aims to provide a target object display method, an object association method, a target object display device, an object association device, a computer-readable storage medium and electronic equipment, which can provide richer and more effective display contents for a user and improve user experience.
Other features and advantages of the present application will be apparent from the following detailed description, or may be learned by practice of the application.
According to an aspect of the present application, there is provided a target object display method, including:
displaying a first interaction control in a playing interface for playing a target video, wherein the first interaction control displays relevant information of a first target object, and the first target object corresponds to a first target entity in the target video;
and in response to the first operation of the first interaction control, displaying an object display interface of the first target object.
According to an aspect of the present application, there is provided an object association method, including:
carrying out entity identification on a target video to identify a first target entity;
and establishing an incidence relation between the first target entity and the first target object so that the terminal equipment displays a first interaction control in a playing interface for playing the target video, wherein the first interaction control displays relevant information of the first target object.
According to an aspect of the present application, there is provided a target object displaying apparatus, including: video playback unit and object display unit, wherein:
the video playing unit is used for displaying a first interaction control in a playing interface for playing the target video, wherein the first interaction control displays relevant information of a first target object, and the first target object corresponds to a first target entity in the target video;
and the object display unit is used for responding to the first operation of the first interaction control and displaying an object display interface of the first target object.
In an exemplary embodiment of the present application, a matching degree between the first target object and the user data of the current user satisfies a preset condition.
In an exemplary embodiment of the present application, the apparatus further includes:
and the operation response unit is used for responding to the second operation on the first interactive control and closing the first interactive control.
In an exemplary embodiment of the present application, a video playing unit displays a first interaction control in a playing interface for playing a target video, including:
when at least one of the following conditions is met, showing a first interactive control in a playing interface of the playing target video:
the display frequency of the first interaction control does not exceed the preset frequency;
the display times of the first interaction control do not exceed a first preset threshold;
the number of second operations on the first interaction control does not exceed a second preset threshold.
In an exemplary embodiment of the present application, the apparatus further includes:
the control display unit is used for highlighting and displaying a second target entity in the target video in the playing interface;
the operation response unit is further configured to, in response to a third operation for the second target entity, display a second interaction control corresponding to the second target entity or an object display interface of the second target object in the play interface, where the second interaction control displays related information of the second target object, and the second target object corresponds to the second target entity.
In an exemplary embodiment of the present application, the apparatus further includes:
the system comprises a tag acquisition unit, a tag analysis unit and a tag analysis unit, wherein the tag acquisition unit is used for performing tagging processing on user preference according to user data of a current user to obtain a user tag used for representing the user preference;
and the matching degree determining unit is used for determining the matching degree between the label of the first target object and the user label as the matching degree between the first target object and the user data of the current user.
In an exemplary embodiment of the present application, the user data includes: at least one of user representation data, user purchase records, user click records, and user browsing records.
In an exemplary embodiment of the present application, the apparatus further includes:
the tag acquisition unit is used for acquiring a browsing object tag corresponding to a browsing record of a user, a click object tag corresponding to a user click record, a purchase object tag corresponding to a user purchase record and a multi-dimensional tag used for describing the user in user portrait data before the video frame selection unit determines the matching degree between the target object and the user data of the current user;
and the data generation unit is used for generating the user data of the current user according to the browsing object tag, the clicking object tag, the purchasing object tag and the multi-dimensional tag.
In an exemplary embodiment of the present application, an object presentation unit presents an object presentation interface of a first target object, including:
and pausing the playing of the target video on the playing interface and jumping to the object display interface.
In an exemplary embodiment of the application, the operation response unit is further configured to close the object presentation interface and continue to play the target video in response to a closing operation for the object presentation interface.
In an exemplary embodiment of the present application, the object showing interface is further configured to show at least one associated object related to the first target object, and a similarity between the associated object and the target object is higher than a preset similarity.
In an exemplary embodiment of the present application, in the object presentation interface, the at least one associated object is arranged based on a degree of correlation with the user data.
In an exemplary embodiment of the present application, the related information of the first target object includes at least one of: the link of the first target object, the identification of the first target object, the graphic representation of the first target object and the description information of the first target object.
According to an aspect of the present application, there is provided an object associating apparatus, including: entity identification unit and incidence relation establishment unit, wherein:
the entity identification unit is used for carrying out entity identification on the target video and identifying a first target entity;
and the association relationship establishing unit is used for establishing an association relationship between the first target entity and the first target object so as to enable the terminal equipment to display a first interaction control in a playing interface for playing the target video, wherein the first interaction control displays relevant information of the first target object.
In an exemplary embodiment of the present application, the first target entity is included in at least one frame of a target video frame in the target video, and the first target object is an object in a preset object set.
In an exemplary embodiment of the application, when a target video picture set is played in a playing interface of a terminal device, a first interaction control is displayed, where the target video set includes at least one of: the video image comprises at least one frame of target video image, at least one frame of video image with preset frame number before the target video image, and at least one frame of video image with preset frame number after the target video image.
According to an aspect of the present application, there is provided an electronic device including: a processor; and a memory for storing executable instructions for the processor; wherein the processor is configured to perform the method of any of the above via execution of the executable instructions.
According to an aspect of the application, there is provided a computer-readable storage medium having stored thereon a computer program which, when executed by a processor, implements the method of any one of the above.
According to an aspect of the application, a computer program product or computer program is provided, comprising computer instructions stored in a computer readable storage medium. The processor of the computer device reads the computer instructions from the computer-readable storage medium, and the processor executes the computer instructions to cause the computer device to perform the method provided in the various alternative implementations described above.
One or more embodiments provided herein may have some or all of the following benefits:
in the target object display method provided in one or more embodiments, a first interaction control may be displayed in a playing interface for playing a target video, where the first interaction control displays relevant information of a first target object, and the first target object corresponds to a first target entity in the target video; and in response to the first operation of the first interaction control, displaying an object display interface of the first target object. The method can provide richer and more effective display contents for the user, and improves the user experience.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the application.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the present application and together with the description, serve to explain the principles of the application. It is obvious that the drawings in the following description are only some embodiments of the application, and that for a person skilled in the art, other drawings can be derived from them without inventive effort.
Fig. 1 is a schematic diagram illustrating an exemplary system architecture of a target object presentation method and a target object presentation apparatus to which the embodiments of the present application may be applied;
FIG. 2 illustrates a schematic structural diagram of a computer system suitable for use in implementing an electronic device of an embodiment of the present application;
FIG. 3 schematically shows a flow chart of a target object presentation method according to an embodiment of the present application;
FIG. 4 schematically illustrates a flow diagram for video frame level association of matching objects with a video file according to one embodiment of the present application;
FIG. 5 schematically illustrates a playback interface diagram according to an embodiment of the present application;
FIG. 6 schematically illustrates an object presentation interface diagram according to an embodiment of the present application;
FIG. 7 schematically shows a sequence diagram of a target object presentation method according to an embodiment of the present application;
FIG. 8 schematically illustrates a flow chart of a target object presentation method according to an embodiment of the present application;
FIG. 9 schematically illustrates a flow diagram of a target object presentation method according to an embodiment of the present application;
FIG. 10 schematically illustrates a flow diagram of an object association method according to an embodiment of the present application;
FIG. 11 schematically illustrates a block diagram of a target object presentation apparatus in an embodiment in accordance with the present application;
fig. 12 schematically shows a block diagram of an object associating apparatus in an embodiment according to the present application.
Detailed Description
Example embodiments will now be described more fully with reference to the accompanying drawings. Example embodiments may, however, be embodied in many different forms and should not be construed as limited to the examples set forth herein; rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the concept of example embodiments to those skilled in the art. The described features, structures, or characteristics may be combined in any suitable manner in one or more embodiments. In the following description, numerous specific details are provided to give a thorough understanding of embodiments of the application. One skilled in the relevant art will recognize, however, that the subject matter of the present application can be practiced without one or more of the specific details, or with other methods, components, devices, steps, and so forth. In other instances, well-known technical solutions have not been shown or described in detail to avoid obscuring aspects of the present application.
Furthermore, the drawings are merely schematic illustrations of the present application and are not necessarily drawn to scale. The same reference numerals in the drawings denote the same or similar parts, and thus their repetitive description will be omitted. Some of the block diagrams shown in the figures are functional entities and do not necessarily correspond to physically or logically separate entities. These functional entities may be implemented in the form of software, or in one or more hardware modules or integrated circuits, or in different networks and/or processor devices and/or microcontroller devices.
Fig. 1 is a schematic diagram illustrating a system architecture of an exemplary application environment to which a target object presentation method and a target object presentation apparatus provided in one or more embodiments of the present application may be applied.
As shown in fig. 1, system architecture 100 may include one or more of end devices 101, 102, 103, a network 104, and a server cluster 105. The network 104 serves to provide a medium of communication links between the terminal devices 101, 102, 103 and the server cluster 105. Network 104 may include various connection types, such as wired, wireless communication links, or fiber optic cables, to name a few. The terminal devices 101, 102, 103 may be various electronic devices having a display screen, including but not limited to desktop computers, portable computers, smart phones, in-vehicle devices, tablet computers, and the like. It should be understood that the number of terminal devices, networks, and servers in fig. 1 is merely illustrative. There may be any number of terminal devices, networks, and servers, as desired for implementation.
The target object display method provided by one or more embodiments of the present application may be executed by any server in the terminal devices 101, 102, and 103 or the server cluster 105. Accordingly, the target object exhibition means are typically arranged in the servers or terminal devices 101, 102, 103 of the server cluster 105. For example, in an exemplary embodiment, any server (or terminal devices 101, 102, 103) in the server cluster 105 may expose a first interaction control in a playing interface for playing a target video, where the first interaction control exposes information about a first target object, and the first target object corresponds to a first target entity in the target video; and in response to the first operation of the first interaction control, displaying an object display interface of the first target object.
FIG. 2 illustrates a schematic structural diagram of a computer system suitable for use in implementing an electronic device of one or more embodiments of the present application.
It should be noted that the computer system 200 of the electronic device shown in fig. 2 is only an example, and should not bring any limitation to the functions and the scope of use of the embodiments of the present application.
As shown in fig. 2, the computer system 200 includes a Central Processing Unit (CPU)201 that can perform various appropriate actions and processes in accordance with a program stored in a Read Only Memory (ROM)202 or a program loaded from a storage section 208 into a Random Access Memory (RAM) 203. In the RAM 203, various programs and data necessary for system operation are also stored. The CPU201, ROM 202, and RAM 203 are connected to each other via a bus 204. An input/output (I/O) interface 205 is also connected to bus 204.
The following components are connected to the I/O interface 205: an input portion 206 including a keyboard, a mouse, and the like; an output section 207 including a display such as a Cathode Ray Tube (CRT), a Liquid Crystal Display (LCD), and the like, and a speaker; a storage section 208 including a hard disk and the like; and a communication section 209 including a network interface card such as a LAN card, a modem, or the like. The communication section 209 performs communication processing via a network such as the internet. A drive 210 is also connected to the I/O interface 205 as needed. A removable medium 211, such as a magnetic disk, an optical disk, a magneto-optical disk, a semiconductor memory, or the like, is mounted on the drive 210 as necessary, so that a computer program read out therefrom is installed into the storage section 208 as necessary.
In one or more embodiments, the processes described below with reference to the flowcharts can be implemented as computer software programs. For example, a computer program product comprising a computer program embodied on a computer-readable medium, the computer program comprising program code for performing the method illustrated in the flow chart. In such an embodiment, the computer program may be downloaded and installed from a network through the communication section 209 and/or installed from the removable medium 211. The computer program, when executed by a Central Processing Unit (CPU)201, performs various functions defined in the methods and apparatus of the present application.
In one or more embodiments, artificial intelligence techniques and computer vision techniques are applied, particularly in video frame level associations for matching objects and video files. Among them, Artificial Intelligence (AI) is a theory, method, technique and application system that simulates, extends and expands human Intelligence using a digital computer or a machine controlled by a digital computer, senses the environment, acquires knowledge and uses the knowledge to obtain the best result. In other words, artificial intelligence is a comprehensive technique of computer science that attempts to understand the essence of intelligence and produce a new intelligent machine that can react in a manner similar to human intelligence. Artificial intelligence is the research of the design principle and the realization method of various intelligent machines, so that the machines have the functions of perception, reasoning and decision making.
The artificial intelligence technology is a comprehensive subject and relates to the field of extensive technology, namely the technology of a hardware level and the technology of a software level. The artificial intelligence infrastructure generally includes technologies such as sensors, dedicated artificial intelligence chips, cloud computing, distributed storage, big data processing technologies, operation/interaction systems, mechatronics, and the like. The artificial intelligence software technology mainly comprises a computer vision technology, a voice processing technology, a natural language processing technology, machine learning/deep learning and the like.
Computer Vision technology (CV) Computer Vision is a science for researching how to make a machine "see", and further refers to that a camera and a Computer are used to replace human eyes to perform machine Vision such as identification, tracking and measurement on a target, and further image processing is performed, so that the Computer processing becomes an image more suitable for human eyes to observe or transmitted to an instrument to detect. As a scientific discipline, computer vision research-related theories and techniques attempt to build artificial intelligence systems that can capture information from images or multidimensional data. Computer vision technologies generally include image processing, image recognition, image semantic understanding, image retrieval, OCR, video processing, video semantic understanding, video content/behavior recognition, three-dimensional object reconstruction, 3D technologies, virtual reality, augmented reality, synchronous positioning, map construction, and other technologies, and also include common biometric technologies such as face recognition and fingerprint recognition.
It is a common practice to advertise by inserting advertisements in videos, such as: advertisements are inserted at the beginning, middle, or end of the video. When the user watches the video, the advertisement is played at the designated progress. However, the user experience brought by this method is not good, for example, the inserted advertisement content cannot meet the user requirement, the video playing is easily interrupted many times when the advertisement insertion is more, and the like.
In one or more embodiments, a target object presentation method is provided. Referring to fig. 3, fig. 3 schematically shows a flowchart of a target object presentation method according to an embodiment of the present application. As shown in fig. 3, the target object presentation method may include: step S310 to step S320.
Step S310: and displaying a first interaction control in a playing interface for playing the target video, wherein the first interaction control displays relevant information of a first target object, and the first target object corresponds to a first target entity in the target video.
Step S320: and in response to the first operation of the first interaction control, displaying an object display interface of the first target object.
Step S310 to step S320 may be executed by a terminal device, and the terminal device may be a user device (e.g., a mobile phone, a tablet computer, a vehicle-mounted device, etc.).
In an exemplary embodiment of the present application, the related information of the first target object includes at least one of: the link of the first target object, the identification of the first target object, the graphic representation of the first target object and the description information of the first target object.
In one or more embodiments, by displaying the first interactive control corresponding to the first target entity in the target video, the object (e.g., advertisement) corresponding to the entity in the video is displayed, so that richer and more effective display content is provided for a user, and the user experience is improved.
In one or more embodiments, the object display interface of the first target object is displayed through the first operation of the user on the first interaction control, so that the frequency of interruption during video playing of the user can be reduced, and the user experience is improved.
In one or more embodiments, the play interface for playing the target video is triggered by a play operation, and the play operation may specifically be: click operation, touch operation, voice control operation, gesture operation, and the like, which are not limited in the embodiments of the present application. The playing interface is used for playing the video file frame by frame, the playing operation corresponds to the target video, and the playing operation can be an operation acting on the unique identifier of the target video.
In one or more embodiments, the first target object is an advertising item or the first target entity is an item in the video content.
In one or more embodiments, the first target object also corresponds to the current playing progress, and the first interaction control may be displayed in a video frame including the first target entity, or may be displayed in the first N frames or the last N frames of the video frame. Wherein N is a positive integer.
In one or more embodiments, further comprising: and closing the first interactive control in response to the second operation on the first interactive control. Therefore, the user can conveniently close the first interaction control in time when the first target object is not interested.
In one or more embodiments, after the playback operation is detected and before the target video is played in the playback interface, the method further includes: reading file information corresponding to the playing operation; requesting a target video corresponding to the file information from the server so that the server selects the target video from the video file library according to the file information; and receiving the target video fed back by the server.
In one or more embodiments, the file information is used to describe the target video.
Referring to fig. 4, fig. 4 schematically illustrates a flow diagram for video frame level association of matching objects with a video file in accordance with one or more embodiments of the present application. As shown in fig. 4, the method includes steps S410 to S440.
Step S410: the server identifies a video item category. Specifically, the server may perform entity identification at a video frame level on each video file in the video file library to determine a category of an entity (e.g., an article) in each video frame as an entity identification result, so as to add a tag to the entity in each video file according to the entity identification result.
Step S420: the server matches the video item with an advertisement library item based on the video item category. In particular, the server may match out from a collection of objects (e.g., an ad library) objects (e.g., ad library items) that are consistent with the entity tags in each video file.
Step S430: and the server sorts the matched advertisement library articles according to the similarity. Specifically, the server may perform similarity ranking on the matched advertisement library articles corresponding to the same entity, where the similarity is used to represent picture similarity/introduction information similarity between the advertisement library articles and the entity, so that ordered object display may be performed conveniently for the user.
Step S440: the server stores the sorting result. It should be noted that the server may store the sorting result locally or in the cloud.
In one or more embodiments, the target video may include one or more target video frames, each target video frame may include one or more target entities, each target entity may be associated with one or more first target objects, and the first target objects may be advertisement goods, term interpretations, and the like; wherein the target entity belongs to an actual object (e.g., a cell phone, earring, pen, etc.) or an actual text (e.g., a slogan, noun) appearing in the video frame. In addition, the candidate video frame is a video frame containing the target entity.
For example, the target video includes 3 target video frames, and the 3 target video frames are respectively an a frame, a B frame, and a C frame. That is, only these 3 frames in the target video have the target entity associated with the first target object. Specifically, frame a includes 1 target entity (e.g., math book), frame B includes 3 target entities (e.g., apple, banana, pineapple), and frame C includes 2 target entities (e.g., earrings, necklaces). The target entity "math book" in the a-frame may correspond to a plurality of first target objects (e.g., high-grade math books, first-grade math books, math out-of-class tutors, term interpretations on "math books", etc.). The target entity "apple" in the B frame may correspond to a plurality of first target objects (e.g., a brand apples, B place of origin apples, etc.).
In one or more embodiments, determining candidate video frames in the target video for which an association exists includes: and detecting whether each video frame in the target video has the associated tag, and screening out the video frames with the associated tags to serve as the target video frames.
In one or more embodiments, obtaining a first target object associated with a target entity in a candidate video frame comprises: sending an object request to a server, so that the server reads a data packet of a first target object according to object information in the object request and feeds the data packet back to the terminal equipment, and the terminal equipment displays the first target object according to the data packet; wherein the data packet may include a link to the first target object.
In one or more embodiments, presenting a first interaction control in a play interface that plays a target video includes: when at least one of the following conditions is met, showing a first interactive control in a playing interface of the playing target video: the display frequency of the first interaction control does not exceed a preset frequency (e.g., 10 times/1 day); the number of times of presentation of the first interaction control does not exceed a first preset threshold (e.g., 5 times); the number of second operations (e.g., close operations) on the first interactive control does not exceed a second preset threshold (e.g., 10).
In one or more embodiments, whether the first interaction control corresponding to the current playing progress is displayed in the playing interface corresponding to the current playing progress is determined based on the display frequency, the display times and the second operation (such as closing operation) times of the previous first interaction control.
In one or more embodiments, the presentation frequency of the first interaction control may be a frequency of the first interaction control being presented in the play interface, and the presentation number may be a total number of presentations of the first interaction control in a unit time (e.g., 1 day) in the play interface. In one or more embodiments, the second interactive control may be displayed in the video frame containing the second target entity, or may be displayed in the first N or last N frames of the video frame. Wherein N is a positive integer.
According to one or more embodiments provided by the application, the condition limit of the control popup can be realized, the user is prevented from being disturbed excessively, and the effect of personalized recommended objects is improved.
In one or more embodiments, before determining the degree of matching between the first target object and the user data of the current user, the method further includes: acquiring a browsing object label corresponding to a user browsing record, a click object label corresponding to a user click record, a purchase object label corresponding to a user purchase record and a multi-dimensional label for describing a user in user portrait data; and generating the user data of the current user according to the browsing object tag, the clicking object tag, the purchasing object tag and the multi-dimensional tag.
In one or more embodiments, the user browsing records include browsing records of the user for goods/web pages/news, etc., for example, the browsing object tags may include sunglasses, gloves, cell phones; the user click record comprises a click record of the user on the commodity, for example, the click object label can comprise sunglasses, gloves and mobile phones; the user purchase record includes all purchasing activities of the user including successful receipt and order placement and order cancellation, for example, the purchase object tag may include sunglasses. The user data of the current user may include a browse object tag, a click object tag, a purchase object tag, and a multi-dimensional tag; the browsing object tag, the clicking object tag, the purchasing object tag and the multidimensional tag may be tag sets including a plurality of tags, respectively.
One or more embodiments provided by the application can acquire data which can describe a user more accurately by collecting various user behaviors, thereby facilitating matching of objects more suitable for the user.
In one or more embodiments, the matching degree between the first target object and the user data of the current user satisfies a preset condition. The user data may describe the online shopping habits, browsing habits, and the like of the current user, and the user data may be represented as a tree structure.
According to one or more embodiments provided by the application, whether the matching degree between the user data (namely, the user portrait) and the first target object meets the preset condition or not can be detected, so that the interactive control display is carried out on the premise that the preset condition is met, personalized object recommendation can be realized, and excessive disturbance to the user is avoided.
In one or more embodiments, the user data for the current user includes: at least one of user portrait data, a user purchase record, a user click record and a user browsing record, optionally, the user data may further include a user historical viewing record, and the like, which is not limited in the embodiment of the present application. Where user representation data may be used to describe the current user, multidimensional labels used to describe the user such as: age, sex, native place, etc. In addition, the number of candidate video frames is greater than or equal to the number of target video frames.
In one or more embodiments, determining a degree of match between the first target object and the user data of the current user includes: performing labeling processing on user preference according to user data of a current user to obtain a user label for representing the user preference; and determining the matching degree between the label of the first target object and the user label as the matching degree between the first target object and the user data of the current user.
In one or more embodiments, the tagging of user preferences based on user data of a current user may enable standardization of user data. Determining the matching degree between the label of the first target object and the label of the user comprises the following steps: extracting the feature vector S1 of the tag of the first target object and the feature vector S2 of the user tag, and calculating the vector distance between S1 and S2 as the matching degree between the tag of the first target object and the user tag.
One or more embodiments provided herein facilitate determining a first target object that is suitable for a user through matching of tags.
In one or more embodiments, if the matching degree does not satisfy the preset condition, the method further includes: highlighting a second target entity in the target video in the playing interface; and responding to a third operation aiming at the second target entity, and displaying a second interaction control corresponding to the second target entity or an object display interface of a second target object in the playing interface, wherein the second interaction control displays related information of the second target object, and the second target object corresponds to the second target entity.
In one or more embodiments, highlighting in the playback interface a second target entity in the target video includes: highlighting a second target entity corresponding to the first target object in the playing interface/selecting the second target entity corresponding to the first target object in the playing interface so as to prompt the user that the second target entity can be interacted; the second target entity and the first target entity may be the same item in the video content or different items. In addition, displaying a second interactive control corresponding to a second target entity in the playing interface includes: popping up a second interaction control corresponding to a second target entity in a non-main screen area (such as a boundary area) of the playing interface, so that adverse effects of displaying the second interaction control on the user viewing experience can be reduced as much as possible.
According to one or more embodiments provided by the application, when the matching degree of the target entity and the current user is not high, the target entity in the video can be highlighted for the user, so that a prompt effect is provided for the user, and if the user wants to purchase an article in the video, the article can be clicked according to the prompt, so that the effect that the interactive control pops up is obtained.
In step S320, in response to a first operation on the first interaction control, an object presentation interface of the first target object is presented.
In one or more embodiments, one or more first target objects may be displayed in the object display interface, and if a plurality of first target objects exist, tags (e.g., earrings) of the plurality of first target objects are consistent with tags (e.g., earrings) of target entities in the target video frame. Furthermore, the object display interface is further configured to display at least one related object related to the first target object, where the related object (e.g., ear nail) is an object having a similarity higher than a preset similarity (e.g., 70%) with the first target object.
Referring to fig. 5, fig. 5 schematically illustrates a playback interface according to one or more embodiments of the present application. As shown in fig. 5, when a play operation is detected, the terminal device may play the target video in the play interface, and determine a candidate video frame (e.g., a video frame in the play interface shown in fig. 5) having an association relationship in the target video.
In one or more embodiments, a first target object corresponding to the target entity 510 and the target entity 520 in the target video frame may be requested from a server, and the server acquires and stores the first target object according to first target object information in the request and feeds the first target object back to the terminal device.
In one or more embodiments, when playing the video frame in the playing interface shown in fig. 5, the terminal device may highlight the entity 510 and the entity 520 in the video frame, where the entity 510 and the entity 520 respectively correspond to different first target objects (e.g., advertisement goods), a matching degree between the first target object corresponding to the entity 510 and the user data of the current user satisfies a preset condition, and then may display an interaction control 511 corresponding to the first target object in the playing interface, where the interaction control 511 is represented as "click to purchase similar goods" in fig. 5.
In one or more embodiments, if the matching degree between the first target object corresponding to the entity 520 and the user data of the current user does not satisfy the preset condition, the interaction control of the first target object corresponding to the entity 520 is not displayed. Therefore, the display efficiency can be improved, and the influence on the video watching experience of the user can be avoided.
FIG. 6 schematically illustrates an object presentation interface diagram in accordance with one or more embodiments of the present application. As shown in fig. 6, when a user operation on interaction control 511 is detected, a jump may be made from the play interface shown in fig. 5 to the object presentation interface shown in fig. 6 for presenting the first target object.
In one or more embodiments, the object representation interface may represent a plurality of target objects, the plurality of target objects including: target object 610, target object 620, target object 630, target object 640. Directly below each target object corresponds to interaction control 611 (denoted as "35 click buy" in fig. 6), interaction control 621 (denoted as "75 click buy" in fig. 6), interaction control 631 (denoted as "105 click buy" in fig. 6), and interaction control 641 (denoted as "95 click buy" in fig. 6), respectively. Where 35, 75, 105, 95 may be used to represent the price of the corresponding target object.
For example, in fig. 6, the target object is represented by an image, and optionally, in a display area of each target object, the target object may also be represented by a video/text, which is not limited in the embodiment of the present application.
In one or more embodiments, when the user clicks on interaction control 611, interaction control 621, interaction control 631, or interaction control 641, the user can jump to the corresponding item detail page so that the user can view the parameters, comments, etc. of the target object before purchasing to improve the user's shopping experience. Optionally, when the user clicks the interactive control 611, the interactive control 621, the interactive control 631, or the interactive control 641, the payment interface may also be directly jumped to receive the payment operation of the user, so as to implement the purchase of the target object.
In one or more embodiments, the method further comprises: the object display interface is further used for displaying at least one associated object related to the first target object, and the similarity between the associated object and the first target object is higher than the preset similarity. Wherein, in the object display interface, at least one associated object is arranged based on the relevance degree of the user data.
In one or more embodiments, an identifying image (e.g., a front view) may be selected from the images of the first target object, a similarity between the identifying image and images of other objects in the set of objects is calculated, and an object with a similarity higher than a preset similarity is determined as an associated object; wherein the similarity can be represented by a cosine distance or a euclidean distance.
One or more embodiments provided by the application can display a plurality of objects (such as advertisement commodities) on the object display interface, and after a user clicks the interactive control and interface jump is realized, the plurality of objects can be browsed, so that the object selection range of the user can be improved. Therefore, the content richness of the object display interface is improved, the purchasing range of the user is not limited to a certain commodity any more, and the use experience of the user is improved to a certain degree.
In one or more embodiments, after determining an object with a similarity higher than a preset similarity as the associated object, the method further includes: determining the correlation degree of each related object with the user data; the user data is used for describing the current user in a multi-dimension mode; and arranging at least one related object for displaying in the object display interface according to the relevance.
In one or more embodiments, the using of the user data to describe multiple dimensions in the current user in multiple dimensions specifically includes: browsing latitude, purchasing latitude, basic information latitude, clicking latitude and the like.
In one or more embodiments, the user data may be used to characterize the shopping habits of the user, and determine the relevance of each relevant object to the user data, including: determining a label (e.g., ear stud) for each related object; determining a label set in the user data (the label set comprises a browsing object label, a clicking object label, a purchasing object label and a multi-dimensional label); and determining whether the label of each related object hits the label set, judging that the related object hitting the label set has high correlation with the current user, and judging that the related object not hitting the label set has low correlation with the current user. Thereafter, the method further comprises the following steps: and screening out the related objects with low correlation degree to obtain at least one related object with high correlation degree. Based on the above, at least one related object for displaying in the object display interface is arranged according to the relevance, which comprises the following steps: and arranging at least one related object with high relevance from high to low according to the label frequency, and displaying the at least one related object with high relevance in an object display interface according to the arrangement sequence. It should be noted that the object display interface may be divided into two parts, where the display priority of the first part is higher than that of the second part, the first part is used for displaying the first target object, and the second part is used for displaying the related object. Therefore, after the user browses all the first target objects (namely, the objects with the highest relevance), the related objects (namely, the objects with the second relevance) can be browsed next, so that the browsing range of the user is expanded, and the use experience of the user is improved.
In one or more embodiments, after the user finishes browsing the related objects, the method may further include: selecting a recommended object according to the label set and displaying the recommended object in an object display interface; the browsing manner of the first target object, the related object and the recommended object by the user can be based on a sliding operation (e.g., an up/down/right/left sliding operation) to continuously change the objects displayed in the display object display interface. It should be noted that the first target object belongs to an object set, and the related object and the recommended object may or may not belong to the object set, which is not limited in the embodiment of the present application. Therefore, the problem of limitation of the object set can be avoided, and when the number of the objects in the object set is small, other objects in the non-set can be displayed to enrich the object display interface, so that the use experience of a user is improved.
One or more embodiments provided by the application can show related objects related to user data on the basis of showing the first target object, so as to enrich an object showing interface, and improve the use experience of a user.
In one or more embodiments, an object representation interface for representing a first target object includes: and pausing the playing of the target video on the playing interface.
In one or more embodiments, further comprising: when the pause playing time length is detected to be longer than the preset time length (for example, 1h), the video playing program in the background can be closed, so that the occupation of the cache resources is reduced.
One or more embodiments provided by the application can pause the video for the user when the user needs to browse the object, thereby improving the use experience of the user.
In one or more embodiments, after the playing of the target video is paused, the method further includes: and in response to the closing operation aiming at the object display interface, closing the object display interface and continuing to play the target video.
In one or more embodiments, continuing to play the target video includes: ending the pause playing state of the target video. In addition, a user operation for closing the object presentation interface may act on a close control in the object presentation interface.
One or more embodiments provided by the application can continue to play the video for the user after the user finishes browsing, so that the use experience of the user is improved.
Referring to fig. 7, fig. 7 schematically shows a sequence diagram of a target object presentation method according to one or more embodiments of the present application. As shown in fig. 7, the sequence diagram includes: step S700 to step S790.
Step S700: and the user behavior system label library extracts user viewing labels (such as beauty makeup) according to the user historical viewing records.
Step S710: and uploading the user film watching tag to a recommendation system by the user behavior system tag library.
Step S720: and uploading the user browsing records, the user clicking records and the user purchasing records to an article purchasing and clicking database by the user side.
Step S730: the item purchase and click database extracts the browse object tag, click object tag, purchase object tag, and multidimensional tag.
Step S740: the item purchase and click database uploads a browse object tag, a click object tag, a purchase object tag and a multi-dimensional tag to the recommendation system.
Step S750: and the video background binds the object set with the video file library at the video frame level.
Step S760: and uploading the binding result to a recommendation system by the video background.
Step S770: the user side plays the target video in the playing interface, determines the candidate video frames with the binding relationship in the target video, and requests the recommendation system for the target object bound with the target entity in the target video frames.
Step S780: the recommendation system feeds back the target object bound with the target entity in the target video frame.
Step S790: the user side determines the matching degree between the target object and the user data of the current user, if the matching degree meets a preset condition, the target video frame is selected from the candidate video frames, when the target video frame is played, the interactive control corresponding to the target entity is displayed in the playing interface, and then when the user operation acting on the interactive control is detected, the playing interface jumps to the object display interface used for displaying the target object.
It should be noted that the user terminal may be run in a terminal device (i.e., a user device), and the user behavior system tag library, the item purchase and click database, the video background, and the recommendation system may be run in a server.
In one or more embodiments, by displaying the first interactive control corresponding to the first target entity in the target video, the object (e.g., advertisement) corresponding to the entity in the video is displayed, so that richer and more effective display content is provided for a user, and the user experience is improved.
Referring to fig. 8, fig. 8 schematically illustrates a flow diagram of a target object presentation method according to one or more embodiments of the present application. As shown in fig. 8, the target object display method includes: step S810 to step S860. The steps S810 to S820 may be performed by the identification association module, and the steps S830 to S860 may be performed by the object recommendation module.
Step S810: and receiving the advertisement information input by the advertiser. The advertisement information is information of any object in the object set.
Step S820: and binding the matched advertisement and the video file at the video frame level.
Step S830: and playing the target video in the playing interface.
Step S840: a target advertisement (i.e., a target object) bound to a target entity in a target video frame is obtained.
Step S850: and determining the matching degree between the target object and the user data of the current user, if the matching degree meets a preset condition, selecting a target video frame from the candidate video frames, and displaying the interactive control corresponding to the target entity in a playing interface when the target video frame is played.
Step S860: and when the user operation acting on the interaction control is detected, jumping to an object display interface for displaying the target advertisement from the playing interface.
It should be noted that steps S810 to S860 correspond to the steps shown in fig. 3 and the embodiment thereof, and for the specific implementation of steps S810 to S860, please refer to the steps shown in fig. 3 and the embodiment thereof, which is not described herein again.
In one or more embodiments, by displaying the first interactive control corresponding to the first target entity in the target video, the object (e.g., advertisement) corresponding to the entity in the video is displayed, so that richer and more effective display content is provided for a user, and the user experience is improved.
Referring to fig. 9, fig. 9 schematically illustrates a flow diagram of a target object presentation method according to one or more embodiments of the present application. As shown in fig. 9, the method for displaying the retargeted object includes: step S900 to step S920.
Step S900: the method comprises the steps that a server matches an object set with a video file library, the object set comprises target objects, feature extraction is conducted on video frames of video files in the video file library, entities in the video frames of the video files are identified according to feature extraction results, labels are added to the entities in the video files according to entity identification results, the entity identification results are used for representing the entities contained in each frame of the corresponding video files, objects which are consistent with the entity labels in the video files are matched from the object set, and video frames to which the entities belong are bound with links corresponding to the objects on the premise that the labels are consistent.
Step S902: when the playing operation is detected, the terminal device reads file information corresponding to the playing operation, and requests the server for a target video corresponding to the file information, so that the server selects the target video from the video file library according to the file information, receives the target video fed back by the server, and plays the target video in a playing interface.
Step S904: the terminal equipment determines candidate video frames with binding relation in the target video and acquires target objects bound with target entities in the candidate video frames.
Step S906: the terminal equipment acquires a browsing object label corresponding to the user browsing record, a click object label corresponding to the user click record, a purchase object label corresponding to the user purchase record and a multi-dimensional label used for describing the user in the user portrait data, and generates user data of the current user according to the browsing object label, the click object label, the purchase object label and the multi-dimensional label.
Step S908: the terminal equipment carries out labeling processing on the user preference according to the user data of the current user to obtain a user label for representing the user preference, and determines the matching degree between the label of the target object and the user label as the matching degree between the target object and the user data of the current user. If the matching degree satisfies the predetermined condition, step S910 is executed. If the matching degree does not satisfy the preset condition, step S912 is executed.
Step S910: and the terminal equipment selects a target video frame from the candidate video frames and displays an interactive control corresponding to the target entity in a playing interface.
Step S912: and the terminal equipment displays a target entity corresponding to the target object in a display interface in a highlighted mode, the target entity belongs to the target video frame, and if the interactive operation acting on the target entity is detected, the interactive control corresponding to the target entity is displayed in the display interface.
Step S914: the terminal device calculates the similarity between the target object and other objects in the object set, determines at least one other object with the similarity larger than a preset similarity as at least one related object, further determines the correlation between each related object and the user data, and arranges at least one related object for displaying in the object display interface according to the correlation.
Step S916: and the terminal equipment detects whether the control pop-up frequency reaches a preset frequency. If so, the flow ends. If not, step S918 is performed.
Step S918: and when detecting the user operation acting on the interactive control, the terminal equipment jumps to an object display interface from the play interface and stops playing the target video, wherein the object display interface is used for displaying the target object and at least one related object.
Step S920: and if the user operation for closing the object display interface is detected, the terminal equipment closes the object display interface and continues to play the target video.
Wherein the user data includes: at least one of user representation data, user purchase records, user click records, and user browsing records.
It should be noted that steps S900 to S920 correspond to the steps shown in fig. 3 and the embodiment thereof, and for the specific implementation of steps S900 to S920, please refer to the steps shown in fig. 3 and the embodiment thereof, which is not described herein again.
In one or more embodiments, by displaying the first interactive control corresponding to the first target entity in the target video, the object (e.g., advertisement) corresponding to the entity in the video is displayed, so that richer and more effective display content is provided for a user, and the user experience is improved.
Further, in one or more embodiments, there is also provided an object association method, including: step S1010 to step S1020. Steps S1010 to S1020 may be executed by a server, a cloud device, or the like.
Step S1010: and carrying out entity identification on the target video to identify a first target entity.
Step S1020: and establishing an incidence relation between the first target entity and the first target object so that the terminal equipment displays a first interaction control in a playing interface for playing the target video, wherein the first interaction control displays relevant information of the first target object.
In one or more embodiments, the first target entity is included in at least one frame of a target video frame in the target video, and the first target object is an object in the preset object set.
In one or more embodiments, when a target video picture set is played in a playing interface of a terminal device, a first interaction control is displayed, where the target video set includes at least one of: the video image comprises at least one frame of target video image, at least one frame of video image with preset frame number before the target video image, and at least one frame of video image with preset frame number after the target video image.
In one or more embodiments, before the terminal device displays the first interaction control in the playing interface for playing the target video, the method may further include: and the server selects a target video from the video file library according to the file information. The specific execution mode is as follows: and the server judges the legality of the terminal equipment according to the request sent by the terminal equipment, and selects the target video from the video file library according to the file information when judging that the terminal equipment has the legality. Specifically, the method further comprises the following steps: the server can read the sender code in the request, detect whether the sender code is a sender code set which is hit to be stored, and if so, judge that the terminal equipment has validity.
In one or more embodiments, after the server selects the target video from the video file library according to the file information, the method may further include: the server encrypts the target video through a public key of the terminal device (i.e., the receiving party), and transmits the encrypted target video to the terminal device. Based on this, the target video fed back by the server is received, and the method comprises the following steps: and the terminal equipment receives the target video fed back by the server and decrypts the target video through a private key of the terminal equipment. Thus, the transmission confidentiality of the video file can be improved.
The one or more embodiments that this application provided can read corresponding video file and feedback according to file information to make things convenient for terminal equipment can play the video file of feedback, the terminal equipment side need not to store the video file, thereby can reduce the occupation to terminal equipment's storage space, promotes the utilization ratio to terminal equipment.
In one or more embodiments, if the first target object is an advertisement product, the set of objects may be advertisement products uploaded by each advertiser that need to be recommended to the user. One or more video files may be included in the video file library, including the target video.
In one or more embodiments, after establishing the association relationship between the first target entity and the first target object, the method may further include: the video file of the completed object association is stored, and the video file of the completed object association is fed back to the legal terminal device when the legal terminal device requests the video file, the terminal device can obtain the corresponding association relation of the video file, and the association relation is used for representing the association relation between the video frame and the object (advertisement commodity). Based on this, can also include: and the server shares the video file associated with the completed object to the cloud so that other video service providers can use the video file.
According to one or more embodiments provided by the application, a plurality of objects can be respectively corresponding to corresponding video frames, so that when a user watches videos, a control for guiding the user to click is output, corresponding objects are displayed, and the purpose of recommending the objects is achieved.
In one or more embodiments, establishing an associative relationship between a first target entity and a first target object includes: the server carries out entity identification of video frame level on each video file in the video file library; the server adds labels to the entities in the video files according to the entity identification result, and the entity identification result is used for representing the entities contained in each frame in the corresponding video files; matching an object consistent with the entity tag in each video file from the object set by the server; and the server associates the video frame to which the entity belongs with the link corresponding to the object on the premise of consistent label.
In one or more embodiments, performing entity identification on a target video, identifying a first target entity, includes: the server carries out framing processing on each video file in the video file library to obtain a video frame set of each video file; the server identifies the entity of each video frame set through a video identification algorithm; the video identification algorithm is based on machine learning and data mining theories, massive pictures and video samples are collected in an off-line mode, characteristic values are extracted after the samples are calibrated artificially, training is carried out based on the characteristic values and the sample calibration, and then an efficient characteristic selection classifier is designed. In particular, feature selection classifiers can be used to accomplish detection, tracking, identification of various entities.
Based on this, the server adds tags to the entities in each video file according to the entity identification result, including: the server adds a label to the entity according to the entity type (such as daily necessities and videos) to which the entity identification result belongs, wherein the label can be a single label or a multi-level label. For example, the first-level label is daily necessities, the second-level label is washing care products, and the third-level label is shampoo.
According to one or more embodiments provided by the application, corresponding tags can be added to the entities through entity identification of the video frames, so that one or more corresponding objects can be associated with the video frames according to the matching degree of the tags. In some applications, the cost of manually associating videos and objects may be reduced. In some application occasions, one object can be associated with a plurality of video frames based on entity identification and tag matching, so that the exposure rate of each object is favorably improved, and the conversion rate of the object as an advertisement commodity is favorably improved.
In one or more embodiments, performing entity identification on a target video, identifying a first target entity, includes: the server extracts the characteristics of the video frames of the video files in the video file library; and the server identifies entities in the video frames of the video files according to the feature extraction result, wherein each video file comprises a target video, and the target video comprises a first target entity.
In one or more embodiments, the server performs feature extraction on video frames of each video file in the video file library, including: and the server calculates the feature vector of the video frame of each video file in the video file library based on a feature extraction algorithm to serve as a feature extraction result. Based on this, the server identifies entities in the video frames of each video file according to the feature extraction result, including: the server inputs each feature vector into a classifier based on a classifier algorithm, so that the classifier outputs an entity recognition result.
One or more embodiments provided by the application can identify the entity in the video frame based on feature extraction, and improve the identification precision of the entity.
Further, in one or more embodiments, a target object display apparatus is also provided. Referring to fig. 11, the target object presentation apparatus 1100 may include: the video playback unit 1101 is an object presentation unit 1102, in which:
the video playing unit 1101 is configured to display a first interaction control in a playing interface for playing a target video, where the first interaction control displays relevant information of a first target object, and the first target object corresponds to a first target entity in the target video;
the object display unit 1102 is configured to display an object display interface of the first target object in response to a first operation on the first interaction control.
And the matching degree between the first target object and the user data of the current user meets a preset condition. In addition, the object display interface is further used for displaying at least one associated object related to the first target object, and the similarity between the associated object and the target object is higher than the preset similarity. In the object presentation interface, at least one associated object is arranged based on a degree of correlation with the user data.
In one or more embodiments, by displaying the first interactive control corresponding to the first target entity in the target video, the object (e.g., advertisement) corresponding to the entity in the video is displayed, so that richer and more effective display content is provided for a user, and the user experience is improved.
In one or more embodiments, the apparatus further comprises:
and an operation response unit (not shown) for responding to the second operation of the first interaction control and closing the first interaction control.
One or more embodiments provided by the application can facilitate a user to close the first interaction control in time when the user is not interested in the first target object.
In an exemplary embodiment of the present application, the video playing unit 1101 displays a first interaction control in a playing interface for playing a target video, including:
when at least one of the following conditions is met, showing a first interactive control in a playing interface of the playing target video:
the display frequency of the first interaction control does not exceed the preset frequency;
the display times of the first interaction control do not exceed a first preset threshold;
the number of second operations on the first interaction control does not exceed a second preset threshold.
According to one or more embodiments provided by the application, the condition limit of the control popup can be realized, the user is prevented from being disturbed excessively, and the effect of personalized recommended objects is improved.
In one or more embodiments, the apparatus further comprises:
a control display unit (not shown) for highlighting the second target entity in the target video in the playing interface;
an operation response unit (not shown) is configured to, in response to a third operation on the second target entity, display a second interaction control corresponding to the second target entity or an object display interface of the second target object in the play interface, where the second interaction control displays related information of the second target object, and the second target object corresponds to the second target entity.
According to one or more embodiments provided by the application, when the matching degree of the target entity and the current user is not high, the target entity in the video can be highlighted for the user, so that a prompt effect is provided for the user, and if the user wants to purchase an item in the video, the item can be clicked according to the prompt, so that the effect that the interaction control pops up is obtained.
In one or more embodiments, the apparatus further comprises:
a tag obtaining unit (not shown) configured to perform tagging processing on user preferences according to user data of a current user to obtain a user tag used for representing the user preferences;
a matching degree determination unit (not shown) for determining a matching degree between the tag of the first target object and the user tag as a matching degree between the first target object and the user data of the current user.
One or more embodiments provided herein facilitate determining a first target object that is suitable for a user through matching of tags.
In one or more embodiments, the apparatus further comprises:
a tag obtaining unit (not shown) configured to obtain a browsing object tag corresponding to a user browsing record, a click object tag corresponding to a user click record, a purchase object tag corresponding to a user purchase record, and a multi-dimensional tag used for describing a user in user portrait data before the video frame selecting unit determines a matching degree between the target object and user data of a current user;
and a data generating unit (not shown) for generating user data of the current user according to the browsing object tag, the clicking object tag, the purchasing object tag and the multi-dimensional tag.
One or more embodiments provided by the application can acquire data which can describe a user more accurately by collecting various user behaviors, thereby facilitating matching of objects more suitable for the user.
In one or more embodiments, the object presentation unit 1102 presents an object presentation interface of the first target object, including:
and pausing the playing of the target video on the playing interface and jumping to the object display interface.
One or more embodiments provided by the application can pause the video for the user when the user needs to browse the object, thereby improving the use experience of the user.
In addition, the operation response unit is further configured to close the object presentation interface and continue to play the target video in response to a closing operation for the object presentation interface.
One or more embodiments provided by the application can continue to play the video for the user after the user finishes browsing, so that the use experience of the user is improved.
Referring to fig. 12, fig. 12 is a block diagram schematically illustrating a structure of an object associating apparatus according to an embodiment of the present application. As shown in fig. 12, the object associating apparatus 1200 may include: an entity identifying unit 1201 and an association relationship establishing unit 1202, wherein:
an entity identification unit 1201, configured to perform entity identification on a target video, and identify a first target entity;
an association relationship establishing unit 1202, configured to establish an association relationship between a first target entity and a first target object, so that the terminal device displays a first interaction control in a playing interface for playing a target video, where the first interaction control displays relevant information of the first target object.
In an exemplary embodiment of the present application, the first target entity is included in at least one frame of a target video frame in the target video, and the first target object is an object in a preset object set.
In an exemplary embodiment of the application, when a target video picture set is played in a playing interface of a terminal device, a first interaction control is displayed, where the target video set includes at least one of: the video image comprises at least one frame of target video image, at least one frame of video image with preset frame number before the target video image, and at least one frame of video image with preset frame number after the target video image.
In one or more embodiments, by displaying the first interactive control corresponding to the first target entity in the target video, the object (e.g., advertisement) corresponding to the entity in the video is displayed, so that richer and more effective display content is provided for a user, and the user experience is improved.
It should be noted that although in the above detailed description several modules or units of the device for action execution are mentioned, such a division is not mandatory. Indeed, the features and functionality of two or more modules or units described above may be embodied in one module or unit, according to embodiments of the application. Conversely, the features and functions of one module or unit described above may be further divided into embodiments by a plurality of modules or units.
For details that are not disclosed in the embodiments of the apparatus of the present application, please refer to the embodiments of the target object displaying method described above for the details that are not disclosed in the embodiments of the apparatus of the present application, because each functional module of the target object displaying apparatus of the exemplary embodiment of the present application corresponds to a step of the exemplary embodiment of the target object displaying method described above.
As another aspect, the present application also provides a computer-readable medium, which may be contained in the electronic device described in the above embodiments; or may exist separately without being assembled into the electronic device. The computer readable medium carries one or more programs which, when executed by an electronic device, cause the electronic device to implement the method described in the above embodiments.
It should be noted that the computer readable medium shown in the present application may be a computer readable signal medium or a computer readable storage medium or any combination of the two. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples of the computer readable storage medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the present application, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. In this application, however, a computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated data signal may take many forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: wireless, wire, fiber optic cable, RF, etc., or any suitable combination of the foregoing.
The flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present application. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams or flowchart illustration, and combinations of blocks in the block diagrams or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The units described in the embodiments of the present application may be implemented by software, or may be implemented by hardware, and the described units may also be disposed in a processor. Wherein the names of the elements do not in some way constitute a limitation on the elements themselves.
Other embodiments of the present application will be apparent to those skilled in the art from consideration of the specification and practice of the invention disclosed herein. This application is intended to cover any variations, uses, or adaptations of the invention following, in general, the principles of the application and including such departures from the present disclosure as come within known or customary practice within the art to which the invention pertains. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the application being indicated by the following claims.
It will be understood that the present application is not limited to the precise arrangements described above and shown in the drawings and that various modifications and changes may be made without departing from the scope thereof. The scope of the application is limited only by the appended claims.

Claims (20)

1. A target object display method is characterized by comprising the following steps:
displaying a first interaction control in a playing interface for playing a target video, wherein the first interaction control displays relevant information of a first target object, and the first target object corresponds to a first target entity in the target video;
and responding to the first operation of the first interaction control, and displaying an object display interface of the first target object.
2. The method according to claim 1, wherein a matching degree between the first target object and the user data of the current user satisfies a preset condition.
3. The method of claim 1, further comprising:
closing the first interaction control in response to a second operation on the first interaction control.
4. The method of claim 2, wherein the presenting the first interactive control in the playing interface for playing the target video comprises:
when at least one of the following conditions is met, showing a first interactive control in a playing interface of the playing target video:
the display frequency of the first interaction control does not exceed a preset frequency;
the display times of the first interaction control do not exceed a first preset threshold;
the number of second operations on the first interaction control does not exceed a second preset threshold.
5. The method of claim 1, further comprising:
highlighting a second target entity in the target video in the playing interface;
and responding to a third operation aiming at the second target entity, and displaying a second interaction control corresponding to the second target entity or an object display interface of a second target object in the playing interface, wherein the second interaction control displays related information of the second target object, and the second target object corresponds to the second target entity.
6. The method of claim 2, further comprising:
performing tagging processing on user preferences according to the user data of the current user to obtain a user tag for representing the user preferences;
and determining the matching degree between the label of the first target object and the user label as the matching degree between the first target object and the user data of the current user.
7. The method of claim 2, wherein the user data comprises: at least one of user representation data, user purchase records, user click records, and user browsing records.
8. The method of claim 7, further comprising:
acquiring a browsing object tag corresponding to the user browsing record, a click object tag corresponding to the user click record, a purchase object tag corresponding to the user purchase record and a multi-dimensional tag used for describing the user in the user portrait data;
and generating the user data of the current user according to the browsing object label, the clicking object label, the purchasing object label and the multi-dimensional label.
9. The method of claim 1, wherein the object representation interface for representing the first target object comprises:
and pausing the playing of the target video on the playing interface and jumping to the object display interface.
10. The method of claim 9, further comprising:
and responding to the closing operation aiming at the object display interface, closing the object display interface and continuing to play the target video.
11. The method according to claim 1, wherein the object presentation interface is further configured to present at least one associated object related to the first target object, and the associated object has a similarity higher than a preset similarity with the first target object.
12. The method according to claim 11, wherein the at least one associated object is arranged in the object presentation interface based on a degree of correlation with the user data.
13. The method of claim 1, wherein the information related to the first target object comprises at least one of: a link to the first target object, an identification of the first target object, a graphical representation of the first target object, description information of the first target object.
14. An object association method, comprising:
carrying out entity identification on the target video to identify a first target entity;
and establishing an incidence relation between the first target entity and the first target object so that the terminal equipment displays a first interaction control in a playing interface for playing the target video, wherein the first interaction control displays relevant information of the first target object.
15. The method according to claim 14, wherein the first target entity is included in at least one frame of a target video frame in the target video, and the first target object is an object in a preset set of objects.
16. The method according to claim 15, wherein when a target video frame set is played in a playing interface of the terminal device, the first interaction control is presented, and the target video frame set includes at least one of: the at least one frame of target video picture, the video picture with the preset frame number before the at least one frame of target video picture and the video picture with the preset frame number after the at least one frame of target video picture.
17. A target object display apparatus, comprising:
the video playing unit is used for displaying a first interaction control in a playing interface for playing a target video, wherein the first interaction control displays relevant information of a first target object, and the first target object corresponds to a first target entity in the target video;
and the object display unit is used for responding to the first operation of the first interaction control and displaying an object display interface of the first target object.
18. An object associating apparatus, comprising:
the entity identification unit is used for carrying out entity identification on the target video and identifying a first target entity;
and the association relationship establishing unit is used for establishing an association relationship between the first target entity and the first target object so as to enable the terminal equipment to display a first interaction control in a playing interface for playing the target video, wherein the first interaction control displays relevant information of the first target object.
19. A computer-readable storage medium, on which a computer program is stored, which, when being executed by a processor, carries out the method of any one of claims 1-16.
20. An electronic device, comprising:
a processor; and
a memory for storing executable instructions of the processor;
wherein the processor is configured to perform the method of any of claims 1-16 via execution of the executable instructions.
CN202110484569.XA 2021-04-30 2021-04-30 Target object display method, object association method, device, medium and equipment Pending CN113760158A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110484569.XA CN113760158A (en) 2021-04-30 2021-04-30 Target object display method, object association method, device, medium and equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110484569.XA CN113760158A (en) 2021-04-30 2021-04-30 Target object display method, object association method, device, medium and equipment

Publications (1)

Publication Number Publication Date
CN113760158A true CN113760158A (en) 2021-12-07

Family

ID=78786967

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110484569.XA Pending CN113760158A (en) 2021-04-30 2021-04-30 Target object display method, object association method, device, medium and equipment

Country Status (1)

Country Link
CN (1) CN113760158A (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114501101A (en) * 2022-01-19 2022-05-13 北京达佳互联信息技术有限公司 Video interaction method and device, electronic equipment and computer program product
CN114546565A (en) * 2022-02-24 2022-05-27 北京字跳网络技术有限公司 Information display method and device, electronic equipment, storage medium and program product
CN114968463A (en) * 2022-05-31 2022-08-30 北京字节跳动网络技术有限公司 Entity display method, device, equipment and medium
CN115334346A (en) * 2022-08-08 2022-11-11 北京达佳互联信息技术有限公司 Interface display method, video publishing method, video editing method and device
WO2023134756A1 (en) * 2022-01-14 2023-07-20 北京有竹居网络技术有限公司 Object recommendation method and apparatus, and electronic device
CN116501223A (en) * 2023-06-16 2023-07-28 北京达佳互联信息技术有限公司 Object display method and device, electronic equipment and storage medium
WO2023202460A1 (en) * 2022-04-19 2023-10-26 北京字跳网络技术有限公司 Page display method and apparatus, electronic device, storage medium and program product

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106202304A (en) * 2016-07-01 2016-12-07 传线网络科技(上海)有限公司 Method of Commodity Recommendation based on video and device
CN107027070A (en) * 2016-02-02 2017-08-08 中国电信股份有限公司 Method, terminal and the system of information are implanted into video
CN109495780A (en) * 2018-10-16 2019-03-19 深圳壹账通智能科技有限公司 A kind of Products Show method, terminal device and computer readable storage medium
CN109688469A (en) * 2018-12-27 2019-04-26 北京爱奇艺科技有限公司 A kind of advertisement demonstration method and show device

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107027070A (en) * 2016-02-02 2017-08-08 中国电信股份有限公司 Method, terminal and the system of information are implanted into video
CN106202304A (en) * 2016-07-01 2016-12-07 传线网络科技(上海)有限公司 Method of Commodity Recommendation based on video and device
CN109495780A (en) * 2018-10-16 2019-03-19 深圳壹账通智能科技有限公司 A kind of Products Show method, terminal device and computer readable storage medium
CN109688469A (en) * 2018-12-27 2019-04-26 北京爱奇艺科技有限公司 A kind of advertisement demonstration method and show device

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2023134756A1 (en) * 2022-01-14 2023-07-20 北京有竹居网络技术有限公司 Object recommendation method and apparatus, and electronic device
CN114501101A (en) * 2022-01-19 2022-05-13 北京达佳互联信息技术有限公司 Video interaction method and device, electronic equipment and computer program product
CN114501101B (en) * 2022-01-19 2024-01-02 北京达佳互联信息技术有限公司 Video interaction method, device, electronic equipment and computer program product
CN114546565A (en) * 2022-02-24 2022-05-27 北京字跳网络技术有限公司 Information display method and device, electronic equipment, storage medium and program product
WO2023160416A1 (en) * 2022-02-24 2023-08-31 北京字跳网络技术有限公司 Information display method and apparatus, electronic device, storage medium, and program product
WO2023202460A1 (en) * 2022-04-19 2023-10-26 北京字跳网络技术有限公司 Page display method and apparatus, electronic device, storage medium and program product
CN114968463A (en) * 2022-05-31 2022-08-30 北京字节跳动网络技术有限公司 Entity display method, device, equipment and medium
CN115334346A (en) * 2022-08-08 2022-11-11 北京达佳互联信息技术有限公司 Interface display method, video publishing method, video editing method and device
CN116501223A (en) * 2023-06-16 2023-07-28 北京达佳互联信息技术有限公司 Object display method and device, electronic equipment and storage medium
CN116501223B (en) * 2023-06-16 2023-11-14 北京达佳互联信息技术有限公司 Object display method and device, electronic equipment and storage medium

Similar Documents

Publication Publication Date Title
CN113760158A (en) Target object display method, object association method, device, medium and equipment
US9495592B2 (en) Image-based information providing device and method using object recognition
WO2023051102A1 (en) Video recommendation method, apparatus, and device, and medium
US10334328B1 (en) Automatic video generation using auto-adaptive video story models
JP2019531547A (en) Object detection with visual search queries
US20170270600A1 (en) Visually generated consumer product presentation
CN111818370B (en) Information recommendation method and device, electronic equipment and computer-readable storage medium
CN108228906B (en) Method and apparatus for generating information
WO2020215977A1 (en) System, method and device for displaying information
US11200241B2 (en) Search query enhancement with context analysis
CN110929138A (en) Recommendation information generation method, device, equipment and storage medium
KR20170021454A (en) Individual products through big data analysis using information collected on the basis of the user's media recommended methods and product recommendation system
CN105893404A (en) Natural information identification based pushing system and method, and client
CN110598095B (en) Method, device and storage medium for identifying article containing specified information
WO2021088468A1 (en) Information pushing method and apparatus
US20200327600A1 (en) Method and system for providing product recommendation to a user
CN113849682A (en) Video searching method, device, equipment and medium
KR20190093755A (en) Big data-based image text recognition and customized foodstuff recommendation methods and system
CN112000823A (en) Function entry display method, electronic device and computer-readable storage medium
CN110827063A (en) Multi-strategy fused commodity recommendation method, device, terminal and storage medium
CN110909154A (en) Abstract generation method and device
Wang et al. Interactive ads recommendation with contextual search on product topic space
CN115860869A (en) Shop information recommendation method, equipment and storage medium
US20210176519A1 (en) System and method for in-video product placement and in-video purchasing capability using augmented reality with automatic continuous user authentication
CN114780790A (en) Content search method, device, equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination