WO2023207463A1 - Voting information generation method and apparatus, and voting information display method and apparatus - Google Patents

Voting information generation method and apparatus, and voting information display method and apparatus Download PDF

Info

Publication number
WO2023207463A1
WO2023207463A1 PCT/CN2023/083979 CN2023083979W WO2023207463A1 WO 2023207463 A1 WO2023207463 A1 WO 2023207463A1 CN 2023083979 W CN2023083979 W CN 2023083979W WO 2023207463 A1 WO2023207463 A1 WO 2023207463A1
Authority
WO
WIPO (PCT)
Prior art keywords
voting
text content
sample
topic
video clip
Prior art date
Application number
PCT/CN2023/083979
Other languages
French (fr)
Chinese (zh)
Inventor
陈小帅
Original Assignee
腾讯科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 腾讯科技(深圳)有限公司 filed Critical 腾讯科技(深圳)有限公司
Publication of WO2023207463A1 publication Critical patent/WO2023207463A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/80Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/14Image acquisition
    • G06V30/148Segmentation of character regions

Definitions

  • the embodiments of the present application relate to the field of computer technology, and in particular to a voting information generation method, voting information display method and device.
  • Voting is a common interaction method that is widely used in various scenarios.
  • a method of voting in videos is currently provided.
  • the operator of the video website or the producer of the video artificially creates voting information in the video, and displays the vote during the video playback process.
  • the message entices those who watch the video to vote.
  • this method requires manual creation of voting information in the video, which is very inefficient.
  • this method is difficult to cover a large number of videos, resulting in insufficient voting interaction in the video.
  • the embodiments of this application provide a voting information generation method, voting information display method and device, which eliminates the need to manually create voting information, improves operating efficiency, saves time, and this method can effectively cover a large number of videos and improve voting interaction. coverage.
  • the technical solutions are as follows:
  • a voting information generation method includes:
  • the computer device obtains text content associated with a video clip in the video, where the text content includes at least one of first text content or second text content, where the first text content is the text content included in the video clip, and the The second text content is the text content contained in the barrage of the video clip;
  • the computer device generates a voting topic for the video clip based on keywords in the text content
  • the computer device generates a plurality of voting candidates for the video clip based on the keywords and the voting topic;
  • the computer device generates voting information for the video clip based on the voting topic and the plurality of voting candidates.
  • a voting information display method includes:
  • the computer device obtains voting information of the video clip based on the video clip, the voting information is generated based on a voting topic and a plurality of voting candidates, and the voting topic is based on keywords in text content associated with the video clip. Generating, the plurality of voting candidates are generated based on the keywords and the voting topic;
  • the computer device determines interaction parameters based on the interest tag of the currently logged-in account and the voting information, where the interaction parameters represent the possibility of the account performing a voting operation based on the voting information;
  • the computer device displays the voting information when playing the video clip when the interaction parameters meet the interaction conditions
  • the text content includes at least one of first text content or second text content
  • the first text content is the text content contained in the video clip
  • the second text content is the bounce of the video clip.
  • the text content contained in the scene is the text content contained in the scene.
  • a voting information generating device which is provided in a computer device, and the device includes:
  • a text content acquisition module configured to acquire text content associated with video clips in the video, where the text content includes at least one of first text content or second text content, and the first text content is the text content contained in the video clip. text content, The second text content is the text content contained in the barrage of the video clip;
  • a topic generation module configured to generate voting topics for the video clips based on keywords in the text content
  • a candidate generation module configured to generate multiple voting candidates for the video clip based on the keywords and the voting topic
  • a voting information generation module configured to generate voting information for the video clip based on the voting topic and the plurality of voting candidates.
  • a voting information display device includes:
  • An information acquisition module configured to obtain voting information of video clips based on video clips in the target video, where the voting information is generated based on a voting topic and a plurality of voting candidates, and the voting topic is based on text associated with the video clip. Keywords in the content are generated, and the plurality of voting candidates are generated based on the keywords and the voting topic;
  • a parameter determination module configured to determine interaction parameters based on the interest tag of the currently logged-in account and the voting information, where the interaction parameters represent the possibility of the account performing a voting operation based on the voting information;
  • An information display module configured to display the voting information when the video clip is played when the interaction parameters meet the interaction conditions
  • the text content includes at least one of first text content or second text content
  • the first text content is the text content contained in the video clip
  • the second text content is the bounce of the video clip.
  • the text content contained in the scene is the text content contained in the scene.
  • a computer device includes a processor and a memory. At least one computer program is stored in the memory. The at least one computer program is loaded and executed by the processor to implement the above. The operations performed by the voting information generation method described in the aspect, or to implement the operations performed by the voting information display method described in the above aspect.
  • a computer-readable storage medium is provided. At least one computer program is stored in the computer-readable storage medium. The at least one computer program is loaded and executed by a processor to implement voting as described in the above aspect. The operations performed by the information generation method, or the operations performed by the voting information display method as described above.
  • a computer program product including a computer program that, when executed by a processor, implements the operations performed by the voting information generation method described in the above aspect, or implements the voting information described in the above aspect. Shows the operations performed by the method.
  • Embodiments of the present application provide a method for automatically generating voting information for video clips, which can automatically generate voting information for video clips based on the text content associated with the video clips, eliminating the need to manually create voting information, improving operating efficiency, saving time, and This method can effectively cover a large number of videos and improve the coverage of voting interaction.
  • Figure 1 is a schematic diagram of an implementation environment provided by an embodiment of the present application.
  • FIG. 2 is a flow chart of a voting information generation method provided by an embodiment of the present application.
  • FIG. 3 is a flow chart of another voting information generation method provided by an embodiment of the present application.
  • Figure 4 is a schematic diagram of voting information provided by an embodiment of the present application.
  • Figure 5 is a schematic flowchart of generating voting topics based on the first generation model provided by an embodiment of the present application.
  • Figure 6 is a schematic diagram of a first generation model provided by an embodiment of the present application.
  • FIG. 7 is a schematic flowchart of generating voting candidates based on the second generation model provided by an embodiment of the present application.
  • Figure 8 is a schematic diagram of a second generation model provided by an embodiment of the present application.
  • Figure 9 is a schematic diagram of an overall process for generating voting information provided by an embodiment of the present application.
  • Figure 10 is a flow chart of a voting information generation method provided by an embodiment of the present application.
  • Figure 11 is a schematic diagram of a voting decision model provided by an embodiment of the present application.
  • Figure 12 is a flow chart of a voting information display method provided by an embodiment of the present application.
  • FIG. 13 is a flow chart of another voting information display method provided by an embodiment of the present application.
  • Figure 14 is a schematic diagram of a voting interaction model provided by an embodiment of the present application.
  • FIG 15 is a schematic structural diagram of a voting information generation device provided by an embodiment of the present application.
  • Figure 16 is a schematic structural diagram of a voting information display device provided by an embodiment of the present application.
  • Figure 17 is a schematic structural diagram of a terminal provided by an embodiment of the present application.
  • Figure 18 is a schematic structural diagram of a server provided by an embodiment of the present application.
  • first, second, etc. used in this application may be used to describe various concepts herein, but these concepts are not limited by these terms unless otherwise specified. These terms are only used to distinguish one concept from another.
  • first text content may be called second text content
  • second text content may be called first text content
  • At least one includes one, two or more than two, and multiple includes two or more, each A refers to each of the corresponding plurality, and any refers to any one of the plurality.
  • multiple keywords include 3 keywords, and each keyword refers to each of these 3 keywords. Any one refers to any one of these 3 keywords, which can be the third keyword. One, it can be the second, it can be the third.
  • Artificial Intelligence is a theory, method, technology and application system that uses digital computers or machines controlled by digital computers to simulate, extend and expand human intelligence, perceive the environment, acquire knowledge and use knowledge to obtain the best results.
  • artificial intelligence is a comprehensive technology of computer science that attempts to understand the essence of intelligence and produce a new intelligent machine that can respond in a similar way to human intelligence.
  • Artificial intelligence is the study of the design principles and implementation methods of various intelligent machines, so that the machines have the functions of perception, reasoning and decision-making.
  • Artificial intelligence technology is a comprehensive subject that covers a wide range of fields, including both hardware-level technology and software-level technology.
  • Basic artificial intelligence technologies generally include technologies such as sensors, dedicated artificial intelligence chips, cloud computing, distributed storage, big data processing technology, operation/interaction systems, mechatronics and other technologies.
  • Artificial intelligence software technology mainly includes computer vision technology, speech processing technology, natural language processing technology, machine learning/deep learning, autonomous driving, smart transportation and other major directions.
  • Computer Vision Technology (Computer Vision, CV) Computer vision is a science that studies how to make machines "see”. Furthermore, it refers to machine vision such as using cameras and computers instead of human eyes to identify and measure targets, and further Perform graphics processing to make computer processing into images more suitable for human eye observation or transmitted to instruments for detection. As a scientific discipline, computer vision studies related theories and technologies, trying to build artificial intelligence systems that can obtain information from images or multi-dimensional data.
  • Computer vision technology usually includes image processing, image recognition, image semantic understanding, image retrieval, OCR (Optical Character Recognition, optical character recognition), video processing, video semantic understanding, video content/behavior recognition, three-dimensional object reconstruction, 3D technology, virtual Reality, augmented reality, simultaneous positioning and map construction, autonomous driving, smart transportation and other technologies, as well as common biometric recognition technologies such as face recognition and fingerprint recognition.
  • OCR Optical Character Recognition, optical character recognition
  • video processing video semantic understanding
  • video content/behavior recognition three-dimensional object reconstruction
  • 3D technology virtual Reality, augmented reality, simultaneous positioning and map construction
  • autonomous driving smart transportation and other technologies
  • biometric recognition technologies such as face recognition and fingerprint recognition.
  • Machine Learning is a multi-field interdisciplinary subject involving probability theory, statistics, approximation theory, convex analysis, algorithm complexity theory and other disciplines. It specializes in studying how computers can simulate or implement human learning behavior to acquire new knowledge or skills, and reorganize existing knowledge structures to continuously improve their performance.
  • Machine learning is the core of artificial intelligence and the fundamental way to make computers intelligent. Its applications cover all fields of artificial intelligence.
  • Machine learning and deep learning usually include artificial neural networks, belief networks, reinforcement learning, transfer learning, inductive learning, teaching learning and other technologies.
  • artificial intelligence technology has been researched and applied in many fields, such as common smart homes, smart wearable devices, virtual assistants, smart speakers, smart marketing, driverless driving, autonomous driving, and drones.
  • Robots smart medical care, smart customer service, Internet of Vehicles, autonomous driving, smart transportation, etc. I believe that with the development of technology, artificial intelligence technology will be applied in more fields and play an increasingly important role.
  • the voting information generation method and voting information display method provided by the embodiments of this application utilize computer vision technology and machine learning technology in artificial intelligence to generate voting information for video clips and display the voting information when the video clips are played.
  • the execution subject of the voting information generation method and voting information display method provided by the embodiments of this application is a computer device, and the computer device is a terminal or a server.
  • the voting information generation method is executed by the server, and the voting information display method is executed by the terminal.
  • the embodiment of the present application provides an implementation environment as shown in Figure 1 below.
  • FIG 1 is a schematic diagram of an implementation environment provided by an embodiment of the present application.
  • the implementation environment includes: a server 101 and a terminal 102.
  • the server 101 and the terminal 102 can be connected directly or indirectly through wired or wireless communication methods, which is not limited in this application.
  • the server 101 is used to store or deliver videos, and is also used to automatically generate voting information for video clips in the video, while the terminal 102 is used to access the server 101, play the video delivered by the server 101, and display the current video when playing the video.
  • the voting information of the clip is thus launched to initiate a voting interaction for the current video clip, attracting users to perform voting operations and participate in the voting interaction.
  • the server 101 is an independent physical server, or a server cluster or distributed system composed of multiple physical servers, or it provides cloud services, cloud databases, cloud computing, cloud functions, cloud storage, and networks. Services, cloud communications, middleware services, domain name services, security services, CDN (Content Delivery Network, content distribution network), and cloud servers for basic cloud computing services such as big data and artificial intelligence platforms.
  • the terminal 102 is a smartphone, a tablet computer, a notebook computer, a desktop computer, a smart speaker, a smart watch, a smart TV, a smart car terminal, etc., but is not limited thereto.
  • a target application provided by the server 101 is installed on the terminal 102, and the terminal 102 can implement functions such as video playback and voting through the target application.
  • the target application is a video sharing application, which has the function of video sharing.
  • the video sharing application can also have other functions, such as the function of posting barrages, the function of voting, etc.
  • FIG. 2 is a flow chart of a voting information generation method provided by an embodiment of the present application.
  • the execution subject of the embodiment of the present application is a computer device, and the computer device is a terminal or a server.
  • the embodiment of the present application explains the process of generating voting information of a video clip. Referring to Figure 2, the method includes:
  • the computer device obtains text content associated with the video clips in the video.
  • the video is any video in the computer device.
  • the computer device is a terminal
  • the video is any video downloaded by the terminal or any video shot, etc.
  • the computer device is a server
  • the server has a video sharing function and can Store the video uploaded by any device and send the video to any device for playback, and the video is any video stored by the server.
  • the video includes one or more video clips, and the playing time of the video clips is not greater than the total playing time of the video.
  • the video is divided into multiple video segments according to a fixed duration, and the playback duration of each video segment is equal to the fixed duration.
  • the text content associated with the video clip includes at least one of first text content or second text content.
  • the first text content is text content contained in the video clip.
  • the first text content includes subtitle text content in the video clip, or includes text content recognized from the voice data in the video clip.
  • the first text content It can represent the content contained in the video clip itself, such as the characters and things that appeared in the video clip or the plot that occurred in the video clip.
  • the second text content is the text content included in the barrage of the video clip, which may be called the barrage text content.
  • the terminal playing the video clip can publish a barrage for the video clip.
  • the text content contained in the barrage can express the end user's views or opinions on the video clip, so the barrage text content can be viewed Create interactive data for video clips.
  • each video clip in the target video has a corresponding playback time period, and the playback time of the video clip If the segment includes the release time point of the barrage, it means that the barrage is the barrage of the video clip.
  • the computer device After the computer device obtains the text content, it can automatically generate the voting information of the video clip based on the text content, without the need for technical personnel to manually generate it, and display the voting information when the video clip is played, and can also attract users to participate in voting interaction , which helps improve the interactive coverage of the video.
  • the process of generating voting information is detailed in steps 202-204 below.
  • the computer device generates a voting topic for the video clip based on the keywords in the text content.
  • the computer device generates multiple voting candidates for the video clip based on the keywords and voting topics.
  • the voting information includes a voting topic and multiple voting candidates.
  • the voting topic represents the question asked of the user, and the multiple voting candidates represent the candidate answers provided to the user.
  • the voting information is displayed, the user understands the question by viewing the voting topic. Select one voting candidate from multiple voting candidates and choose your own answer, which is to implement the voting operation.
  • the text content includes at least one word
  • the keywords in the text content may include every word in the text content, or only include words extracted from the text content through a keyword extraction algorithm. Since the keyword can represent the content of the video clip, the voting topic generated based on the keyword is related to the content of the video clip. Moreover, the multiple voting candidates generated based on keywords and voting topics are also related to the content of the video clip and consistent with the generated voting topics.
  • the computer device generates voting information of the video clip based on the voting topic and multiple voting candidates.
  • the embodiment of the present application only takes one video clip in a video as an example, and the process of generating voting information for other video clips is similar to the embodiment of the present application and will not be described again here.
  • voting information in videos is mainly created by operators of video sites or creators of videos, which takes a long time and has low operating efficiency. Moreover, this method is difficult to cover a large number of videos, resulting in The voting interaction in the video is not sufficient, which prohibits users from participating in voting interaction.
  • Embodiments of the present application provide a method for automatically generating voting information for video clips, which can automatically generate voting information for video clips based on the text content associated with the video clips, eliminating the need to manually create voting information, improving operating efficiency, saving time, and This method can effectively cover a large number of videos and improve the coverage of voting interaction.
  • FIG. 3 is a flow chart of another voting information generation method provided by an embodiment of the present application.
  • the execution subject of the embodiment of the present application is a computer device, and the computer device is a terminal or a server. Referring to Figure 3, the method includes:
  • the computer device obtains the text content associated with the video clips in the video.
  • the text content includes first text content, and the first text content is text content included in the video clip.
  • the process of obtaining the first text content includes at least one of the following:
  • one or more video frames are extracted from the video clip, and an OCR (Optical Character Recognition) algorithm is used to extract subtitle text content from the one or more video frames.
  • OCR Optical Character Recognition
  • the ASR Automatic Speech Recognition
  • the speech data is recognized and obtain the text content corresponding to the speech data, which is the text content of the dialogue in the video clip.
  • the text content includes second text content
  • the second text content is the text content contained in the bullet comments of the video clip.
  • the process of obtaining the second text content includes: extracting the bullet comments of the video clip from the video clip collection. Screen, extract text content from the barrage of video clips.
  • the voting information includes a voting topic and multiple voting candidates. After obtaining the text content, the computer device first needs to generate a voting topic. For specific steps to generate a voting topic, please refer to the following steps 302-303.
  • the computer device encodes the keywords in the text content and obtains the keyword characteristics of the keywords.
  • the keyword feature is used to describe keywords, and the keywords are converted into the form of keyword features to facilitate subsequent processing based on quantifiable keyword features, and to generate voting information related to the keywords.
  • the encoding model can be a Transformer model (a model based on a self-attention mechanism) or other types of models.
  • the keyword extraction model may be a TextRank model (a model that extracts keywords based on text ranking) or other types of models.
  • the computer device decodes the keyword features to obtain the voting topic, which is composed of multiple voting topic words.
  • the computer device decodes the keyword features each time to obtain a voting topic word, and then continues to decode the keyword features and the last determined voting topic word to obtain the next voting topic word until the target number is obtained.
  • voting topic words, and the target number of voting topic words constitute the voting topic.
  • the computer device obtains the first keyword, which is a keyword in the first text content.
  • the computer device obtains the first text content and the second text content, where the keywords in the first text content are called first keywords, and the keywords in the second text content are called second keywords. word.
  • the step of obtaining the first keyword in the first text content is similar to the method of extracting the keyword in the above-mentioned step 302, and will not be described again here.
  • the computer device clusters the second text content to obtain multiple text categories, each text category contains at least one piece of second text content, and extracts the second keyword from each text category.
  • the video may include multiple barrages. Accordingly, the computer device will obtain the second text content included in the multiple barrages, thereby extracting the second keywords from the plurality of second text contents.
  • multiple second text contents can be clustered first, and semantically related second text contents can be divided into one text category, and then different texts can be classified into Category, extract the second keyword in each text category separately.
  • the computer device generates voting candidates for each text category based on the first keyword, the voting topic, and the second keyword of each text category.
  • i is a positive integer and i is not greater than the number of text categories.
  • the process of generating the i-th voting candidate includes: comparing the first keyword, voting topic and i-th text
  • the second keyword of the category is encoded to obtain keyword features, and the keyword features are decoded to obtain the i-th voting candidate.
  • the i-th voting candidate is composed of multiple voting candidate words.
  • the specific process is similar to the above steps 302-303. The difference is that the keyword characteristics determined this time are used to describe the first keyword, the keywords in the voting topic and the second keyword. Multiple keywords are obtained based on the keyword characteristics.
  • the voting candidate words comprehensively consider the influence of the first keyword, the keywords in the voting topic, and the second keyword, ensuring that the voting candidate words are related to the first keyword, the voting topic, and the second keyword.
  • the computer device generates voting information of the video clip based on the voting topic and multiple voting candidates.
  • the voting topic and multiple voting candidates constitute the voting information of the video clip.
  • the voting topic, multiple voting candidates, and associated information constitute the voting information of the video clip.
  • the associated information includes text or images used to prompt users to vote, etc., and may also include other types of information.
  • the video screen of the video clip and voting information are displayed in the playback interface.
  • the voting information is divided into two parts, one part is the voting topic "What did you come to see?", and the other part There are three voting candidates for users to choose from.
  • the computer device After generating the voting information, the computer device stores the video clips in association with the voting information, displays the voting information when playing the video clips, or delivers the voting information each time the video clips are delivered to other devices. Alternatively, the computer device adds the voting information to the video clip so that the voting information is displayed when the video clip is played.
  • the specific process of displaying voting information is detailed in the embodiment shown in FIG. 12 and FIG. 13 below, and will not be described here.
  • the embodiment of the present application only takes voting information of a video clip as an example for explanation.
  • the computer device can generate multiple voting topics and each voting topic by repeatedly executing the above steps.
  • the corresponding voting candidates constitute multiple voting information, and then the multiple voting information can be displayed when the video clip is played, or one or more of the multiple voting information can be displayed.
  • the embodiment of the present application does not limit this. .
  • Embodiments of the present application provide a method for automatically generating voting information for video clips, which can automatically generate voting information for video clips based on the text content associated with the video clips, eliminating the need to manually create voting information, improving operating efficiency, saving time, and This method can effectively cover a large number of videos and improve the coverage of voting interaction.
  • the generated voting information is related to the text content of the video clip, which meets the interactive function requirements of the video, helps to increase the enthusiasm of users to participate in voting interaction while watching the video clip, and thereby enhances the interactive atmosphere.
  • Using the above method makes it easy to cover a large number of videos, improves the interactive coverage and richness of the videos, and increases the user interaction activity on the video platform.
  • the process of generating voting topics in the above steps 302-303 can be performed based on the first generation model, the first generation model includes the first encoding sub-model and a first decoding sub-model, wherein the first encoding sub-model is used to encode keywords into keyword features, and the first decoding sub-model is used to decode keyword features into voting topic words.
  • the process of generating voting topics based on the first generation model includes:
  • the computer device calls the first encoding sub-model, encodes the N keywords, and obtains the keyword characteristics of the N keywords, where N is an integer greater than 1.
  • the first encoding sub-model is a Transformer Encoder model (the encoder in the Transformer model) or other types of encoding models.
  • the N keywords include the first keywords in the first text content, such as dialogue text keywords or alphabetical text keywords of the video clip, etc., and also include the second keywords in the second text content, such as the video clip's Text keywords in the barrage.
  • the computer device obtains N keywords of the video clip, and needs to comprehensively consider these N keywords to generate a voting topic. Then N keywords are input into the first encoding sub-model of the first generative model, so that the N keywords are encoded respectively based on the first encoding sub-model to obtain N keyword features, each keyword corresponding to a key word features.
  • the computer device calls the first decoding sub-model, decodes the N keyword features, obtains the first decoding feature, and determines the first voting topic word based on the first decoding feature and the N keyword features.
  • the first decoding sub-model is a Transformer Decoder model (the decoder in the Transformer model) or other types of coding models.
  • the computer device calls the first decoding sub-model, decodes the N keyword features and the first voting topic words, obtains the second decoding feature, and determines the reference voting topic based on the second decoding feature and the N keyword features. , the reference voting topic includes the first voting topic word and the second voting topic word, until after N times of decoding, the reference voting topic obtained by the Nth decoding is determined as the voting topic of the video clip.
  • a method of sequential decoding is used to generate voting topic words.
  • the current voting topic words are determined.
  • the current voting topic words are combined with the previously determined voting topic words in order to obtain the current voting topic words.
  • Reference voting topic As decoding is performed multiple times, the number of voting topic words contained in the reference voting topic gradually increases until, after N decodings, the reference voting topic obtained by decoding for the Nth time contains N voting topic words, thus obtaining the result containing N voting topic words. Poll topic for topic words.
  • the N keyword features are considered during each decoding, but also the previously determined voting topic words are considered. This can ensure that the voting topic words determined this time are associated with the previously determined voting topic words. This ensures that the associations of different voting topic words in the voting topic composed of the determined N voting topic words can be combined to form a sentence with clear semantics.
  • the first generation model also includes a first classification layer and a preset word library.
  • the preset word library includes a plurality of words. Based on the first decoding feature and the N keyword features, the first voting topic word is determined, including :
  • the probability is the probability of using the jth keyword in the voting topic, j is a positive integer, and j is not greater than N.
  • the jth first usage probability is also the probability of determining the jth keyword as the first voting topic word in the voting topic.
  • the usage conditions refer to the conditions that need to be met when using keywords in voting topics.
  • the first classification layer is called to classify based on the first decoding feature and the preset word library, and the classification probability of each word in the preset word library is obtained. , based on the classification probability of each word, determine the first voting topic word. Among them, the classification probability of a word represents the probability of determining the word as the voting subject word.
  • the embodiments of this application provide two ways of determining voting topic words.
  • One is to use a copy mechanism to copy keywords into voting topic words. That is, if the first usage probability of the keyword meets the usage conditions, then the key The words were directly used as the voting topic words. The other is to generate new voting topic words based on the preset word library. Then every time the first decoding sub-model performs decoding, the usage probability is first determined, and based on whether the usage probability meets the usage conditions, it is determined whether to use the keyword as a voting topic word or to generate a new voting topic word.
  • the usage condition includes a usage probability threshold.
  • the jth keyword is determined as the first voting topic word, wherein the plurality of first usage probabilities are equal to If it is greater than the usage probability threshold, the keyword with the largest first usage probability will be determined as the first voting topic word, and if each first usage probability is not greater than the usage probability threshold, then determine the keyword in the preset word library
  • the classification probability of each word represents the possibility of determining the word as a voting topic word. Then based on the classification probability of each word, the first voting topic word can be selected from multiple words.
  • determining the first voting topic word based on the classification probability of each word includes: selecting the word with the highest classification probability from multiple words in the preset word library as the first voting topic word, or selecting from Among the multiple words in the preset word library, the target number of words with the highest classification probability are selected as the candidate first voting topic words. Among them, the number of targets is greater than 1.
  • the reference voting topic includes the first voting topic word and the second voting topic word, including:
  • the jth second usage probability is the probability of using the jth keyword in the voting topic, j is a positive integer, and j is not greater than N.
  • the j-th second usage probability is also the probability of determining the j-th keyword as the second voting topic word in the voting topic.
  • the usage conditions refer to the conditions that need to be met when using keywords in voting topics.
  • the first classification layer is called to classify based on the second decoding feature and the preset word library to obtain the classification probabilities of multiple candidate voting topics.
  • Each candidate The voting topic includes a first voting topic word and a second voting topic word.
  • the classification probability of a candidate voting topic represents the probability of determining the candidate voting topic as a reference voting topic.
  • determine the reference voting topic based on the classification probability of each candidate voting topic including: selecting the candidate voting topic with the highest classification probability from multiple candidate voting topics as the reference voting topic, or voting from multiple candidate voting topics Among the topics, the target number of candidate voting topics with the highest classification probability are selected as reference voting topics. Among them, the number of targets is greater than 1. Subsequently, voting candidates can be generated for each voting topic to form a target number of voting information.
  • the process of the second decoding is similar to the process of the first decoding.
  • the first voting topic words obtained by the first decoding will also be input into the first decoding sub-model, so that The first voting topic word and each word in the preset word library constitute a candidate voting topic.
  • the classification probability of the candidate voting topics can also reflect the candidate voting topics obtained by combining the words in the preset word library and the first voting topic words. to a reasonable degree, thereby ensuring that voting topics with clear semantics and logical logic can be generated.
  • P i e ⁇ (X*w i *sv i )/sum j (e ⁇ (X*w i *sv i )), where P i represents the i-th usage probability, X represents the current decoding feature, w i represents the weight parameter of the i-th keyword feature, sv i represents the i-th keyword feature, j is any positive integer, j is not greater than N, X j represents the j-th decoding feature, w j represents the j-th key The weight parameter of the word feature, sv j represents the jth keyword feature.
  • Figure 6 is a schematic diagram of a first generation model provided by an embodiment of the present application.
  • the N keywords obtained by the computer device include dialogue text keywords, subtitle text keywords, and barrage text keywords.
  • N keywords are input into the first encoding sub-model, and N keyword features are obtained by calling the first encoding sub-model: keyword feature 1, keyword feature 2...keyword feature N, and during each decoding, You can choose to copy text keywords or redefine new voting topic words.
  • the corresponding text keywords can be directly determined as the voting topic words. If it is determined not to copy the text keywords, then the corresponding text keywords can be determined based on the last determined voting topic words. The current voting topic word.
  • the training process of the first generation model includes:
  • the positive sample video clip is a video clip that contains sample voting information and the participation rate of the sample voting information reaches the target threshold; obtain the sample voting topic included in the sample voting information; based on the sample text content and Sample voting topics, adjusting model parameters in the first generative model.
  • the positive sample video clip is a video clip for which sample voting information has been created, and when the sample voting information is displayed when the positive sample video clip is played, the participation rate of the sample voting information reaches the target threshold, indicating that there are many The user participated in voting while watching the positive sample video clip.
  • the sample voting information has a strong correlation with the positive sample video clip.
  • the sample voting topic contained in the sample voting information also has a strong correlation with the positive sample video clip. Strong correlation, so the first generative model is trained based on the sample text content associated with the positive sample video clip and the sample voting topic, so that the trained first generative model is based on the voting topic obtained from the sample text content and the sample voting topic.
  • the similarity increases, thereby improving the accuracy of the first generation model, so that the function of the first generation model in generating voting topics based on text content is improved.
  • the model parameters in the first generation model may include weight parameters or other parameters in each layer of the first generation model.
  • the model parameters include the weight parameter w used to determine the probability of use. Through one or more Training and adjusting the model parameters can improve the accuracy of the weight parameter w, improve the accuracy of the determined usage probability, and thereby improve the accuracy of the final determined voting topic.
  • negative sample video clips can also be introduced when training the first generation model.
  • Negative sample video clips are video clips for which sample voting information has been created but the participation rate has not reached the target threshold, or video clips for which no sample voting information has been created.
  • sample voting information is displayed when the negative sample video clip is played, and the participation rate of the sample voting information does not reach the target threshold, it means that many users did not participate in voting when watching the negative sample video clip.
  • the sample voting information If the correlation with the negative sample video clip is not strong, the sample voting topic contained in the sample voting information will not be strongly correlated with the negative sample video clip.
  • the sample voting topic should not be used as the vote corresponding to the negative sample video clip. topic, so the first generative model is trained based on the sample text content associated with the negative sample video clip and the sample voting topic, so that the trained first generative model is based on the similarity between the voting topic obtained from the sample text content and the sample voting topic. By reducing, the accuracy of the first generative model can be improved to prevent the first generative model from generating inappropriate voting topics based on text content.
  • Embodiments of the present application provide a method for automatically generating voting topics for video clips.
  • a first generative model is trained, and a voting topic is generated based on the first generative model, which enables the video to be
  • the content of the clip is deeply understood, and the voting topic is generated based on the characteristics of the deep representation.
  • the generated voting topic is related to the text content of the video clip, which meets the interactive functional requirements of the video and helps to improve user participation in the process of watching the video clip. The enthusiasm of voting interaction further enhances the interactive atmosphere.
  • the process of voting candidates may be performed based on a second generative model, which includes a second encoding sub-model and a second decoding sub-model, wherein the second encoding sub-model is used to encode keywords into keyword features, and the second encoding sub-model is used to encode keywords into keyword features.
  • the binary decoding sub-model is used to decode keyword features into voting candidate words.
  • the process of generating voting candidates based on the second generation model includes:
  • the computer device obtains the first keyword in the first text content, clusters the second text content, and obtains multiple text categories. Each text category contains at least one piece of second text content. From each text category, Extract the second keyword.
  • this step 701 is similar to the above-mentioned steps 304-305, and will not be described again here.
  • the computer device calls the second encoding sub-model to encode the first keyword, the voting topic and the second keyword of the i-th text category to obtain keyword features.
  • the second encoding sub-model is a Transformer Encoder model or other types of encoding models.
  • the computer device obtains the first keyword of the video clip, the voting topic, and the second keyword of the i-th text category, and needs to comprehensively consider these keywords to generate the i-th voting candidate. Then input the first keyword, the voting topic and the second keyword of the i-th text category into the second encoding sub-model of the second generation model, and take the number of input keywords as M as an example, so that The M keywords are encoded respectively based on the second encoding sub-model to obtain M keyword features.
  • the computer device calls the second decoding sub-model, decodes the M keyword features, obtains the first decoding feature, and determines the first voting candidate word based on the first decoding feature and the M keyword features.
  • the computer device calls the second decoding sub-model, decodes the M keyword features and the first voting candidate words, obtains the second decoding feature, and determines the reference vote based on the second decoding feature and the M keyword features.
  • Candidates, reference voting candidates include the first voting candidate word and the second voting candidate word, until after M times of decoding, the reference voting candidate obtained by the M-th decoding is determined as the i-th voting candidate of the video clip item.
  • a method of sequential decoding is used to generate voting candidate words.
  • the current voting candidate words are determined.
  • the current voting candidate words are combined with the previously determined voting candidate words in order. You can get the current reference voting candidates.
  • the number of voting candidate words contained in the voting candidates gradually increases, until after M times of decoding, the reference voting candidate obtained by decoding for the Mth time contains M voting candidate words, thus obtaining A voting candidate containing M voting candidate words.
  • the M keyword features are considered during each decoding, but also the previously determined voting candidate words are considered. This ensures that the voting candidate words determined this time are consistent with the previously determined voting candidate words. Correlation, thereby ensuring that the association of different voting candidate words in the voting candidates composed of the determined M voting candidate words can be combined to form a sentence with clear semantics.
  • the second generation model also includes a second classification layer and a preset word library.
  • the preset word library includes a plurality of words. Based on the first decoding feature and the M keyword features, the first voting candidate word is determined, include:
  • the jth third usage probability is the probability of using the jth keyword in the voting candidate, j is a positive integer, And j is not greater than M.
  • the j-th third usage probability is also the probability of determining the j-th keyword as the first voting candidate word among the voting candidates.
  • the usage conditions refer to the conditions that need to be met when using keywords in voting candidates.
  • the embodiments of this application provide two ways to determine voting candidate words.
  • One is to use a copy mechanism to copy keywords into voting candidate words. That is, if the third usage probability of the keyword meets the usage conditions, then Keywords are directly used as voting candidates.
  • the other is to generate new voting candidate words based on the preset word library. Then every time the first decoding sub-model performs decoding, the usage probability is first determined, and based on whether the usage probability meets the usage conditions, it is determined whether to Using keywords as voting candidate words still requires generating new voting candidate words.
  • the usage condition includes a usage probability threshold.
  • the jth third usage probability is greater than the usage probability threshold, the jth keyword is determined as the first voting candidate word, wherein among multiple third usage probabilities If both are greater than the usage probability threshold, the keyword with the largest third usage probability will be determined as the first voting topic word, and if each third usage probability is not greater than the usage probability threshold, then the preset word library will be determined.
  • the classification probability of each word in the classification probability represents the possibility of determining the word as a voting candidate word, then based on the classification probability of each word, the first voting candidate word can be selected from multiple words.
  • determining the first voting candidate word based on the classification probability of each word includes: selecting the word with the highest classification probability from multiple words in the preset word library as the first voting candidate word, or , from the multiple words in the preset word library, select the target number of words with the highest classification probability as the first voting candidate words. Among them, the number of targets is greater than 1.
  • the reference voting candidates are determined, and the reference voting candidates include the first voting candidate words and the second voting candidate words, including:
  • the jth fourth usage probability is the probability of using the jth keyword in the voting candidate, j is a positive integer, And j is not greater than N.
  • the j-th fourth usage probability is also the probability of determining the j-th keyword as the second voting candidate word among the voting candidates.
  • the usage conditions refer to the conditions that need to be met when using keywords in voting candidates.
  • the first classification layer is called to classify based on the second decoding feature and the preset word library to obtain the classification probabilities of multiple candidate voting candidates.
  • Each The candidate voting candidates include a first voting candidate word and a second voting candidate word.
  • the classification probability of the voting candidate represents the probability of determining the voting candidate as the reference voting candidate.
  • determining the reference voting candidate based on the classification probability of each candidate voting candidate includes: selecting the voting candidate with the highest classification probability from multiple voting candidates as the reference voting candidate, or, From multiple candidate voting candidates, a target number of candidate voting candidates with the highest classification probability are selected as reference voting candidates respectively. Among them, the number of targets is greater than 1. Subsequently, corresponding voting candidates can be generated for each voting candidate, thereby forming a target number of voting information.
  • the process of the second decoding is similar to the process of the first decoding.
  • the first voting candidate words obtained by the first decoding will also be input into the first decoding sub-model.
  • the first voting candidate word and each word in the preset word library constitute alternative voting candidates.
  • the candidate voting candidates are selected from multiple alternative voting candidates.
  • the current voting candidate is determined in the item, which ensures that the correlation between the words in the preset word library and the first voting candidate word is taken into account, and the classification probability of the alternative voting candidate can also reflect the preset word
  • the reasonable degree of the alternative voting candidates obtained by combining the words in the library with the first voting candidate words ensures that voting candidates with clear semantics and smooth logic can be generated.
  • the following formula is used to determine the i-th usage probability:
  • P i e ⁇ (X*w i *sv i )/sum j (e ⁇ (X*w i *sv i )), where P i represents the i-th usage probability, X represents the current decoding feature, w i represents the weight parameter of the i-th keyword feature, sv i represents the i-th keyword feature, j is any positive integer, j is not greater than M, X j represents the j-th decoding feature, w j represents the j-th key The weight parameter of the word feature, sv j represents the jth keyword feature.
  • Figure 8 is a schematic diagram of a second generation model provided by an embodiment of the present application.
  • the computer device clusters the text content of the barrage to obtain multiple text categories, and the second keywords extracted from each text category are , which is the barrage text keyword.
  • the M keywords obtained by the computer device include dialogue text keywords, subtitle text keywords, and voting topics.
  • the keywords in and the barrage text keywords corresponding to the i-th text category, input M keywords into the second encoding sub-model, and obtain M keyword features by calling the second encoding sub-model: keyword features 1. Keyword feature 2...keyword feature M, and by calling the second decoding sub-model, during each decoding, you can choose whether to copy the text keyword or re-determine new voting candidate words.
  • the corresponding text keywords can be directly determined as voting candidate words. If it is determined not to copy the text keywords, then based on the last determined voting candidate words , determine the current voting candidate words.
  • the training process of the second generative model includes:
  • the positive sample video clip is a video clip that contains sample voting information and the participation rate of the sample voting information reaches the target threshold; obtain the sample voting topic and multiple sample voting candidates included in the sample voting information. items; based on the correlation between each sample text content and each sample voting candidate, determine the text category corresponding to each sample voting candidate.
  • the text category includes the sample text content associated with the sample voting candidate; respectively, from each sample voting candidate Extract sample keywords from the text category corresponding to the sample voting candidate; adjust the model parameters in the second generation model based on the sample text content, the sample voting topic, the multiple sample voting candidates, and the sample keywords of each sample voting candidate.
  • the positive sample video clip is a video clip for which sample voting information has been created, and when the sample voting information is displayed when the positive sample video clip is played, the participation rate of the sample voting information reaches the target threshold, indicating that there are many The user participated in voting while watching the positive sample video clip, and the sample voting information has a strong correlation with the positive sample video clip, then the sample voting topic and multiple sample voting candidates included in the sample voting information are related to the sample voting information.
  • Positive sample video clips also have strong correlation, so the second generative model is trained based on the sample text content associated with the positive sample video clip and the sample voting information, so that the trained second generative model is based on the votes obtained from the sample text content.
  • the similarity between the candidate and the sample voting candidate increases, thereby improving the accuracy of the second generation model, so that the function of the second generation model in generating voting candidates based on text content is improved.
  • the sample text content associated with the positive sample video clip contains text content associated with each sample voting candidate.
  • each sample text is The content is divided into the text categories of each sample voting candidate, which can distinguish the sample text content of different text categories, so that the sample keywords corresponding to each sample voting candidate can be extracted according to different text categories to avoid being affected by other texts.
  • Category interference so that the second generative model has the function of generating voting candidates based on keywords of different text categories.
  • the model parameters in the second generation model may include weight parameters or other parameters in each layer of the second generation model.
  • negative sample video clips can also be introduced when training the second generation model.
  • Negative sample video clips are video clips for which sample voting information has been created but the participation rate has not reached the target threshold, or video clips for which no sample voting information has been created.
  • sample voting information is displayed when the negative sample video clip is played, and the participation rate of the sample voting information does not reach the target threshold, it means that many users did not participate in voting when watching the negative sample video clip.
  • the sample voting information If the correlation with the negative sample video clip is not strong, the sample voting topic and sample voting candidate included in the sample voting information will not be strongly correlated with the negative sample video clip, and the sample voting candidate should not be used as the negative sample video clip.
  • the voting candidate corresponding to the sample video clip, or the sample voting candidate should not be used as a candidate for the sample voting topic, so based on the sample text content associated with the negative sample video clip, the sample voting topic and the multiple sample voting candidates
  • the item trains the second generative model, so that the similarity between the voting candidates obtained by the trained second generative model based on the sample text content and the sample voting candidates is reduced, thereby improving the accuracy of the second generative model to avoid the second generation model.
  • the second generative model generates inappropriate voting candidates based on text content and voting topics.
  • Embodiments of the present application provide a method for automatically generating voting candidates for video clips.
  • a first generation model is trained, and voting candidates are generated based on the first generation model. It can deeply understand the content of video clips and generate voting information based on deep representation features, improving the accuracy of voting information.
  • the generated voting information is related to the text content of the video clip, which meets the interactive functional requirements of the video, helps to increase the enthusiasm of users to participate in voting interaction while watching the video clip, and thereby enhances the interactive atmosphere.
  • FIG. 9 is a schematic diagram of an overall process for generating voting information provided by embodiments of the present application. See Figure 9.
  • the voting information In the generation method, it can first be determined whether voting information needs to be generated for the video clip. Only when it is determined that voting information needs to be generated for the video clip, the steps of generating voting information, including generating voting topics and generating voting candidates, will be performed. After the voting information is generated, personalized voting information is displayed based on the login account.
  • the embodiment shown in Figure 10 below will explain the process of determining whether voting information needs to be generated, and the process of displaying personalized voting information based on the login account is detailed in the implementation shown in Figures 12 and 13 below. For example, the embodiment shown in FIG. 10 below will not be described for the time being.
  • FIG 10 is a flow chart of a voting information generation method provided by an embodiment of the present application.
  • the execution subject of the embodiment of the present application is a computer device, and the computer device is a terminal or a server. Referring to Figure 10, the method includes:
  • the computer device acquires the first text content, the second text content, the first popularity parameter and the second popularity parameter.
  • the first text content is the text content contained in the video clip
  • the second text content is the text content contained in the barrage of the video clip.
  • the process of obtaining the first text content and the second text content is detailed in step 301 above, and will not be described again here.
  • the first popularity parameter represents the popularity of the video clip
  • the second popularity parameter represents the popularity of the barrages of the video clip
  • the first popularity parameter may be determined based on the number of times the video clip is played, for example, the first popularity parameter is positively correlated with the number of times the video clip is played.
  • the first popularity parameter of the video segment may be determined based on the number of times the video segment is played and the maximum number of times the multiple video segments are played by the computer device.
  • Q1 represents the first popularity parameter
  • num1 represents the number of plays of the video clip
  • num2 represents the maximum number of plays
  • R1 represents the maximum value of the first popularity parameter interval.
  • R1 is a preset value.
  • R1 is 0.1, which means that the heat parameter interval is divided into 10 levels.
  • a popularity parameter Q1 is the popularity level of the video clip.
  • the second popularity parameter can be determined based on at least one of the number of barrages or the number of likes on barrages included in the video clip, such as the second popularity parameter and the number of barrages or the number of likes on barrages included in the video clip. at least one positive correlation.
  • the video clip may be based on at least one of the number of barrages or the number of barrage likes included in the video clip, and the number of barrage likes among the multiple video clips of the computer device. At least one of the maximum number of barrages or the maximum number of likes on barrages determines the second popularity parameter.
  • the number of likes of the barrage of the video clip is the sum of the number of likes of the barrage of the video clip or the maximum value of the number of likes of the barrage of the video clip.
  • Q2 represents the second popularity parameter
  • num3 represents the number of barrages included in the video clip
  • num4 represents the number of barrage likes of the video clip
  • num5 represents the maximum number of barrages in multiple video clips, that is, the one containing the most barrages
  • num6 represents the maximum number of barrage likes in multiple video clips, that is, the maximum number of barrage likes in multiple video clips
  • R2 represents the maximum value of the second popularity parameter interval.
  • R2 is a preset value. For example, R2 is 0.1, which means that the second heat parameter interval is divided into 10 levels.
  • the second popularity parameter Q2 determined by /R2 is the barrage popularity level of the video clip.
  • the computer device determines a voting identifier of the video clip based on the first text content, the second text content, the first popularity parameter, and the second popularity parameter.
  • the voting identifier indicates whether to generate voting information for the video clip.
  • the first text content and the second text content represent the content of the video clip
  • the first popularity parameter and the second popularity parameter represent the popularity of the video clip.
  • the first popularity parameter and the second popularity parameter determine the voting identification of the video clip.
  • the content and popularity of the video clip are comprehensively considered.
  • the determined voting identification can better reflect whether the average user will participate in the voting interaction of the video clip and whether it will be based on
  • the voting information is used to perform a voting operation to ensure that the determined voting identification can accurately represent whether voting information is to be generated for the video clip.
  • step 1002 includes:
  • the first text feature of the first text content and the second text feature of the second text content are used to describe the first text content and the second text content respectively, and then obtain the first text feature.
  • the first heat feature of the heat parameter and the second heat feature of the second heat parameter are used to describe the first heat parameter and the second heat parameter respectively.
  • the first text feature is , the second text feature, the first popularity feature and the second popularity feature are spliced to obtain the video clip features, so that the video clip features can accurately describe the video clip, then the video clip features are classified based on the video clip features to obtain the voting identification.
  • the computer device may first determine a heat feature table, where the heat feature table includes heat features corresponding to at least one heat parameter.
  • the heat feature table includes a first heat feature table and a second heat feature table.
  • the first heat feature table contains heat features of the heat parameters of the video clips.
  • the first heat feature table can be determined by querying the first heat feature table.
  • the first popularity feature and the second popularity feature table include the popularity feature of the popularity parameter of the barrage of the video clip.
  • the second popularity feature of the second popularity parameter can be determined by querying the second popularity feature table.
  • the computer device When the voting identifier indicates that the voting information is generated for the video clip, the computer device generates a voting topic for the video clip based on the keywords in the text content, and generates multiple voting candidates for the video clip based on the keywords and the voting topic. Based on the voting topic and multiple voting candidates, the voting information of the video clip is generated.
  • the voting identification includes a first voting identification and a second voting identification.
  • the first voting identification indicates that voting information is generated for the video clip
  • the second voting identification indicates that voting information is not generated for the video clip.
  • the step of generating voting information is performed.
  • the first voting ID is 1 and the second voting ID is 0.
  • the voting identifier is a probability of generating voting information. The determined probability is greater than the probability threshold, indicating that voting information is generated for the video clip. The determined probability is not greater than the probability threshold, indicating that voting information is not generated for the video clip.
  • the step of generating the voting information of the video clip includes the steps of generating a voting topic, generating a plurality of voting candidates, and generating voting information based on the video topic and the voting candidates.
  • the specific process can be seen in the above embodiments and will not be described again here.
  • the method provided by the embodiment of the present application determines the voting identifier of the video clip based on the characteristics corresponding to the first text content, the second text content, the first popularity parameter, and the second popularity parameter, taking into account the content and popularity of the video clip, that is, It fully captures the content characteristics of the video clip and also takes into account the interaction data of the video clip.
  • the determined voting identification can better reflect whether the general user will participate in the voting interaction of the video clip and whether the voting operation will be based on the voting information, thus effectively It can accurately determine whether to generate voting information for this video clip, avoiding the need to directly generate voting information for each video clip. On the basis of improving the interactivity of the video, it also saves processing and reduces the amount of video data.
  • the process of determining the voting identification can be performed based on the voting decision model.
  • the voting decision model includes a first feature extraction sub-model, a first splicing layer and a second classification layer.
  • the voting decision model also It includes a heat feature table, and the heat feature table includes heat features corresponding to at least one heat parameter.
  • the heat feature table includes the above-mentioned first heat feature table and second heat feature table.
  • the process of determining voting identification includes:
  • the first feature extraction sub-model is a BERT (Bidirectional Encoder Representations from Transformers) model or other types of models.
  • multiple barrage text contents can be obtained, thereby obtaining the barrage text features of the multiple barrage text contents, and the multiple barrage text features can be obtained.
  • Maximum pooling is performed to obtain the second text feature, and then the second text feature is spliced with the first text feature, the first popularity feature, and the second popularity feature to obtain the video clip feature.
  • FIG 11 is a schematic diagram of a voting decision model provided by the embodiment of the present application.
  • the sub-model is the BERT model as an example.
  • the voting decision model includes the BERT model, the first splicing layer and the second classification layer.
  • the computer device obtains the dialogue text content and subtitle text content of the video clip, as well as the first popularity parameter, and also obtains the K barrage text content of the video clip and the second popularity parameter, where K is an integer greater than 1.
  • the BERT model is used to extract the first text features corresponding to the dialogue text content and subtitle text content, as well as the barrage text features corresponding to K barrage text contents, and perform maximum pooling on the K barrage text features to obtain the second text feature.
  • the voting decision model also includes a first heat feature table and a second heat feature table, and the first heat feature and the second heat feature are determined respectively by querying the first heat feature table and the second heat feature table.
  • the obtained features can then be spliced and classified to determine the voting identifier.
  • the training process of the voting decision model includes:
  • the sample video clip includes at least one of a positive sample video clip or a negative sample video clip.
  • the positive sample video clip contains sample voting information and the sample Video clips whose participation rate of voting information reaches the target threshold.
  • Negative sample video clips are: video clips that contain sample voting information and the participation rate of sample voting information does not reach the target threshold, or video clips that do not contain sample voting information. ;
  • the model parameters include a popularity feature table.
  • the voting decision model is trained, which enables the voting decision model to learn the impact of the text content and popularity parameters of the video clips on whether to generate voting information, thereby improving the voting decision.
  • the accuracy of the model can accurately determine the voting identification based on the voting determination model, avoiding the need to directly generate voting information for each video segment of the video.
  • it also saves processing and reduces the data of the video. quantity.
  • the training objectives of the voting determination model are: the voting identification of the positive sample video clips obtained by the voting determination model represents the voting information generated for the positive sample video clips, and the voting identification of the negative sample video clips obtained by the voting determination model Indicates that no voting information will be generated for this negative sample video clip.
  • the heat feature table is also used as one of the model parameters of the voting decision model.
  • the heat feature table will be adjusted based on the training samples so that the heat feature table changes with the voting decision model. Gradually update through training to improve the accuracy of the hot feature table.
  • Figure 12 is a flow chart of a voting information display method provided by an embodiment of the present application.
  • the execution subject of the embodiment of the present application is a terminal.
  • the embodiment of the present application describes the process of generating the voting information of the video clip and then displaying the voting information. Referring to Figure 12, the method includes:
  • the terminal obtains the voting information of the video clip based on the video clip.
  • the voting information is generated based on the voting topic and multiple voting candidates
  • the voting topic is generated based on the keywords in the text content associated with the video clip
  • the multiple voting candidates are generated based on the keywords and voting topics.
  • the text content includes at least one of first text content or second text content.
  • the first text content is the text content contained in the video clip
  • the second text content is the text content contained in the barrage of the video clip.
  • the voting information may be generated by the terminal or by other devices other than the terminal.
  • a server used to provide the video generates voting information for one or more video clips.
  • the terminal requests the video from the server, the server sends the video and the generated voting information to the terminal, and the terminal can Get the video to be played and the voting information of one or more video clips.
  • the terminal determines interaction parameters based on the interest tag and voting information of the currently logged-in account.
  • the interaction parameters represent the possibility of the account performing a voting operation based on the voting information.
  • the terminal is currently logged in with an account, and the interest tag of the account indicates the interest of the user to whom the account belongs.
  • the terminal determines the interaction parameters based on the interest tag and the voting information, and can accurately determine whether the current user is interested in the voting information, thus ensuring that all The determined interaction parameters can accurately measure the likelihood of the current user participating in voting and whether there is a need to generate voting information for this video clip.
  • the terminal displays voting information when playing video clips.
  • the interaction condition refers to the condition for displaying voting information when playing video clips.
  • the interaction parameters do not meet the interaction conditions, it can be considered that the current user is not interested in the voting information and the probability of participating in voting is low. At this time, in order to avoid causing interference to the playback process of the video clip, do not play the video clip. Show voting information again.
  • the interaction condition includes an interaction parameter threshold. If the interaction parameter is greater than the interaction parameter threshold, the voting information is displayed when the video clip is played. When the interaction parameter is not greater than the interaction parameter threshold, the voting information will not be displayed when the video clip is played.
  • the embodiment of this application only takes one video clip as an example to illustrate the display process of voting information of this video clip, and the video includes multiple video clips, and the multiple video information has voting information.
  • each video clip is played in sequence according to the playback order of each video clip.
  • the voting information corresponding to the video clip determines the interaction parameters.
  • the voting information is displayed. Otherwise, playback continues until the next video clip is played.
  • a video clip has multiple voting information, and the voting information whose corresponding interaction parameters satisfy the interaction conditions is selected from them and displayed when the video clip is played. Or, when there are multiple voting information whose corresponding interaction parameters satisfy the interaction conditions, select the voting information with the largest interaction parameter and display it when the video clip is played.
  • the method provided by the embodiment of this application determines the interaction parameters that can measure the possibility of the user participating in voting based on the interest tags and voting information of the currently logged-in account. Only when the interaction parameters meet the interaction conditions will they be displayed when playing the video clip. Voting information makes it easier to attract users to participate in voting, realizes personalized display of voting information, enhances users' active participation in interaction, improves interactive experience, and avoids interference to users with a low probability of participating in voting.
  • FIG. 13 is a flow chart of another voting information display method provided by an embodiment of the present application.
  • the execution subject of the embodiment of this application is a terminal. Referring to Figure 13, the method includes:
  • the terminal obtains the voting information of the video clip based on the video clip.
  • the terminal obtains the interest characteristics of the account's interest tag, the voting topic characteristics of the voting topic, and the voting candidate characteristics of multiple voting candidates.
  • the interest feature is used to describe the interest tag
  • the voting topic feature is used to describe the voting subject
  • the voting candidate feature is used to describe the voting candidate.
  • each of the above features can be in the form of vectors, matrices or other forms.
  • the terminal splices the interest features, voting topic features and multiple voting candidate features to obtain interactive features.
  • the terminal classifies based on interaction features and obtains interaction parameters.
  • the interaction feature is spliced by the characteristics of the interest tag, the characteristics of the voting topic and the characteristics of multiple voting candidates, and integrates the information of the interest tag, voting topic and multiple voting candidates
  • the interaction obtained by classifying based on the interaction parameters Parameters can consider information about interest tags, voting topics, and multiple voting candidates, and can more accurately measure the user's interest in voting information, thereby accurately determining the user's likelihood of participating in voting.
  • the terminal displays voting information when playing video clips.
  • This step is similar to the above-mentioned step 1203 and will not be described again.
  • the method provided by the embodiment of this application determines the interaction parameters that can measure the possibility of the user participating in voting based on the interest tags and voting information of the currently logged-in account. Only when the interaction parameters meet the interaction conditions will they be displayed when playing the video clip. Voting information, thereby attracting users to participate in voting, realizing personalized display of voting information, and improving interactive coverage. Coverage rate also avoids causing interference to users with a low probability of participating in voting.
  • the process of determining interaction parameters in the above steps 1302-1304 can be performed based on a voting interaction model, which includes a second feature extraction sub-model, a second splicing layer, and a third classification layer. Accordingly, the process of determining interaction parameters includes:
  • the second feature extraction sub-model to obtain the interest features of the interest tag, the voting topic features of the voting topic, and the voting candidate features corresponding to multiple voting candidates; call the second splicing layer to combine the interest features, voting topic features, and multiple voting candidates.
  • the features of voting candidates are spliced together to obtain interactive features; the third classification layer is called to classify based on the interactive features to obtain interactive parameters.
  • the interest feature, the voting topic feature and the multiple voting candidate features are respectively used to describe the current user's interest tag, voting topic and multiple voting candidates, and the interest feature, voting topic feature and multiple voting candidate features can be integrated Taking into account the influence of the current user's interest tags, voting topics and multiple voting candidates, the interaction parameters are obtained through classification based on interaction features, which improves the accuracy of the interaction parameters.
  • the second feature extraction sub-model is a BERT model or other types of models, which is not limited in the embodiments of the present application.
  • FIG 14 is a schematic diagram of a voting interaction model provided by an embodiment of the present application.
  • the voting interaction model includes the BERT model, the second splicing layer and the third classification layer.
  • the computer device obtains the account's interest tag, voting topic, and M voting candidates, where M is an integer greater than 1.
  • the BERT model is used to extract the features corresponding to the interest tags, voting topics and M voting candidates, and then splicing and classifying are performed through the second splicing layer and the third classification layer respectively to obtain the interaction parameters.
  • the training process of the voting interaction model includes:
  • the sample account has performed a voting operation based on the sample voting information; based on the sample interest tag, the sample voting topic and multiple sample voting candidates in the sample voting information Term, adjust the model parameters in the voting interaction model.
  • the sample account is any account
  • the sample interest tag indicates the interest of the user to whom the sample account belongs
  • the sample video clip includes sample voting information
  • the sample account has performed a voting operation based on the sample voting information, indicating that the sample video is being played
  • the sample voting information is displayed in the clip
  • the user to whom the sample account belongs is interested in the sample voting information and has participated in the voting interaction. Therefore, adjusting the model parameters in the voting interaction model based on the sample interest labels, the sample voting topics in the sample voting information, and multiple sample voting candidates can enable the voting interaction model to learn the association between the sample interest labels and the sample voting information. relationship, thus having the function of determining the corresponding interaction parameters based on any interest tag and any voting information, which improves the accuracy of the voting interaction model.
  • the training goal of the voting interaction model is to increase the interaction parameters obtained by the voting interaction model, that is, to enable the voting interaction model to predict based on sample interest labels, sample voting topics and multiple sample voting candidates.
  • This sample account is more likely to conduct voting operations.
  • the interaction parameter represents the probability of the account performing a voting operation based on voting information
  • the training goal is to make the interaction parameter obtained by the voting interaction model equal to 1.
  • each set of sample data includes sample interest tags corresponding to the sample account and samples in the sample video clips.
  • the voting interaction model is iteratively trained based on the multiple sets of sample data, thereby improving the accuracy of the voting interaction model, and thereby improving the accuracy of the interaction parameters determined by the voting interaction model.
  • FIG 15 is a schematic structural diagram of a voting information generation device provided by an embodiment of the present application.
  • the device is installed in a computer device.
  • the device includes:
  • the text content acquisition module 1501 is used to acquire text content associated with video clips in the video.
  • the text content includes at least one of first text content or second text content.
  • the first text content is the text content contained in the video clip
  • the second text content is the text content contained in the video clip.
  • the text content is the text content contained in the barrage of the video clip;
  • the topic generation module 1502 is used to generate voting topics for video clips based on keywords in the text content
  • the candidate generation module 1503 is used to generate multiple voting candidates for video clips based on keywords and voting topics;
  • the voting information generation module 1504 is used to generate voting information for video clips based on the voting topic and multiple voting candidates.
  • the topic generation module 1502 includes:
  • the encoding unit is used to encode keywords and obtain the keyword characteristics of the keywords
  • the decoding unit is used to decode the keyword features to obtain the voting topic.
  • the voting topic is composed of multiple voting topic words.
  • the first generation model includes a first encoding sub-model and a first decoding sub-model
  • the encoding unit is used to: call the first encoding sub-model, encode the N keywords, and obtain the keyword features of the N keywords, where N is an integer greater than 1;
  • Decoding unit for:
  • the first decoding sub-model to decode the N keyword features and the first voting topic words to obtain the second decoding feature. Based on the second decoding feature and the N keyword features, determine the reference voting topic and the reference voting topic. Including the first voting topic words and the second voting topic words, until after N decodings, the reference voting topic obtained by the Nth decoding is determined as the voting topic of the video clip.
  • the first generation model also includes a first classification layer and a preset word library.
  • the preset word library includes a plurality of words and a decoding unit for:
  • the jth first usage probability is the probability of using the jth keyword in the voting topic. j is a positive integer and j is not greater than N;
  • the jth keyword is determined as the first voting topic word
  • the first classification layer is called to classify based on the first decoding feature and the preset word library, and the classification probability of each word in the preset word library is obtained. Based on each The classification probability of each word determines the first voting topic word.
  • a decoding unit is used for:
  • the jth second usage probability is the probability of using the jth keyword in the voting topic.
  • j is a positive integer, and j is not greater than N;
  • the j-th keyword is determined as the second voting topic word
  • the first classification layer is called to classify based on the second decoding feature and the preset word library to obtain the classification probabilities of multiple candidate voting topics.
  • Each candidate voting topic includes One first voting topic word and one second voting topic word;
  • the reference voting topic is determined.
  • the training process of the first generation model includes:
  • the positive sample video clip is a video clip that contains sample voting information and the participation rate of the sample voting information reaches the target threshold;
  • model parameters in the first generative model are adjusted.
  • the text content includes first text content and second text content
  • the candidate generation module 1503 includes:
  • the first acquisition unit is used to acquire the first keyword, which is a keyword in the first text content
  • a clustering unit is used to cluster the second text content to obtain multiple text categories, each text category containing at least one piece of second text content;
  • the second acquisition unit is used to extract the second keyword from each text category respectively;
  • the generation unit is used to respectively generate each keyword based on the first keyword, the voting topic and the second keyword of each text category. Voting candidates for text categories.
  • the keyword features are decoded to obtain the i-th voting candidate, which consists of multiple voting candidate words.
  • the second generation model includes a second encoding sub-model and a second decoding sub-model; a generation unit, used for:
  • the second decoding sub-model is called to decode the M keyword features and the first voting candidate words to obtain the second decoding feature. Based on the second decoding feature and the M keyword features, the reference voting candidates are determined. Refer to The voting candidates include the first voting candidate words and the second voting candidate words, until after M times of decoding, the reference voting candidate obtained by the M-th decoding is determined as the i-th voting candidate of the video clip.
  • the training process of the second generative model includes:
  • the positive sample video clip is a video clip that contains sample voting information and the participation rate of the sample voting information reaches the target threshold;
  • each sample text content Based on the correlation between each sample text content and each sample voting candidate, determine the text category of each sample voting candidate, where the text category includes the sample text content associated with the sample voting candidate;
  • the model parameters in the second generation model are adjusted.
  • the text content acquisition module 1501 is used for:
  • the first popularity parameter represents the popularity of the video clip
  • the second popularity parameter represents the popularity of the barrage of the video clip
  • Topic generation module 1502 used for:
  • the voting identification Based on the first text content, the second text content, the first popularity parameter and the second popularity parameter, determine the voting identification of the video clip, where the voting identification indicates whether to generate voting information for the video clip;
  • a voting topic of the video clip is generated based on the keywords in the obtained text content.
  • topic generation module 1502 is used for:
  • the voting decision model includes a first feature extraction layer, a first splicing layer and a second classification layer, the voting decision model also includes a heat feature table, the heat feature table includes heat features of at least one heat parameter;
  • Topic generation module 1502 used for:
  • the second classification layer is called to classify based on the characteristics of the video clips and obtain the voting identification.
  • the training process of the voting decision model includes:
  • the sample video clip includes at least one of a positive sample video clip or a negative sample video clip.
  • the positive sample video clip contains sample voting information and the sample Video clips whose participation rate of voting information reaches the target threshold.
  • Negative sample video clips are: video clips that contain sample voting information and the participation rate of sample voting information does not reach the target threshold, or video clips that do not contain sample voting information. ;
  • the model parameters in the voting decision model include a popularity feature table.
  • the device also includes:
  • the interaction parameter determination module is used to determine the interaction parameters based on the account's interest tags and voting information.
  • the interaction parameters represent the possibility of the account performing voting operations based on the voting information;
  • the sending module is used to send voting information to the terminal of the account when the interaction parameters meet the interaction conditions.
  • the voting information is used to display when the terminal plays the video clip.
  • FIG 16 is a schematic structural diagram of a voting information display device provided by an embodiment of the present application.
  • the device includes:
  • the information acquisition module 1601 is used to obtain the voting information of the video clip based on the video clip in the video.
  • the voting information is generated based on the voting topic and multiple voting candidates.
  • the voting topic is generated based on the keywords in the text content associated with the video clip, and more Voting candidates are generated based on keywords and voting topics;
  • the parameter determination module 1602 is used to determine interaction parameters based on the interest tags and voting information of the currently logged-in account.
  • the interaction parameters represent the possibility of the account performing voting operations based on the voting information;
  • the text content includes at least one of first text content or second text content.
  • the first text content is the text content contained in the video clip
  • the second text content is the text content contained in the barrage of the video clip.
  • the voting information includes a voting topic and multiple voting candidates.
  • the parameter determination module 1602 includes:
  • the feature acquisition unit is used to acquire the interest features of the account's interest tag, the voting topic features of the voting topic, and the voting candidate features of multiple voting candidates;
  • the splicing unit is used to splice interest features, voting topic features and multiple voting candidate features to obtain interactive features;
  • Classification unit is used to classify based on interaction features and obtain interaction parameters.
  • the voting interaction model includes a second feature extraction sub-model, a second splicing layer and a third classification layer;
  • the feature acquisition unit is used for: calling the second feature extraction sub-model to obtain the interest features of the interest tag, the voting topic features of the voting topic, and the voting candidate features of multiple voting candidates;
  • the splicing unit is used to: call the second splicing layer to splice the interest features, voting topic features and multiple voting candidate features to obtain interactive features;
  • Classification unit is used to: call the third classification layer, classify based on interaction features, and obtain interaction parameters.
  • the training process of the voting interaction model includes:
  • sample voting topics in the sample voting information Based on the sample interest tags, sample voting topics in the sample voting information, and multiple sample voting candidates, adjust the model parameters in the voting interaction model.
  • voting information generation device or the voting information display device provided in the above embodiments generates or displays voting information
  • only the division of the above functional modules is used as an example. In practical applications, the above functions can be used as needed.
  • the above function allocation is completed by different functional modules, that is, the internal structure of the computer device is divided into different functional modules to complete all or part of the functions described above.
  • the voting information generation device and the voting information generation method embodiments provided by the above embodiments belong to the same concept.
  • the voting information display device and the voting information display method embodiments provided by the above embodiments belong to the same concept.
  • we won’t go into details here.
  • Embodiments of the present application also provide a computer device.
  • the computer device includes a processor and a memory. At least one computer program is stored in the memory. The at least one computer program is loaded and executed by the processor to implement the voting information of the above embodiments. The operation performed by the generate method or voting information display method.
  • the computer device is provided as a terminal.
  • Figure 17 shows a schematic structural diagram of a terminal 1700 provided by an exemplary embodiment of the present application.
  • the terminal 1700 includes: a processor 1701 and a memory 1702.
  • the processor 1701 may include one or more processing cores, such as a 4-core processor, an 8-core processor, etc.
  • the processor 1701 can be implemented using at least one hardware form among DSP (Digital Signal Processing, digital signal processing), FPGA (Field Programmable Gate Array, field programmable gate array), and PLA (Programmable Logic Array, programmable logic array).
  • the processor 1701 can also include a main processor and a co-processor.
  • the main processor is a processor used to process data in the wake-up state, also called CPU (Central Processing Unit, central processing unit); the co-processor is A low-power processor used to process data in standby mode.
  • the processor 1701 may be integrated with a GPU (Graphics Processing Unit, an image processing interface), and the GPU is responsible for rendering and drawing the content that needs to be displayed on the display screen.
  • the processor 1701 may also include an AI (Artificial Intelligence, artificial intelligence) processor, which is used to process computing operations related to machine learning.
  • AI Artificial Intelligence, artificial intelligence
  • Memory 1702 may include one or more computer-readable storage media, which may be non-transitory. Memory 1702 may also include high-speed random access memory, as well as non-volatile memory, such as one or more disk storage devices, flash memory storage devices. In some embodiments, the non-transitory computer-readable storage medium in the memory 1702 is used to store at least one computer program, and the at least one computer program is used to be possessed by the processor 1701 to implement the methods provided by the method embodiments in this application. Voting information generation method.
  • the terminal 1700 optionally further includes: a peripheral device interface 1703 and at least one peripheral device.
  • the processor 1701, the memory 1702 and the peripheral device interface 1703 may be connected through a bus or a signal line.
  • Each peripheral device can be connected to the peripheral device interface 1703 through a bus, a signal line, or a circuit board.
  • the peripheral device includes: at least one of a radio frequency circuit 1704, a display screen 1705, and a camera assembly 1706.
  • the peripheral device interface 1703 may be used to connect at least one I/O (Input/Output) related peripheral device to the processor 1701 and the memory 1702 .
  • the processor 1701, the memory 1702, and the peripheral device interface 1703 are integrated on the same chip or circuit board; in some other embodiments, any one of the processor 1701, the memory 1702, and the peripheral device interface 1703 or Both of them can be implemented on separate chips or circuit boards, which is not limited in this embodiment.
  • the radio frequency circuit 1704 is used to receive and transmit RF (Radio Frequency, radio frequency) signals, also called electromagnetic signals.
  • Radio frequency circuitry 1704 communicates with communication networks and other communication devices through electromagnetic signals.
  • the radio frequency circuit 1704 converts electrical signals into electromagnetic signals for transmission, or converts received electromagnetic signals into electrical signals.
  • the radio frequency circuit 1704 includes: an antenna system, an RF transceiver, one or more amplifiers, a tuner, an oscillator, a digital signal processor, a codec chipset, a user identity module card, and the like.
  • Radio frequency circuitry 1704 can communicate with other devices through at least one wireless communication protocol.
  • the wireless communication protocol includes but is not limited to: metropolitan area network, mobile communication networks of all generations (2G, 3G, 4G and 5G), wireless LAN and/or WiFi (Wireless Fidelity, wireless fidelity) network.
  • the radio frequency circuit 1704 may also include NFC (Near Field Communication) related circuits, which is not limited in this application.
  • the display screen 1705 is used to display UI (User Interface, user interface).
  • the UI can include graphics, text, icons, videos, and any combination thereof.
  • display screen 1705 also has the ability to collect touch signals on or above the surface of display screen 1705 .
  • the touch signal can be input to the processor as a control signal 1701 for processing.
  • the display screen 1705 can also be used to provide virtual buttons and/or virtual keyboards, also called soft buttons and/or soft keyboards.
  • the display screen 1705 there may be one display screen 1705, which is provided on the front panel of the terminal 1700; in other embodiments, there may be at least two display screens 1705, which are respectively provided on different surfaces of the terminal 1700 or have a folding design; In other embodiments, the display screen 1705 may be a flexible display screen disposed on a curved surface or a folding surface of the terminal 1700. Even, the display screen 1705 can be set into a non-rectangular irregular shape, that is, a special-shaped screen.
  • the display screen 1705 can be made of materials such as LCD (Liquid Crystal Display) and OLED (Organic Light-Enitting Diode).
  • the camera component 1706 is used to capture images or videos.
  • camera assembly 1706 includes a front camera and a rear camera.
  • the front camera is installed on the front panel of the terminal 1700
  • the rear camera is installed on the back of the terminal 1700 .
  • there are at least two rear cameras one of which is a main camera, a depth-of-field camera, a wide-angle camera, and a telephoto camera, so as to realize the integration of the main camera and the depth-of-field camera to realize the background blur function.
  • camera assembly 1706 may also include a flash.
  • the flash can be a single color temperature flash or a dual color temperature flash. Dual color temperature flash refers to a combination of warm light flash and cold light flash, which can be used for light compensation under different color temperatures.
  • FIG. 17 does not constitute a limitation on the terminal 1700, and may include more or fewer components than shown, or combine certain components, or adopt different component arrangements.
  • the computer device is provided as a server.
  • Figure 18 is a schematic structural diagram of a server provided by an embodiment of the present application.
  • the server 1800 may vary greatly due to different configurations or performance, and may include one or more processors (Central Processing Units, CPU) 1801 and a Or one or more memories 1802, wherein at least one computer program is stored in the memory 1802, and the at least one computer program is loaded and executed by the processor 1801 to implement the methods provided by the above method embodiments.
  • the server may also have components such as wired or wireless network interfaces and input and output interfaces to facilitate input and output.
  • the server may also include other components for implementing device functions, which will not be described in detail here.
  • Embodiments of the present application also provide a computer-readable storage medium.
  • the computer-readable storage medium stores at least one computer program.
  • the at least one computer program is loaded and executed by the processor to implement the voting information generation of the above embodiments. The operation performed by the method.
  • Embodiments of the present application also provide a computer program product.
  • the computer program product includes a computer program.
  • the operations performed by the voting information generation method or the voting information display method of the above embodiments are implemented. operate.
  • the computer program involved in the embodiments of the present application may be deployed and executed on one computer device, or executed on multiple computer devices located in one location, or distributed in multiple locations and communicated through Executed on multiple computer devices interconnected by a network, multiple computer devices distributed in multiple locations and interconnected through a communication network can form a blockchain system.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Databases & Information Systems (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Computational Linguistics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Data Mining & Analysis (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

A voting information generation method and apparatus, and a voting information display method and apparatus, which belong to the technical field of computers. The voting information generation method comprises: acquiring text content associated with a video clip in a video, wherein the text content comprises at least one of first text content or second text content, the first text content is text content that is included in the video clip, and the second text content is text content that is included in bullet-screen comments of the video clip (201); generating a voting theme of the video clip on the basis of keywords in the text content (202); generating a plurality of voting candidate items of the video clip on the basis of the keywords and the voting theme (203); and generating voting information of the video clip on the basis of the voting theme and the plurality of voting candidate items (204). Provided in the embodiments of the present application is a method for automatically generating voting information of a video clip. It is not necessary to manually create voting information, thereby improving the operation efficiency and saving on time.

Description

投票信息生成方法、投票信息显示方法及装置Voting information generation method, voting information display method and device
本申请要求于2022年04月29日提交、申请号为202210473861.6、发明名称为“投票信息生成方法、投票信息显示方法及装置”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims priority to the Chinese patent application submitted on April 29, 2022, with the application number 202210473861.6 and the invention title "Voting Information Generation Method, Voting Information Display Method and Device", the entire content of which is incorporated into this application by reference. middle.
技术领域Technical field
本申请实施例涉及计算机技术领域,特别涉及一种投票信息生成方法、投票信息显示方法及装置。The embodiments of the present application relate to the field of computer technology, and in particular to a voting information generation method, voting information display method and device.
背景技术Background technique
投票是一种常用的互动方式,广泛应用于各种场景中。而随着视频的广泛传播,目前提供了一种在视频中进行投票的方法,由视频网站的运营人员或者视频的制作者在视频中人为创建投票信息,在播放视频的过程中通过显示该投票信息吸引观看视频的对象进行投票。但是,这种方式需要人工在视频中创建投票信息,操作效率很低,且这种方式难以覆盖大量的视频,导致视频的投票互动不够充分。Voting is a common interaction method that is widely used in various scenarios. With the widespread spread of videos, a method of voting in videos is currently provided. The operator of the video website or the producer of the video artificially creates voting information in the video, and displays the vote during the video playback process. The message entices those who watch the video to vote. However, this method requires manual creation of voting information in the video, which is very inefficient. Moreover, this method is difficult to cover a large number of videos, resulting in insufficient voting interaction in the video.
发明内容Contents of the invention
本申请实施例提供了一种投票信息生成方法、投票信息显示方法及装置,无需人工创建投票信息,提高了操作效率,节省了时间,且这种方法能够有效覆盖大量的视频,提高了投票互动的覆盖率。所述技术方案如下:The embodiments of this application provide a voting information generation method, voting information display method and device, which eliminates the need to manually create voting information, improves operating efficiency, saves time, and this method can effectively cover a large number of videos and improve voting interaction. coverage. The technical solutions are as follows:
一方面,提供了一种投票信息生成方法,所述方法包括:On the one hand, a voting information generation method is provided, and the method includes:
计算机设备获取视频中的视频片段关联的文本内容,所述文本内容包括第一文本内容或第二文本内容的至少一种,所述第一文本内容为所述视频片段包含的文本内容,所述第二文本内容为所述视频片段的弹幕包含的文本内容;The computer device obtains text content associated with a video clip in the video, where the text content includes at least one of first text content or second text content, where the first text content is the text content included in the video clip, and the The second text content is the text content contained in the barrage of the video clip;
所述计算机设备基于所述文本内容中的关键词,生成所述视频片段的投票主题;The computer device generates a voting topic for the video clip based on keywords in the text content;
所述计算机设备基于所述关键词和所述投票主题,生成所述视频片段的多个投票候选项;The computer device generates a plurality of voting candidates for the video clip based on the keywords and the voting topic;
所述计算机设备基于所述投票主题和所述多个投票候选项,生成所述视频片段的投票信息。The computer device generates voting information for the video clip based on the voting topic and the plurality of voting candidates.
另一方面,提供了一种投票信息显示方法,所述方法包括:On the other hand, a voting information display method is provided, and the method includes:
计算机设备基于视频中的视频片段,获取所述视频片段的投票信息,所述投票信息基于投票主题和多个投票候选项生成,所述投票主题基于所述视频片段关联的文本内容中的关键词生成,所述多个投票候选项基于所述关键词和所述投票主题生成;The computer device obtains voting information of the video clip based on the video clip, the voting information is generated based on a voting topic and a plurality of voting candidates, and the voting topic is based on keywords in text content associated with the video clip. Generating, the plurality of voting candidates are generated based on the keywords and the voting topic;
所述计算机设备基于当前登录的账号的兴趣标签和所述投票信息,确定互动参数,所述互动参数表示所述账号基于所述投票信息进行投票操作的可能性;The computer device determines interaction parameters based on the interest tag of the currently logged-in account and the voting information, where the interaction parameters represent the possibility of the account performing a voting operation based on the voting information;
所述计算机设备在所述互动参数满足互动条件的情况下,在播放所述视频片段时显示所述投票信息;The computer device displays the voting information when playing the video clip when the interaction parameters meet the interaction conditions;
其中,所述文本内容包括第一文本内容或第二文本内容的至少一种,所述第一文本内容为所述视频片段包含的文本内容,所述第二文本内容为所述视频片段的弹幕包含的文本内容。Wherein, the text content includes at least one of first text content or second text content, the first text content is the text content contained in the video clip, and the second text content is the bounce of the video clip. The text content contained in the scene.
另一方面,提供了一种投票信息生成装置,设置于计算机设备中,所述装置包括:On the other hand, a voting information generating device is provided, which is provided in a computer device, and the device includes:
文本内容获取模块,用于获取视频中的视频片段关联的文本内容,所述文本内容包括第一文本内容或第二文本内容的至少一种,所述第一文本内容为所述视频片段包含的文本内容, 所述第二文本内容为所述视频片段的弹幕包含的文本内容;A text content acquisition module, configured to acquire text content associated with video clips in the video, where the text content includes at least one of first text content or second text content, and the first text content is the text content contained in the video clip. text content, The second text content is the text content contained in the barrage of the video clip;
主题生成模块,用于基于所述文本内容中的关键词,生成所述视频片段的投票主题;A topic generation module, configured to generate voting topics for the video clips based on keywords in the text content;
候选项生成模块,用于基于所述关键词和所述投票主题,生成所述视频片段的多个投票候选项;A candidate generation module, configured to generate multiple voting candidates for the video clip based on the keywords and the voting topic;
投票信息生成模块,用于基于所述投票主题和所述多个投票候选项,生成所述视频片段的投票信息。A voting information generation module, configured to generate voting information for the video clip based on the voting topic and the plurality of voting candidates.
另一方面,提供了一种投票信息显示装置,所述装置包括:On the other hand, a voting information display device is provided, and the device includes:
信息获取模块,用于基于目标视频中的视频片段,获取所述视频片段的投票信息,所述投票信息基于投票主题和多个投票候选项生成,所述投票主题基于所述视频片段关联的文本内容中的关键词生成,所述多个投票候选项基于所述关键词和所述投票主题生成;An information acquisition module, configured to obtain voting information of video clips based on video clips in the target video, where the voting information is generated based on a voting topic and a plurality of voting candidates, and the voting topic is based on text associated with the video clip. Keywords in the content are generated, and the plurality of voting candidates are generated based on the keywords and the voting topic;
参数确定模块,用于基于当前登录的账号的兴趣标签和所述投票信息,确定互动参数,所述互动参数表示所述账号基于所述投票信息进行投票操作的可能性;A parameter determination module, configured to determine interaction parameters based on the interest tag of the currently logged-in account and the voting information, where the interaction parameters represent the possibility of the account performing a voting operation based on the voting information;
信息显示模块,用于在所述互动参数满足互动条件的情况下,在播放所述视频片段时显示所述投票信息;An information display module, configured to display the voting information when the video clip is played when the interaction parameters meet the interaction conditions;
其中,所述文本内容包括第一文本内容或第二文本内容的至少一种,所述第一文本内容为所述视频片段包含的文本内容,所述第二文本内容为所述视频片段的弹幕包含的文本内容。Wherein, the text content includes at least one of first text content or second text content, the first text content is the text content contained in the video clip, and the second text content is the bounce of the video clip. The text content contained in the scene.
另一方面,提供了一种计算机设备,所述计算机设备包括处理器和存储器,所述存储器中存储有至少一条计算机程序,所述至少一条计算机程序由所述处理器加载并执行以实现如上述方面所述的投票信息生成方法所执行的操作,或者以实现如上述方面所述的投票信息显示方法所执行的操作。On the other hand, a computer device is provided. The computer device includes a processor and a memory. At least one computer program is stored in the memory. The at least one computer program is loaded and executed by the processor to implement the above. The operations performed by the voting information generation method described in the aspect, or to implement the operations performed by the voting information display method described in the above aspect.
另一方面,提供了一种计算机可读存储介质,所述计算机可读存储介质中存储有至少一条计算机程序,所述至少一条计算机程序由处理器加载并执行以实现如上述方面所述的投票信息生成方法所执行的操作,或者以实现如上述方面所述的投票信息显示方法所执行的操作。On the other hand, a computer-readable storage medium is provided. At least one computer program is stored in the computer-readable storage medium. The at least one computer program is loaded and executed by a processor to implement voting as described in the above aspect. The operations performed by the information generation method, or the operations performed by the voting information display method as described above.
另一方面,提供了一种计算机程序产品,包括计算机程序,所述计算机程序被处理器执行时实现上述方面所述的投票信息生成方法所执行的操作,或者实现如上述方面所述的投票信息显示方法所执行的操作。On the other hand, a computer program product is provided, including a computer program that, when executed by a processor, implements the operations performed by the voting information generation method described in the above aspect, or implements the voting information described in the above aspect. Shows the operations performed by the method.
本申请实施例提供了自动生成视频片段的投票信息的方法,能够基于视频片段关联的文本内容,自动地为视频片段生成投票信息,无需人工创建投票信息,提高了操作效率,节省了时间,且这种方法能够有效覆盖大量的视频,提高了投票互动的覆盖率。Embodiments of the present application provide a method for automatically generating voting information for video clips, which can automatically generate voting information for video clips based on the text content associated with the video clips, eliminating the need to manually create voting information, improving operating efficiency, saving time, and This method can effectively cover a large number of videos and improve the coverage of voting interaction.
附图说明Description of the drawings
图1是本申请实施例提供的一种实施环境的示意图。Figure 1 is a schematic diagram of an implementation environment provided by an embodiment of the present application.
图2是本申请实施例提供的一种投票信息生成方法的流程图。Figure 2 is a flow chart of a voting information generation method provided by an embodiment of the present application.
图3是本申请实施例提供的另一种投票信息生成方法的流程图。Figure 3 is a flow chart of another voting information generation method provided by an embodiment of the present application.
图4是本申请实施例提供的一种投票信息的示意图。Figure 4 is a schematic diagram of voting information provided by an embodiment of the present application.
图5是本申请实施例提供的一种基于第一生成模型生成投票主题的流程示意图。Figure 5 is a schematic flowchart of generating voting topics based on the first generation model provided by an embodiment of the present application.
图6是本申请实施例提供的一种第一生成模型的示意图。Figure 6 is a schematic diagram of a first generation model provided by an embodiment of the present application.
图7是本申请实施例提供的一种基于第二生成模型生成投票候选项的流程示意图。FIG. 7 is a schematic flowchart of generating voting candidates based on the second generation model provided by an embodiment of the present application.
图8是本申请实施例提供的一种第二生成模型的示意图。Figure 8 is a schematic diagram of a second generation model provided by an embodiment of the present application.
图9是本申请实施例提供的一种生成投票信息的整体流程的示意图。Figure 9 is a schematic diagram of an overall process for generating voting information provided by an embodiment of the present application.
图10是本申请实施例提供的一种投票信息生成方法的流程图。Figure 10 is a flow chart of a voting information generation method provided by an embodiment of the present application.
图11是本申请实施例提供的一种投票判定模型的示意图。 Figure 11 is a schematic diagram of a voting decision model provided by an embodiment of the present application.
图12是本申请实施例提供的一种投票信息显示方法的流程图。Figure 12 is a flow chart of a voting information display method provided by an embodiment of the present application.
图13是本申请实施例提供的另一种投票信息显示方法的流程图。Figure 13 is a flow chart of another voting information display method provided by an embodiment of the present application.
图14是本申请实施例提供的一种投票互动模型的示意图。Figure 14 is a schematic diagram of a voting interaction model provided by an embodiment of the present application.
图15是本申请实施例提供的一种投票信息生成装置的结构示意图。Figure 15 is a schematic structural diagram of a voting information generation device provided by an embodiment of the present application.
图16是本申请实施例提供的一种投票信息显示装置的结构示意图。Figure 16 is a schematic structural diagram of a voting information display device provided by an embodiment of the present application.
图17是本申请实施例提供的一种终端的结构示意图。Figure 17 is a schematic structural diagram of a terminal provided by an embodiment of the present application.
图18是本申请实施例提供的一种服务器的结构示意图。Figure 18 is a schematic structural diagram of a server provided by an embodiment of the present application.
具体实施方式Detailed ways
为使本申请实施例的目的、技术方案和优点更加清楚,下面将结合附图对本申请实施方式作进一步地详细描述。In order to make the objectives, technical solutions, and advantages of the embodiments of the present application clearer, the embodiments of the present application will be described in further detail below with reference to the accompanying drawings.
可以理解,本申请所使用的术语“第一”、“第二”等可在本文中用于描述各种概念,但除非特别说明,这些概念不受这些术语限制。这些术语仅用于将一个概念与另一个概念区分。举例来说,在不脱离本申请的范围的情况下,可以将第一文本内容称为第二文本内容,将第二文本内容称为第一文本内容。It will be understood that the terms "first", "second", etc. used in this application may be used to describe various concepts herein, but these concepts are not limited by these terms unless otherwise specified. These terms are only used to distinguish one concept from another. For example, without departing from the scope of the present application, the first text content may be called second text content, and the second text content may be called first text content.
本申请所使用的术语“至少一个”、“多个”、“每个”、“任一”等,至少一个包括一个、两个或两个以上,多个包括两个或两个以上,每个是指对应的多个中的每一个,任一是指多个中的任意一个。举例来说,多个关键词包括3个关键词,而每个关键词是指这3个关键词中的每一个关键词,任一是指这3个关键词中的任意一个,可以是第一个,可以是第二个,也可以是第三个。The terms "at least one", "multiple", "each", "any", etc. used in this application, at least one includes one, two or more than two, and multiple includes two or more, each A refers to each of the corresponding plurality, and any refers to any one of the plurality. For example, multiple keywords include 3 keywords, and each keyword refers to each of these 3 keywords. Any one refers to any one of these 3 keywords, which can be the third keyword. One, it can be the second, it can be the third.
人工智能(Artificial Intelligence,AI)是利用数字计算机或者数字计算机控制的机器模拟、延伸和扩展人的智能,感知环境、获取知识并使用知识获得最佳结果的理论、方法、技术及应用系统。换句话说,人工智能是计算机科学的一个综合技术,它企图了解智能的实质,并生产出一种新的能以人类智能相似的方式做出反应的智能机器。人工智能也就是研究各种智能机器的设计原理与实现方法,使机器具有感知、推理与决策的功能。Artificial Intelligence (AI) is a theory, method, technology and application system that uses digital computers or machines controlled by digital computers to simulate, extend and expand human intelligence, perceive the environment, acquire knowledge and use knowledge to obtain the best results. In other words, artificial intelligence is a comprehensive technology of computer science that attempts to understand the essence of intelligence and produce a new intelligent machine that can respond in a similar way to human intelligence. Artificial intelligence is the study of the design principles and implementation methods of various intelligent machines, so that the machines have the functions of perception, reasoning and decision-making.
人工智能技术是一门综合学科,涉及领域广泛,既有硬件层面的技术也有软件层面的技术。人工智能基础技术一般包括如传感器、专用人工智能芯片、云计算、分布式存储、大数据处理技术、操作/交互系统、机电一体化等技术。人工智能软件技术主要包括计算机视觉技术、语音处理技术、自然语言处理技术以及机器学习/深度学习、自动驾驶、智慧交通等几大方向。Artificial intelligence technology is a comprehensive subject that covers a wide range of fields, including both hardware-level technology and software-level technology. Basic artificial intelligence technologies generally include technologies such as sensors, dedicated artificial intelligence chips, cloud computing, distributed storage, big data processing technology, operation/interaction systems, mechatronics and other technologies. Artificial intelligence software technology mainly includes computer vision technology, speech processing technology, natural language processing technology, machine learning/deep learning, autonomous driving, smart transportation and other major directions.
计算机视觉技术(Computer Vision,CV)计算机视觉是一门研究如何使机器“看”的科学,更进一步的说,就是指用摄影机和电脑代替人眼对目标进行识别和测量等机器视觉,并进一步做图形处理,使电脑处理成为更适合人眼观察或传送给仪器检测的图像。作为一个科学学科,计算机视觉研究相关的理论和技术,试图建立能够从图像或者多维数据中获取信息的人工智能系统。计算机视觉技术通常包括图像处理、图像识别、图像语义理解、图像检索、OCR(Optical Character Recognition,光学字符识别)、视频处理、视频语义理解、视频内容/行为识别、三维物体重建、3D技术、虚拟现实、增强现实、同步定位与地图构建、自动驾驶、智慧交通等技术,还包括常见的人脸识别、指纹识别等生物特征识别技术。Computer Vision Technology (Computer Vision, CV) Computer vision is a science that studies how to make machines "see". Furthermore, it refers to machine vision such as using cameras and computers instead of human eyes to identify and measure targets, and further Perform graphics processing to make computer processing into images more suitable for human eye observation or transmitted to instruments for detection. As a scientific discipline, computer vision studies related theories and technologies, trying to build artificial intelligence systems that can obtain information from images or multi-dimensional data. Computer vision technology usually includes image processing, image recognition, image semantic understanding, image retrieval, OCR (Optical Character Recognition, optical character recognition), video processing, video semantic understanding, video content/behavior recognition, three-dimensional object reconstruction, 3D technology, virtual Reality, augmented reality, simultaneous positioning and map construction, autonomous driving, smart transportation and other technologies, as well as common biometric recognition technologies such as face recognition and fingerprint recognition.
机器学习(Machine Learning,ML)是一门多领域交叉学科,涉及概率论、统计学、逼近论、凸分析、算法复杂度理论等多门学科。专门研究计算机怎样模拟或实现人类的学习行为,以获取新的知识或技能,重新组织已有的知识结构使之不断改善自身的性能。机器学习是人工智能的核心,是使计算机具有智能的根本途径,其应用遍及人工智能的各个领域。机器学习和深度学习通常包括人工神经网络、置信网络、强化学习、迁移学习、归纳学习、示教学习等技术。Machine Learning (ML) is a multi-field interdisciplinary subject involving probability theory, statistics, approximation theory, convex analysis, algorithm complexity theory and other disciplines. It specializes in studying how computers can simulate or implement human learning behavior to acquire new knowledge or skills, and reorganize existing knowledge structures to continuously improve their performance. Machine learning is the core of artificial intelligence and the fundamental way to make computers intelligent. Its applications cover all fields of artificial intelligence. Machine learning and deep learning usually include artificial neural networks, belief networks, reinforcement learning, transfer learning, inductive learning, teaching learning and other technologies.
随着人工智能技术研究和进步,人工智能技术在多个领域展开研究和应用,例如常见的智能家居、智能穿戴设备、虚拟助理、智能音箱、智能营销、无人驾驶、自动驾驶、无人机、 机器人、智能医疗、智能客服、车联网、自动驾驶、智慧交通等,相信随着技术的发展,人工智能技术将在更多的领域得到应用,并发挥越来越重要的价值。With the research and progress of artificial intelligence technology, artificial intelligence technology has been researched and applied in many fields, such as common smart homes, smart wearable devices, virtual assistants, smart speakers, smart marketing, driverless driving, autonomous driving, and drones. , Robots, smart medical care, smart customer service, Internet of Vehicles, autonomous driving, smart transportation, etc. I believe that with the development of technology, artificial intelligence technology will be applied in more fields and play an increasingly important role.
本申请实施例提供的投票信息生成方法和投票信息显示方法,利用人工智能中的计算机视觉技术以及机器学习等技术,能够生成视频片段的投票信息,并在播放视频片段时显示投票信息。The voting information generation method and voting information display method provided by the embodiments of this application utilize computer vision technology and machine learning technology in artificial intelligence to generate voting information for video clips and display the voting information when the video clips are played.
本申请实施例提供的投票信息生成方法和投票信息显示方法的执行主体为计算机设备,该计算机设备为终端或服务器。在一种可能实现方式中,该投票信息生成方法由服务器执行,该投票信息显示方法由终端执行,则本申请实施例提供如下图1所示的实施环境。The execution subject of the voting information generation method and voting information display method provided by the embodiments of this application is a computer device, and the computer device is a terminal or a server. In one possible implementation, the voting information generation method is executed by the server, and the voting information display method is executed by the terminal. Then the embodiment of the present application provides an implementation environment as shown in Figure 1 below.
图1是本申请实施例提供的一种实施环境的示意图,参见图1,该实施环境包括:服务器101和终端102。服务器101以及终端102可以通过有线或无线通信方式进行直接或间接地连接,本申请在此不做限制。Figure 1 is a schematic diagram of an implementation environment provided by an embodiment of the present application. Refer to Figure 1. The implementation environment includes: a server 101 and a terminal 102. The server 101 and the terminal 102 can be connected directly or indirectly through wired or wireless communication methods, which is not limited in this application.
服务器101用于存储或下发视频,并且还用于为视频中的视频片段自动生成投票信息,而终端102用于访问服务器101,播放服务器101下发的视频,并在播放视频时展示当前视频片段的投票信息,从而发起了针对当前视频片段的投票互动,吸引用户进行投票操作,参与到投票互动中。The server 101 is used to store or deliver videos, and is also used to automatically generate voting information for video clips in the video, while the terminal 102 is used to access the server 101, play the video delivered by the server 101, and display the current video when playing the video. The voting information of the clip is thus launched to initiate a voting interaction for the current video clip, attracting users to perform voting operations and participate in the voting interaction.
在一种可能实现方式中,服务器101是独立的物理服务器,或者是多个物理服务器构成的服务器集群或者分布式系统,或者是提供云服务、云数据库、云计算、云函数、云存储、网络服务、云通信、中间件服务、域名服务、安全服务、CDN(Content Delivery Network,内容分发网络)、以及大数据和人工智能平台等基础云计算服务的云服务器。终端102是智能手机、平板电脑、笔记本电脑、台式计算机、智能音箱、智能手表、智能电视、智能车载终端等,但并不局限于此。In one possible implementation, the server 101 is an independent physical server, or a server cluster or distributed system composed of multiple physical servers, or it provides cloud services, cloud databases, cloud computing, cloud functions, cloud storage, and networks. Services, cloud communications, middleware services, domain name services, security services, CDN (Content Delivery Network, content distribution network), and cloud servers for basic cloud computing services such as big data and artificial intelligence platforms. The terminal 102 is a smartphone, a tablet computer, a notebook computer, a desktop computer, a smart speaker, a smart watch, a smart TV, a smart car terminal, etc., but is not limited thereto.
在另一种可能实现方式中,终端102上安装由服务器101提供服务的目标应用,终端102能够通过该目标应用实现例如视频播放、投票等功能。例如,目标应用为视频分享应用,该视频分享应用具有视频分享的功能,当然,该视频分享应用还能够具有其他功能,例如,发布弹幕的功能、进行投票的功能等。In another possible implementation, a target application provided by the server 101 is installed on the terminal 102, and the terminal 102 can implement functions such as video playback and voting through the target application. For example, the target application is a video sharing application, which has the function of video sharing. Of course, the video sharing application can also have other functions, such as the function of posting barrages, the function of voting, etc.
图2是本申请实施例提供的一种投票信息生成方法的流程图。本申请实施例的执行主体为计算机设备,该计算机设备为终端或服务器,本申请实施例对生成视频片段的投票信息的过程进行说明。参见图2,该方法包括:Figure 2 is a flow chart of a voting information generation method provided by an embodiment of the present application. The execution subject of the embodiment of the present application is a computer device, and the computer device is a terminal or a server. The embodiment of the present application explains the process of generating voting information of a video clip. Referring to Figure 2, the method includes:
201、计算机设备获取视频中的视频片段关联的文本内容。201. The computer device obtains text content associated with the video clips in the video.
其中,该视频为计算机设备中的任一视频,例如,计算机设备为终端,该视频为终端下载的任一视频或者拍摄的任一视频等,或者计算机设备为服务器,服务器具有视频分享功能,能存储任一设备上传的视频,并将该视频发送给任一设备进行播放,而该视频为服务器存储的任一视频。Wherein, the video is any video in the computer device. For example, the computer device is a terminal, and the video is any video downloaded by the terminal or any video shot, etc., or the computer device is a server, and the server has a video sharing function and can Store the video uploaded by any device and send the video to any device for playback, and the video is any video stored by the server.
该视频中包括一个或多个视频片段,视频片段的播放时长不大于该视频的总播放时长。例如,该视频按照固定时长被划分为多个视频片段,每个视频片段的播放时长等于该固定时长。The video includes one or more video clips, and the playing time of the video clips is not greater than the total playing time of the video. For example, the video is divided into multiple video segments according to a fixed duration, and the playback duration of each video segment is equal to the fixed duration.
该视频片段关联的文本内容包括第一文本内容或第二文本内容的至少一种。第一文本内容为视频片段包含的文本内容,例如,该第一文本内容包括视频片段中的字幕文本内容,或者包括从该视频片段中的语音数据中识别出的文本内容,该第一文本内容能够表示该视频片段本身所包含的内容,如该视频片段中出现过的人物、事物或者该视频片段中发生的情节等。The text content associated with the video clip includes at least one of first text content or second text content. The first text content is text content contained in the video clip. For example, the first text content includes subtitle text content in the video clip, or includes text content recognized from the voice data in the video clip. The first text content It can represent the content contained in the video clip itself, such as the characters and things that appeared in the video clip or the plot that occurred in the video clip.
该第二文本内容为视频片段的弹幕包含的文本内容,可称为弹幕文本内容。在播放该视频片段时,播放该视频片段的终端可以发布针对该视频片段的弹幕,弹幕中包含的文本内容能够表示终端用户对该视频片段的看法或观点,因此弹幕文本内容可以看做视频片段的互动数据。其中,在目标视频中的每个视频片段具有对应的播放时间段,而视频片段的播放时间 段包括弹幕的发布时间点,则表示该弹幕是该视频片段的弹幕。The second text content is the text content included in the barrage of the video clip, which may be called the barrage text content. When playing the video clip, the terminal playing the video clip can publish a barrage for the video clip. The text content contained in the barrage can express the end user's views or opinions on the video clip, so the barrage text content can be viewed Create interactive data for video clips. Among them, each video clip in the target video has a corresponding playback time period, and the playback time of the video clip If the segment includes the release time point of the barrage, it means that the barrage is the barrage of the video clip.
计算机设备在获取到文本内容后,即可自动地基于该文本内容生成视频片段的投票信息,无需技术人员手动生成,而且在播放该视频片段时显示该投票信息,还能够吸引用户参与到投票互动中,有助于提高视频的互动覆盖率。其中生成投票信息的过程详见下述步骤202-204。After the computer device obtains the text content, it can automatically generate the voting information of the video clip based on the text content, without the need for technical personnel to manually generate it, and display the voting information when the video clip is played, and can also attract users to participate in voting interaction , which helps improve the interactive coverage of the video. The process of generating voting information is detailed in steps 202-204 below.
202、计算机设备基于文本内容中的关键词,生成视频片段的投票主题。202. The computer device generates a voting topic for the video clip based on the keywords in the text content.
203、计算机设备基于关键词和投票主题,生成视频片段的多个投票候选项。203. The computer device generates multiple voting candidates for the video clip based on the keywords and voting topics.
投票信息包括投票主题和多个投票候选项,投票主题代表了询问用户的问题,而多个投票候选项代表了为用户提供的候选答案,显示投票信息时,用户通过查看投票主题了解问题,通过从多个投票候选项中选取一个投票候选项,选择自己的答案,也即是实现了投票操作。The voting information includes a voting topic and multiple voting candidates. The voting topic represents the question asked of the user, and the multiple voting candidates represent the candidate answers provided to the user. When the voting information is displayed, the user understands the question by viewing the voting topic. Select one voting candidate from multiple voting candidates and choose your own answer, which is to implement the voting operation.
文本内容中包括至少一个词语,该文本内容中的关键词可以包括该文本内容中的每个词语,或者仅包括通过关键词抽取算法从该文本内容中抽取的词语。由于该关键词能够表示该视频片段的内容,因此基于该关键词生成的投票主题,与该视频片段的内容相关。而且基于关键词和投票主题生成的多个投票候选项,也与该视频片段的内容相关,且与所生成的投票主题相符。The text content includes at least one word, and the keywords in the text content may include every word in the text content, or only include words extracted from the text content through a keyword extraction algorithm. Since the keyword can represent the content of the video clip, the voting topic generated based on the keyword is related to the content of the video clip. Moreover, the multiple voting candidates generated based on keywords and voting topics are also related to the content of the video clip and consistent with the generated voting topics.
204、计算机设备基于投票主题和多个投票候选项,生成视频片段的投票信息。204. The computer device generates voting information of the video clip based on the voting topic and multiple voting candidates.
需要说明的是,本申请实施例仅是以一个视频中的一个视频片段为例进行说明,而生成其他的视频片段的投票信息的过程与本申请实施例类似,在此不再赘述。It should be noted that the embodiment of the present application only takes one video clip in a video as an example, and the process of generating voting information for other video clips is similar to the embodiment of the present application and will not be described again here.
相关技术中,视频中的投票信息主要是由视频站点的运营人员创建,或者由视频的创作者创建,需要花费较长的时间,操作效率较低,且这种方式难以覆盖大量的视频,导致视频的投票互动不够充分,禁锢了用户参与投票互动。In related technologies, voting information in videos is mainly created by operators of video sites or creators of videos, which takes a long time and has low operating efficiency. Moreover, this method is difficult to cover a large number of videos, resulting in The voting interaction in the video is not sufficient, which prohibits users from participating in voting interaction.
本申请实施例提供了自动生成视频片段的投票信息的方法,能够基于视频片段关联的文本内容,自动地为视频片段生成投票信息,无需人工创建投票信息,提高了操作效率,节省了时间,且这种方法能够有效覆盖大量的视频,提高了投票互动的覆盖率。Embodiments of the present application provide a method for automatically generating voting information for video clips, which can automatically generate voting information for video clips based on the text content associated with the video clips, eliminating the need to manually create voting information, improving operating efficiency, saving time, and This method can effectively cover a large number of videos and improve the coverage of voting interaction.
在上述图2所示实施例的基础上,本申请实施例还提供了另一种投票信息生成方法,对生成投票主题和投票候选项的具体过程进行了说明。图3是本申请实施例提供的另一种投票信息生成方法的流程图。本申请实施例的执行主体为计算机设备,该计算机设备为终端或服务器。参见图3,该方法包括:Based on the above embodiment shown in Figure 2, the embodiment of the present application also provides another voting information generation method, and explains the specific process of generating voting topics and voting candidates. Figure 3 is a flow chart of another voting information generation method provided by an embodiment of the present application. The execution subject of the embodiment of the present application is a computer device, and the computer device is a terminal or a server. Referring to Figure 3, the method includes:
301、计算机设备获取视频中的视频片段关联的文本内容。301. The computer device obtains the text content associated with the video clips in the video.
可选地,该文本内容包括第一文本内容,第一文本内容为视频片段包含的文本内容。获取第一文本内容的过程包括如下至少一项:Optionally, the text content includes first text content, and the first text content is text content included in the video clip. The process of obtaining the first text content includes at least one of the following:
(1)从该视频片段中提取字幕文本内容。(1) Extract subtitle text content from the video clip.
例如,从该视频片段中提取一个或多个视频帧,采用OCR(Optical Character Recognition,光学字符识别)算法,从该一个或多个视频帧中提取字幕文本内容。For example, one or more video frames are extracted from the video clip, and an OCR (Optical Character Recognition) algorithm is used to extract subtitle text content from the one or more video frames.
(2)从该视频片段中提取语音数据,对该语音数据进行文本识别,得到该语音数据对应的文本内容。(2) Extract voice data from the video clip, perform text recognition on the voice data, and obtain the text content corresponding to the voice data.
例如,采用ASR(Automatic Speech Recognition,自动语音识别)算法,对语音数据进行识别,得到语音数据对应的文本内容,该文本内容即为该视频片段中的对白的文本内容。For example, the ASR (Automatic Speech Recognition) algorithm is used to recognize the speech data and obtain the text content corresponding to the speech data, which is the text content of the dialogue in the video clip.
可选地,该文本内容包括第二文本内容,第二文本内容为视频片段的弹幕包含的文本内容,获取第二文本内容的过程包括:从视频的弹幕集合中,提取视频片段的弹幕,从视频片段的弹幕中提取文本内容。Optionally, the text content includes second text content, and the second text content is the text content contained in the bullet comments of the video clip. The process of obtaining the second text content includes: extracting the bullet comments of the video clip from the video clip collection. Screen, extract text content from the barrage of video clips.
投票信息包括投票主题和多个投票候选项,计算机设备在获取到文本内容之后,首先需要生成投票主题,生成投票主题的具体步骤详见下述步骤302-303。The voting information includes a voting topic and multiple voting candidates. After obtaining the text content, the computer device first needs to generate a voting topic. For specific steps to generate a voting topic, please refer to the following steps 302-303.
302、计算机设备对文本内容中的关键词进行编码,得到关键词的关键词特征。302. The computer device encodes the keywords in the text content and obtains the keyword characteristics of the keywords.
其中,该关键词特征用于描述关键词,将关键词转换为关键词特征的形式,方便后续基于可量化的关键词特征进行处理,能够生成与该关键词相关的投票信息。 Among them, the keyword feature is used to describe keywords, and the keywords are converted into the form of keyword features to facilitate subsequent processing based on quantifiable keyword features, and to generate voting information related to the keywords.
可选地,采用编码算法对该关键词进行编码,得到关键词特征,或者调用编码模型对该关键词进行编码,得到关键词特征。其中,该编码模型可以为Transformer模型(一种基于自注意力机制的模型)或者其他类型的模型。Optionally, use a coding algorithm to code the keyword to obtain keyword features, or call a coding model to code the keyword to obtain keyword features. Among them, the encoding model can be a Transformer model (a model based on a self-attention mechanism) or other types of models.
而在对关键词进行编码之前,首先需要从文本内容中抽取关键词。可选地,采用关键词抽取算法抽取关键词,或者调用关键词抽取模型抽取关键词。其中,该关键词抽取模型可以为TextRank模型(一种基于文本排名抽取关键词的模型)或者其他类型的模型。Before encoding keywords, you first need to extract keywords from the text content. Optionally, use a keyword extraction algorithm to extract keywords, or call a keyword extraction model to extract keywords. The keyword extraction model may be a TextRank model (a model that extracts keywords based on text ranking) or other types of models.
303、计算机设备对关键词特征进行解码,得到投票主题,投票主题由多个投票主题词语构成。303. The computer device decodes the keyword features to obtain the voting topic, which is composed of multiple voting topic words.
可选地,计算机设备每次对关键词特征进行解码,得到一个投票主题词语,之后继续对关键词特征和已确定的最后一个投票主题词语进行解码,得到下一个投票主题词语,直至得到目标数量的投票主题词语,该目标数量的投票主题词语构成投票主题。Optionally, the computer device decodes the keyword features each time to obtain a voting topic word, and then continues to decode the keyword features and the last determined voting topic word to obtain the next voting topic word until the target number is obtained. of voting topic words, and the target number of voting topic words constitute the voting topic.
计算机设备生成多个投票候选项的具体步骤详见下述步骤304-306。The specific steps for the computer device to generate multiple voting candidates are detailed in steps 304-306 below.
304、计算机设备获取第一关键词,第一关键词为第一文本内容中的关键词。304. The computer device obtains the first keyword, which is a keyword in the first text content.
在本申请实施例中,计算机设备获取到第一文本内容和第二文本内容,其中第一文本内容中的关键词称为第一关键词,第二文本内容中的关键词称为第二关键词。In this embodiment of the present application, the computer device obtains the first text content and the second text content, where the keywords in the first text content are called first keywords, and the keywords in the second text content are called second keywords. word.
其中,获取第一文本内容中的第一关键词的步骤与上述步骤302中抽取关键词的方式类似,在此不再赘述。The step of obtaining the first keyword in the first text content is similar to the method of extracting the keyword in the above-mentioned step 302, and will not be described again here.
305、计算机设备对第二文本内容进行聚类,得到多个文本类别,每个文本类别包含至少一条第二文本内容,分别从每个文本类别中抽取第二关键词。305. The computer device clusters the second text content to obtain multiple text categories, each text category contains at least one piece of second text content, and extracts the second keyword from each text category.
视频中可以包括多个弹幕,相应地,计算机设备会获取到多个弹幕分别包括的第二文本内容,从而从多个第二文本内容中抽取第二关键词。而为了避免抽取到重复的关键词,且为了节省处理量,可以先对多个第二文本内容进行聚类,将语义相关的第二文本内容划分到一个文本类别中,之后再按照不同的文本类别,分别抽取每个文本类别中的第二关键词。The video may include multiple barrages. Accordingly, the computer device will obtain the second text content included in the multiple barrages, thereby extracting the second keywords from the plurality of second text contents. In order to avoid extracting duplicate keywords and save processing time, multiple second text contents can be clustered first, and semantically related second text contents can be divided into one text category, and then different texts can be classified into Category, extract the second keyword in each text category separately.
306、计算机设备基于第一关键词、投票主题和每个文本类别的第二关键词,分别生成每个文本类别的投票候选项。306. The computer device generates voting candidates for each text category based on the first keyword, the voting topic, and the second keyword of each text category.
在划分不同的文本类别后,由于不同文本类别的第二关键词的语义相差较大,而同一文本类别的第二关键词的语义相差较小,因此可以按照不同的文本类别,分别生成不同的投票候选项。则所生成的投票候选项的数量等于聚类得到的文本类别的数量。After dividing different text categories, since the semantic differences of the second keywords of different text categories are large, and the semantic differences of the second keywords of the same text category are small, different text categories can be generated according to different text categories. Vote candidates. Then the number of generated voting candidates is equal to the number of text categories obtained by clustering.
以生成第i个投票候选项的过程为例,i为正整数,i不大于文本类别的数量,生成第i个投票候选项的过程包括:对第一关键词、投票主题和第i个文本类别的第二关键词进行编码,得到关键词特征,对关键词特征进行解码,得到第i个投票候选项,第i个投票候选项由多个投票候选项词语构成。其具体过程与上述步骤302-303类似,区别在于:本次确定的关键词特征用于描述第一关键词、投票主题中的关键词和第二关键词,根据该关键词特征得到的多个投票候选项词语综合考虑了第一关键词、投票主题中的关键词和第二关键词的影响,保证了投票候选项词语是与第一关键词、投票主题和第二关键词相关。Taking the process of generating the i-th voting candidate as an example, i is a positive integer and i is not greater than the number of text categories. The process of generating the i-th voting candidate includes: comparing the first keyword, voting topic and i-th text The second keyword of the category is encoded to obtain keyword features, and the keyword features are decoded to obtain the i-th voting candidate. The i-th voting candidate is composed of multiple voting candidate words. The specific process is similar to the above steps 302-303. The difference is that the keyword characteristics determined this time are used to describe the first keyword, the keywords in the voting topic and the second keyword. Multiple keywords are obtained based on the keyword characteristics. The voting candidate words comprehensively consider the influence of the first keyword, the keywords in the voting topic, and the second keyword, ensuring that the voting candidate words are related to the first keyword, the voting topic, and the second keyword.
307、计算机设备基于投票主题和多个投票候选项,生成视频片段的投票信息。307. The computer device generates voting information of the video clip based on the voting topic and multiple voting candidates.
可选地,将投票主题和多个投票候选项构成视频片段的投票信息。或者,将投票主题和多个投票候选项以及关联信息构成视频片段的投票信息。该关联信息包括用于提示用户进行投票的文本或图像等,也可以包括其他类型的信息。Optionally, the voting topic and multiple voting candidates constitute the voting information of the video clip. Alternatively, the voting topic, multiple voting candidates, and associated information constitute the voting information of the video clip. The associated information includes text or images used to prompt users to vote, etc., and may also include other types of information.
如图4所示,在播放视频片段时,播放界面中显示有视频片段的视频画面以及投票信息,该投票信息分为两部分,一部分是投票主题“你是来看什么的?”,另一部分是三个投票候选项,供用户选择。As shown in Figure 4, when a video clip is played, the video screen of the video clip and voting information are displayed in the playback interface. The voting information is divided into two parts, one part is the voting topic "What did you come to see?", and the other part There are three voting candidates for users to choose from.
生成投票信息之后,计算机设备将视频片段与投票信息关联存储,在播放视频片段时显示投票信息,或者在每次向其他设备下发视频片段时下发投票信息。或者,计算机设备将投票信息添加到视频片段中,以便在播放视频片段时显示投票信息。其中,显示投票信息的具体过程详见下述图12和图13所示的实施例,在此暂不进行说明。 After generating the voting information, the computer device stores the video clips in association with the voting information, displays the voting information when playing the video clips, or delivers the voting information each time the video clips are delivered to other devices. Alternatively, the computer device adds the voting information to the video clip so that the voting information is displayed when the video clip is played. The specific process of displaying voting information is detailed in the embodiment shown in FIG. 12 and FIG. 13 below, and will not be described here.
需要说明的是,本申请实施例仅是以视频片段的一个投票信息为例进行说明,而在另一实施例中,计算机设备通过重复执行上述步骤,可以生成多个投票主题和每个投票主题对应的投票候选项,从而构成多个投票信息,那么在播放视频片段时可以显示该多个投票信息,或者显示该多个投票信息中的一个或多个,本申请实施例对此不做限定。It should be noted that the embodiment of the present application only takes voting information of a video clip as an example for explanation. In another embodiment, the computer device can generate multiple voting topics and each voting topic by repeatedly executing the above steps. The corresponding voting candidates constitute multiple voting information, and then the multiple voting information can be displayed when the video clip is played, or one or more of the multiple voting information can be displayed. The embodiment of the present application does not limit this. .
本申请实施例提供了自动生成视频片段的投票信息的方法,能够基于视频片段关联的文本内容,自动地为视频片段生成投票信息,无需人工创建投票信息,提高了操作效率,节省了时间,且这种方法能够有效覆盖大量的视频,提高了投票互动的覆盖率。而且所生成的投票信息与视频片段的文本内容相关,满足了视频的互动功能需求,有助于提升用户在观看视频片段过程中参与投票互动的积极性,进而提升了互动氛围。采用上述方式便于覆盖大量的视频,提升了视频的互动覆盖率与丰富度,提高了视频平台上的用户互动活跃度。Embodiments of the present application provide a method for automatically generating voting information for video clips, which can automatically generate voting information for video clips based on the text content associated with the video clips, eliminating the need to manually create voting information, improving operating efficiency, saving time, and This method can effectively cover a large number of videos and improve the coverage of voting interaction. Moreover, the generated voting information is related to the text content of the video clip, which meets the interactive function requirements of the video, helps to increase the enthusiasm of users to participate in voting interaction while watching the video clip, and thereby enhances the interactive atmosphere. Using the above method makes it easy to cover a large number of videos, improves the interactive coverage and richness of the videos, and increases the user interaction activity on the video platform.
在上述图3所示实施例的基础上,在一种可能实现方式中,上述步骤302-303中生成投票主题的过程可以基于第一生成模型执行,该第一生成模型包括第一编码子模型和第一解码子模型,其中,第一编码子模型用于将关键词编码为关键词特征,第一解码子模型用于将关键词特征解码为投票主题词语。如图5所示,基于第一生成模型生成投票主题的过程包括:Based on the above embodiment shown in Figure 3, in one possible implementation, the process of generating voting topics in the above steps 302-303 can be performed based on the first generation model, the first generation model includes the first encoding sub-model and a first decoding sub-model, wherein the first encoding sub-model is used to encode keywords into keyword features, and the first decoding sub-model is used to decode keyword features into voting topic words. As shown in Figure 5, the process of generating voting topics based on the first generation model includes:
501、计算机设备调用第一编码子模型,对N个关键词进行编码,得到N个关键词的关键词特征,N为大于1的整数。501. The computer device calls the first encoding sub-model, encodes the N keywords, and obtains the keyword characteristics of the N keywords, where N is an integer greater than 1.
其中,第一编码子模型为Transformer Encoder模型(Transformer模型中的编码器)或者其他类型的编码模型。Among them, the first encoding sub-model is a Transformer Encoder model (the encoder in the Transformer model) or other types of encoding models.
其中,N个关键词包括第一文本内容中的第一关键词,如视频片段的对白文本关键词或字母文本关键词等,还包括第二文本内容中的第二关键词,如视频片段的弹幕中的文本关键词。Among them, the N keywords include the first keywords in the first text content, such as dialogue text keywords or alphabetical text keywords of the video clip, etc., and also include the second keywords in the second text content, such as the video clip's Text keywords in the barrage.
本申请实施例中,计算机设备获取到了视频片段的N个关键词,需要综合考虑这N个关键词来生成投票主题。则将N个关键词输入至第一生成模型的第一编码子模型中,从而基于第一编码子模型对N个关键词分别进行编码,得到N个关键词特征,每个关键词对应一个关键词特征。In the embodiment of this application, the computer device obtains N keywords of the video clip, and needs to comprehensively consider these N keywords to generate a voting topic. Then N keywords are input into the first encoding sub-model of the first generative model, so that the N keywords are encoded respectively based on the first encoding sub-model to obtain N keyword features, each keyword corresponding to a key word features.
502、计算机设备调用第一解码子模型,对N个关键词特征进行解码,得到第1个解码特征,基于第1个解码特征和N个关键词特征,确定第一投票主题词语。502. The computer device calls the first decoding sub-model, decodes the N keyword features, obtains the first decoding feature, and determines the first voting topic word based on the first decoding feature and the N keyword features.
其中,第一解码子模型为Transformer Decoder模型(Transformer模型中的解码器)或者其他类型的编码模型。Among them, the first decoding sub-model is a Transformer Decoder model (the decoder in the Transformer model) or other types of coding models.
503、计算机设备调用第一解码子模型,对N个关键词特征和第一投票主题词语进行解码,得到第2个解码特征,基于第2个解码特征和N个关键词特征,确定参考投票主题,参考投票主题包括第一投票主题词语和第二投票主题词语,直至通过N次解码后,将第N次解码得到的参考投票主题确定为视频片段的投票主题。503. The computer device calls the first decoding sub-model, decodes the N keyword features and the first voting topic words, obtains the second decoding feature, and determines the reference voting topic based on the second decoding feature and the N keyword features. , the reference voting topic includes the first voting topic word and the second voting topic word, until after N times of decoding, the reference voting topic obtained by the Nth decoding is determined as the voting topic of the video clip.
在本申请实施例中,采用了逐次解码生成投票主题词语的方式,每进行一次解码,确定当前的投票主题词语,当前的投票主题词语与之前确定的投票主题词语按照顺序组合,即可得到当前的参考投票主题。随着解码的多次进行,参考投票主题中包含的投票主题词语逐渐增多,直至进行N次解码后,第N次解码得到的参考投票主题包含了N个投票主题词语,从而得到包含N个投票主题词语的投票主题。In the embodiment of this application, a method of sequential decoding is used to generate voting topic words. Each time decoding is performed, the current voting topic words are determined. The current voting topic words are combined with the previously determined voting topic words in order to obtain the current voting topic words. Reference voting topic. As decoding is performed multiple times, the number of voting topic words contained in the reference voting topic gradually increases until, after N decodings, the reference voting topic obtained by decoding for the Nth time contains N voting topic words, thus obtaining the result containing N voting topic words. Poll topic for topic words.
本申请实施例中,每次解码时不仅考虑N个关键词特征,还会考虑之前确定的投票主题词语,这样能保证本次确定的投票主题词语与之前确定的投票主题词语之间相关联,从而保证基于所确定的N个投票主题词语构成的投票主题中的不同投票主题词语关联,能够组合构成一条语义清楚的语句。In the embodiment of this application, not only the N keyword features are considered during each decoding, but also the previously determined voting topic words are considered. This can ensure that the voting topic words determined this time are associated with the previously determined voting topic words. This ensures that the associations of different voting topic words in the voting topic composed of the determined N voting topic words can be combined to form a sentence with clear semantics.
可选地,第一生成模型还包括第一分类层和预设词语库,预设词语库包括多个词语,基于第1个解码特征和N个关键词特征,确定第一投票主题词语,包括:Optionally, the first generation model also includes a first classification layer and a preset word library. The preset word library includes a plurality of words. Based on the first decoding feature and the N keyword features, the first voting topic word is determined, including :
(1)基于第1个解码特征和N个关键词特征确定N个第一使用概率,第j个第一使用 概率为在投票主题中使用第j个关键词的概率,j为正整数,且j不大于N。该第j个第一使用概率也即是将该第j个关键词确定为投票主题中的第一投票主题词语的概率。(1) Determine the N first usage probabilities based on the 1st decoding feature and N keyword features, and the jth first usage The probability is the probability of using the jth keyword in the voting topic, j is a positive integer, and j is not greater than N. The jth first usage probability is also the probability of determining the jth keyword as the first voting topic word in the voting topic.
(2)在第j个第一使用概率满足使用条件的情况下,将第j个关键词确定为第一投票主题词语。其中,使用条件是指在投票主题中使用关键词所需满足的条件。(2) When the jth first usage probability satisfies the usage condition, determine the jth keyword as the first voting topic word. Among them, the usage conditions refer to the conditions that need to be met when using keywords in voting topics.
(3)在每个第一使用概率不满足使用条件的情况下,调用第一分类层基于第1个解码特征和预设词语库进行分类,得到预设词语库中的每个词语的分类概率,基于每个词语的分类概率,确定第一投票主题词语。其中,词语的分类概率表示将该词语确定为投票主体词语的概率。(3) When each first usage probability does not meet the usage conditions, the first classification layer is called to classify based on the first decoding feature and the preset word library, and the classification probability of each word in the preset word library is obtained. , based on the classification probability of each word, determine the first voting topic word. Among them, the classification probability of a word represents the probability of determining the word as the voting subject word.
本申请实施例中提供了两种确定投票主题词语的方式,一种是采用复制机制,将关键词复制为投票主题词语,也即是如果关键词的第一使用概率满足使用条件,则将关键词直接作为了投票主题词语。另一种是根据预设词语库生成新的投票主题词语。则在第一解码子模型每次进行解码时,首先确定使用概率,根据使用概率是否满足使用条件,确定是要使用关键词作为投票主题词语还是要生成新的投票主题词语。The embodiments of this application provide two ways of determining voting topic words. One is to use a copy mechanism to copy keywords into voting topic words. That is, if the first usage probability of the keyword meets the usage conditions, then the key The words were directly used as the voting topic words. The other is to generate new voting topic words based on the preset word library. Then every time the first decoding sub-model performs decoding, the usage probability is first determined, and based on whether the usage probability meets the usage conditions, it is determined whether to use the keyword as a voting topic word or to generate a new voting topic word.
可选地,使用条件包括使用概率阈值,在第j个第一使用概率大于使用概率阈值的情况下,将第j个关键词确定为第一投票主题词语,其中在多个第一使用概率均大于使用概率阈值的情况下,将最大的第一使用概率的关键词确定为第一投票主题词语,而在每个第一使用概率不大于使用概率阈值的情况下,再确定预设词语库中的每个词语的分类概率,该分类概率表示将该词语确定为投票主题词语的可能性,则基于每个词语的分类概率,能够从多个词语中选取出第一投票主题词语。Optionally, the usage condition includes a usage probability threshold. In the case where the jth first usage probability is greater than the usage probability threshold, the jth keyword is determined as the first voting topic word, wherein the plurality of first usage probabilities are equal to If it is greater than the usage probability threshold, the keyword with the largest first usage probability will be determined as the first voting topic word, and if each first usage probability is not greater than the usage probability threshold, then determine the keyword in the preset word library The classification probability of each word represents the possibility of determining the word as a voting topic word. Then based on the classification probability of each word, the first voting topic word can be selected from multiple words.
可选地,基于每个词语的分类概率,确定第一投票主题词语,包括:从预设词语库中的多个词语中,选取分类概率最大的词语,作为第一投票主题词语,或者,从预设词语库中的多个词语中,选取分类概率最大的目标数量个词语,分别作为候选的第一投票主题词语。其中,目标数量大于1。Optionally, determining the first voting topic word based on the classification probability of each word includes: selecting the word with the highest classification probability from multiple words in the preset word library as the first voting topic word, or selecting from Among the multiple words in the preset word library, the target number of words with the highest classification probability are selected as the candidate first voting topic words. Among them, the number of targets is greater than 1.
可选地,基于第2个解码特征和N个关键词特征,确定参考投票主题,参考投票主题包括第一投票主题词语和第二投票主题词语,包括:Optionally, based on the second decoding feature and the N keyword features, determine the reference voting topic. The reference voting topic includes the first voting topic word and the second voting topic word, including:
(1)基于第2个解码特征和N个关键词特征确定N个第二使用概率,第j个第二使用概率为在投票主题中使用第j个关键词的概率,j为正整数,且j不大于N。该第j个第二使用概率也即是将该第j个关键词确定为投票主题中的第二投票主题词语的概率。(1) Determine N second usage probabilities based on the 2nd decoding feature and N keyword features. The jth second usage probability is the probability of using the jth keyword in the voting topic, j is a positive integer, and j is not greater than N. The j-th second usage probability is also the probability of determining the j-th keyword as the second voting topic word in the voting topic.
(2)在第j个第二使用概率满足使用条件的情况下,将第j个关键词确定为第二投票主题词语。其中,使用条件是指在投票主题中使用关键词所需满足的条件。(2) When the j-th second usage probability satisfies the usage conditions, determine the j-th keyword as the second voting topic word. Among them, the usage conditions refer to the conditions that need to be met when using keywords in voting topics.
(3)在每个第二使用概率不满足使用条件的情况下,调用第一分类层基于第2个解码特征和预设词语库进行分类,得到多个候选投票主题的分类概率,每个候选投票主题包括一个第一投票主题词语和一个第二投票主题词语。其中,候选投票主题的分类概率表示将该候选投票主题确定为参考投票主题的概率。(3) When each second usage probability does not meet the usage conditions, the first classification layer is called to classify based on the second decoding feature and the preset word library to obtain the classification probabilities of multiple candidate voting topics. Each candidate The voting topic includes a first voting topic word and a second voting topic word. Among them, the classification probability of a candidate voting topic represents the probability of determining the candidate voting topic as a reference voting topic.
(4)基于每个候选投票主题的分类概率,确定参考投票主题。(4) Based on the classification probability of each candidate voting topic, determine the reference voting topic.
可选地,基于每个候选投票主题的分类概率,确定参考投票主题,包括:从多个候选投票主题中,选取分类概率最大的候选投票主题,作为参考投票主题,或者,从多个候选投票主题中,选取分类概率最大的目标数量个候选投票主题,分别作为参考投票主题。其中,目标数量大于1。后续可以针对每个投票主题生成投票候选项,进而构成目标数量个投票信息。Optionally, determine the reference voting topic based on the classification probability of each candidate voting topic, including: selecting the candidate voting topic with the highest classification probability from multiple candidate voting topics as the reference voting topic, or voting from multiple candidate voting topics Among the topics, the target number of candidate voting topics with the highest classification probability are selected as reference voting topics. Among them, the number of targets is greater than 1. Subsequently, voting candidates can be generated for each voting topic to form a target number of voting information.
第二次解码的过程与第一次解码的过程类似,区别在于:在第二次进行解码时,会将第一次解码得到的第一投票主题词语也输入到第一解码子模型中,从而将第一投票主题词语,与预设词语库中的每个词语构成了候选投票主题,通过确定多个候选投票主题的分类概率,从多个候选投票主题中确定当前的投票主题,这样能保证考虑到预设词语库中的词语与第一投票主题词语之间的关联性,而候选投票主题的分类概率也能体现预设词语库中的词语与第一投票主题词语组合得到的候选投票主题的合理程度,从而保证能生成语义清楚、逻辑通顺的投票主题。 The process of the second decoding is similar to the process of the first decoding. The difference is that during the second decoding, the first voting topic words obtained by the first decoding will also be input into the first decoding sub-model, so that The first voting topic word and each word in the preset word library constitute a candidate voting topic. By determining the classification probabilities of multiple candidate voting topics, the current voting topic is determined from multiple candidate voting topics, which can ensure Taking into account the correlation between the words in the preset word library and the first voting topic words, the classification probability of the candidate voting topics can also reflect the candidate voting topics obtained by combining the words in the preset word library and the first voting topic words. to a reasonable degree, thereby ensuring that voting topics with clear semantics and logical logic can be generated.
需要说明的是,本申请实施例仅是对第一次解码和第二次解码进行说明,而在N大于2的情况下,在第二次解码之后的解码过程均与第二次解码过程类似,在此不再赘述。It should be noted that the embodiment of the present application only explains the first decoding and the second decoding. When N is greater than 2, the decoding process after the second decoding is similar to the second decoding process. , which will not be described in detail here.
而在解码过程中,基于解码特征和N个关键词特征,采用如下公式确定第i个使用概率:During the decoding process, based on the decoding features and N keyword features, the following formula is used to determine the i-th usage probability:
Pi=e^(X*wi*svi)/sumj(e^(X*wi*svi)),其中,Pi表示第i个使用概率,X表示当前的解码特征,wi表示第i个关键词特征的权重参数,svi表示第i个关键词特征,j为任一正整数,j不大于N,Xj表示第j个解码特征,wj表示第j个关键词特征的权重参数,svj表示第j个关键词特征。P i =e^(X*w i *sv i )/sum j (e^(X*w i *sv i )), where P i represents the i-th usage probability, X represents the current decoding feature, w i represents the weight parameter of the i-th keyword feature, sv i represents the i-th keyword feature, j is any positive integer, j is not greater than N, X j represents the j-th decoding feature, w j represents the j-th key The weight parameter of the word feature, sv j represents the jth keyword feature.
图6是本申请实施例提供的一种第一生成模型的示意图,参见图6,计算机设备获取到的N个关键词包括对白文本关键词及字幕文本关键词,以及弹幕文本关键词,将N个关键词输入到第一编码子模型中,通过调用第一编码子模型得到了N个关键词特征:关键词特征1、关键词特征2…关键词特征N,而在每次解码时,可以选择是复制文本关键词,还是重新确定新的投票主题词语。如图6右半部分所示,如果确定要复制文本关键词,可直接将对应的文本关键词确定为投票主题词语,如果确定不复制文本关键词,则根据上一次确定的投票主题词语,确定当前的投票主题词语。Figure 6 is a schematic diagram of a first generation model provided by an embodiment of the present application. Referring to Figure 6, the N keywords obtained by the computer device include dialogue text keywords, subtitle text keywords, and barrage text keywords. N keywords are input into the first encoding sub-model, and N keyword features are obtained by calling the first encoding sub-model: keyword feature 1, keyword feature 2...keyword feature N, and during each decoding, You can choose to copy text keywords or redefine new voting topic words. As shown in the right half of Figure 6, if it is determined to copy the text keywords, the corresponding text keywords can be directly determined as the voting topic words. If it is determined not to copy the text keywords, then the corresponding text keywords can be determined based on the last determined voting topic words. The current voting topic word.
可选地,第一生成模型的训练过程包括:Optionally, the training process of the first generation model includes:
获取正样本视频片段关联的样本文本内容,正样本视频片段为包含样本投票信息且样本投票信息的参与率达到目标阈值的视频片段;获取样本投票信息中包含的样本投票主题;基于样本文本内容和样本投票主题,调整第一生成模型中的模型参数。Obtain the sample text content associated with the positive sample video clip. The positive sample video clip is a video clip that contains sample voting information and the participation rate of the sample voting information reaches the target threshold; obtain the sample voting topic included in the sample voting information; based on the sample text content and Sample voting topics, adjusting model parameters in the first generative model.
其中,正样本视频片段为已创建好样本投票信息的视频片段,而且在播放该正样本视频片段时显示该样本投票信息的情况下,该样本投票信息的参与率达到了目标阈值,表示有很多用户在观看该正样本视频片段时参与了投票,该样本投票信息与该正样本视频片段具有较强的关联性,则该样本投票信息中包含的样本投票主题与该正样本视频片段也具有较强的关联性,因此基于正样本视频片段关联的样本文本内容和该样本投票主题训练第一生成模型,使得训练后的第一生成模型基于样本文本内容所得到的投票主题与该样本投票主题的相似度增大,从而提高第一生成模型的准确性,以使第一生成模型基于文本内容生成投票主题的功能得到提升。Among them, the positive sample video clip is a video clip for which sample voting information has been created, and when the sample voting information is displayed when the positive sample video clip is played, the participation rate of the sample voting information reaches the target threshold, indicating that there are many The user participated in voting while watching the positive sample video clip. The sample voting information has a strong correlation with the positive sample video clip. Then the sample voting topic contained in the sample voting information also has a strong correlation with the positive sample video clip. Strong correlation, so the first generative model is trained based on the sample text content associated with the positive sample video clip and the sample voting topic, so that the trained first generative model is based on the voting topic obtained from the sample text content and the sample voting topic. The similarity increases, thereby improving the accuracy of the first generation model, so that the function of the first generation model in generating voting topics based on text content is improved.
其中,该第一生成模型中的模型参数可以包括该第一生成模型中各个层中的权重参数或其他参数等,如该模型参数包括用于确定使用概率的权重参数w,通过一次或多次训练调整模型参数,能够提高权重参数w的准确性,提高所确定的使用概率的准确性,进而提高最终确定的投票主题的准确性。The model parameters in the first generation model may include weight parameters or other parameters in each layer of the first generation model. For example, the model parameters include the weight parameter w used to determine the probability of use. Through one or more Training and adjusting the model parameters can improve the accuracy of the weight parameter w, improve the accuracy of the determined usage probability, and thereby improve the accuracy of the final determined voting topic.
可选地,训练第一生成模型时还可以引入负样本视频片段,负样本视频片段为已创建好样本投票信息但参与率未达到目标阈值的视频片段,或者未创建样本投票信息的视频片段。Optionally, negative sample video clips can also be introduced when training the first generation model. Negative sample video clips are video clips for which sample voting information has been created but the participation rate has not reached the target threshold, or video clips for which no sample voting information has been created.
若是在播放该负样本视频片段时显示该样本投票信息的情况下,该样本投票信息的参与率未达到目标阈值,表示有很多用户在观看该负样本视频片段时未参与投票,该样本投票信息与该负样本视频片段关联性不强,则该样本投票信息中包含的样本投票主题与该负样本视频片段的关联性也不强,该样本投票主题不应当作为该负样本视频片段对应的投票主题,因此基于负样本视频片段关联的样本文本内容和该样本投票主题训练第一生成模型,使得训练后的第一生成模型基于该样本文本内容所得到的投票主题与该样本投票主题的相似度减小,能够提高第一生成模型的准确性,以避免第一生成模型基于文本内容生成不合适的投票主题。If the sample voting information is displayed when the negative sample video clip is played, and the participation rate of the sample voting information does not reach the target threshold, it means that many users did not participate in voting when watching the negative sample video clip. The sample voting information If the correlation with the negative sample video clip is not strong, the sample voting topic contained in the sample voting information will not be strongly correlated with the negative sample video clip. The sample voting topic should not be used as the vote corresponding to the negative sample video clip. topic, so the first generative model is trained based on the sample text content associated with the negative sample video clip and the sample voting topic, so that the trained first generative model is based on the similarity between the voting topic obtained from the sample text content and the sample voting topic. By reducing, the accuracy of the first generative model can be improved to prevent the first generative model from generating inappropriate voting topics based on text content.
本申请实施例提供了自动生成视频片段的投票主题的方法,通过对视频片段的文本内容和投票主题进行建模,从而训练出第一生成模型,基于第一生成模型生成投票主题,能够对视频片段的内容进行深度理解,基于深度表示的特征生成投票主题,,所生成的投票主题与视频片段的文本内容相关,满足了视频的互动功能需求,有助于提升用户在观看视频片段过程中参与投票互动的积极性,进而提升了互动氛围。Embodiments of the present application provide a method for automatically generating voting topics for video clips. By modeling the text content and voting topics of video clips, a first generative model is trained, and a voting topic is generated based on the first generative model, which enables the video to be The content of the clip is deeply understood, and the voting topic is generated based on the characteristics of the deep representation. The generated voting topic is related to the text content of the video clip, which meets the interactive functional requirements of the video and helps to improve user participation in the process of watching the video clip. The enthusiasm of voting interaction further enhances the interactive atmosphere.
在上述图3所示实施例的基础上,在一种可能实现方式中,上述步骤304-306中生成投 票候选项的过程可以基于第二生成模型执行,该第二生成模型包括第二编码子模型和第二解码子模型,其中,第二编码子模型用于将关键词编码为关键词特征,第二解码子模型用于将关键词特征解码为投票候选项词语。如图7所示,基于第二生成模型生成投票候选项的过程包括:Based on the above embodiment shown in Figure 3, in one possible implementation manner, the investment generated in the above steps 304-306 The process of voting candidates may be performed based on a second generative model, which includes a second encoding sub-model and a second decoding sub-model, wherein the second encoding sub-model is used to encode keywords into keyword features, and the second encoding sub-model is used to encode keywords into keyword features. The binary decoding sub-model is used to decode keyword features into voting candidate words. As shown in Figure 7, the process of generating voting candidates based on the second generation model includes:
701、计算机设备获取第一文本内容中的第一关键词,对第二文本内容进行聚类,得到多个文本类别,每个文本类别包含至少一条第二文本内容,分别从每个文本类别中抽取第二关键词。701. The computer device obtains the first keyword in the first text content, clusters the second text content, and obtains multiple text categories. Each text category contains at least one piece of second text content. From each text category, Extract the second keyword.
本步骤701的具体过程与上述步骤304-305类似,在此不再赘述。The specific process of this step 701 is similar to the above-mentioned steps 304-305, and will not be described again here.
702、计算机设备调用第二编码子模型,对第一关键词、投票主题和第i个文本类别的第二关键词进行编码,得到关键词特征。702. The computer device calls the second encoding sub-model to encode the first keyword, the voting topic and the second keyword of the i-th text category to obtain keyword features.
其中,第二编码子模型为Transformer Encoder模型或者其他类型的编码模型。Among them, the second encoding sub-model is a Transformer Encoder model or other types of encoding models.
本申请实施例中,计算机设备获取到了视频片段的第一关键词、投票主题和第i个文本类别的第二关键词,需要综合考虑这些关键词来生成第i个投票候选项。则将第一关键词、投票主题和第i个文本类别的第二关键词输入至第二生成模型的第二编码子模型中,且以所输入的关键词的数量为M为例,从而能够基于第二编码子模型对M个关键词分别进行编码,得到M个关键词特征。In the embodiment of this application, the computer device obtains the first keyword of the video clip, the voting topic, and the second keyword of the i-th text category, and needs to comprehensively consider these keywords to generate the i-th voting candidate. Then input the first keyword, the voting topic and the second keyword of the i-th text category into the second encoding sub-model of the second generation model, and take the number of input keywords as M as an example, so that The M keywords are encoded respectively based on the second encoding sub-model to obtain M keyword features.
703、计算机设备调用第二解码子模型,对M个关键词特征进行解码,得到第1个解码特征,基于第1个解码特征和M个关键词特征,确定第一投票候选项词语。703. The computer device calls the second decoding sub-model, decodes the M keyword features, obtains the first decoding feature, and determines the first voting candidate word based on the first decoding feature and the M keyword features.
704、计算机设备调用第二解码子模型,对M个关键词特征和第一投票候选项词语进行解码,得到第2个解码特征,基于第2个解码特征和M个关键词特征,确定参考投票候选项,参考投票候选项包括第一投票候选项词语和第二投票候选项词语,直至通过M次解码后,将第M次解码得到的参考投票候选项确定为视频片段的第i个投票候选项。704. The computer device calls the second decoding sub-model, decodes the M keyword features and the first voting candidate words, obtains the second decoding feature, and determines the reference vote based on the second decoding feature and the M keyword features. Candidates, reference voting candidates include the first voting candidate word and the second voting candidate word, until after M times of decoding, the reference voting candidate obtained by the M-th decoding is determined as the i-th voting candidate of the video clip item.
在本申请实施例中,采用了逐次解码生成投票候选项词语的方式,每进行一次解码,确定当前的投票候选项词语,当前的投票候选项词语与之前确定的投票候选项词语按照顺序组合,即可得到当前的参考投票候选项。随着解码的多次进行,投票候选项中包含的投票候选项词语逐渐增多,直至进行M次解码后,第M次解码得到的参考投票候选项包含了M个投票候选项词语,从而得到了包含M个投票候选项词语的投票候选项。In the embodiment of this application, a method of sequential decoding is used to generate voting candidate words. Each time decoding is performed, the current voting candidate words are determined. The current voting candidate words are combined with the previously determined voting candidate words in order. You can get the current reference voting candidates. As decoding is carried out multiple times, the number of voting candidate words contained in the voting candidates gradually increases, until after M times of decoding, the reference voting candidate obtained by decoding for the Mth time contains M voting candidate words, thus obtaining A voting candidate containing M voting candidate words.
本申请实施例中,每次解码时不仅考虑M个关键词特征,还会考虑之前确定的投票候选项词语,这样能保证本次确定的投票候选项词语与之前确定的投票候选项词语之间相关联,从而保证基于所确定的M个投票候选项词语构成的投票候选项中的不同投票候选项词语关联,能够组合构成一条语义清楚的语句。In the embodiment of this application, not only the M keyword features are considered during each decoding, but also the previously determined voting candidate words are considered. This ensures that the voting candidate words determined this time are consistent with the previously determined voting candidate words. Correlation, thereby ensuring that the association of different voting candidate words in the voting candidates composed of the determined M voting candidate words can be combined to form a sentence with clear semantics.
可选地,第二生成模型还包括第二分类层和预设词语库,预设词语库包括多个词语,基于第1个解码特征和M个关键词特征,确定第一投票候选项词语,包括:Optionally, the second generation model also includes a second classification layer and a preset word library. The preset word library includes a plurality of words. Based on the first decoding feature and the M keyword features, the first voting candidate word is determined, include:
(1)基于第1个解码特征和M个关键词特征确定M个第三使用概率,第j个第三使用概率为在投票候选项中使用第j个关键词的概率,j为正整数,且j不大于M。该第j个第三使用概率也即是将该第j个关键词确定为投票候选项中的第一投票候选项词语的概率。(1) Determine M third usage probabilities based on the 1st decoding feature and M keyword features. The jth third usage probability is the probability of using the jth keyword in the voting candidate, j is a positive integer, And j is not greater than M. The j-th third usage probability is also the probability of determining the j-th keyword as the first voting candidate word among the voting candidates.
(2)在第j个第三使用概率满足使用条件的情况下,将第j个关键词确定为第一投票候选项词语。其中,使用条件是指在投票候选项中使用关键词所需满足的条件。(2) When the j-th third usage probability satisfies the usage condition, determine the j-th keyword as the first voting candidate word. Among them, the usage conditions refer to the conditions that need to be met when using keywords in voting candidates.
(3)在每个第三使用概率不满足使用条件的情况下,调用第二分类层基于第1个解码特征和预设词语库进行分类,得到预设词语库中的每个词语的分类概率,基于每个词语的分类概率,确定第一投票候选项词语。其中,词语的分类概率表示表示将该词语确定为第一投票候选项词语的概率。(3) When each third usage probability does not meet the usage conditions, call the second classification layer to classify based on the first decoding feature and the preset word library to obtain the classification probability of each word in the preset word library , based on the classification probability of each word, determine the first voting candidate word. The classification probability of a word represents the probability of determining the word as the first voting candidate word.
本申请实施例中提供了两种确定投票候选项词语的方式,一种是采用复制机制,将关键词复制为投票候选项词语,也即是如果关键词的第三使用概率满足使用条件,则将关键词直接作为了投票候选项词语。另一种是根据预设词语库生成新的投票候选项词语。则在第一解码子模型每次进行解码时,首先确定使用概率,根据使用概率是否满足使用条件,确定是要 使用关键词作为投票候选项词语还是要生成新的投票候选项词语。The embodiments of this application provide two ways to determine voting candidate words. One is to use a copy mechanism to copy keywords into voting candidate words. That is, if the third usage probability of the keyword meets the usage conditions, then Keywords are directly used as voting candidates. The other is to generate new voting candidate words based on the preset word library. Then every time the first decoding sub-model performs decoding, the usage probability is first determined, and based on whether the usage probability meets the usage conditions, it is determined whether to Using keywords as voting candidate words still requires generating new voting candidate words.
可选地,使用条件包括使用概率阈值,在第j个第三使用概率大于使用概率阈值的情况下,将第j个关键词确定为第一投票候选项词语,其中在多个第三使用概率均大于使用概率阈值的情况下,将最大的第三使用概率的关键词确定为第一投票主题词语,而在每个第三使用概率不大于使用概率阈值的情况下,再确定预设词语库中的每个词语的分类概率,该分类概率表示将该词语确定为投票候选项词语的可能性,则基于每个词语的分类概率,能够从多个词语中选取出第一投票候选项词语。Optionally, the usage condition includes a usage probability threshold. When the jth third usage probability is greater than the usage probability threshold, the jth keyword is determined as the first voting candidate word, wherein among multiple third usage probabilities If both are greater than the usage probability threshold, the keyword with the largest third usage probability will be determined as the first voting topic word, and if each third usage probability is not greater than the usage probability threshold, then the preset word library will be determined The classification probability of each word in , the classification probability represents the possibility of determining the word as a voting candidate word, then based on the classification probability of each word, the first voting candidate word can be selected from multiple words.
可选地,基于每个词语的分类概率,确定第一投票候选项词语,包括:从预设词语库中的多个词语中,选取分类概率最大的词语,作为第一投票候选项词语,或者,从预设词语库中的多个词语中,选取分类概率最大的目标数量个词语,分别作为候选的第一投票候选项词语。其中,目标数量大于1。Optionally, determining the first voting candidate word based on the classification probability of each word includes: selecting the word with the highest classification probability from multiple words in the preset word library as the first voting candidate word, or , from the multiple words in the preset word library, select the target number of words with the highest classification probability as the first voting candidate words. Among them, the number of targets is greater than 1.
可选地,基于第2个解码特征和M个关键词特征,确定参考投票候选项,参考投票候选项包括第一投票候选项词语和第二投票候选项词语,包括:Optionally, based on the second decoding feature and the M keyword features, the reference voting candidates are determined, and the reference voting candidates include the first voting candidate words and the second voting candidate words, including:
(1)基于第2个解码特征和M个关键词特征确定M个第四使用概率,第j个第四使用概率为在投票候选项中使用第j个关键词的概率,j为正整数,且j不大于N。该第j个第四使用概率也即是将该第j个关键词确定为投票候选项中的第二投票候选项词语的概率。(1) Determine M fourth usage probabilities based on the second decoding feature and M keyword features. The jth fourth usage probability is the probability of using the jth keyword in the voting candidate, j is a positive integer, And j is not greater than N. The j-th fourth usage probability is also the probability of determining the j-th keyword as the second voting candidate word among the voting candidates.
(2)在第j个第四使用概率满足使用条件的情况下,将第j个关键词确定为第二投票候选项词语。其中,该使用条件是指在投票候选项中使用关键词所需满足的条件。(2) When the jth fourth usage probability satisfies the usage condition, determine the jth keyword as the second voting candidate word. Among them, the usage conditions refer to the conditions that need to be met when using keywords in voting candidates.
(3)在每个第四使用概率不满足使用条件的情况下,调用第一分类层基于第2个解码特征和预设词语库进行分类,得到多个候选投票候选项的分类概率,每个候选投票候选项包括一个第一投票候选项词语和一个第二投票候选项词语。其中,选投票候选项的分类概率表示将该选投票候选项确定为参考投票候选项的概率。(3) When each fourth usage probability does not meet the usage conditions, the first classification layer is called to classify based on the second decoding feature and the preset word library to obtain the classification probabilities of multiple candidate voting candidates. Each The candidate voting candidates include a first voting candidate word and a second voting candidate word. Among them, the classification probability of the voting candidate represents the probability of determining the voting candidate as the reference voting candidate.
(4)基于每个候选投票候选项的分类概率,确定参考投票候选项。(4) Based on the classification probability of each candidate voting candidate, determine the reference voting candidate.
可选地,基于每个候选投票候选项的分类概率,确定参考投票候选项,包括:从多个候选投票候选项中,选取分类概率最大的候选投票候选项,作为参考投票候选项,或者,从多个候选投票候选项中,选取分类概率最大的目标数量个候选投票候选项,分别作为参考投票候选项。其中,目标数量大于1。后续可以针对每个投票候选项生成对应的投票候选项,进而构成目标数量个投票信息。Optionally, determining the reference voting candidate based on the classification probability of each candidate voting candidate includes: selecting the voting candidate with the highest classification probability from multiple voting candidates as the reference voting candidate, or, From multiple candidate voting candidates, a target number of candidate voting candidates with the highest classification probability are selected as reference voting candidates respectively. Among them, the number of targets is greater than 1. Subsequently, corresponding voting candidates can be generated for each voting candidate, thereby forming a target number of voting information.
第二次解码的过程与第一次解码的过程类似,区别在于:在第二次进行解码时,会将第一次解码得到的第一投票候选项词语也输入到第一解码子模型中,从而将第一投票候选项词语,与预设词语库中的每个词语构成了备选的投票候选项,通过确定多个备选的投票候选项的分类概率,从多个备选的投票候选项中确定当前的投票候选项,这样能保证考虑到预设词语库中的词语与第一投票候选项词语之间的关联性,而备选的投票候选项的分类概率也能体现预设词语库中的词语与第一投票候选项词语组合得到的备选的投票候选项的合理程度,从而保证能生成语义清楚、逻辑通顺的投票候选项。The process of the second decoding is similar to the process of the first decoding. The difference is that during the second decoding, the first voting candidate words obtained by the first decoding will also be input into the first decoding sub-model. Thus, the first voting candidate word and each word in the preset word library constitute alternative voting candidates. By determining the classification probabilities of multiple alternative voting candidates, the candidate voting candidates are selected from multiple alternative voting candidates. The current voting candidate is determined in the item, which ensures that the correlation between the words in the preset word library and the first voting candidate word is taken into account, and the classification probability of the alternative voting candidate can also reflect the preset word The reasonable degree of the alternative voting candidates obtained by combining the words in the library with the first voting candidate words ensures that voting candidates with clear semantics and smooth logic can be generated.
需要说明的是,本申请实施例仅是对第一次解码和第二次解码进行说明,而在N大于2的情况下,在第二次解码之后的解码过程均与第二次解码过程类似,在此不再赘述。It should be noted that the embodiment of the present application only explains the first decoding and the second decoding. When N is greater than 2, the decoding process after the second decoding is similar to the second decoding process. , which will not be described in detail here.
而在解码过程中,基于第i个解码特征和M个关键词特征,采用如下公式确定第i个使用概率:During the decoding process, based on the i-th decoding feature and M keyword features, the following formula is used to determine the i-th usage probability:
Pi=e^(X*wi*svi)/sumj(e^(X*wi*svi)),其中,Pi表示第i个使用概率,X表示当前的解码特征,wi表示第i个关键词特征的权重参数,svi表示第i个关键词特征,j为任一正整数,j不大于M,Xj表示第j个解码特征,wj表示第j个关键词特征的权重参数,svj表示第j个关键词特征。P i =e^(X*w i *sv i )/sum j (e^(X*w i *sv i )), where P i represents the i-th usage probability, X represents the current decoding feature, w i represents the weight parameter of the i-th keyword feature, sv i represents the i-th keyword feature, j is any positive integer, j is not greater than M, X j represents the j-th decoding feature, w j represents the j-th key The weight parameter of the word feature, sv j represents the jth keyword feature.
图8是本申请实施例提供的一种第二生成模型的示意图,参见图8,计算机设备对弹幕文本内容进行聚类得到多个文本类别,从每个文本类别中抽取的第二关键词,即为弹幕文本关键词。计算机设备获取到的M个关键词包括对白文本关键词及字幕文本关键词,投票主题 中的关键词以及第i个文本类别对应的弹幕文本关键词,将M个关键词输入到第二编码子模型中,通过调用第二编码子模型得到了M个关键词特征:关键词特征1、关键词特征2…关键词特征M,而通过调用第二解码子模型,在每次解码时,可以选择是复制文本关键词,还是重新确定新的投票候选项词语。如图8右半部分所示,如果确定要复制文本关键词,可直接将对应的文本关键词确定为投票候选项词语,如果确定不复制文本关键词,则根据上一次确定的投票候选项词语,确定当前的投票候选项词语。Figure 8 is a schematic diagram of a second generation model provided by an embodiment of the present application. Referring to Figure 8, the computer device clusters the text content of the barrage to obtain multiple text categories, and the second keywords extracted from each text category are , which is the barrage text keyword. The M keywords obtained by the computer device include dialogue text keywords, subtitle text keywords, and voting topics. The keywords in and the barrage text keywords corresponding to the i-th text category, input M keywords into the second encoding sub-model, and obtain M keyword features by calling the second encoding sub-model: keyword features 1. Keyword feature 2...keyword feature M, and by calling the second decoding sub-model, during each decoding, you can choose whether to copy the text keyword or re-determine new voting candidate words. As shown in the right half of Figure 8, if it is determined to copy text keywords, the corresponding text keywords can be directly determined as voting candidate words. If it is determined not to copy the text keywords, then based on the last determined voting candidate words , determine the current voting candidate words.
可选地,第二生成模型的训练过程包括:Optionally, the training process of the second generative model includes:
获取正样本视频片段关联的样本文本内容,正样本视频片段为包含样本投票信息且样本投票信息的参与率达到目标阈值的视频片段;获取样本投票信息中包含的样本投票主题和多个样本投票候选项;基于每个样本文本内容与每个样本投票候选项之间的关联度,确定每个样本投票候选项对应的文本类别,文本类别包括样本投票候选项关联的样本文本内容;分别从每个样本投票候选项对应的文本类别中抽取样本关键词;基于样本文本内容、样本投票主题、多个样本投票候选项以及每个样本投票候选项的样本关键词,调整第二生成模型中的模型参数。Obtain the sample text content associated with the positive sample video clip. The positive sample video clip is a video clip that contains sample voting information and the participation rate of the sample voting information reaches the target threshold; obtain the sample voting topic and multiple sample voting candidates included in the sample voting information. items; based on the correlation between each sample text content and each sample voting candidate, determine the text category corresponding to each sample voting candidate. The text category includes the sample text content associated with the sample voting candidate; respectively, from each sample voting candidate Extract sample keywords from the text category corresponding to the sample voting candidate; adjust the model parameters in the second generation model based on the sample text content, the sample voting topic, the multiple sample voting candidates, and the sample keywords of each sample voting candidate. .
其中,正样本视频片段为已创建好样本投票信息的视频片段,而且在播放该正样本视频片段时显示该样本投票信息的情况下,该样本投票信息的参与率达到了目标阈值,表示有很多用户在观看该正样本视频片段时参与了投票,该样本投票信息与该正样本视频片段具有较强的关联性,则该样本投票信息中包含的样本投票主题和多个样本投票候选项与该正样本视频片段也具有较强的关联性,因此基于正样本视频片段关联的样本文本内容和该样本投票信息训练第二生成模型,使得训练后的第二生成模型基于样本文本内容所得到的投票候选项与该样本投票候选项的相似度增大,从而提高第二生成模型的准确性,以使第二生成模型基于文本内容生成投票候选项的功能得到提升。Among them, the positive sample video clip is a video clip for which sample voting information has been created, and when the sample voting information is displayed when the positive sample video clip is played, the participation rate of the sample voting information reaches the target threshold, indicating that there are many The user participated in voting while watching the positive sample video clip, and the sample voting information has a strong correlation with the positive sample video clip, then the sample voting topic and multiple sample voting candidates included in the sample voting information are related to the sample voting information. Positive sample video clips also have strong correlation, so the second generative model is trained based on the sample text content associated with the positive sample video clip and the sample voting information, so that the trained second generative model is based on the votes obtained from the sample text content. The similarity between the candidate and the sample voting candidate increases, thereby improving the accuracy of the second generation model, so that the function of the second generation model in generating voting candidates based on text content is improved.
而且,正样本视频片段关联的样本文本内容中包含与每个样本投票候选项关联的文本内容,通过确定每个样本文本内容与每个样本投票候选项之间的关联度,将每个样本文本内容划分到每个样本投票候选项的文本类别,能够区分出不同文本类别的样本文本内容,从而按照不同的文本类别,分别抽取出每个样本投票候选项对应的样本关键词,避免受到其他文本类别的干扰,以使第二生成模型具备根据不同文本类别的关键词生成投票候选项的功能。Moreover, the sample text content associated with the positive sample video clip contains text content associated with each sample voting candidate. By determining the correlation between each sample text content and each sample voting candidate, each sample text is The content is divided into the text categories of each sample voting candidate, which can distinguish the sample text content of different text categories, so that the sample keywords corresponding to each sample voting candidate can be extracted according to different text categories to avoid being affected by other texts. Category interference, so that the second generative model has the function of generating voting candidates based on keywords of different text categories.
其中,该第二生成模型中的模型参数可以包括该第二生成模型中各个层中的权重参数或其他参数等。The model parameters in the second generation model may include weight parameters or other parameters in each layer of the second generation model.
可选地,训练第二生成模型时还可以引入负样本视频片段,负样本视频片段为已创建好样本投票信息但参与率未达到目标阈值的视频片段,或者未创建样本投票信息的视频片段。Optionally, negative sample video clips can also be introduced when training the second generation model. Negative sample video clips are video clips for which sample voting information has been created but the participation rate has not reached the target threshold, or video clips for which no sample voting information has been created.
若是在播放该负样本视频片段时显示该样本投票信息的情况下,该样本投票信息的参与率未达到目标阈值,表示有很多用户在观看该负样本视频片段时未参与投票,该样本投票信息与该负样本视频片段关联性不强,则该样本投票信息中包含的样本投票主题以及样本投票候选项与该负样本视频片段的关联性也不强,该样本投票候选项不应当作为该负样本视频片段对应的投票候选项,或者,该样本投票候选项不应该作为该样本投票主题的候选项,因此基于负样本视频片段关联的样本文本内容、该样本投票主题和该多个样本投票候选项训练第二生成模型,使得训练后的第二生成模型基于样本文本内容所得到的投票候选项与该样本投票候选项的相似度减小,从而提高第二生成模型的准确性,以避免第二生成模型基于文本内容和投票主题生成不合适的投票候选项。If the sample voting information is displayed when the negative sample video clip is played, and the participation rate of the sample voting information does not reach the target threshold, it means that many users did not participate in voting when watching the negative sample video clip. The sample voting information If the correlation with the negative sample video clip is not strong, the sample voting topic and sample voting candidate included in the sample voting information will not be strongly correlated with the negative sample video clip, and the sample voting candidate should not be used as the negative sample video clip. The voting candidate corresponding to the sample video clip, or the sample voting candidate should not be used as a candidate for the sample voting topic, so based on the sample text content associated with the negative sample video clip, the sample voting topic and the multiple sample voting candidates The item trains the second generative model, so that the similarity between the voting candidates obtained by the trained second generative model based on the sample text content and the sample voting candidates is reduced, thereby improving the accuracy of the second generative model to avoid the second generation model. The second generative model generates inappropriate voting candidates based on text content and voting topics.
本申请实施例提供了自动生成视频片段的投票候选项的方法,通过对视频片段的文本内容和投票候选项进行建模,从而训练出第一生成模型,基于第一生成模型生成投票候选项,能够对视频片段的内容进行深度理解,基于深度表示的特征生成投票信息,提高了投票信息的准确性。所生成的投票信息与视频片段的文本内容相关,满足了视频的互动功能需求,有助于提升用户在观看视频片段过程中参与投票互动的积极性,进而提升了互动氛围。 Embodiments of the present application provide a method for automatically generating voting candidates for video clips. By modeling the text content and voting candidates of the video clips, a first generation model is trained, and voting candidates are generated based on the first generation model. It can deeply understand the content of video clips and generate voting information based on deep representation features, improving the accuracy of voting information. The generated voting information is related to the text content of the video clip, which meets the interactive functional requirements of the video, helps to increase the enthusiasm of users to participate in voting interaction while watching the video clip, and thereby enhances the interactive atmosphere.
在上述实施例的基础上,本申请实施例还提供了另一种投票信息生成方法,图9是本申请实施例提供的一种生成投票信息的整体流程的示意图,参见图9,该投票信息生成方法中可以先判定是否需要为视频片段生成投票信息,只有在确定需要为视频片段生成投票信息的情况下,才会执行生成投票信息的步骤,包括生成投票主题和生成投票候选项。在生成投票信息之后再基于登录账号进行个性化的投票信息显示。其中,下述图10所示的实施例将对判定是否需要生成投票信息的过程进行说明,而基于登录账号进行个性化的投票信息显示的过程详见下述图12和图13所示的实施例,下述图10所示的实施例暂不进行说明。On the basis of the above embodiments, embodiments of the present application also provide another method for generating voting information. Figure 9 is a schematic diagram of an overall process for generating voting information provided by embodiments of the present application. See Figure 9. The voting information In the generation method, it can first be determined whether voting information needs to be generated for the video clip. Only when it is determined that voting information needs to be generated for the video clip, the steps of generating voting information, including generating voting topics and generating voting candidates, will be performed. After the voting information is generated, personalized voting information is displayed based on the login account. Among them, the embodiment shown in Figure 10 below will explain the process of determining whether voting information needs to be generated, and the process of displaying personalized voting information based on the login account is detailed in the implementation shown in Figures 12 and 13 below. For example, the embodiment shown in FIG. 10 below will not be described for the time being.
图10是本申请实施例提供的一种投票信息生成方法的流程图。本申请实施例的执行主体为计算机设备,该计算机设备为终端或服务器。参见图10,该方法包括:Figure 10 is a flow chart of a voting information generation method provided by an embodiment of the present application. The execution subject of the embodiment of the present application is a computer device, and the computer device is a terminal or a server. Referring to Figure 10, the method includes:
1001、计算机设备获取第一文本内容、第二文本内容、第一热度参数和第二热度参数。1001. The computer device acquires the first text content, the second text content, the first popularity parameter and the second popularity parameter.
第一文本内容为视频片段包含的文本内容,第二文本内容为视频片段的弹幕包含的文本内容。获取第一文本内容和第二文本内容的过程详见上述步骤301,在此不再赘述。The first text content is the text content contained in the video clip, and the second text content is the text content contained in the barrage of the video clip. The process of obtaining the first text content and the second text content is detailed in step 301 above, and will not be described again here.
第一热度参数表示视频片段的热门程度,第二热度参数表示视频片段的弹幕的热门程度。The first popularity parameter represents the popularity of the video clip, and the second popularity parameter represents the popularity of the barrages of the video clip.
其中,该第一热度参数可以根据视频片段的播放次数确定,如该第一热度参数与该视频片段的播放次数正相关。或者,为了准确衡量在多个视频片段中该视频片段的热门程度,可以根据该视频片段的播放次数以及该计算机设备的多个视频片段的最大播放次数,确定该视频片段的第一热度参数。The first popularity parameter may be determined based on the number of times the video clip is played, for example, the first popularity parameter is positively correlated with the number of times the video clip is played. Alternatively, in order to accurately measure the popularity of the video segment among multiple video segments, the first popularity parameter of the video segment may be determined based on the number of times the video segment is played and the maximum number of times the multiple video segments are played by the computer device.
例如,采用以下公式确定第一热度参数:
Q1=log(1.0+num1)/log(1.0+num2),或者,Q1=log(1.0+num1)/log(1.0+num2)/R1;
For example, the following formula is used to determine the first heat parameter:
Q1=log(1.0+num1)/log(1.0+num2), or Q1=log(1.0+num1)/log(1.0+num2)/R1;
其中,Q1表示第一热度参数,num1表示该视频片段的播放次数,num2表示最大播放次数,R1表示第一热度参数区间的最大值。其中R1为预先设定的数值,如R1为0.1,表示将热度参数区间划分成了10个等级,采用上述公式Q1=log(1.0+num1)/log(1.0+num2)/R1所确定的第一热度参数Q1即为该视频片段的热度等级。Among them, Q1 represents the first popularity parameter, num1 represents the number of plays of the video clip, num2 represents the maximum number of plays, and R1 represents the maximum value of the first popularity parameter interval. Among them, R1 is a preset value. For example, R1 is 0.1, which means that the heat parameter interval is divided into 10 levels. The first level is determined by using the above formula Q1=log(1.0+num1)/log(1.0+num2)/R1. A popularity parameter Q1 is the popularity level of the video clip.
其中,该第二热度参数可以根据视频片段包括的弹幕数量或弹幕点赞次数中的至少一个确定,如该第二热度参数与该视频片段包括的弹幕数量或弹幕点赞次数中的至少一个正相关。或者,为了准确衡量在多个视频片段中该视频片段的热门程度,可以根据该视频片段包括的弹幕数量或弹幕点赞次数中的至少一个,以及该计算机设备的多个视频片段中的最大弹幕数量或最大弹幕点赞次数中的至少一个,确定第二热度参数。其中,视频片段的弹幕点赞次数为视频片段的弹幕的点赞次数的总和或视频片段的弹幕的点赞次数中的最大值。Wherein, the second popularity parameter can be determined based on at least one of the number of barrages or the number of likes on barrages included in the video clip, such as the second popularity parameter and the number of barrages or the number of likes on barrages included in the video clip. at least one positive correlation. Alternatively, in order to accurately measure the popularity of the video clip among multiple video clips, the video clip may be based on at least one of the number of barrages or the number of barrage likes included in the video clip, and the number of barrage likes among the multiple video clips of the computer device. At least one of the maximum number of barrages or the maximum number of likes on barrages determines the second popularity parameter. Among them, the number of likes of the barrage of the video clip is the sum of the number of likes of the barrage of the video clip or the maximum value of the number of likes of the barrage of the video clip.
例如,采用以下公式确定第二热度参数:
Q2=log(1.0+num3+num4)/log(1.0+num5+num6);
For example, the following formula is used to determine the second heat parameter:
Q2=log(1.0+num3+num4)/log(1.0+num5+num6);
或者,Q2=log(1.0+num3+num4)/log(1.0+num5+num6)/R2;Or, Q2=log(1.0+num3+num4)/log(1.0+num5+num6)/R2;
其中,Q2表示第二热度参数,num3表示该视频片段包括的弹幕数量,num4表示视频片段的弹幕点赞次数,num5表示多个视频片段中的最大弹幕数量,即包含弹幕最多的视频片段包括的弹幕数量,num6表示多个视频片段中的最大弹幕点赞次数,即多个视频片段的弹幕点赞次数中的最大值,R2表示第二热度参数区间的最大值。其中R2为预先设定的数值,如R2为0.1,表示将第二热度参数区间划分成了10个等级,采用上述公式Q2=log(1.0+num3+num4)/log(1.0+num5+num6)/R2所确定的第二热度参数Q2即为该视频片段的弹幕热度等级。Among them, Q2 represents the second popularity parameter, num3 represents the number of barrages included in the video clip, num4 represents the number of barrage likes of the video clip, and num5 represents the maximum number of barrages in multiple video clips, that is, the one containing the most barrages The number of barrages included in the video clip, num6 represents the maximum number of barrage likes in multiple video clips, that is, the maximum number of barrage likes in multiple video clips, and R2 represents the maximum value of the second popularity parameter interval. R2 is a preset value. For example, R2 is 0.1, which means that the second heat parameter interval is divided into 10 levels. The above formula Q2=log(1.0+num3+num4)/log(1.0+num5+num6) is used The second popularity parameter Q2 determined by /R2 is the barrage popularity level of the video clip.
1002、计算机设备基于第一文本内容、第二文本内容、第一热度参数和第二热度参数,确定视频片段的投票标识,投票标识表示是否为视频片段生成投票信息。1002. The computer device determines a voting identifier of the video clip based on the first text content, the second text content, the first popularity parameter, and the second popularity parameter. The voting identifier indicates whether to generate voting information for the video clip.
在本申请实施例中,第一文本内容和第二文本内容表示视频片段的内容,第一热度参数和第二热度参数表示视频片段的热门程度,基于第一文本内容、第二文本内容、第一热度参数和第二热度参数确定视频片段的投票标识,综合考虑了视频片段的内容和热门程度,所确定的投票标识更能体现出一般用户是否会参与该视频片段的投票互动,是否会基于投票信息进行投票操作,从而保证所确定的投票标识能够准确地表示是否要为该视频片段生成投票信息。 In this embodiment of the present application, the first text content and the second text content represent the content of the video clip, and the first popularity parameter and the second popularity parameter represent the popularity of the video clip. Based on the first text content, the second text content, and the third The first popularity parameter and the second popularity parameter determine the voting identification of the video clip. The content and popularity of the video clip are comprehensively considered. The determined voting identification can better reflect whether the average user will participate in the voting interaction of the video clip and whether it will be based on The voting information is used to perform a voting operation to ensure that the determined voting identification can accurately represent whether voting information is to be generated for the video clip.
可选地,步骤1002包括:Optionally, step 1002 includes:
获取第一文本内容的第一文本特征和第二文本内容的第二文本特征,该第一文本特征和该第二文本特征分别用于描述第一文本内容和第二文本内容,之后获取第一热度参数的第一热度特征和第二热度参数的第二热度特征,该第一热度特征和该第二热度特征分别用于描述该第一热度参数和该第二热度参数,将第一文本特征、第二文本特征、第一热度特征和第二热度特征进行拼接,得到视频片段特征,以使该视频片段特征能够准确描述该视频片段,则基于视频片段特征进行分类,得到投票标识。Obtain the first text feature of the first text content and the second text feature of the second text content. The first text feature and the second text feature are used to describe the first text content and the second text content respectively, and then obtain the first text feature. The first heat feature of the heat parameter and the second heat feature of the second heat parameter. The first heat feature and the second heat feature are used to describe the first heat parameter and the second heat parameter respectively. The first text feature is , the second text feature, the first popularity feature and the second popularity feature are spliced to obtain the video clip features, so that the video clip features can accurately describe the video clip, then the video clip features are classified based on the video clip features to obtain the voting identification.
其中,计算机设备可以先确定热度特征表,热度特征表包含至少一种热度参数对应的热度特征。通过查询该热度特征表即可确定第一热度参数的第一热度特征和第二热度参数的第二热度特征。例如,该热度特征表包括第一热度特征表和第二热度特征表,第一热度特征表包含视频片段的热度参数的热度特征,通过查询该第一热度特征表即可确定第一热度参数的第一热度特征,第二热度特征表包含视频片段的弹幕的热度参数的热度特征,通过查询该第二热度特征表即可确定第二热度参数的第二热度特征。Wherein, the computer device may first determine a heat feature table, where the heat feature table includes heat features corresponding to at least one heat parameter. By querying the heat feature table, the first heat feature of the first heat parameter and the second heat feature of the second heat parameter can be determined. For example, the heat feature table includes a first heat feature table and a second heat feature table. The first heat feature table contains heat features of the heat parameters of the video clips. The first heat feature table can be determined by querying the first heat feature table. The first popularity feature and the second popularity feature table include the popularity feature of the popularity parameter of the barrage of the video clip. The second popularity feature of the second popularity parameter can be determined by querying the second popularity feature table.
1003、计算机设备在投票标识表示为视频片段生成投票信息的情况下,基于文本内容中的关键词,生成视频片段的投票主题,基于关键词和投票主题,生成视频片段的多个投票候选项,基于投票主题和多个投票候选项,生成视频片段的投票信息。1003. When the voting identifier indicates that the voting information is generated for the video clip, the computer device generates a voting topic for the video clip based on the keywords in the text content, and generates multiple voting candidates for the video clip based on the keywords and the voting topic. Based on the voting topic and multiple voting candidates, the voting information of the video clip is generated.
可选地,投票标识包括第一投票标识和第二投票标识,第一投票标识表示为视频片段生成投票信息,第二投票标识表示不为视频片段生成投票信息。则在所确定的投票标识为第一投票标识的情况下执行生成投票信息的步骤。例如第一投票标识为1,第二投票标识为0。或者,投票标识为生成投票信息的概率,所确定的概率大于概率阈值表示为视频片段生成投票信息,所确定的概率不大于概率阈值,表示不为视频片段生成投票信息。Optionally, the voting identification includes a first voting identification and a second voting identification. The first voting identification indicates that voting information is generated for the video clip, and the second voting identification indicates that voting information is not generated for the video clip. Then, when the determined voting identification is the first voting identification, the step of generating voting information is performed. For example, the first voting ID is 1 and the second voting ID is 0. Alternatively, the voting identifier is a probability of generating voting information. The determined probability is greater than the probability threshold, indicating that voting information is generated for the video clip. The determined probability is not greater than the probability threshold, indicating that voting information is not generated for the video clip.
其中,生成该视频片段的投票信息的步骤包括生成投票主题、生成多个投票候选项以及基于视频主题和投票候选项生成投票信息的步骤,具体过程详见上述实施例,在此不再赘述。The step of generating the voting information of the video clip includes the steps of generating a voting topic, generating a plurality of voting candidates, and generating voting information based on the video topic and the voting candidates. The specific process can be seen in the above embodiments and will not be described again here.
本申请实施例提供的方法,基于第一文本内容、第二文本内容、第一热度参数和第二热度参数对应的特征确定视频片段的投票标识,综合考虑了视频片段的内容和热门程度,即充分捕获了视频片段的内容特征同时还兼顾了视频片段的互动数据,所确定的投票标识更能体现出一般用户是否会参与该视频片段的投票互动,是否会基于投票信息进行投票操作,从而有效地确定是否要为该视频片段生成投票信息,避免了直接为视频的每个视频片段生成投票信息,在提升视频互动性的基础上还节省了处理量,减小了视频的数据量。The method provided by the embodiment of the present application determines the voting identifier of the video clip based on the characteristics corresponding to the first text content, the second text content, the first popularity parameter, and the second popularity parameter, taking into account the content and popularity of the video clip, that is, It fully captures the content characteristics of the video clip and also takes into account the interaction data of the video clip. The determined voting identification can better reflect whether the general user will participate in the voting interaction of the video clip and whether the voting operation will be based on the voting information, thus effectively It can accurately determine whether to generate voting information for this video clip, avoiding the need to directly generate voting information for each video clip. On the basis of improving the interactivity of the video, it also saves processing and reduces the amount of video data.
在上述图10所示的实施例的基础上,确定投票标识的过程可以基于投票判定模型执行,投票判定模型包括第一特征提取子模型、第一拼接层和第二分类层,投票判定模型还包括热度特征表,热度特征表包含至少一种热度参数对应的热度特征。可选地,该热度特征表包括上述第一热度特征表和第二热度特征表。Based on the above embodiment shown in Figure 10, the process of determining the voting identification can be performed based on the voting decision model. The voting decision model includes a first feature extraction sub-model, a first splicing layer and a second classification layer. The voting decision model also It includes a heat feature table, and the heat feature table includes heat features corresponding to at least one heat parameter. Optionally, the heat feature table includes the above-mentioned first heat feature table and second heat feature table.
相应地,确定投票标识的过程包括:Accordingly, the process of determining voting identification includes:
调用第一特征提取子模型,获取第一文本内容的第一文本特征和第二文本内容的第二文本特征;基于第一热度参数和第二热度参数查询热度特征表,得到第一热度特征和第二热度特征;调用第一拼接层,将第一文本特征、第二文本特征、第一热度特征和第二热度特征进行拼接,得到视频片段特征;调用第二分类层,基于视频片段特征进行分类,得到投票标识。Call the first feature extraction sub-model to obtain the first text feature of the first text content and the second text feature of the second text content; query the heat feature table based on the first heat parameter and the second heat parameter to obtain the first heat feature and The second hotness feature; call the first splicing layer to splice the first text feature, the second text feature, the first hotness feature and the second hotness feature to obtain the video clip features; call the second classification layer to perform the classification based on the video clip features Classify and get the voting ID.
可选地,该第一特征提取子模型为BERT(Bidirectional Encoder Representations from Transformers,基于变换的双向编码表示)模型或者其他类型的模型。Optionally, the first feature extraction sub-model is a BERT (Bidirectional Encoder Representations from Transformers) model or other types of models.
可选地,该视频片段包括多条弹幕的情况下,可以获取到多条弹幕文本内容,从而获取到多条弹幕文本内容的弹幕文本特征,对该多个弹幕文本特征进行最大值池化,得到第二文本特征,再将该第二文本特征与第一文本特征、第一热度特征和第二热度特征进行拼接,得到视频片段特征。Optionally, when the video clip includes multiple barrages, multiple barrage text contents can be obtained, thereby obtaining the barrage text features of the multiple barrage text contents, and the multiple barrage text features can be obtained. Maximum pooling is performed to obtain the second text feature, and then the second text feature is spliced with the first text feature, the first popularity feature, and the second popularity feature to obtain the video clip feature.
图11是本申请实施例提供的一种投票判定模型的示意图,参见图11,以第一特征提取 子模型为BERT模型为例,投票判定模型包括BERT模型、第一拼接层和第二分类层。计算机设备获取到视频片段的对白文本内容和字幕文本内容,以及第一热度参数,还获取到了视频片段的K条弹幕文本内容以及第二热度参数,K为大于1的整数。通过BERT模型提取对白文本内容和字幕文本内容对应的第一文本特征,以及K个弹幕文本内容对应的弹幕文本特征,对K个弹幕文本特征进行最大值池化得到第二文本特征。投票判定模型还包括第一热度特征表和第二热度特征表,通过查询第一热度特征表和第二热度特征表分别确定第一热度特征和第二热度特征。之后即可将获取到的特征进行拼接和分类,从而确定投票标识。Figure 11 is a schematic diagram of a voting decision model provided by the embodiment of the present application. Referring to Figure 11, the first feature extraction The sub-model is the BERT model as an example. The voting decision model includes the BERT model, the first splicing layer and the second classification layer. The computer device obtains the dialogue text content and subtitle text content of the video clip, as well as the first popularity parameter, and also obtains the K barrage text content of the video clip and the second popularity parameter, where K is an integer greater than 1. The BERT model is used to extract the first text features corresponding to the dialogue text content and subtitle text content, as well as the barrage text features corresponding to K barrage text contents, and perform maximum pooling on the K barrage text features to obtain the second text feature. The voting decision model also includes a first heat feature table and a second heat feature table, and the first heat feature and the second heat feature are determined respectively by querying the first heat feature table and the second heat feature table. The obtained features can then be spliced and classified to determine the voting identifier.
投票判定模型的训练过程包括:The training process of the voting decision model includes:
获取样本视频片段、样本视频片段关联的样本文本内容及样本文本内容的热度参数,样本视频片段包括正样本视频片段或负样本视频片段的至少一种,正样本视频片段为包含样本投票信息且样本投票信息的参与率达到目标阈值的视频片段,负样本视频片段为:包含样本投票信息且样本投票信息的参与率未达到目标阈值的视频片段,或不包含样本投票信息的视频片段的至少一种;基于样本视频片段、样本视频片段关联的样本文本内容及样本文本内容的热度参数,调整投票判定模型中的模型参数,模型参数包括热度特征表。Obtain the sample video clip, the sample text content associated with the sample video clip, and the popularity parameters of the sample text content. The sample video clip includes at least one of a positive sample video clip or a negative sample video clip. The positive sample video clip contains sample voting information and the sample Video clips whose participation rate of voting information reaches the target threshold. Negative sample video clips are: video clips that contain sample voting information and the participation rate of sample voting information does not reach the target threshold, or video clips that do not contain sample voting information. ; Based on the sample video clip, the sample text content associated with the sample video clip, and the popularity parameters of the sample text content, adjust the model parameters in the voting decision model. The model parameters include a popularity feature table.
通过对视频片段的文本内容和热度参数进行建模,从而训练出投票判定模型,能够使投票判定模型学习视频片段的文本内容及热度参数对是否要生成投票信息的结果的影响,从而提升投票判定模型的准确性,以便基于投票判定模型准确地确定投票标识,避免了直接为视频的每个视频片段生成投票信息,在提升视频互动性的基础上还节省了处理量,减小了视频的数据量。其中,该投票判定模型的训练目标为:该投票判定模型所得到的正样本视频片段的投票标识表示为该正样本视频片段生成投票信息,该投票判定模型所得到的负样本视频片段的投票标识表示不为该负样本视频片段生成投票信息。By modeling the text content and popularity parameters of the video clips, the voting decision model is trained, which enables the voting decision model to learn the impact of the text content and popularity parameters of the video clips on whether to generate voting information, thereby improving the voting decision. The accuracy of the model can accurately determine the voting identification based on the voting determination model, avoiding the need to directly generate voting information for each video segment of the video. On the basis of improving the interactivity of the video, it also saves processing and reduces the data of the video. quantity. Among them, the training objectives of the voting determination model are: the voting identification of the positive sample video clips obtained by the voting determination model represents the voting information generated for the positive sample video clips, and the voting identification of the negative sample video clips obtained by the voting determination model Indicates that no voting information will be generated for this negative sample video clip.
需要说明的是,热度特征表也作为投票判定模型的其中一种模型参数,则在投票判定模型的训练过程中,会基于训练样本调整热度特征表,以使热度特征表随着投票判定模型的训练而逐渐更新,提高热度特征表的准确性。It should be noted that the heat feature table is also used as one of the model parameters of the voting decision model. During the training process of the voting decision model, the heat feature table will be adjusted based on the training samples so that the heat feature table changes with the voting decision model. Gradually update through training to improve the accuracy of the hot feature table.
图12是本申请实施例提供的一种投票信息显示方法的流程图。本申请实施例的执行主体为终端,本申请实施例对生成视频片段的投票信息后显示投票信息的过程进行说明。参见图12,该方法包括:Figure 12 is a flow chart of a voting information display method provided by an embodiment of the present application. The execution subject of the embodiment of the present application is a terminal. The embodiment of the present application describes the process of generating the voting information of the video clip and then displaying the voting information. Referring to Figure 12, the method includes:
1201、终端基于视频中的视频片段,获取视频片段的投票信息。1201. The terminal obtains the voting information of the video clip based on the video clip.
其中,投票信息基于投票主题和多个投票候选项生成,投票主题基于视频片段关联的文本内容中的关键词生成,多个投票候选项基于关键词和投票主题生成。而文本内容包括第一文本内容或第二文本内容的至少一种,第一文本内容为视频片段包含的文本内容,第二文本内容为视频片段的弹幕包含的文本内容。生成视频片段的投票信息的过程详见上述投票信息生成方法的实施例,本申请实施例不再赘述。Among them, the voting information is generated based on the voting topic and multiple voting candidates, the voting topic is generated based on the keywords in the text content associated with the video clip, and the multiple voting candidates are generated based on the keywords and voting topics. The text content includes at least one of first text content or second text content. The first text content is the text content contained in the video clip, and the second text content is the text content contained in the barrage of the video clip. The process of generating voting information for video clips is detailed in the above embodiments of the voting information generation method, and will not be described in detail in the embodiments of this application.
并且,该投票信息可以由该终端生成,或者由除该终端以外的其他设备生成。例如,用于提供该视频的服务器生成一个或多个视频片段的投票信息,在该终端向该服务器请求该视频时,该服务器向该终端下发该视频以及所生成的投票信息,终端即可获取到待播放的视频以及一个或多个视频片段的投票信息。Furthermore, the voting information may be generated by the terminal or by other devices other than the terminal. For example, a server used to provide the video generates voting information for one or more video clips. When the terminal requests the video from the server, the server sends the video and the generated voting information to the terminal, and the terminal can Get the video to be played and the voting information of one or more video clips.
1202、终端基于当前登录的账号的兴趣标签和投票信息,确定互动参数,互动参数表示账号基于投票信息进行投票操作的可能性。1202. The terminal determines interaction parameters based on the interest tag and voting information of the currently logged-in account. The interaction parameters represent the possibility of the account performing a voting operation based on the voting information.
终端当前登录有账号,该账号的兴趣标签表示该账号所属用户的兴趣,终端基于该兴趣标签和该投票信息来确定互动参数,可以准确地判断当前用户是否对该投票信息感兴趣,从而保证所确定的互动参数能够准确地衡量当前用户参与投票的可能性,以及是否有为该视频片段生成投票信息的必要性。The terminal is currently logged in with an account, and the interest tag of the account indicates the interest of the user to whom the account belongs. The terminal determines the interaction parameters based on the interest tag and the voting information, and can accurately determine whether the current user is interested in the voting information, thus ensuring that all The determined interaction parameters can accurately measure the likelihood of the current user participating in voting and whether there is a need to generate voting information for this video clip.
1203、终端在互动参数满足互动条件的情况下,在播放视频片段时显示投票信息。1203. When the interaction parameters meet the interaction conditions, the terminal displays voting information when playing video clips.
只有在互动参数满足互动条件的情况下,即在当前用户对该投票信息感兴趣,可能会参 与投票的情况下,终端才会在播放视频片段时显示投票信息。其中,该互动条件是指播放视频片段时显示投票信息的条件。Only when the interaction parameters meet the interaction conditions, that is, when the current user is interested in the voting information, he may participate In the case of voting, the terminal will only display voting information when playing video clips. Among them, the interaction condition refers to the condition for displaying voting information when playing video clips.
而在互动参数不满足互动条件的情况下,即可认为当前用户对该投票信息不感兴趣,参与投票的概率较低,此时为避免对视频片段的播放过程造成干扰,在播放视频片段时不再显示投票信息。When the interaction parameters do not meet the interaction conditions, it can be considered that the current user is not interested in the voting information and the probability of participating in voting is low. At this time, in order to avoid causing interference to the playback process of the video clip, do not play the video clip. Show voting information again.
可选地,该互动条件包括互动参数阈值,在互动参数大于互动参数阈值的情况下,在播放视频片段时显示投票信息。而在互动参数不大于互动参数阈值的情况下,在播放该视频片段时不显示投票信息。Optionally, the interaction condition includes an interaction parameter threshold. If the interaction parameter is greater than the interaction parameter threshold, the voting information is displayed when the video clip is played. When the interaction parameter is not greater than the interaction parameter threshold, the voting information will not be displayed when the video clip is played.
需要说明的一点是,本申请实施例仅是以一个视频片段为例,对该视频片段的投票信息的显示过程进行说明,而在该视频包括多个视频片段,且多个视频信息具有投票信息的情况下,终端播放视频的过程中,按照各个视频片段的播放顺序,依次播放每个视频片段,在当前播放的视频片段具有对应的投票信息的情况下,基于该账号的兴趣标签和当前播放的视频片段对应的投票信息确定互动参数,在互动参数满足互动条件的情况下显示该投票信息,否则继续进行播放,直至播放到下一个视频片段。It should be noted that the embodiment of this application only takes one video clip as an example to illustrate the display process of voting information of this video clip, and the video includes multiple video clips, and the multiple video information has voting information. In the case of , during the video playback process of the terminal, each video clip is played in sequence according to the playback order of each video clip. When the currently played video clip has corresponding voting information, based on the interest tag of the account and the currently played The voting information corresponding to the video clip determines the interaction parameters. When the interaction parameters meet the interaction conditions, the voting information is displayed. Otherwise, playback continues until the next video clip is played.
需要说明的另一点是,本申请实施例仅是以一个视频片段具有一个投票信息为例进行说明。而在另一实施例中,一个视频片段具有多个投票信息,则从中选取对应的互动参数满足互动条件的投票信息,在播放该视频片段时进行显示。或者,对应的互动参数满足互动条件的投票信息为多个的情况下,选取互动参数最大的投票信息,在播放该视频片段时进行显示。Another point that needs to be noted is that the embodiment of the present application only takes one video clip with one voting information as an example for explanation. In another embodiment, a video clip has multiple voting information, and the voting information whose corresponding interaction parameters satisfy the interaction conditions is selected from them and displayed when the video clip is played. Or, when there are multiple voting information whose corresponding interaction parameters satisfy the interaction conditions, select the voting information with the largest interaction parameter and display it when the video clip is played.
本申请实施例提供的方法,基于当前登录的账号的兴趣标签和投票信息确定能够衡量用户参与投票可能性的互动参数,只有在互动参数满足互动条件的情况下,才会在播放视频片段时显示投票信息,从而更容易吸引用户参与投票,实现了投票信息的个性化显示,提升了用户的互动主动参与性,提升了互动体验,也避免了对参与投票概率较低的用户造成干扰。The method provided by the embodiment of this application determines the interaction parameters that can measure the possibility of the user participating in voting based on the interest tags and voting information of the currently logged-in account. Only when the interaction parameters meet the interaction conditions will they be displayed when playing the video clip. Voting information makes it easier to attract users to participate in voting, realizes personalized display of voting information, enhances users' active participation in interaction, improves interactive experience, and avoids interference to users with a low probability of participating in voting.
在上述图12所示实施例的基础上,本申请实施例还提供了另一种投票信息显示方法,对确定互动参数的过程进行具体说明。图13是本申请实施例提供的另一种投票信息显示方法的流程图。本申请实施例的执行主体为终端。参见图13,该方法包括:Based on the above-mentioned embodiment shown in Figure 12, the embodiment of the present application also provides another method of displaying voting information, and provides a detailed description of the process of determining interaction parameters. Figure 13 is a flow chart of another voting information display method provided by an embodiment of the present application. The execution subject of the embodiment of this application is a terminal. Referring to Figure 13, the method includes:
1301、终端基于视频中的视频片段,获取视频片段的投票信息。1301. The terminal obtains the voting information of the video clip based on the video clip.
其中,生成视频片段的投票信息的过程详见上述投票信息生成方法的实施例,本申请实施例不再赘述。The process of generating voting information for video clips can be found in the above-mentioned embodiments of the voting information generation method, and will not be described in detail in the embodiments of this application.
1302、终端获取账号的兴趣标签的兴趣特征、投票主题的投票主题特征和多个投票候选项的投票候选项特征。1302. The terminal obtains the interest characteristics of the account's interest tag, the voting topic characteristics of the voting topic, and the voting candidate characteristics of multiple voting candidates.
其中,该兴趣特征用于描述该兴趣标签,该投票主题特征用于描述该投票主体,该投票候选项特征用于描述投票候选项。且上述各个特征可以为向量、矩阵或其他形式。通过获取兴趣标签、投票主题和多个投票候选项的特征,能够将兴趣标签、投票主题和多个投票候选项进行准确量化,以便后续基于量化后的特征进行运算得到互动参数。The interest feature is used to describe the interest tag, the voting topic feature is used to describe the voting subject, and the voting candidate feature is used to describe the voting candidate. And each of the above features can be in the form of vectors, matrices or other forms. By obtaining the characteristics of interest tags, voting topics, and multiple voting candidates, the interest tags, voting topics, and multiple voting candidates can be accurately quantified, so that subsequent calculations can be performed based on the quantified features to obtain interaction parameters.
1303、终端将兴趣特征、投票主题特征和多个投票候选项特征进行拼接,得到互动特征。1303. The terminal splices the interest features, voting topic features and multiple voting candidate features to obtain interactive features.
1304、终端基于互动特征进行分类,得到互动参数。1304. The terminal classifies based on interaction features and obtains interaction parameters.
由于该互动特征由兴趣标签的特征、投票主题的特征和多个投票候选项的特征拼接得到,融合了兴趣标签、投票主题和多个投票候选项的信息,因此基于互动参数进行分类得到的互动参数能够考虑兴趣标签、投票主题和多个投票候选项的信息,更能准确衡量用户对投票信息的感兴趣程度,从而准确确定用户参与投票的可能性。Since the interaction feature is spliced by the characteristics of the interest tag, the characteristics of the voting topic and the characteristics of multiple voting candidates, and integrates the information of the interest tag, voting topic and multiple voting candidates, the interaction obtained by classifying based on the interaction parameters Parameters can consider information about interest tags, voting topics, and multiple voting candidates, and can more accurately measure the user's interest in voting information, thereby accurately determining the user's likelihood of participating in voting.
1305、终端在互动参数满足互动条件的情况下,在播放视频片段时显示投票信息。1305. When the interaction parameters meet the interaction conditions, the terminal displays voting information when playing video clips.
该步骤与上述步骤1203类似,在此不再赘述。This step is similar to the above-mentioned step 1203 and will not be described again.
本申请实施例提供的方法,基于当前登录的账号的兴趣标签和投票信息确定能够衡量用户参与投票可能性的互动参数,只有在互动参数满足互动条件的情况下,才会在播放视频片段时显示投票信息,从而吸引用户参与投票,实现了投票信息的个性化显示,提高了互动覆 盖率,也避免了对参与投票概率较低的用户造成干扰。The method provided by the embodiment of this application determines the interaction parameters that can measure the possibility of the user participating in voting based on the interest tags and voting information of the currently logged-in account. Only when the interaction parameters meet the interaction conditions will they be displayed when playing the video clip. Voting information, thereby attracting users to participate in voting, realizing personalized display of voting information, and improving interactive coverage. Coverage rate also avoids causing interference to users with a low probability of participating in voting.
在一种可能实现方式中,上述步骤1302-1304确定互动参数的过程可以基于投票互动模型执行,该投票互动模型包括第二特征提取子模型、第二拼接层和第三分类层。相应地,确定互动参数的过程包括:In one possible implementation, the process of determining interaction parameters in the above steps 1302-1304 can be performed based on a voting interaction model, which includes a second feature extraction sub-model, a second splicing layer, and a third classification layer. Accordingly, the process of determining interaction parameters includes:
调用第二特征提取子模型,获取兴趣标签的兴趣特征、投票主题的投票主题特征和多个投票候选项分别对应的投票候选项特征;调用第二拼接层,将兴趣特征、投票主题特征和多个投票候选项特征进行拼接,得到互动特征;调用第三分类层,基于互动特征进行分类,得到互动参数。Call the second feature extraction sub-model to obtain the interest features of the interest tag, the voting topic features of the voting topic, and the voting candidate features corresponding to multiple voting candidates; call the second splicing layer to combine the interest features, voting topic features, and multiple voting candidates. The features of voting candidates are spliced together to obtain interactive features; the third classification layer is called to classify based on the interactive features to obtain interactive parameters.
该兴趣特征、该投票主题特征和多个投票候选项特征分别用于描述当前用户的兴趣标签、投票主题和多个投票候选项,而兴趣特征、投票主题特征和多个投票候选项特征能够综合考虑当前用户的兴趣标签、投票主题和多个投票候选项的影响,从而基于互动特征进行分类得到互动参数,提高了互动参数的准确性。The interest feature, the voting topic feature and the multiple voting candidate features are respectively used to describe the current user's interest tag, voting topic and multiple voting candidates, and the interest feature, voting topic feature and multiple voting candidate features can be integrated Taking into account the influence of the current user's interest tags, voting topics and multiple voting candidates, the interaction parameters are obtained through classification based on interaction features, which improves the accuracy of the interaction parameters.
其中,该第二特征提取子模型为BERT模型或其他类型的模型,本申请实施例对此不做限定。Wherein, the second feature extraction sub-model is a BERT model or other types of models, which is not limited in the embodiments of the present application.
图14是本申请实施例提供的一种投票互动模型的示意图,参见图14,以第二特征提取子模型为BERT模型为例,投票互动模型包括BERT模型、第二拼接层和第三分类层。计算机设备获取到账号的兴趣标签、投票主题和M个投票候选项,M为大于1的整数。通过BERT模型提取兴趣标签、投票主题和M个投票候选项对应的特征,再通过第二拼接层和第三分类层分别进行拼接和分类,从而得到互动参数。Figure 14 is a schematic diagram of a voting interaction model provided by an embodiment of the present application. Refer to Figure 14. Taking the second feature extraction sub-model as the BERT model as an example, the voting interaction model includes the BERT model, the second splicing layer and the third classification layer. . The computer device obtains the account's interest tag, voting topic, and M voting candidates, where M is an integer greater than 1. The BERT model is used to extract the features corresponding to the interest tags, voting topics and M voting candidates, and then splicing and classifying are performed through the second splicing layer and the third classification layer respectively to obtain the interaction parameters.
可选地,投票互动模型的训练过程包括:Optionally, the training process of the voting interaction model includes:
获取样本账号的样本兴趣标签以及样本视频片段中的样本投票信息,其中,样本账号已执行基于样本投票信息进行投票操作;基于样本兴趣标签、样本投票信息中的样本投票主题和多个样本投票候选项,调整投票互动模型中的模型参数。Obtain the sample interest tag of the sample account and the sample voting information in the sample video clip. The sample account has performed a voting operation based on the sample voting information; based on the sample interest tag, the sample voting topic and multiple sample voting candidates in the sample voting information Term, adjust the model parameters in the voting interaction model.
其中,该样本账号为任意账号,该样本兴趣标签表示该样本账号所属用户的兴趣,该样本视频片段包括样本投票信息,且样本账号已执行基于样本投票信息进行投票操作,表示在播放该样本视频片段时显示了该样本投票信息,且该样本账号所属用户对该样本投票信息感兴趣,已经参与了投票互动。因此,基于样本兴趣标签、样本投票信息中的样本投票主题和多个样本投票候选项,调整投票互动模型中的模型参数,能够使得投票互动模型学习到样本兴趣标签与样本投票信息之间的关联关系,从而具备基于任一兴趣标签和任一投票信息确定对应的互动参数的功能,提高了投票互动模型的准确性。其中,该投票互动模型的训练目标为:使得该投票互动模型所得到的互动参数增大,也即是使得该投票互动模型能够基于样本兴趣标签、样本投票主题和多个样本投票候选项预测出该样本账号进行投票操作的可能性较大。例如该互动参数表示账号基于投票信息进行投票操作的概率,则训练目标为使得投票互动模型所得到的互动参数为1。Among them, the sample account is any account, the sample interest tag indicates the interest of the user to whom the sample account belongs, the sample video clip includes sample voting information, and the sample account has performed a voting operation based on the sample voting information, indicating that the sample video is being played The sample voting information is displayed in the clip, and the user to whom the sample account belongs is interested in the sample voting information and has participated in the voting interaction. Therefore, adjusting the model parameters in the voting interaction model based on the sample interest labels, the sample voting topics in the sample voting information, and multiple sample voting candidates can enable the voting interaction model to learn the association between the sample interest labels and the sample voting information. relationship, thus having the function of determining the corresponding interaction parameters based on any interest tag and any voting information, which improves the accuracy of the voting interaction model. Among them, the training goal of the voting interaction model is to increase the interaction parameters obtained by the voting interaction model, that is, to enable the voting interaction model to predict based on sample interest labels, sample voting topics and multiple sample voting candidates. This sample account is more likely to conduct voting operations. For example, the interaction parameter represents the probability of the account performing a voting operation based on voting information, and the training goal is to make the interaction parameter obtained by the voting interaction model equal to 1.
需要说明的是,上述可选方案仅是以一次训练为例,而在实际训练过程中,需要获取多组样本数据,每组样本数据包括样本账号对应的样本兴趣标签以及样本视频片段中的样本投票信息,基于该多组样本数据分别对投票互动模型进行迭代训练,从而提高投票互动模型的准确性,进而提高投票互动模型所确定的互动参数的准确性。It should be noted that the above optional solutions only take one training as an example. In the actual training process, multiple sets of sample data need to be obtained. Each set of sample data includes sample interest tags corresponding to the sample account and samples in the sample video clips. Based on the voting information, the voting interaction model is iteratively trained based on the multiple sets of sample data, thereby improving the accuracy of the voting interaction model, and thereby improving the accuracy of the interaction parameters determined by the voting interaction model.
图15是本申请实施例提供的一种投票信息生成装置的结构示意图,该装置设置于计算机设备中,参见图15,该装置包括:Figure 15 is a schematic structural diagram of a voting information generation device provided by an embodiment of the present application. The device is installed in a computer device. Referring to Figure 15, the device includes:
文本内容获取模块1501,用于获取视频中的视频片段关联的文本内容,文本内容包括第一文本内容或第二文本内容的至少一种,第一文本内容为视频片段包含的文本内容,第二文本内容为视频片段的弹幕包含的文本内容;The text content acquisition module 1501 is used to acquire text content associated with video clips in the video. The text content includes at least one of first text content or second text content. The first text content is the text content contained in the video clip, and the second text content is the text content contained in the video clip. The text content is the text content contained in the barrage of the video clip;
主题生成模块1502,用于基于文本内容中的关键词,生成视频片段的投票主题; The topic generation module 1502 is used to generate voting topics for video clips based on keywords in the text content;
候选项生成模块1503,用于基于关键词和投票主题,生成视频片段的多个投票候选项;The candidate generation module 1503 is used to generate multiple voting candidates for video clips based on keywords and voting topics;
投票信息生成模块1504,用于基于投票主题和多个投票候选项,生成视频片段的投票信息。The voting information generation module 1504 is used to generate voting information for video clips based on the voting topic and multiple voting candidates.
可选地,主题生成模块1502,包括:Optionally, the topic generation module 1502 includes:
编码单元,用于对关键词进行编码,得到关键词的关键词特征;The encoding unit is used to encode keywords and obtain the keyword characteristics of the keywords;
解码单元,用于对关键词特征进行解码,得到投票主题,投票主题由多个投票主题词语构成。The decoding unit is used to decode the keyword features to obtain the voting topic. The voting topic is composed of multiple voting topic words.
可选地,第一生成模型包括第一编码子模型和第一解码子模型;Optionally, the first generation model includes a first encoding sub-model and a first decoding sub-model;
编码单元,用于:调用第一编码子模型,对N个关键词进行编码,得到N个关键词的关键词特征,N为大于1的整数;The encoding unit is used to: call the first encoding sub-model, encode the N keywords, and obtain the keyword features of the N keywords, where N is an integer greater than 1;
解码单元,用于:Decoding unit for:
调用第一解码子模型,对N个关键词特征进行解码,得到第1个解码特征,基于第1个解码特征和N个关键词特征,确定第一投票主题词语;Call the first decoding sub-model to decode the N keyword features to obtain the first decoding feature. Based on the first decoding feature and the N keyword features, determine the first voting topic word;
调用第一解码子模型,对N个关键词特征和第一投票主题词语进行解码,得到第2个解码特征,基于第2个解码特征和N个关键词特征,确定参考投票主题,参考投票主题包括第一投票主题词语和第二投票主题词语,直至通过N次解码后,将第N次解码得到的参考投票主题确定为视频片段的投票主题。Call the first decoding sub-model to decode the N keyword features and the first voting topic words to obtain the second decoding feature. Based on the second decoding feature and the N keyword features, determine the reference voting topic and the reference voting topic. Including the first voting topic words and the second voting topic words, until after N decodings, the reference voting topic obtained by the Nth decoding is determined as the voting topic of the video clip.
可选地,第一生成模型还包括第一分类层和预设词语库,预设词语库包括多个词语,解码单元,用于:Optionally, the first generation model also includes a first classification layer and a preset word library. The preset word library includes a plurality of words and a decoding unit for:
基于第1个解码特征和N个关键词特征确定N个第一使用概率,第j个第一使用概率为在投票主题中使用第j个关键词的概率,j为正整数,且j不大于N;Determine N first usage probabilities based on the 1st decoding feature and N keyword features. The jth first usage probability is the probability of using the jth keyword in the voting topic. j is a positive integer and j is not greater than N;
在第j个第一使用概率满足使用条件的情况下,将第j个关键词确定为第一投票主题词语;When the jth first usage probability satisfies the usage condition, the jth keyword is determined as the first voting topic word;
在每个第一使用概率不满足使用条件的情况下,调用第一分类层基于第1个解码特征和预设词语库进行分类,得到预设词语库中的每个词语的分类概率,基于每个词语的分类概率,确定第一投票主题词语。When each first usage probability does not meet the usage conditions, the first classification layer is called to classify based on the first decoding feature and the preset word library, and the classification probability of each word in the preset word library is obtained. Based on each The classification probability of each word determines the first voting topic word.
可选地,解码单元,用于:Optionally, a decoding unit is used for:
基于第2个解码特征和N个关键词特征确定N个第二使用概率,第j个第二使用概率为在投票主题中使用第j个关键词的概率,j为正整数,且j不大于N;Determine N second usage probabilities based on the 2nd decoding feature and N keyword features. The jth second usage probability is the probability of using the jth keyword in the voting topic. j is a positive integer, and j is not greater than N;
在第j个第二使用概率满足使用条件的情况下,将第j个关键词确定为第二投票主题词语;When the j-th second usage probability satisfies the usage conditions, the j-th keyword is determined as the second voting topic word;
在每个第二使用概率不满足使用条件的情况下,调用第一分类层基于第2个解码特征和预设词语库进行分类,得到多个候选投票主题的分类概率,每个候选投票主题包括一个第一投票主题词语和一个第二投票主题词语;When each second usage probability does not meet the usage conditions, the first classification layer is called to classify based on the second decoding feature and the preset word library to obtain the classification probabilities of multiple candidate voting topics. Each candidate voting topic includes One first voting topic word and one second voting topic word;
基于每个候选投票主题的分类概率,确定参考投票主题。Based on the classification probability of each candidate voting topic, the reference voting topic is determined.
可选地,第一生成模型的训练过程包括:Optionally, the training process of the first generation model includes:
获取正样本视频片段关联的样本文本内容,正样本视频片段为包含样本投票信息且样本投票信息的参与率达到目标阈值的视频片段;Obtain the sample text content associated with the positive sample video clip. The positive sample video clip is a video clip that contains sample voting information and the participation rate of the sample voting information reaches the target threshold;
获取样本投票信息中包含的样本投票主题;Get the sample voting topic contained in the sample voting information;
基于样本文本内容和样本投票主题,调整第一生成模型中的模型参数。Based on the sample text content and the sample voting topic, model parameters in the first generative model are adjusted.
可选地,文本内容包括第一文本内容和第二文本内容,候选项生成模块1503,包括:Optionally, the text content includes first text content and second text content, and the candidate generation module 1503 includes:
第一获取单元,用于获取第一关键词,第一关键词为第一文本内容中的关键词;The first acquisition unit is used to acquire the first keyword, which is a keyword in the first text content;
聚类单元,用于对第二文本内容进行聚类,得到多个文本类别,每个文本类别包含至少一条第二文本内容;A clustering unit is used to cluster the second text content to obtain multiple text categories, each text category containing at least one piece of second text content;
第二获取单元,用于分别从每个文本类别中抽取第二关键词;The second acquisition unit is used to extract the second keyword from each text category respectively;
生成单元,用于基于第一关键词、投票主题和每个文本类别的第二关键词,分别生成每 个文本类别的投票候选项。The generation unit is used to respectively generate each keyword based on the first keyword, the voting topic and the second keyword of each text category. Voting candidates for text categories.
可选地,生成单元,用于:Optionally, generate units for:
对第一关键词、投票主题和第i个文本类别的第二关键词进行编码,得到关键词特征,i为正整数,i不大于文本类别的数量;Encode the first keyword, voting topic and the second keyword of the i-th text category to obtain keyword features, i is a positive integer, and i is not greater than the number of text categories;
对关键词特征进行解码,得到第i个投票候选项,第i个投票候选项由多个投票候选项词语构成。The keyword features are decoded to obtain the i-th voting candidate, which consists of multiple voting candidate words.
可选地,第二生成模型包括第二编码子模型和第二解码子模型;生成单元,用于:Optionally, the second generation model includes a second encoding sub-model and a second decoding sub-model; a generation unit, used for:
调用第二编码子模型,对第一关键词、投票主题和第i个文本类别的第二关键词进行编码,得到关键词特征;Call the second encoding sub-model to encode the first keyword, voting topic and the second keyword of the i-th text category to obtain keyword features;
调用第二解码子模型,对M个关键词特征进行解码,得到第1个解码特征,基于第1个解码特征和M个关键词特征,确定第一投票候选项词语,M为大于1的整数;Call the second decoding sub-model to decode the M keyword features to obtain the first decoding feature. Based on the first decoding feature and the M keyword features, determine the first voting candidate word, where M is an integer greater than 1. ;
调用第二解码子模型,对M个关键词特征和第一投票候选项词语进行解码,得到第2个解码特征,基于第2个解码特征和M个关键词特征,确定参考投票候选项,参考投票候选项包括第一投票候选项词语和第二投票候选项词语,直至通过M次解码后,将第M次解码得到的参考投票候选项确定为视频片段的第i个投票候选项。The second decoding sub-model is called to decode the M keyword features and the first voting candidate words to obtain the second decoding feature. Based on the second decoding feature and the M keyword features, the reference voting candidates are determined. Refer to The voting candidates include the first voting candidate words and the second voting candidate words, until after M times of decoding, the reference voting candidate obtained by the M-th decoding is determined as the i-th voting candidate of the video clip.
可选地,第二生成模型的训练过程包括:Optionally, the training process of the second generative model includes:
获取正样本视频片段关联的样本文本内容,正样本视频片段为包含样本投票信息且样本投票信息的参与率达到目标阈值的视频片段;Obtain the sample text content associated with the positive sample video clip. The positive sample video clip is a video clip that contains sample voting information and the participation rate of the sample voting information reaches the target threshold;
获取样本投票信息中包含的样本投票主题和多个样本投票候选项;Obtain the sample voting topic and multiple sample voting candidates contained in the sample voting information;
基于每个样本文本内容与每个样本投票候选项之间的关联度,确定每个样本投票候选项的文本类别,文本类别包括样本投票候选项关联的样本文本内容;Based on the correlation between each sample text content and each sample voting candidate, determine the text category of each sample voting candidate, where the text category includes the sample text content associated with the sample voting candidate;
分别从每个样本投票候选项的文本类别中抽取样本关键词;Extract sample keywords from the text category of each sample voting candidate respectively;
基于样本文本内容、样本投票主题、多个样本投票候选项以及每个样本投票候选项的样本关键词,调整第二生成模型中的模型参数。Based on the sample text content, the sample voting topic, the plurality of sample voting candidates, and the sample keywords of each sample voting candidate, the model parameters in the second generation model are adjusted.
可选地,文本内容获取模块1501,用于:Optionally, the text content acquisition module 1501 is used for:
获取第一文本内容、第二文本内容、第一热度参数和第二热度参数,第一热度参数表示视频片段的热门程度,第二热度参数表示视频片段的弹幕的热门程度;Obtain the first text content, the second text content, the first popularity parameter and the second popularity parameter. The first popularity parameter represents the popularity of the video clip, and the second popularity parameter represents the popularity of the barrage of the video clip;
主题生成模块1502,用于:Topic generation module 1502, used for:
基于第一文本内容、第二文本内容、第一热度参数和第二热度参数,确定视频片段的投票标识,投票标识表示是否为视频片段生成投票信息;Based on the first text content, the second text content, the first popularity parameter and the second popularity parameter, determine the voting identification of the video clip, where the voting identification indicates whether to generate voting information for the video clip;
在投票标识表示为视频片段生成投票信息的情况下,基于所获取的文本内容中的关键词,生成视频片段的投票主题。In the case where the voting identification indicates that voting information is generated for the video clip, a voting topic of the video clip is generated based on the keywords in the obtained text content.
可选地,主题生成模块1502,用于:Optionally, the topic generation module 1502 is used for:
获取第一文本内容的第一文本特征和第二文本内容的第二文本特征;Obtain the first text feature of the first text content and the second text feature of the second text content;
获取第一热度参数的第一热度特征和第二热度参数的第二热度特征;Obtain the first heat feature of the first heat parameter and the second heat feature of the second heat parameter;
将第一文本特征、第二文本特征、第一热度特征和第二热度特征进行拼接,得到视频片段特征;Splice the first text feature, the second text feature, the first popularity feature and the second popularity feature to obtain video clip features;
基于视频片段特征进行分类,得到投票标识。Classify based on the characteristics of the video clips to obtain the voting identification.
可选地,投票判定模型包括第一特征提取层、第一拼接层和第二分类层,投票判定模型还包括热度特征表,热度特征表包含至少一种热度参数的热度特征;Optionally, the voting decision model includes a first feature extraction layer, a first splicing layer and a second classification layer, the voting decision model also includes a heat feature table, the heat feature table includes heat features of at least one heat parameter;
主题生成模块1502,用于:Topic generation module 1502, used for:
调用第一特征提取子模型,获取第一文本内容的第一文本特征和第二文本内容的第二文本特征;Call the first feature extraction sub-model to obtain the first text feature of the first text content and the second text feature of the second text content;
基于第一热度参数和第二热度参数查询热度特征表,得到第一热度特征和第二热度特征;Query the heat feature table based on the first heat parameter and the second heat parameter to obtain the first heat feature and the second heat feature;
调用第一拼接层,将第一文本特征、第二文本特征、第一热度特征和第二热度特征进行拼接,得到视频片段特征; Call the first splicing layer to splice the first text feature, the second text feature, the first popularity feature and the second popularity feature to obtain the video clip features;
调用第二分类层,基于视频片段特征进行分类,得到投票标识。The second classification layer is called to classify based on the characteristics of the video clips and obtain the voting identification.
可选地,投票判定模型的训练过程包括:Optionally, the training process of the voting decision model includes:
获取样本视频片段、样本视频片段关联的样本文本内容及样本文本内容的热度参数,样本视频片段包括正样本视频片段或负样本视频片段的至少一种,正样本视频片段为包含样本投票信息且样本投票信息的参与率达到目标阈值的视频片段,负样本视频片段为:包含样本投票信息且样本投票信息的参与率未达到目标阈值的视频片段,或不包含样本投票信息的视频片段的至少一种;Obtain the sample video clip, the sample text content associated with the sample video clip, and the popularity parameters of the sample text content. The sample video clip includes at least one of a positive sample video clip or a negative sample video clip. The positive sample video clip contains sample voting information and the sample Video clips whose participation rate of voting information reaches the target threshold. Negative sample video clips are: video clips that contain sample voting information and the participation rate of sample voting information does not reach the target threshold, or video clips that do not contain sample voting information. ;
基于样本视频片段、样本视频片段关联的样本文本内容及样本文本内容的热度参数,调整投票判定模型中的模型参数,模型参数包括热度特征表。Based on the sample video clip, the sample text content associated with the sample video clip, and the popularity parameters of the sample text content, adjust the model parameters in the voting decision model, and the model parameters include a popularity feature table.
可选地,装置还包括:Optionally, the device also includes:
互动参数确定模块,用于基于账号的兴趣标签和投票信息,确定互动参数,互动参数表示账号基于投票信息进行投票操作的可能性;The interaction parameter determination module is used to determine the interaction parameters based on the account's interest tags and voting information. The interaction parameters represent the possibility of the account performing voting operations based on the voting information;
发送模块,用于在互动参数满足互动条件的情况下,向账号的终端发送投票信息,投票信息用于在终端播放视频片段时展示。The sending module is used to send voting information to the terminal of the account when the interaction parameters meet the interaction conditions. The voting information is used to display when the terminal plays the video clip.
图16是本申请实施例提供的一种投票信息显示装置的结构示意图。参见图16,该装置包括:Figure 16 is a schematic structural diagram of a voting information display device provided by an embodiment of the present application. Referring to Figure 16, the device includes:
信息获取模块1601,用于基于视频中的视频片段,获取视频片段的投票信息,投票信息基于投票主题和多个投票候选项生成,投票主题基于视频片段关联的文本内容中的关键词生成,多个投票候选项基于关键词和投票主题生成;The information acquisition module 1601 is used to obtain the voting information of the video clip based on the video clip in the video. The voting information is generated based on the voting topic and multiple voting candidates. The voting topic is generated based on the keywords in the text content associated with the video clip, and more Voting candidates are generated based on keywords and voting topics;
参数确定模块1602,用于基于当前登录的账号的兴趣标签和投票信息,确定互动参数,互动参数表示账号基于投票信息进行投票操作的可能性;The parameter determination module 1602 is used to determine interaction parameters based on the interest tags and voting information of the currently logged-in account. The interaction parameters represent the possibility of the account performing voting operations based on the voting information;
信息显示模块1603,用于在互动参数满足互动条件的情况下,在播放视频片段时显示投票信息;Information display module 1603, used to display voting information when playing video clips when the interaction parameters meet the interaction conditions;
其中,文本内容包括第一文本内容或第二文本内容的至少一种,第一文本内容为视频片段包含的文本内容,第二文本内容为视频片段的弹幕包含的文本内容。The text content includes at least one of first text content or second text content. The first text content is the text content contained in the video clip, and the second text content is the text content contained in the barrage of the video clip.
可选地,投票信息包括投票主题和多个投票候选项,参数确定模块1602,包括:Optionally, the voting information includes a voting topic and multiple voting candidates. The parameter determination module 1602 includes:
特征获取单元,用于获取账号的兴趣标签的兴趣特征、投票主题的投票主题特征和多个投票候选项的投票候选项特征;The feature acquisition unit is used to acquire the interest features of the account's interest tag, the voting topic features of the voting topic, and the voting candidate features of multiple voting candidates;
拼接单元,用于将兴趣特征、投票主题特征和多个投票候选项特征进行拼接,得到互动特征;The splicing unit is used to splice interest features, voting topic features and multiple voting candidate features to obtain interactive features;
分类单元,用于基于互动特征进行分类,得到互动参数。Classification unit is used to classify based on interaction features and obtain interaction parameters.
可选地,投票互动模型包括第二特征提取子模型、第二拼接层和第三分类层;Optionally, the voting interaction model includes a second feature extraction sub-model, a second splicing layer and a third classification layer;
特征获取单元,用于:调用第二特征提取子模型,获取兴趣标签的兴趣特征、获取投票主题的投票主题特征和多个投票候选项的投票候选项特征;The feature acquisition unit is used for: calling the second feature extraction sub-model to obtain the interest features of the interest tag, the voting topic features of the voting topic, and the voting candidate features of multiple voting candidates;
拼接单元,用于:调用第二拼接层,将兴趣特征、投票主题特征和多个投票候选项特征进行拼接,得到互动特征;The splicing unit is used to: call the second splicing layer to splice the interest features, voting topic features and multiple voting candidate features to obtain interactive features;
分类单元,用于:调用第三分类层,基于互动特征进行分类,得到互动参数。Classification unit is used to: call the third classification layer, classify based on interaction features, and obtain interaction parameters.
可选地,投票互动模型的训练过程包括:Optionally, the training process of the voting interaction model includes:
获取样本账号的样本兴趣标签以及样本视频片段中的样本投票信息,其中,样本账号已执行基于样本投票信息进行投票操作;Obtain the sample interest tag of the sample account and the sample voting information in the sample video clip, where the sample account has performed a voting operation based on the sample voting information;
基于样本兴趣标签、样本投票信息中的样本投票主题和多个样本投票候选项,调整投票互动模型中的模型参数。Based on the sample interest tags, sample voting topics in the sample voting information, and multiple sample voting candidates, adjust the model parameters in the voting interaction model.
需要说明的是:上述实施例提供的投票信息生成装置或投票信息显示装置在生成或显示投票信息时,仅以上述各功能模块的划分进行举例说明,实际应用中,可以根据需要而将上 述功能分配由不同的功能模块完成,即将计算机设备的内部结构划分成不同的功能模块,以完成以上描述的全部或者部分功能。另外,上述实施例提供的投票信息生成装置与投票信息生成方法实施例属于同一构思,上述实施例提供的投票信息显示装置与投票信息显示方法实施例属于同一构思,其具体实现过程详见方法实施例,这里不再赘述。It should be noted that when the voting information generation device or the voting information display device provided in the above embodiments generates or displays voting information, only the division of the above functional modules is used as an example. In practical applications, the above functions can be used as needed. The above function allocation is completed by different functional modules, that is, the internal structure of the computer device is divided into different functional modules to complete all or part of the functions described above. In addition, the voting information generation device and the voting information generation method embodiments provided by the above embodiments belong to the same concept. The voting information display device and the voting information display method embodiments provided by the above embodiments belong to the same concept. For details of the implementation process, please refer to the method implementation. For example, we won’t go into details here.
本申请实施例还提供了一种计算机设备,该计算机设备包括处理器和存储器,存储器中存储有至少一条计算机程序,该至少一条计算机程序由处理器加载并执行,以实现上述实施例的投票信息生成方法或投票信息显示方法所执行的操作。Embodiments of the present application also provide a computer device. The computer device includes a processor and a memory. At least one computer program is stored in the memory. The at least one computer program is loaded and executed by the processor to implement the voting information of the above embodiments. The operation performed by the generate method or voting information display method.
可选地,该计算机设备提供为终端,图17示出了本申请一个示例性实施例提供的终端1700的结构示意图。Optionally, the computer device is provided as a terminal. Figure 17 shows a schematic structural diagram of a terminal 1700 provided by an exemplary embodiment of the present application.
终端1700包括有:处理器1701和存储器1702。The terminal 1700 includes: a processor 1701 and a memory 1702.
处理器1701可以包括一个或多个处理核心,比如4核心处理器、8核心处理器等。处理器1701可以采用DSP(Digital Signal Processing,数字信号处理)、FPGA(Field Programmable Gate Array,现场可编程门阵列)、PLA(Programmable Logic Array,可编程逻辑阵列)中的至少一种硬件形式来实现。处理器1701也可以包括主处理器和协处理器,主处理器是用于对在唤醒状态下的数据进行处理的处理器,也称CPU(Central Processing Unit,中央处理器);协处理器是用于对在待机状态下的数据进行处理的低功耗处理器。在一些实施例中,处理器1701可以集成有GPU(Graphics Processing Unit,图像处理的交互器),GPU用于负责显示屏所需要显示的内容的渲染和绘制。一些实施例中,处理器1701还可以包括AI(Artificial Intelligence,人工智能)处理器,该AI处理器用于处理有关机器学习的计算操作。The processor 1701 may include one or more processing cores, such as a 4-core processor, an 8-core processor, etc. The processor 1701 can be implemented using at least one hardware form among DSP (Digital Signal Processing, digital signal processing), FPGA (Field Programmable Gate Array, field programmable gate array), and PLA (Programmable Logic Array, programmable logic array). . The processor 1701 can also include a main processor and a co-processor. The main processor is a processor used to process data in the wake-up state, also called CPU (Central Processing Unit, central processing unit); the co-processor is A low-power processor used to process data in standby mode. In some embodiments, the processor 1701 may be integrated with a GPU (Graphics Processing Unit, an image processing interface), and the GPU is responsible for rendering and drawing the content that needs to be displayed on the display screen. In some embodiments, the processor 1701 may also include an AI (Artificial Intelligence, artificial intelligence) processor, which is used to process computing operations related to machine learning.
存储器1702可以包括一个或多个计算机可读存储介质,该计算机可读存储介质可以是非暂态的。存储器1702还可包括高速随机存取存储器,以及非易失性存储器,比如一个或多个磁盘存储设备、闪存存储设备。在一些实施例中,存储器1702中的非暂态的计算机可读存储介质用于存储至少一条计算机程序,该至少一条计算机程序用于被处理器1701所具有以实现本申请中方法实施例提供的投票信息生成方法。Memory 1702 may include one or more computer-readable storage media, which may be non-transitory. Memory 1702 may also include high-speed random access memory, as well as non-volatile memory, such as one or more disk storage devices, flash memory storage devices. In some embodiments, the non-transitory computer-readable storage medium in the memory 1702 is used to store at least one computer program, and the at least one computer program is used to be possessed by the processor 1701 to implement the methods provided by the method embodiments in this application. Voting information generation method.
在一些实施例中,终端1700还可选包括有:外围设备接口1703和至少一个外围设备。处理器1701、存储器1702和外围设备接口1703之间可以通过总线或信号线相连。各个外围设备可以通过总线、信号线或电路板与外围设备接口1703相连。可选地,外围设备包括:射频电路1704、显示屏1705和摄像头组件1706中的至少一种。In some embodiments, the terminal 1700 optionally further includes: a peripheral device interface 1703 and at least one peripheral device. The processor 1701, the memory 1702 and the peripheral device interface 1703 may be connected through a bus or a signal line. Each peripheral device can be connected to the peripheral device interface 1703 through a bus, a signal line, or a circuit board. Optionally, the peripheral device includes: at least one of a radio frequency circuit 1704, a display screen 1705, and a camera assembly 1706.
外围设备接口1703可被用于将I/O(Input/Output,输入/输出)相关的至少一个外围设备连接到处理器1701和存储器1702。在一些实施例中,处理器1701、存储器1702和外围设备接口1703被集成在同一芯片或电路板上;在一些其他实施例中,处理器1701、存储器1702和外围设备接口1703中的任意一个或两个可以在单独的芯片或电路板上实现,本实施例对此不加以限定。The peripheral device interface 1703 may be used to connect at least one I/O (Input/Output) related peripheral device to the processor 1701 and the memory 1702 . In some embodiments, the processor 1701, the memory 1702, and the peripheral device interface 1703 are integrated on the same chip or circuit board; in some other embodiments, any one of the processor 1701, the memory 1702, and the peripheral device interface 1703 or Both of them can be implemented on separate chips or circuit boards, which is not limited in this embodiment.
射频电路1704用于接收和发射RF(Radio Frequency,射频)信号,也称电磁信号。射频电路1704通过电磁信号与通信网络以及其他通信设备进行通信。射频电路1704将电信号转换为电磁信号进行发送,或者,将接收到的电磁信号转换为电信号。可选地,射频电路1704包括:天线系统、RF收发器、一个或多个放大器、调谐器、振荡器、数字信号处理器、编解码芯片组、用户身份模块卡等等。射频电路1704可以通过至少一种无线通信协议来与其它设备进行通信。该无线通信协议包括但不限于:城域网、各代移动通信网络(2G、3G、4G及5G)、无线局域网和/或WiFi(Wireless Fidelity,无线保真)网络。在一些实施例中,射频电路1704还可以包括NFC(Near Field Communication,近距离无线通信)有关的电路,本申请对此不加以限定。The radio frequency circuit 1704 is used to receive and transmit RF (Radio Frequency, radio frequency) signals, also called electromagnetic signals. Radio frequency circuitry 1704 communicates with communication networks and other communication devices through electromagnetic signals. The radio frequency circuit 1704 converts electrical signals into electromagnetic signals for transmission, or converts received electromagnetic signals into electrical signals. Optionally, the radio frequency circuit 1704 includes: an antenna system, an RF transceiver, one or more amplifiers, a tuner, an oscillator, a digital signal processor, a codec chipset, a user identity module card, and the like. Radio frequency circuitry 1704 can communicate with other devices through at least one wireless communication protocol. The wireless communication protocol includes but is not limited to: metropolitan area network, mobile communication networks of all generations (2G, 3G, 4G and 5G), wireless LAN and/or WiFi (Wireless Fidelity, wireless fidelity) network. In some embodiments, the radio frequency circuit 1704 may also include NFC (Near Field Communication) related circuits, which is not limited in this application.
显示屏1705用于显示UI(User Interface,用户界面)。该UI可以包括图形、文本、图标、视频及其它们的任意组合。当显示屏1705是触摸显示屏时,显示屏1705还具有采集在显示屏1705的表面或表面上方的触摸信号的能力。该触摸信号可以作为控制信号输入至处理器 1701进行处理。此时,显示屏1705还可以用于提供虚拟按钮和/或虚拟键盘,也称软按钮和/或软键盘。在一些实施例中,显示屏1705可以为一个,设置在终端1700的前面板;在另一些实施例中,显示屏1705可以为至少两个,分别设置在终端1700的不同表面或呈折叠设计;在另一些实施例中,显示屏1705可以是柔性显示屏,设置在终端1700的弯曲表面上或折叠面上。甚至,显示屏1705还可以设置成非矩形的不规则图形,也即异形屏。显示屏1705可以采用LCD(Liquid Crystal Display,液晶显示屏)、OLED(Organic Light-Enitting Diode,有机发光二极管)等材质制备。The display screen 1705 is used to display UI (User Interface, user interface). The UI can include graphics, text, icons, videos, and any combination thereof. When display screen 1705 is a touch display screen, display screen 1705 also has the ability to collect touch signals on or above the surface of display screen 1705 . The touch signal can be input to the processor as a control signal 1701 for processing. At this time, the display screen 1705 can also be used to provide virtual buttons and/or virtual keyboards, also called soft buttons and/or soft keyboards. In some embodiments, there may be one display screen 1705, which is provided on the front panel of the terminal 1700; in other embodiments, there may be at least two display screens 1705, which are respectively provided on different surfaces of the terminal 1700 or have a folding design; In other embodiments, the display screen 1705 may be a flexible display screen disposed on a curved surface or a folding surface of the terminal 1700. Even, the display screen 1705 can be set into a non-rectangular irregular shape, that is, a special-shaped screen. The display screen 1705 can be made of materials such as LCD (Liquid Crystal Display) and OLED (Organic Light-Enitting Diode).
摄像头组件1706用于采集图像或视频。可选地,摄像头组件1706包括前置摄像头和后置摄像头。前置摄像头设置在终端1700的前面板,后置摄像头设置在终端1700的背面。在一些实施例中,后置摄像头为至少两个,分别为主摄像头、景深摄像头、广角摄像头、长焦摄像头中的任意一种,以实现主摄像头和景深摄像头融合实现背景虚化功能、主摄像头和广角摄像头融合实现全景拍摄以及VR(Virtual Reality,虚拟现实)拍摄功能或者其它融合拍摄功能。在一些实施例中,摄像头组件1706还可以包括闪光灯。闪光灯可以是单色温闪光灯,也可以是双色温闪光灯。双色温闪光灯是指暖光闪光灯和冷光闪光灯的组合,可以用于不同色温下的光线补偿。The camera component 1706 is used to capture images or videos. Optionally, camera assembly 1706 includes a front camera and a rear camera. The front camera is installed on the front panel of the terminal 1700 , and the rear camera is installed on the back of the terminal 1700 . In some embodiments, there are at least two rear cameras, one of which is a main camera, a depth-of-field camera, a wide-angle camera, and a telephoto camera, so as to realize the integration of the main camera and the depth-of-field camera to realize the background blur function. Integrated with a wide-angle camera to achieve panoramic shooting and VR (Virtual Reality, virtual reality) shooting functions or other integrated shooting functions. In some embodiments, camera assembly 1706 may also include a flash. The flash can be a single color temperature flash or a dual color temperature flash. Dual color temperature flash refers to a combination of warm light flash and cold light flash, which can be used for light compensation under different color temperatures.
本领域技术人员可以理解,图17中示出的结构并不构成对终端1700的限定,可以包括比图示更多或更少的组件,或者组合某些组件,或者采用不同的组件布置。Those skilled in the art can understand that the structure shown in FIG. 17 does not constitute a limitation on the terminal 1700, and may include more or fewer components than shown, or combine certain components, or adopt different component arrangements.
可选地,该计算机设备提供为服务器。图18是本申请实施例提供的一种服务器的结构示意图,该服务器1800可因配置或性能不同而产生比较大的差异,可以包括一个或一个以上处理器(Central Processing Units,CPU)1801和一个或一个以上的存储器1802,其中,所述存储器1802中存储有至少一条计算机程序,所述至少一条计算机程序由所述处理器1801加载并执行以实现上述各个方法实施例提供的方法。当然,该服务器还可以具有有线或无线网络接口以及输入输出接口等部件,以便进行输入输出,该服务器还可以包括其他用于实现设备功能的部件,在此不做赘述。Optionally, the computer device is provided as a server. Figure 18 is a schematic structural diagram of a server provided by an embodiment of the present application. The server 1800 may vary greatly due to different configurations or performance, and may include one or more processors (Central Processing Units, CPU) 1801 and a Or one or more memories 1802, wherein at least one computer program is stored in the memory 1802, and the at least one computer program is loaded and executed by the processor 1801 to implement the methods provided by the above method embodiments. Of course, the server may also have components such as wired or wireless network interfaces and input and output interfaces to facilitate input and output. The server may also include other components for implementing device functions, which will not be described in detail here.
本申请实施例还提供了一种计算机可读存储介质,该计算机可读存储介质中存储有至少一条计算机程序,该至少一条计算机程序由处理器加载并执行,以实现上述实施例的投票信息生成方法所执行的操作。Embodiments of the present application also provide a computer-readable storage medium. The computer-readable storage medium stores at least one computer program. The at least one computer program is loaded and executed by the processor to implement the voting information generation of the above embodiments. The operation performed by the method.
本申请实施例还提供了一种计算机程序产品,该计算机程序产品包括计算机程序,该计算机程序被处理器执行时实现上述实施例的投票信息生成方法所执行的操作或投票信息显示方法所执行的操作。Embodiments of the present application also provide a computer program product. The computer program product includes a computer program. When the computer program is executed by a processor, the operations performed by the voting information generation method or the voting information display method of the above embodiments are implemented. operate.
在一些实施例中,本申请实施例所涉及的计算机程序可被部署在一个计算机设备上执行,或者在位于一个地点的多个计算机设备上执行,又或者,在分布在多个地点且通过通信网络互连的多个计算机设备上执行,分布在多个地点且通过通信网络互连的多个计算机设备可以组成区块链系统。In some embodiments, the computer program involved in the embodiments of the present application may be deployed and executed on one computer device, or executed on multiple computer devices located in one location, or distributed in multiple locations and communicated through Executed on multiple computer devices interconnected by a network, multiple computer devices distributed in multiple locations and interconnected through a communication network can form a blockchain system.
可以理解的是,在本申请的具体实施方式中,涉及到用户信息等相关的数据,当本申请以上实施例运用到具体产品或技术中时,需要获得用户许可或者同意,且相关数据的收集、使用和处理需要遵守相关国家和地区的相关法律法规和标准。例如,本申请中涉及到的兴趣标签等都是在充分授权的情况下获取的。It can be understood that in the specific implementation of this application, user information and other related data are involved. When the above embodiments of this application are applied to specific products or technologies, user permission or consent needs to be obtained, and the collection of relevant data , use and processing need to comply with relevant laws, regulations and standards of relevant countries and regions. For example, the interest tags involved in this application were all obtained with full authorization.
本领域普通技术人员可以理解实现上述实施例的全部或部分步骤可以通过硬件来完成,也可以通过程序来指令相关的硬件完成,所述的程序可以存储于一种计算机可读存储介质中,上述提到的存储介质可以是只读存储器,磁盘或光盘等。Those of ordinary skill in the art can understand that all or part of the steps to implement the above embodiments can be completed by hardware, or can be completed by instructing relevant hardware through a program. The program can be stored in a computer-readable storage medium. The above-mentioned The storage media mentioned can be read-only memory, magnetic disks or optical disks, etc.
以上所述仅为本申请实施例的可选实施例,并不用以限制本申请实施例,凡在本申请实施例的精神和原则之内,所作的任何修改、等同替换、改进等,均应包含在本申请的保护范围之内。 The above are only optional embodiments of the embodiments of the present application and are not intended to limit the embodiments of the present application. Any modifications, equivalent substitutions, improvements, etc. made within the spirit and principles of the embodiments of the present application shall be included within the protection scope of this application.

Claims (24)

  1. 一种投票信息生成方法,所述方法包括:A method for generating voting information, the method includes:
    计算机设备获取视频中的视频片段关联的文本内容,所述文本内容包括第一文本内容或第二文本内容的至少一种,所述第一文本内容为所述视频片段包含的文本内容,所述第二文本内容为所述视频片段的弹幕包含的文本内容;The computer device obtains text content associated with a video clip in the video, where the text content includes at least one of first text content or second text content, where the first text content is the text content included in the video clip, and the The second text content is the text content contained in the barrage of the video clip;
    所述计算机设备基于所述文本内容中的关键词,生成所述视频片段的投票主题;The computer device generates a voting topic for the video clip based on keywords in the text content;
    所述计算机设备基于所述关键词和所述投票主题,生成所述视频片段的多个投票候选项;The computer device generates a plurality of voting candidates for the video clip based on the keywords and the voting topic;
    所述计算机设备基于所述投票主题和所述多个投票候选项,生成所述视频片段的投票信息。The computer device generates voting information for the video clip based on the voting topic and the plurality of voting candidates.
  2. 根据权利要求1所述的方法,其中,所述计算机设备基于所述文本内容中的关键词,生成所述视频片段的投票主题,包括:The method of claim 1, wherein the computer device generates a voting topic for the video clip based on keywords in the text content, including:
    所述计算机设备对所述关键词进行编码,得到所述关键词的关键词特征;The computer device encodes the keywords to obtain keyword features of the keywords;
    所述计算机设备对所述关键词特征进行解码,得到所述投票主题,所述投票主题由多个投票主题词语构成。The computer device decodes the keyword features to obtain the voting topic, which is composed of a plurality of voting topic words.
  3. 根据权利要求2所述的方法,其中,第一生成模型包括第一编码子模型和第一解码子模型;The method of claim 2, wherein the first generative model includes a first encoding sub-model and a first decoding sub-model;
    所述计算机设备对所述关键词进行编码,得到所述关键词的关键词特征,包括:The computer device encodes the keywords to obtain keyword features of the keywords, including:
    所述计算机设备调用所述第一编码子模型,对N个所述关键词进行编码,得到N个所述关键词的关键词特征,N为大于1的整数;The computer device calls the first encoding sub-model, encodes the N keywords, and obtains keyword features of the N keywords, where N is an integer greater than 1;
    所述计算机设备对所述关键词特征进行解码,得到所述投票主题,包括:The computer device decodes the keyword features to obtain the voting topic, including:
    所述计算机设备调用所述第一解码子模型,对N个所述关键词特征进行解码,得到第1个解码特征,基于所述第1个解码特征和N个所述关键词特征,确定第一投票主题词语;The computer device calls the first decoding sub-model, decodes the N keyword features to obtain the first decoding feature, and determines the first decoding feature and the N keyword features based on the first decoding feature and the N keyword features. One voting topic word;
    所述计算机设备调用所述第一解码子模型,对N个所述关键词特征和所述第一投票主题词语进行解码,得到第2个解码特征,基于所述第2个解码特征和N个所述关键词特征,确定参考投票主题,所述参考投票主题包括所述第一投票主题词语和第二投票主题词语,直至通过N次解码后,将第N次解码得到的参考投票主题确定为所述视频片段的投票主题。The computer device calls the first decoding sub-model, decodes the N keyword features and the first voting topic words, and obtains the second decoding feature, based on the second decoding feature and the N The keyword features determine the reference voting topic, and the reference voting topic includes the first voting topic words and the second voting topic words, until after N times of decoding, the reference voting topic obtained by the Nth decoding is determined as The poll topic for said video clip.
  4. 根据权利要求3所述的方法,其中,所述第一生成模型还包括第一分类层和预设词语库,所述预设词语库包括多个词语,所述基于所述第1个解码特征和N个所述关键词特征,确定第一投票主题词语,包括:The method according to claim 3, wherein the first generation model further includes a first classification layer and a preset word library, the preset word library includes a plurality of words, the first decoding feature based on the and N keyword features to determine the first voting topic word, including:
    基于所述第1个解码特征和N个所述关键词特征确定N个第一使用概率,第j个所述第一使用概率为在投票主题中使用第j个所述关键词的概率,j为正整数,且j不大于N;N first usage probabilities are determined based on the 1st decoding feature and N keyword features, and the jth first usage probability is the probability of using the jth keyword in the voting topic, j is a positive integer, and j is not greater than N;
    在第j个所述第一使用概率满足使用条件的情况下,将第j个所述关键词确定为所述第一投票主题词语;In the case where the jth first usage probability satisfies the usage condition, determine the jth keyword as the first voting topic word;
    在每个所述第一使用概率不满足所述使用条件的情况下,调用所述第一分类层基于所述第1个解码特征和所述预设词语库进行分类,得到所述预设词语库中的每个词语的分类概率,基于所述每个词语的分类概率,确定所述第一投票主题词语。In each case where the first usage probability does not meet the usage conditions, the first classification layer is called to perform classification based on the first decoding feature and the preset word library to obtain the preset words The first voting topic word is determined based on the classification probability of each word in the library.
  5. 根据权利要求3所述的方法,其中,所述基于所述第2个解码特征和N个所述关键词特征,确定参考投票主题,包括:The method according to claim 3, wherein determining the reference voting topic based on the second decoding feature and the N keyword features includes:
    基于所述第2个解码特征和N个所述关键词特征确定N个第二使用概率,第j个所述第二使用概率为在投票主题中使用第j个所述关键词的概率,j为正整数,且j不大于N; N second usage probabilities are determined based on the 2nd decoding feature and N keyword features, and the jth second usage probability is the probability of using the jth keyword in the voting topic, j is a positive integer, and j is not greater than N;
    在第j个所述第二使用概率满足所述使用条件的情况下,将第j个所述关键词确定为所述第二投票主题词语;In the case where the jth second usage probability satisfies the usage condition, determine the jth keyword as the second voting topic word;
    在每个所述第二使用概率不满足所述使用条件的情况下,调用所述第一分类层基于所述第2个解码特征和所述预设词语库进行分类,得到多个候选投票主题的分类概率,每个所述候选投票主题包括一个所述第一投票主题词语和一个所述第二投票主题词语;In the case where each of the second usage probabilities does not meet the usage conditions, the first classification layer is called to perform classification based on the second decoding feature and the preset word library to obtain multiple candidate voting topics. The classification probability, each of the candidate voting topics includes one of the first voting topic words and one of the second voting topic words;
    基于每个所述候选投票主题的分类概率,确定所述参考投票主题。The reference voting topic is determined based on the classification probability of each of the candidate voting topics.
  6. 根据权利要求3所述的方法,其中,所述第一生成模型的训练过程包括:The method according to claim 3, wherein the training process of the first generative model includes:
    所述计算机设备获取正样本视频片段关联的样本文本内容,所述正样本视频片段为包含样本投票信息且所述样本投票信息的参与率达到目标阈值的视频片段;The computer device obtains sample text content associated with a positive sample video clip, where the positive sample video clip is a video clip that contains sample voting information and the participation rate of the sample voting information reaches a target threshold;
    所述计算机设备获取所述样本投票信息中包含的样本投票主题;The computer device obtains a sample voting topic contained in the sample voting information;
    所述计算机设备基于所述样本文本内容和所述样本投票主题,调整所述第一生成模型中的模型参数。The computer device adjusts model parameters in the first generative model based on the sample text content and the sample voting topic.
  7. 根据权利要求1所述的方法,其中,所述文本内容包括所述第一文本内容和所述第二文本内容,所述计算机设备基于所述关键词和所述投票主题,生成所述视频片段的多个投票候选项,包括:The method of claim 1, wherein the text content includes the first text content and the second text content, and the computer device generates the video clip based on the keywords and the voting topic Multiple voting candidates, including:
    所述计算机设备获取第一关键词,所述第一关键词为所述第一文本内容中的关键词;The computer device obtains a first keyword, where the first keyword is a keyword in the first text content;
    所述计算机设备对所述第二文本内容进行聚类,得到多个文本类别,每个所述文本类别包含至少一条所述第二文本内容;The computer device performs clustering on the second text content to obtain a plurality of text categories, each of the text categories containing at least one piece of the second text content;
    所述计算机设备分别从每个所述文本类别中抽取第二关键词;The computer device extracts second keywords from each of the text categories respectively;
    所述计算机设备基于所述第一关键词、所述投票主题和每个所述文本类别的第二关键词,分别生成每个所述文本类别的投票候选项。The computer device generates voting candidates for each of the text categories based on the first keyword, the voting topic, and the second keyword of each of the text categories.
  8. 根据权利要求7所述的方法,其中,所述计算机设备基于所述第一关键词、所述投票主题和每个所述文本类别的第二关键词,分别生成每个所述文本类别的投票候选项,包括:The method of claim 7, wherein the computer device generates votes for each of the text categories based on the first keyword, the voting topic, and a second keyword for each of the text categories. Candidates include:
    所述计算机设备对所述第一关键词、所述投票主题和第i个所述文本类别的第二关键词进行编码,得到关键词特征,i为正整数,i不大于所述文本类别的数量;The computer device encodes the first keyword, the voting topic and the i-th second keyword of the text category to obtain keyword features, i is a positive integer, and i is not greater than the text category quantity;
    所述计算机设备对所述关键词特征进行解码,得到第i个所述投票候选项,第i个所述投票候选项由多个投票候选项词语构成。The computer device decodes the keyword features to obtain the i-th voting candidate, where the i-th voting candidate is composed of multiple voting candidate words.
  9. 根据权利要求8所述的方法,其中,第二生成模型包括第二编码子模型和第二解码子模型;The method of claim 8, wherein the second generative model includes a second encoding sub-model and a second decoding sub-model;
    所述计算机设备对所述第一关键词、所述投票主题和第i个所述文本类别的第二关键词进行编码,得到关键词特征,包括:The computer device encodes the first keyword, the voting topic and the second keyword of the i-th text category to obtain keyword features, including:
    所述计算机设备调用所述第二编码子模型,对所述第一关键词、所述投票主题和第i个所述文本类别的第二关键词进行编码,得到所述关键词特征;The computer device calls the second encoding sub-model to encode the first keyword, the voting topic and the i-th second keyword of the text category to obtain the keyword feature;
    所述计算机设备对所述关键词特征进行解码,得到第i个所述投票候选项,包括:The computer device decodes the keyword features to obtain the i-th voting candidate, including:
    所述计算机设备调用所述第二解码子模型,对M个所述关键词特征进行解码,得到第1个解码特征,基于所述第1个解码特征和M个所述关键词特征,确定第一投票候选项词语,M为大于1的整数;The computer device calls the second decoding sub-model, decodes the M keyword features to obtain the first decoding feature, and determines the first decoding feature based on the first decoding feature and the M keyword features. A voting candidate word, M is an integer greater than 1;
    所述计算机设备调用所述第二解码子模型,对M个所述关键词特征和所述第一投票候选项词语进行解码,得到第2个解码特征,基于所述第2个解码特征和M个所述关键词特征,确定参考投票候选项,所述参考投票候选项包括所述第一投票候选项词语和第二投票候选项词语,直至通过M次解码后,将第M次解码得到的参考投票候选项确定为所述视频片段的第i个所述投票候选项。 The computer device calls the second decoding sub-model to decode the M keyword features and the first voting candidate words to obtain a second decoding feature. Based on the second decoding feature and M Each of the keyword features is used to determine a reference voting candidate. The reference voting candidate includes the first voting candidate word and the second voting candidate word. Until after M times of decoding, the Mth decoding is obtained. The reference voting candidate is determined to be the i-th voting candidate of the video clip.
  10. 根据权利要求9所述的方法,其中,所述第二生成模型的训练过程包括:The method according to claim 9, wherein the training process of the second generative model includes:
    所述计算机设备获取正样本视频片段关联的样本文本内容,所述正样本视频片段为包含样本投票信息且所述样本投票信息的参与率达到目标阈值的视频片段;The computer device obtains sample text content associated with a positive sample video clip, where the positive sample video clip is a video clip that contains sample voting information and the participation rate of the sample voting information reaches a target threshold;
    所述计算机设备获取所述样本投票信息中包含的样本投票主题和多个样本投票候选项;The computer device obtains a sample voting topic and a plurality of sample voting candidates included in the sample voting information;
    所述计算机设备基于每个所述样本文本内容与每个所述样本投票候选项之间的关联度,确定每个所述样本投票候选项的文本类别,所述文本类别包括所述样本投票候选项关联的样本文本内容;The computer device determines a text category of each of the sample voting candidates based on a correlation between each of the sample text content and each of the sample voting candidates, the text category including the sample voting candidates Sample text content associated with the item;
    所述计算机设备分别从每个所述样本投票候选项的所述文本类别中抽取样本关键词;The computer device extracts sample keywords from the text category of each of the sample voting candidates respectively;
    所述计算机设备基于所述样本文本内容、所述样本投票主题、所述多个样本投票候选项以及每个所述样本投票候选项的样本关键词,调整所述第二生成模型中的模型参数。The computer device adjusts model parameters in the second generation model based on the sample text content, the sample voting topic, the plurality of sample voting candidates, and the sample keywords of each of the sample voting candidates. .
  11. 根据权利要求1-10任一项所述的方法,其中,所述计算机设备获取视频中的视频片段关联的文本内容,包括:The method according to any one of claims 1-10, wherein the computer device obtains text content associated with video clips in the video, including:
    所述计算机设备获取所述第一文本内容、所述第二文本内容、第一热度参数和第二热度参数,所述第一热度参数表示所述视频片段的热门程度,所述第二热度参数表示所述视频片段的弹幕的热门程度;The computer device acquires the first text content, the second text content, a first popularity parameter, and a second popularity parameter. The first popularity parameter represents the popularity of the video clip, and the second popularity parameter Indicates the popularity of the barrage of the video clip;
    所述计算机设备基于所述文本内容中的关键词,生成所述视频片段的投票主题,包括:The computer device generates a voting topic for the video clip based on the keywords in the text content, including:
    所述计算机设备基于所述第一文本内容、所述第二文本内容、所述第一热度参数和所述第二热度参数,确定所述视频片段的投票标识,所述投票标识表示是否为所述视频片段生成投票信息;The computer device determines a voting identification of the video clip based on the first text content, the second text content, the first popularity parameter and the second popularity parameter, and the voting identification indicates whether the video clip is the The above video clips are used to generate voting information;
    所述计算机设备在所述投票标识表示为所述视频片段生成投票信息的情况下,基于所述文本内容中的关键词,生成所述投票主题。The computer device generates the voting topic based on keywords in the text content when the voting identification indicates that voting information is generated for the video clip.
  12. 根据权利要求11所述的方法,其中,所述计算机设备基于所述第一文本内容、所述第二文本内容、所述第一热度参数和所述第二热度参数,确定所述视频片段的投票标识,包括:The method of claim 11, wherein the computer device determines the video segment based on the first text content, the second text content, the first popularity parameter, and the second popularity parameter. Voting signs, including:
    所述计算机设备获取所述第一文本内容的第一文本特征和所述第二文本内容的第二文本特征;The computer device obtains a first text feature of the first text content and a second text feature of the second text content;
    所述计算机设备获取所述第一热度参数的第一热度特征和所述第二热度参数的第二热度特征;The computer device acquires a first heat characteristic of the first heat parameter and a second heat characteristic of the second heat parameter;
    所述计算机设备将所述第一文本特征、所述第二文本特征、所述第一热度特征和所述第二热度特征进行拼接,得到视频片段特征;The computer device splices the first text feature, the second text feature, the first heat feature and the second heat feature to obtain video clip features;
    所述计算机设备基于所述视频片段特征进行分类,得到所述投票标识。The computer device performs classification based on the characteristics of the video clips to obtain the voting identification.
  13. 根据权利要求12所述的方法,其中,投票判定模型包括第一特征提取子模型、第一拼接层和第二分类层,所述投票判定模型还包括热度特征表,所述热度特征表包含至少一种热度参数的热度特征;The method according to claim 12, wherein the voting decision model includes a first feature extraction sub-model, a first splicing layer and a second classification layer, the voting decision model further includes a popularity feature table, the popularity feature table includes at least The heat characteristic of a heat parameter;
    所述计算机设备获取所述第一文本内容的第一文本特征和所述第二文本内容的第二文本特征,包括:所述计算机设备调用所述第一特征提取子模型,获取所述第一文本内容的第一文本特征和所述第二文本内容的第二文本特征;The computer device obtains the first text feature of the first text content and the second text feature of the second text content, including: the computer device calls the first feature extraction sub-model to obtain the first a first text feature of the text content and a second text feature of the second text content;
    所述计算机设备获取所述第一热度参数的第一热度特征和所述第二热度参数的第二热度特征,包括:所述计算机设备基于所述第一热度参数和所述第二热度参数查询所述热度特征表,得到所述第一热度特征和所述第二热度特征;The computer device obtains the first heat characteristic of the first heat parameter and the second heat feature of the second heat parameter, including: the computer device queries based on the first heat parameter and the second heat parameter. The heat characteristic table obtains the first heat characteristic and the second heat characteristic;
    所述计算机设备将所述第一文本特征、所述第二文本特征、所述第一热度特征和所述第二热度特征进行拼接,得到视频片段特征,包括:所述计算机设备调用所述第一拼接层,将所述第一文本特征、所述第二文本特征、所述第一热度特征和所述第二热度特征进行拼接, 得到所述视频片段特征;The computer device splices the first text feature, the second text feature, the first heat feature and the second heat feature to obtain video clip features, including: the computer device calls the third A splicing layer that splices the first text feature, the second text feature, the first heat feature and the second heat feature, Obtain the characteristics of the video clip;
    所述计算机设备基于所述视频片段特征进行分类,得到所述投票标识,包括:所述计算机设备调用所述第二分类层,基于所述视频片段特征进行分类,得到所述投票标识。The computer device performs classification based on the characteristics of the video clips to obtain the voting identification, including: the computer equipment calls the second classification layer, performs classification based on the characteristics of the video clips, and obtains the voting identification.
  14. 根据权利要求13所述的方法,其中,所述投票判定模型的训练过程包括:The method according to claim 13, wherein the training process of the voting decision model includes:
    所述计算机设备获取样本视频片段、所述样本视频片段关联的样本文本内容及所述样本文本内容的热度参数,所述样本视频片段包括正样本视频片段或负样本视频片段的至少一种,所述正样本视频片段为包含样本投票信息且所述样本投票信息的参与率达到目标阈值的视频片段,所述负样本视频片段为:包含样本投票信息且所述样本投票信息的参与率未达到所述目标阈值的视频片段,或不包含样本投票信息的视频片段的至少一种;The computer device obtains a sample video clip, a sample text content associated with the sample video clip, and a popularity parameter of the sample text content. The sample video clip includes at least one of a positive sample video clip or a negative sample video clip, so The positive sample video clip is a video clip that contains sample voting information and the participation rate of the sample voting information reaches the target threshold. The negative sample video clip is a video clip that contains sample voting information and the participation rate of the sample voting information does not reach the target threshold. At least one of the video clips with the target threshold, or the video clips that do not contain sample voting information;
    所述计算机设备基于所述样本视频片段、所述样本视频片段关联的样本文本内容及所述样本文本内容的热度参数,调整所述投票判定模型中的模型参数,所述模型参数包括所述热度特征表。The computer device adjusts model parameters in the voting decision model based on the sample video clip, the sample text content associated with the sample video clip, and the popularity parameters of the sample text content, where the model parameters include the popularity Feature table.
  15. 根据权利要求1-10任一项所述的方法,其中,所述计算机设备基于所述投票主题和所述多个投票候选项,生成所述视频片段的投票信息之后,所述方法还包括:The method according to any one of claims 1 to 10, wherein after the computer device generates the voting information of the video clip based on the voting topic and the plurality of voting candidates, the method further includes:
    所述计算机设备基于账号的兴趣标签和所述投票信息,确定互动参数,所述互动参数表示所述账号基于所述投票信息进行投票操作的可能性;The computer device determines interaction parameters based on the interest tag of the account and the voting information, where the interaction parameters represent the possibility of the account performing a voting operation based on the voting information;
    所述计算机设备在所述互动参数满足互动条件的情况下,向所述账号的终端发送所述投票信息,所述投票信息用于在所述终端播放所述视频片段时展示。When the interaction parameters meet the interaction conditions, the computer device sends the voting information to the terminal of the account, and the voting information is used to display when the terminal plays the video clip.
  16. 一种投票信息显示方法,所述方法包括:A method for displaying voting information, the method comprising:
    计算机设备基于视频中的视频片段,获取所述视频片段的投票信息,所述投票信息基于投票主题和多个投票候选项生成,所述投票主题基于所述视频片段关联的文本内容中的关键词生成,所述多个投票候选项基于所述关键词和所述投票主题生成;The computer device obtains voting information of the video clip based on the video clip, the voting information is generated based on a voting topic and a plurality of voting candidates, and the voting topic is based on keywords in text content associated with the video clip. Generating, the plurality of voting candidates are generated based on the keywords and the voting topic;
    所述计算机设备基于当前登录的账号的兴趣标签和所述投票信息,确定互动参数,所述互动参数表示所述账号基于所述投票信息进行投票操作的可能性;The computer device determines interaction parameters based on the interest tag of the currently logged-in account and the voting information, where the interaction parameters represent the possibility of the account performing a voting operation based on the voting information;
    所述计算机设备在所述互动参数满足互动条件的情况下,在播放所述视频片段时显示所述投票信息;The computer device displays the voting information when playing the video clip when the interaction parameters meet the interaction conditions;
    其中,所述文本内容包括第一文本内容或第二文本内容的至少一种,所述第一文本内容为所述视频片段包含的文本内容,所述第二文本内容为所述视频片段的弹幕包含的文本内容。Wherein, the text content includes at least one of first text content or second text content, the first text content is the text content contained in the video clip, and the second text content is the bounce of the video clip. The text content contained in the scene.
  17. 根据权利要求16所述的方法,其中,所述计算机设备基于当前登录的账号的兴趣标签和所述投票信息,确定互动参数,包括:The method according to claim 16, wherein the computer device determines interaction parameters based on the interest tag of the currently logged-in account and the voting information, including:
    所述计算机设备获取所述兴趣标签的兴趣特征、所述投票主题的投票主题特征和所述多个投票候选项的投票候选项特征;The computer device acquires the interest characteristics of the interest tag, the voting topic characteristics of the voting topic, and the voting candidate characteristics of the plurality of voting candidates;
    所述计算机设备将所述兴趣特征、所述投票主题特征和多个投票候选项特征进行拼接,得到互动特征;The computer device splices the interest characteristics, the voting topic characteristics and the plurality of voting candidate characteristics to obtain interactive characteristics;
    所述计算机设备基于所述互动特征进行分类,得到所述互动参数。The computer device performs classification based on the interaction characteristics to obtain the interaction parameters.
  18. 根据权利要求17所述的方法,其中,投票互动模型包括第二特征提取子模型、第二拼接层和第三分类层;The method according to claim 17, wherein the voting interaction model includes a second feature extraction sub-model, a second splicing layer and a third classification layer;
    所述计算机设备获取所述兴趣标签的兴趣特征、所述投票主题的投票主题特征和所述多个投票候选项的投票候选项特征,包括:所述计算机设备调用所述第二特征提取子模型,获取所述兴趣标签的兴趣特征、所述投票主题的投票主题特征和所述多个投票候选项的投票候选项特征; The computer device obtains the interest characteristics of the interest tag, the voting topic characteristics of the voting topic, and the voting candidate characteristics of the plurality of voting candidates, including: the computer equipment calls the second feature extraction sub-model. , obtain the interest characteristics of the interest tag, the voting topic characteristics of the voting topic, and the voting candidate characteristics of the plurality of voting candidates;
    所述计算机设备将所述兴趣特征、所述投票主题特征和多个投票候选项特征进行拼接,得到互动特征,包括:所述计算机设备调用所述第二拼接层,将所述兴趣特征、所述投票主题特征和所述多个投票候选项特征进行拼接,得到所述互动特征;The computer device splices the interest features, the voting topic features and multiple voting candidate features to obtain interactive features, including: the computer device calls the second splicing layer to combine the interest features, all voting candidate features Splicing the voting topic characteristics and the plurality of voting candidate characteristics to obtain the interactive characteristics;
    所述计算机设备基于所述互动特征进行分类,得到所述互动参数,包括:所述计算机设备调用所述第三分类层,基于所述互动特征进行分类,得到所述互动参数。The computer device performs classification based on the interaction characteristics to obtain the interaction parameters, including: the computer equipment calls the third classification layer, performs classification based on the interaction characteristics, and obtains the interaction parameters.
  19. 根据权利要求18所述的方法,其中,所述投票互动模型的训练过程包括:The method according to claim 18, wherein the training process of the voting interaction model includes:
    所述计算机设备获取样本账号的样本兴趣标签以及样本视频片段中的样本投票信息,其中,所述样本账号已执行基于所述样本投票信息进行投票操作;The computer device obtains the sample interest tag of the sample account and the sample voting information in the sample video clip, wherein the sample account has performed a voting operation based on the sample voting information;
    所述计算机设备基于所述样本兴趣标签、所述样本投票信息中的样本投票主题和多个样本投票候选项,调整所述投票互动模型中的模型参数。The computer device adjusts model parameters in the voting interaction model based on the sample interest tag, the sample voting topic in the sample voting information, and a plurality of sample voting candidates.
  20. 一种投票信息生成装置,设置于计算机设备中,所述装置包括:A voting information generation device, installed in computer equipment, the device includes:
    文本内容获取模块,用于获取视频中的视频片段关联的文本内容,所述文本内容包括第一文本内容或第二文本内容的至少一种,所述第一文本内容为所述视频片段包含的文本内容,所述第二文本内容为所述视频片段的弹幕包含的文本内容;A text content acquisition module, configured to acquire text content associated with video clips in the video, where the text content includes at least one of first text content or second text content, where the first text content is included in the video clip. Text content, the second text content is the text content contained in the barrage of the video clip;
    主题生成模块,用于基于所述文本内容中的关键词,生成所述视频片段的投票主题;A topic generation module, configured to generate voting topics for the video clips based on keywords in the text content;
    候选项生成模块,用于基于所述关键词和所述投票主题,生成所述视频片段的多个投票候选项;A candidate generation module, configured to generate multiple voting candidates for the video clip based on the keywords and the voting topic;
    投票信息生成模块,用于基于所述投票主题和所述多个投票候选项,生成所述视频片段的投票信息。A voting information generation module, configured to generate voting information for the video clip based on the voting topic and the plurality of voting candidates.
  21. 一种投票信息显示装置,设置于计算机设备中,所述装置包括:A voting information display device, installed in computer equipment, the device includes:
    信息获取模块,用于基于视频中的视频片段,获取所述视频片段的投票信息,所述投票信息基于投票主题和多个投票候选项生成,所述投票主题基于所述视频片段关联的文本内容中的关键词生成,所述多个投票候选项基于所述关键词和所述投票主题生成;An information acquisition module, configured to obtain voting information of video clips based on video clips in the video, where the voting information is generated based on a voting topic and a plurality of voting candidates, and the voting topic is based on text content associated with the video clips. The keywords in are generated, and the plurality of voting candidates are generated based on the keywords and the voting topic;
    参数确定模块,用于基于当前登录的账号的兴趣标签和所述投票信息,确定互动参数,所述互动参数表示所述账号基于所述投票信息进行投票操作的可能性;A parameter determination module, configured to determine interaction parameters based on the interest tag of the currently logged-in account and the voting information, where the interaction parameters represent the possibility of the account performing a voting operation based on the voting information;
    信息显示模块,用于在所述互动参数满足互动条件的情况下,在播放所述视频片段时显示所述投票信息;An information display module, configured to display the voting information when the video clip is played when the interaction parameters meet the interaction conditions;
    其中,所述文本内容包括第一文本内容或第二文本内容的至少一种,所述第一文本内容为所述视频片段包含的文本内容,所述第二文本内容为所述视频片段的弹幕包含的文本内容。Wherein, the text content includes at least one of first text content or second text content, the first text content is the text content contained in the video clip, and the second text content is the bounce of the video clip. The text content contained in the scene.
  22. 一种计算机设备,所述计算机设备包括处理器和存储器,所述存储器中存储有至少一条计算机程序,所述至少一条计算机程序由所述处理器加载并执行,以实现如权利要求1至15任一项所述的投票信息生成方法所执行的操作,或者以实现如权利要求16至19任一项所述的投票信息显示方法所执行的操作。A computer device. The computer device includes a processor and a memory. At least one computer program is stored in the memory. The at least one computer program is loaded and executed by the processor to implement any of claims 1 to 15. The operations performed by the voting information generating method described in one of the claims, or to implement the operations performed by the voting information display method described in any one of claims 16 to 19.
  23. 一种计算机可读存储介质,所述计算机可读存储介质中存储有至少一条计算机程序,所述至少一条计算机程序由处理器加载并执行,以实现如权利要求1至15任一项所述的投票信息生成方法所执行的操作,或者以实现如权利要求16至19任一项所述的投票信息显示方法所执行的操作。A computer-readable storage medium, in which at least one computer program is stored, and the at least one computer program is loaded and executed by a processor to implement the method described in any one of claims 1 to 15 The operations performed by the voting information generation method, or the operations performed by the voting information display method according to any one of claims 16 to 19.
  24. 一种计算机程序产品,包括计算机程序,所述计算机程序被处理器执行时实现如权利要求1至15任一项所述的投票信息生成方法所执行的操作,或者实现如权利要求16至19任一项所述的投票信息显示方法所执行的操作。 A computer program product, including a computer program that, when executed by a processor, implements the operations performed by the voting information generation method according to any one of claims 1 to 15, or implements the operations performed by any one of claims 16 to 19. An operation performed by the voting information display method described in one item.
PCT/CN2023/083979 2022-04-29 2023-03-27 Voting information generation method and apparatus, and voting information display method and apparatus WO2023207463A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210473861.6A CN117037012A (en) 2022-04-29 2022-04-29 Voting information generation method, voting information display method and device
CN202210473861.6 2022-04-29

Publications (1)

Publication Number Publication Date
WO2023207463A1 true WO2023207463A1 (en) 2023-11-02

Family

ID=88517319

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/083979 WO2023207463A1 (en) 2022-04-29 2023-03-27 Voting information generation method and apparatus, and voting information display method and apparatus

Country Status (2)

Country Link
CN (1) CN117037012A (en)
WO (1) WO2023207463A1 (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104317804A (en) * 2014-09-23 2015-01-28 小米科技有限责任公司 Voting information publishing method and device
CN108174247A (en) * 2017-12-27 2018-06-15 优酷网络技术(北京)有限公司 Video interaction method and device
US20180211467A1 (en) * 2017-01-23 2018-07-26 Smartmatic International Corporation Means to create a physical audit trail verifiable by remote voters in electronic elections
CN111708915A (en) * 2020-06-12 2020-09-25 腾讯科技(深圳)有限公司 Content recommendation method and device, computer equipment and storage medium
CN112800339A (en) * 2021-04-06 2021-05-14 腾讯科技(深圳)有限公司 Information stream searching method, device and equipment
CN112866787A (en) * 2021-04-12 2021-05-28 上海哔哩哔哩科技有限公司 Bullet screen setting method, device and system

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104317804A (en) * 2014-09-23 2015-01-28 小米科技有限责任公司 Voting information publishing method and device
US20180211467A1 (en) * 2017-01-23 2018-07-26 Smartmatic International Corporation Means to create a physical audit trail verifiable by remote voters in electronic elections
CN108174247A (en) * 2017-12-27 2018-06-15 优酷网络技术(北京)有限公司 Video interaction method and device
CN111708915A (en) * 2020-06-12 2020-09-25 腾讯科技(深圳)有限公司 Content recommendation method and device, computer equipment and storage medium
CN112800339A (en) * 2021-04-06 2021-05-14 腾讯科技(深圳)有限公司 Information stream searching method, device and equipment
CN112866787A (en) * 2021-04-12 2021-05-28 上海哔哩哔哩科技有限公司 Bullet screen setting method, device and system

Also Published As

Publication number Publication date
CN117037012A (en) 2023-11-10

Similar Documents

Publication Publication Date Title
CN111444357B (en) Content information determination method, device, computer equipment and storage medium
US20150243279A1 (en) Systems and methods for recommending responses
CN110234018B (en) Multimedia content description generation method, training method, device, equipment and medium
CN112215171B (en) Target detection method, device, equipment and computer readable storage medium
CN103970791B (en) A kind of method, apparatus for recommending video from video library
CN111414506B (en) Emotion processing method and device based on artificial intelligence, electronic equipment and storage medium
US20200012862A1 (en) Multi-model Techniques to Generate Video Metadata
CN111615002B (en) Video background playing control method, device and system and electronic equipment
CN111432282B (en) Video recommendation method and device
CN114390217A (en) Video synthesis method and device, computer equipment and storage medium
CN115510194A (en) Question and answer sentence retrieval method and device, electronic equipment and storage medium
CN114969282A (en) Intelligent interaction method based on rich media knowledge graph multi-modal emotion analysis model
CN112702613A (en) Live video recording method and device, storage medium and electronic equipment
CN110990632B (en) Video processing method and device
CN116977701A (en) Video classification model training method, video classification method and device
WO2023207463A1 (en) Voting information generation method and apparatus, and voting information display method and apparatus
CN116955591A (en) Recommendation language generation method, related device and medium for content recommendation
CN116956183A (en) Multimedia resource recommendation method, model training method, device and storage medium
CN117009577A (en) Video data processing method, device, equipment and readable storage medium
CN116955707A (en) Content tag determination method, device, equipment, medium and program product
CN112242140A (en) Intelligent device control method and device, electronic device and storage medium
CN117093785B (en) Method, system, equipment and storage medium for guiding user based on social contact
CN111159472A (en) Multi-modal chat techniques
US20220198141A1 (en) System and method for identifying and displaying information related to an off screen plot element or character in a media stream
CN116994169A (en) Label prediction method, label prediction device, computer equipment and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23794898

Country of ref document: EP

Kind code of ref document: A1