CN112818984A - Title generation method and device, electronic equipment and storage medium - Google Patents

Title generation method and device, electronic equipment and storage medium Download PDF

Info

Publication number
CN112818984A
CN112818984A CN202110114237.2A CN202110114237A CN112818984A CN 112818984 A CN112818984 A CN 112818984A CN 202110114237 A CN202110114237 A CN 202110114237A CN 112818984 A CN112818984 A CN 112818984A
Authority
CN
China
Prior art keywords
title
candidate
image frames
text
target
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110114237.2A
Other languages
Chinese (zh)
Other versions
CN112818984B (en
Inventor
姚晓宇
李海
谭颖
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing QIYI Century Science and Technology Co Ltd
Original Assignee
Beijing QIYI Century Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing QIYI Century Science and Technology Co Ltd filed Critical Beijing QIYI Century Science and Technology Co Ltd
Priority to CN202110114237.2A priority Critical patent/CN112818984B/en
Publication of CN112818984A publication Critical patent/CN112818984A/en
Application granted granted Critical
Publication of CN112818984B publication Critical patent/CN112818984B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/62Text, e.g. of license plates, overlay texts or captions on TV images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/253Grammatical analysis; Style critique
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/46Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/28Character recognition specially adapted to the type of the alphabet, e.g. Latin alphabet
    • G06V30/287Character recognition specially adapted to the type of the alphabet, e.g. Latin alphabet of Kanji, Hiragana or Katakana characters
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Abstract

The invention discloses a title generation method, a title generation device, electronic equipment and a storage medium, wherein the method comprises the following steps: the method comprises the steps of obtaining a target video, respectively detecting areas with title characteristics from a plurality of image frames of the target video to obtain area positions, determining the positions of title candidate areas in the image frames according to the area positions corresponding to the image frames, carrying out text recognition on the title candidate areas in the image frames to obtain a target title of the target video, analyzing and obtaining the title candidate areas where the titles of the videos may appear according to the image frames of the target video, eliminating numerous confusion information in the video, improving the accuracy when the titles are determined, and then carrying out text recognition on the title candidate areas, so that the target title of the target video is automatically generated, the manual intervention degree when the titles are generated is reduced, and the efficiency of generating the titles for the video is improved.

Description

Title generation method and device, electronic equipment and storage medium
Technical Field
The present invention relates to the field of video technologies, and in particular, to a title generation method, a title generation apparatus, an electronic device, and a computer-readable storage medium.
Background
Currently, many local channels in various locations have news coverage programs. For the current fast-paced life style, news reports are broken into a plurality of news videos one by one, and the requirement for watching network videos by users is met.
In order to enable users to quickly find out interesting contents in a plurality of news videos for understanding, an important task in the process of dividing a news report into a plurality of segments is to allocate a proper title to each news video.
Since a lot of news reports are reported every day, generating a proper title for each news video is a work with a large workload, a lot of confusion information such as subtitles and characters in picture content appears in the video, and the title is found from the lot of confusion information, so that the problem that the time and the labor are consumed for finding the title from the video is caused.
Disclosure of Invention
An object of embodiments of the present invention is to provide a title generating method, a title generating apparatus, an electronic device, and a computer-readable storage medium, so as to solve the technical problem that it takes time and effort to find a title from a video because many pieces of confusion information, such as subtitles and characters in picture content, appear in the video and the title is found from many pieces of confusion information.
In order to solve the above problem, in a first aspect of the present invention, there is provided a title generating method, including:
acquiring a target video;
respectively detecting areas with title characteristics from a plurality of image frames of the target video to obtain area positions;
determining the position of a title candidate region in the image frames according to the region positions corresponding to the image frames;
and performing text recognition on the title candidate areas in the plurality of image frames to obtain a target title of the target video.
Optionally, the detecting, from a plurality of image frames of the target video, regions having a title characteristic respectively, and obtaining the region position includes at least one of:
obtaining the area position of which the difference value of pixel values in different image frames is smaller than a preset threshold value by comparing different image frames;
respectively detecting areas containing texts in the image frames to obtain the positions of the areas containing the texts in the image frames;
and respectively carrying out edge detection on the plurality of image frames to obtain the area positions of the areas surrounded by the edges in the plurality of image frames.
Optionally, the determining, according to the region positions corresponding to the plurality of image frames, the position of the title candidate region in the image frame includes:
counting the area positions corresponding to the plurality of image frames to generate frequency distribution data of the area positions;
performing gradient operation on the frequency distribution data to obtain a gradient operation result;
and determining the position of the title candidate region in the image frame according to the gradient operation result.
Optionally, before the text recognition is performed on the title candidate regions in the image frames to obtain the target title of the target video, the method further includes:
respectively detecting whether the image change rate of the title candidate area is smaller than a preset threshold value or not for each image frame;
and eliminating the title candidate area with the image change rate larger than a preset threshold value.
Optionally, the performing text recognition on the title candidate regions in the image frames to obtain a target title of the target video includes:
respectively carrying out text recognition on the title candidate regions in the plurality of image frames to obtain candidate texts;
and selecting a target title of the target video according to the candidate text.
Optionally, the performing text recognition on the title candidate regions in the plurality of image frames respectively to obtain candidate texts includes:
when the title candidate area is subjected to text recognition, generating the appearance duration and/or time distribution of the candidate text in the target video according to the image frame of the candidate text;
before the selecting a target title of the target video according to the candidate text, the method further includes:
and eliminating the candidate texts of which the occurrence durations and/or time distributions do not accord with the preset time condition.
Optionally, before the selecting the target title of the target video according to the candidate text, the method further includes:
detecting at least one of attribute information of text elements in the candidate text, text length of the candidate text and character categories contained in the candidate text;
and eliminating at least one candidate text which does not accord with a preset rule in the attribute information, the text length and the character category.
Optionally, the selecting a target title of the target video according to the candidate text includes:
inputting the candidate texts into a title detection network; the title detection network is used for detecting whether the text can be used as a title or not, and is obtained by adopting a title text sample and a non-title text sample for training;
detecting whether the candidate text can be used as a title or not by the title detection network, and outputting a title confidence coefficient;
and selecting the candidate text with the highest title confidence as a target title.
According to a second aspect of the present invention, there is also provided a title generating apparatus, including:
the video acquisition module is used for acquiring a target video;
the position detection module is used for respectively detecting areas with title characteristics from a plurality of image frames of the target video to obtain area positions;
a region determining module, configured to determine, according to the region positions corresponding to the plurality of image frames, positions of the title candidate regions in the image frames;
and the title generation module is used for performing text recognition on the title candidate areas in the image frames to obtain the target title of the target video.
Optionally, the position detection module comprises at least one of:
the comparison submodule is used for obtaining the area position of which the difference value of the pixel values between different image frames is smaller than a preset threshold value by comparing different image frames;
the text detection submodule is used for respectively detecting the areas containing the texts in the image frames to obtain the positions of the areas containing the texts in the image frames;
and the edge detection submodule is used for respectively carrying out edge detection on the plurality of image frames to obtain the area positions of the areas surrounded by the edges in the plurality of image frames.
Optionally, the region determining module includes:
the data generation submodule is used for counting the area positions corresponding to the image frames and generating frequency distribution data of the area positions;
the gradient operation submodule is used for carrying out gradient operation on the frequency distribution data to obtain a gradient operation result;
and the area determining submodule is used for determining the position of the title candidate area in the image frame according to the gradient operation result.
Optionally, the apparatus further comprises:
a detection module, configured to detect whether an image change rate of the title candidate region is smaller than a preset threshold for each image frame before performing text recognition on the title candidate region in the image frames to obtain a target title of the target video;
and the area removing module is used for removing the title candidate area with the image change rate larger than a preset threshold value.
Optionally, the title generating module includes:
the text recognition submodule is used for respectively carrying out text recognition on the title candidate areas in the plurality of image frames to obtain candidate texts;
and the title selection submodule is used for selecting a target title of the target video according to the candidate text.
Optionally, the text recognition sub-module comprises:
the time generation unit is used for generating the appearance duration and/or time distribution of the candidate text in the target video according to the image frame of the candidate text identified when the title candidate area is subjected to text identification;
the title selection submodule comprises:
and the first eliminating unit is used for eliminating the candidate texts of which the occurrence durations and/or time distributions do not accord with the preset time condition before the target titles of the target videos are selected according to the candidate texts.
Optionally, the title selecting sub-module includes:
the attribute detection unit is used for detecting at least one of attribute information of text elements in the candidate text, text length of the candidate text and character types contained in the candidate text before the target title of the target video is selected according to the candidate text;
and the second eliminating unit is used for eliminating at least one candidate text which does not accord with a preset rule in the attribute information, the text length and the character category.
Optionally, the title selecting sub-module includes:
an input unit configured to input the candidate text into a title detection network; the title detection network is used for detecting whether the text can be used as a title or not, and is obtained by adopting a title text sample and a non-title text sample for training;
an output unit, configured to detect, by the title detection network, whether the candidate text can be used as a title, and output a title confidence;
and the selecting unit is used for selecting the candidate text with the highest title confidence coefficient as the target title.
In another aspect of the present invention, there is also provided an electronic device, including a processor, a communication interface, a memory and a communication bus, where the processor, the communication interface, and the memory complete communication with each other through the communication bus;
a memory for storing a computer program;
a processor for implementing any of the above method steps when executing a program stored in the memory.
In yet another aspect of the present invention, there is also provided a computer-readable storage medium having stored therein instructions, which when run on a computer, cause the computer to perform any of the methods described above.
In yet another aspect of the present invention, there is also provided a computer program product containing instructions which, when run on a computer, cause the computer to perform any of the methods described above.
In summary, according to the embodiments of the present invention, by acquiring a target video, regions having a title characteristic are respectively detected from a plurality of image frames of the target video, and a region position is obtained, determining the position of a title candidate region in the image frames according to the region positions corresponding to the plurality of image frames, performing text recognition on the title candidate regions in the plurality of image frames to obtain a target title of the target video, so that the title candidate area which is possibly appeared in the title of the video is obtained by analyzing according to a plurality of image frames of the target video, thereby eliminating a plurality of confusion information in the video, improving the accuracy when the title is determined, and then, text recognition is carried out on the title candidate area, so that the target title of the target video is automatically generated, the manual intervention degree during title generation is reduced, and the title generation efficiency for the video is improved.
Drawings
FIG. 1 is a flow chart illustrating the steps of one embodiment of a title generation method of the present invention;
FIG. 2 is a flow chart illustrating the steps of another title generation method embodiment of the present invention;
FIG. 3 illustrates a title hot zone feature diagram;
FIG. 4 illustrates a cut-off point hot-zone signature diagram;
FIG. 5 is a block diagram illustrating an embodiment of a title generation apparatus according to the present invention;
fig. 6 shows a schematic view of an electronic device of the invention.
Detailed Description
In order to make the aforementioned objects, features and advantages of the present invention comprehensible, embodiments accompanied with figures are described in further detail below.
Referring to fig. 1, a flowchart illustrating steps of an embodiment of a title generating method according to the present invention is shown, which may specifically include the following steps:
step 101, acquiring a target video.
In the embodiment of the present invention, the target video includes a video submitted by a user, a plurality of video segments split from one video, and the like, for example, a video of one news report may include a plurality of news events, the video may be split into video segments corresponding to the plurality of news events, and each video segment serves as the target video.
Step 102, detecting areas with title characteristics from a plurality of image frames of the target video respectively to obtain area positions.
In the embodiment of the present invention, the areas where the titles appear in the videos are relatively concentrated, but the specific areas in different videos are different, for example, in many videos of news reports, the titles usually appear in the lower half area in the videos, but the areas where the titles appear in different videos are different from the lower edge of the video image.
In the embodiment of the present invention, in each area of the video, there is an area having a title characteristic. A region in a video has a title property, and does not necessarily mean that a title exists in the region, but means that the region has a specific property, and may be a region where a title exists. For example, because the title does not change in the video playing process, if the image content in a certain area is not changed all the time, the certain area has a title characteristic, or a text is detected to exist in the certain area, the certain area has a title characteristic, or because a special title box is designed in some videos, if a straight line or a neat edge is detected in the certain area, the certain area has a title characteristic, and any applicable title characteristic may be included.
In the embodiment of the invention, the video is composed of image frames, and if a certain area in the image frames has a title characteristic, the area position of the area in the image frames is obtained. When the area with the title characteristic is detected and the area position is obtained, the area with the title characteristic can be detected through the same image frame in the target video to obtain the area position, or the area with the title characteristic can be detected through different image frames in the target video to obtain the area position. The specific implementation manner may include multiple manners, for example, by comparing different image frames, obtaining the position of an area where a difference value between pixel values of different image frames is smaller than a preset threshold, or respectively detecting areas containing texts in the image frames, obtaining the position of an area containing a text in the image frames, or respectively performing edge detection on the image frames, obtaining the position of an area at an edge in the image frames, or any other suitable manner, which is not limited in this embodiment of the present invention.
For example, in a news video, a title usually appears in the lower half of a video image, and in order to reduce the workload of detecting the position of an area with a title characteristic, only the lower half of a plurality of image frames in a target video is detected, specifically, a stable area with unchanged content in different image frames can be detected by a frame difference method, the position of the area of the stable area is detected, a text area is detected for the image frames, the position of the area containing text is detected, edge detection is performed for the image frames by using a Canny edge detection algorithm, the position of the area of an edge is detected, and the like.
Step 103, determining the position of the title candidate area in the image frame according to the area positions corresponding to the plurality of image frames.
In the embodiment of the invention, the area positions respectively detected from different image frames of the target video are the same and different. The more image frames a certain region position is detected, the more likely it is that the region the title is located in. Therefore, according to the corresponding region positions of a plurality of image frames, the positions of one or more regions in the image frames can be determined by using the principle of statistical analysis, and the one or more regions are marked as title candidate regions.
In this embodiment of the present invention, the implementation manner for determining the position of the title candidate region in the image frame according to the region positions corresponding to the plurality of image frames may include multiple implementations, for example, counting the region positions corresponding to the plurality of image frames, generating frequency distribution data of the region positions, performing a gradient operation on the frequency distribution data, and determining the position of the title candidate region in the image frame according to a result of the gradient operation, or any other suitable implementation manner, which is not limited in this embodiment of the present invention.
In this embodiment of the present invention, optionally, before performing text recognition on the title candidate regions in the plurality of image frames to obtain a target title of the target video, the method may further include: respectively detecting whether the image change rate of the title candidate area is smaller than a preset threshold value or not for each image frame; and eliminating the title candidate area with the image change rate larger than a preset threshold value.
Generally, in a video, the region where the title is located includes only text and background color, and the image change rate of the region where the title is located should be low relative to other images of the video. Before performing text recognition on each image frame, it is detected whether the image change rate of the title candidate area is smaller than a preset threshold, for example, a variance of the image in the title candidate area is calculated, and the variance may represent the image change rate. If the image change rate is less than the preset threshold, the title candidate area is represented to be in accordance with the general background features of the title, and the title candidate area can be reserved.
And 104, performing text recognition on the title candidate areas in the plurality of image frames to obtain a target title of the target video.
In the embodiment of the invention, text recognition is carried out on the title candidate area in each image frame to obtain the text in the title candidate area, and then the recognized text is directly used as the title of the target video, or whether the recognized text can be used as the title of the target video is judged firstly, or after the recognized text is processed, the processed text is used as the title of the target video. And recording the finally obtained title of the target video as the target title.
In one case, the same text is identified in the title candidate regions in the multiple image frames, and the identified text may be directly used as the target title of the target video, or whether the identified text can be used as the target title of the target video is determined according to a preset rule, and finally the target title of the target video is obtained, or any other suitable manner, which is not limited in the embodiment of the present invention.
In another case, different texts are identified in candidate areas of titles in a plurality of image frames, and a text that can be a target title is selected from the candidate texts. The specific implementation manner of selecting the target title may include multiple manners, for example, removing a part of candidate texts that do not meet a preset rule, inputting the remaining candidate texts into a title detection network, where the title detection network is used to detect whether texts can be used as titles, the title detection network is obtained by training a title text sample and a non-title text sample, and the title detection network detects whether candidate texts can be used as titles, outputs a title confidence, and selects a candidate text with the highest title confidence as a target title, or any other suitable manner.
In summary, according to the embodiments of the present invention, by acquiring a target video, regions having a title characteristic are respectively detected from a plurality of image frames of the target video, and a region position is obtained, determining the position of the title candidate region in the image frame according to the region positions corresponding to the plurality of image frames, performing text recognition on the title candidate regions in the plurality of image frames to obtain a target title of the target video, so that the title candidate area which is possibly appeared in the title of the video is obtained by analyzing according to a plurality of image frames of the target video, thereby eliminating a plurality of confusion information in the video, improving the accuracy when the title is determined, and then, text recognition is carried out on the title candidate area, so that the target title of the target video is automatically generated, the manual intervention degree during title generation is reduced, and the title generation efficiency for the video is improved.
Referring to fig. 2, a flowchart illustrating steps of an embodiment of a title generating method according to the present invention is shown, which may specifically include the following steps:
step 201, acquiring a target video.
Step 202, comparing different image frames to obtain the area position where the difference value of the pixel values between the different image frames is smaller than a preset threshold value.
In the embodiment of the invention, the image of the area where the title is located in the video does not change in a period of time, so that one method for detecting the area position with the title characteristic is to obtain the area position with the difference value of the pixel values smaller than the preset threshold value by comparing different image frames. For the target video, one image frame can be taken from the target video every set time length, each taken image frame is compared with the last taken image frame respectively, and the region position where the difference value of the pixel value corresponding to each image frame is smaller than the preset threshold value is obtained.
For example, a frame difference method is used to perform difference operation on two frames of images, pixel points corresponding to different image frames are subtracted to determine an absolute value of a gray difference, and when the absolute value is smaller than a certain threshold, a static target with an unchanged image can be determined, so that detection of an area with an unchanged image is realized. In particular, different image frames may be compared in any suitable manner, which is not limited in this embodiment of the present invention.
Step 203, detecting the regions containing the text in the plurality of image frames respectively, and obtaining the positions of the regions containing the text in the plurality of image frames.
In the embodiment of the present invention, another method for detecting the location of the area having the title characteristic is to detect the area containing the text in the image frame, and obtain the location of the area containing the text in the image frame. For the target video, one image frame can be taken from the target video every set time length, the area containing the text in each image frame is respectively detected, and the area position containing the text in each image frame is obtained. For example, the location of the region containing text in the image can be roughly found by MSER (maximum Stable extreme region) algorithm. Any suitable text detection mode may be specifically adopted, and the embodiment of the present invention is not limited thereto.
Step 204, respectively performing edge detection on the plurality of image frames to obtain the area positions of the areas surrounded by the edges in the plurality of image frames.
In the embodiment of the present invention, the area where the title is located in the video has a straight line or a neat edge, and therefore, another method for detecting the position of the area with the title characteristic is to perform edge detection on the image frame to obtain the edge in the image frame, and then obtain the position of the area surrounded by the edge. For the target video, one image frame can be taken from the target video every set time length, and edge detection is performed on each image frame respectively to obtain the area position of the area surrounded by the edge in each image frame. Specifically, any suitable edge detection method may be adopted, for example, a Canny edge detection algorithm, a Sobel edge detection algorithm, and the like, which is not limited in this embodiment of the present invention.
Step 205, counting the area positions corresponding to the plurality of image frames, and generating frequency distribution data of the area positions.
In an embodiment of the present invention, the detected region positions may also be different for different image frames. After detecting the region positions from each of the plurality of image frames, the region positions are counted to generate frequency distribution data of the region positions. The time distribution data may characterize the time distribution of the location of the region, for example, as shown in the title hot zone feature map of fig. 3, and the time distribution data may be presented in the form of the title hot zone feature map, wherein the higher the brightness of the region in the map, the higher the statistical time, that is, the higher the probability that the region is the title region. For another example, as shown in the cutoff point hot zone feature map shown in fig. 4, the number distribution data may be shown in the form of a cutoff point hot zone feature map, in which the higher the brightness of the vertical line, the higher the statistical number, that is, the higher the probability that the region is the edge of the title frame.
And step 206, performing gradient operation on the frequency distribution data to obtain a gradient operation result.
In the embodiment of the invention, the gradient of the pixel point is a vector with size and direction, for the frequency distribution data, the direction of the gradient is the direction with the fastest frequency change of one pixel point, and the size of the gradient is the frequency change rate of one pixel point. And performing gradient operation on the frequency distribution data, wherein the obtained gradient operation result comprises gradients at all positions corresponding to the pixel points on the image frame.
Step 207, determining the position of the title candidate region in the image frame according to the gradient operation result.
In the embodiment of the present invention, the manner of determining the position of the title candidate region in the image frame according to the gradient operation result may include multiple manners, for example, dividing the title candidate region by taking the pixel point with the maximum gradient as the edge; or selecting a pixel point with a gradient between a preset first gradient threshold and a preset second gradient threshold as an edge, and dividing a title candidate region, or any other suitable manner.
And 208, respectively performing text recognition on the title candidate regions in the plurality of image frames to obtain candidate texts.
In the embodiment of the invention, the title candidate areas in each image frame are respectively subjected to text recognition to obtain the candidate texts corresponding to each image frame. For example, the text in the candidate region of the title is recognized by using an OCR (Optical Character Recognition) technique to obtain a candidate text.
Step 209, selecting a target title of the target video according to the candidate text.
In the embodiment of the present invention, if the candidate texts obtained from the plurality of image frames are all the same, the candidate text may be directly used as the target title, or whether the candidate text can be used as the target title may be determined first. If the candidate texts obtained from the plurality of image frames are not all the same, one of the candidate texts needs to be selected as a target title.
In this embodiment of the present invention, optionally, performing text recognition on the title candidate regions in the plurality of image frames respectively, and an implementation manner of obtaining candidate texts may include: when the title candidate area is subjected to text recognition, generating the appearance duration and/or time distribution of the candidate text in the target video according to the image frame of the candidate text; correspondingly, before selecting the target title of the target video according to the candidate text, the method may further include: and eliminating the candidate texts of which the occurrence durations and/or time distributions do not accord with the preset time condition.
When the image frames are subjected to text recognition, the time stamps of the image frames are extracted, and the occurrence duration and/or time distribution of the candidate text in the target video can be generated according to the image frames in which the candidate text is recognized, for example, firstly, the candidate text A is recognized in the 1 st to x th image frames, then, the candidate text B is recognized in the (x +1) th to y th image frames, then, the candidate text A is recognized in the (y +1) th to z th image frames, then, the occurrence duration and time distribution of the candidate text A can be generated according to the time stamps of the 1 st to x th image frames and the time stamps of the (y +1) th to z th image frames, and the occurrence duration and time distribution of the candidate text B can be generated according to the time stamps of the (x +1) th to y image frames.
The occurrence duration and/or time distribution of the title in the video need to meet a preset time condition, where the preset time condition may be set according to an actual situation, and the embodiment of the present invention is not limited to this. And eliminating the candidate texts which do not meet the preset time condition, and taking the remaining candidate texts as target titles or further selecting the target titles from the remaining candidate texts. According to the occurrence duration and/or the time distribution, many candidate texts which are not the titles can be filtered, and the accuracy of title generation is improved.
For example, title selection is performed on a plurality of candidate texts, the candidate texts are tracked, occurrence duration and time distribution of the candidate texts are recorded, time sequence analysis is performed on the candidate texts, the preset time condition comprises that the occurrence duration of the candidate titles is higher than the preset duration, the candidate texts not higher than the preset duration are removed, the preset time condition also comprises that the time distribution of the candidate titles meets the requirement that the duration of interruption is not higher than the preset interruption duration, and the candidate texts higher than the preset interruption duration are removed.
In this embodiment of the present invention, optionally, before selecting the target title of the target video according to the candidate text, the method may further include: detecting at least one of attribute information of text elements in the candidate text, text length of the candidate text and character categories contained in the candidate text; and eliminating at least one candidate text which does not accord with a preset rule in the attribute information, the text length and the character category.
The text elements include words, phrases, sentences, and the like, and the attribute information of the text elements includes parts of speech, semantics, and the like, or any other applicable attribute information, which is not limited in this embodiment of the present invention. The character category includes a number category, a language category, a punctuation category, and the like, or any other suitable character category, which is not limited in the embodiment of the present invention.
Attribute information, text length, character type and the like of text elements of a title in a video need to meet preset rules, wherein the preset rules can be set according to actual conditions, and the embodiment of the invention does not limit the preset rules. And eliminating the candidate texts which do not accord with the preset rule, and taking the remaining candidate texts as target titles or further selecting the target titles from the remaining candidate texts. According to the preset rule, many candidate texts which are not the titles can be filtered, and the accuracy of title generation is improved.
For example, after the candidate text is obtained by text recognition, a part of the candidate text with a high probability of not being the title may be initially filtered according to the text length and the character type, and the remaining candidate text may be used as an alternative. The preset rules include that the number of words is within a preset number range, the character categories cannot include several preset categories, and the like. And carrying out title selection on the candidate texts, and further filtering the candidate texts after filtering according to the occurrence duration and time distribution. And performing semantic analysis on the alternative titles, detecting the parts of speech, the semantics and the like of the alternative titles, and filtering the candidate texts which do not meet the preset rules. The preset rule includes that verbs cannot be included, names of people cannot be included, and the like in the text.
In this embodiment of the present invention, optionally, an implementation manner of selecting the target title of the target video according to the candidate text may include: inputting the candidate texts into a title detection network; detecting whether the candidate text can be used as a title or not by the title detection network, and outputting a title confidence coefficient; and selecting the candidate text with the highest title confidence as a target title.
The title detection network is used for detecting whether the text can be used as a title or not, and the title detection network is obtained by adopting a title text sample and a non-title text sample for training. For example, a two-class network model, i.e., a headline detection network, is trained based on a large number of news headline corpora and non-headline corpora using supervised learning. The trained title detection network can detect whether the candidate texts can be used as titles or not, the candidate texts are input into the title detection network, title confidence degrees are output by the title detection network, the title confidence degrees can represent the probability that the candidate texts are used as the titles, the candidate texts are ranked according to the title confidence degrees, and the candidate texts with the highest title confidence degrees are selected as target titles.
To sum up, according to the embodiment of the present invention, a target video is obtained, region positions where differences between pixel values in different image frames are smaller than a preset threshold are obtained by comparing different image frames, regions including texts in the image frames are respectively detected, region positions including texts in the image frames are obtained, edge detection is performed on the image frames, region positions of regions surrounded by edges in the image frames are obtained, the region positions corresponding to the image frames are counted, frequency distribution data of the region positions are generated, gradient operation is performed on the frequency distribution data, a gradient operation result is obtained, the position of a title candidate region in an image frame is determined according to the gradient operation result, text recognition is performed on the title candidate regions in the image frames, the candidate text is obtained, the target title of the target video is selected according to the candidate text, so that the title candidate area where the title of the video is likely to appear is obtained through analysis according to a plurality of image frames of the target video, a plurality of confusion information in the video is eliminated, the accuracy in title determination is improved, and then the text recognition is carried out on the title candidate area, so that the target title of the target video is automatically generated, the manual intervention degree in title generation is reduced, and the title generation efficiency for the video is improved.
It should be noted that, for simplicity of description, the method embodiments are described as a series of acts or combination of acts, but those skilled in the art will recognize that the present invention is not limited by the illustrated order of acts, as some steps may occur in other orders or concurrently in accordance with the embodiments of the present invention. Further, those skilled in the art will appreciate that the embodiments described in the specification are presently preferred and that no particular act is required to implement the invention.
Referring to fig. 5, a block diagram of a title generation apparatus according to an embodiment of the present invention is shown, which may specifically include the following modules:
a video obtaining module 301, configured to obtain a target video;
a position detection module 302, configured to detect regions with a title characteristic from a plurality of image frames of the target video, respectively, to obtain region positions;
a region determining module 303, configured to determine, according to the region positions corresponding to the plurality of image frames, positions of title candidate regions in the image frames;
a title generating module 304, configured to perform text recognition on the title candidate regions in the multiple image frames to obtain a target title of the target video.
Optionally, the position detection module comprises at least one of:
the comparison submodule is used for obtaining the area position of which the difference value of the pixel values between different image frames is smaller than a preset threshold value by comparing different image frames;
the text detection submodule is used for respectively detecting the areas containing the texts in the image frames to obtain the positions of the areas containing the texts in the image frames;
and the edge detection submodule is used for respectively carrying out edge detection on the plurality of image frames to obtain the area positions of the areas surrounded by the edges in the plurality of image frames.
Optionally, the region determining module includes:
the data generation submodule is used for counting the area positions corresponding to the image frames and generating frequency distribution data of the area positions;
the gradient operation submodule is used for carrying out gradient operation on the frequency distribution data to obtain a gradient operation result;
and the area determining submodule is used for determining the position of the title candidate area in the image frame according to the gradient operation result.
Optionally, the apparatus further comprises:
a detection module, configured to detect whether an image change rate of the title candidate region is smaller than a preset threshold for each image frame before performing text recognition on the title candidate region in the image frames to obtain a target title of the target video;
and the area removing module is used for removing the title candidate area with the image change rate larger than a preset threshold value.
Optionally, the title generating module includes:
the text recognition submodule is used for respectively carrying out text recognition on the title candidate areas in the plurality of image frames to obtain candidate texts;
and the title selection submodule is used for selecting a target title of the target video according to the candidate text.
Optionally, the text recognition sub-module comprises:
the time generation unit is used for generating the appearance duration and/or time distribution of the candidate text in the target video according to the image frame of the candidate text identified when the title candidate area is subjected to text identification;
the title selection submodule comprises:
and the first eliminating unit is used for eliminating the candidate texts of which the occurrence durations and/or time distributions do not accord with the preset time condition before the target titles of the target videos are selected according to the candidate texts.
Optionally, the title selecting sub-module includes:
the attribute detection unit is used for detecting at least one of attribute information of text elements in the candidate text, text length of the candidate text and character types contained in the candidate text before the target title of the target video is selected according to the candidate text;
and the second eliminating unit is used for eliminating at least one candidate text which does not accord with a preset rule in the attribute information, the text length and the character category.
Optionally, the title selecting sub-module includes:
an input unit configured to input the candidate text into a title detection network; the title detection network is used for detecting whether the text can be used as a title or not, and is obtained by adopting a title text sample and a non-title text sample for training;
an output unit, configured to detect, by the title detection network, whether the candidate text can be used as a title, and output a title confidence;
and the selecting unit is used for selecting the candidate text with the highest title confidence coefficient as the target title.
In summary, according to the embodiments of the present invention, by acquiring a target video, regions having a title characteristic are respectively detected from a plurality of image frames of the target video, and a region position is obtained, determining the position of a title candidate region in the image frames according to the region positions corresponding to the plurality of image frames, performing text recognition on the title candidate regions in the plurality of image frames to obtain a target title of the target video, so that the title candidate area which is possibly appeared in the title of the video is obtained by analyzing according to a plurality of image frames of the target video, thereby eliminating a plurality of confusion information in the video, improving the accuracy when the title is determined, and then, text recognition is carried out on the title candidate area, so that the target title of the target video is automatically generated, the manual intervention degree during title generation is reduced, and the title generation efficiency for the video is improved.
An embodiment of the present invention further provides an electronic device, as shown in fig. 6, including a processor 601, a communication interface 602, a memory 603, and a communication bus 604, where the processor 601, the communication interface 602, and the memory 603 complete mutual communication through the communication bus 604,
a memory 603 for storing a computer program;
the processor 601 is configured to implement the following steps when executing the program stored in the memory 603:
acquiring a target video;
respectively detecting areas with title characteristics from a plurality of image frames of the target video to obtain area positions;
determining the position of a title candidate region in the image frames according to the region positions corresponding to the image frames;
and performing text recognition on the title candidate areas in the plurality of image frames to obtain a target title of the target video.
Optionally, the detecting, from a plurality of image frames of the target video, a region having a title characteristic, and obtaining a region position includes at least one of:
obtaining the area position of which the difference value of pixel values between different image frames is smaller than a preset threshold value by comparing different image frames;
respectively detecting areas containing texts in the image frames to obtain the positions of the areas containing the texts in the image frames;
and respectively carrying out edge detection on the plurality of image frames to obtain the area positions of the areas surrounded by the edges in the plurality of image frames.
Optionally, the generating the title candidate region according to the region positions corresponding to the plurality of image frames includes:
counting the area positions corresponding to the plurality of image frames to generate frequency distribution data of the area positions;
performing gradient operation on the frequency distribution data to obtain a gradient operation result;
and determining the position of the title candidate region in the image frame according to the gradient operation result.
Optionally, before the text recognition is performed on the title candidate regions in the image frames to obtain the target title of the target video, the method further includes:
respectively detecting whether the image change rate of the title candidate area is smaller than a preset threshold value or not for each image frame;
and eliminating the title candidate area with the image change rate larger than a preset threshold value.
Optionally, the performing text recognition on the title candidate regions in the image frames to obtain a target title of the target video includes:
respectively carrying out text recognition on the title candidate regions in the plurality of image frames to obtain candidate texts;
and selecting a target title of the target video according to the candidate text.
Optionally, the performing text recognition on the title candidate regions in the plurality of image frames respectively to obtain candidate texts includes:
when the title candidate area is subjected to text recognition, generating the appearance duration and/or time distribution of the candidate text in the target video according to the image frame of the candidate text;
before the selecting a target title of the target video according to the candidate text, the method further includes:
and eliminating the candidate texts of which the occurrence durations and/or time distributions do not accord with the preset time condition.
Optionally, before the selecting the target title of the target video according to the candidate text, the method further includes:
detecting at least one of attribute information of text elements in the candidate text, text length of the candidate text and character categories contained in the candidate text;
and eliminating at least one candidate text which does not accord with a preset rule in the attribute information, the text length and the character category.
Optionally, the selecting a target title of the target video according to the candidate text includes:
inputting the candidate texts into a title detection network; the title detection network is used for detecting whether the text can be used as a title or not, and is obtained by adopting a title text sample and a non-title text sample for training;
detecting whether the candidate text can be used as a title or not by the title detection network, and outputting a title confidence coefficient;
and selecting the candidate text with the highest title confidence as a target title.
The communication bus mentioned in the above terminal may be a Peripheral Component Interconnect (PCI) bus, an Extended Industry Standard Architecture (EISA) bus, or the like. The communication bus may be divided into an address bus, a data bus, a control bus, etc. For ease of illustration, only one thick line is shown, but this does not mean that there is only one bus or one type of bus.
The communication interface is used for communication between the terminal and other equipment.
The Memory may include a Random Access Memory (RAM) or a non-volatile Memory (non-volatile Memory), such as at least one disk Memory. Optionally, the memory may also be at least one memory device located remotely from the processor.
The Processor may be a general-purpose Processor, and includes a Central Processing Unit (CPU), a Network Processor (NP), and the like; the Integrated Circuit may also be a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA) or other Programmable logic device, a discrete Gate or transistor logic device, or a discrete hardware component.
In yet another embodiment, a computer-readable storage medium is provided, having stored thereon instructions, which, when executed on a computer, cause the computer to perform the method of any of the above embodiments.
In a further embodiment provided by the present invention, there is also provided a computer program product comprising instructions which, when run on a computer, cause the computer to perform the method of any of the above embodiments.
In the above embodiments, the implementation may be wholly or partially realized by software, hardware, firmware, or any combination thereof. When implemented in software, may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer instructions. When loaded and executed on a computer, cause the processes or functions described in accordance with the embodiments of the invention to occur, in whole or in part. The computer may be a general purpose computer, a special purpose computer, a network of computers, or other programmable device. The computer instructions may be stored in a computer readable storage medium or transmitted from one computer readable storage medium to another, for example, from one website site, computer, server, or data center to another website site, computer, server, or data center via wired (e.g., coaxial cable, fiber optic, Digital Subscriber Line (DSL)) or wireless (e.g., infrared, wireless, microwave, etc.). The computer-readable storage medium can be any available medium that can be accessed by a computer or a data storage device, such as a server, a data center, etc., that incorporates one or more of the available media. The usable medium may be a magnetic medium (e.g., floppy Disk, hard Disk, magnetic tape), an optical medium (e.g., DVD), or a semiconductor medium (e.g., Solid State Disk (SSD)), among others.
It is noted that, herein, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other identical elements in a process, method, article, or apparatus that comprises the element.
All the embodiments in the present specification are described in a related manner, and the same and similar parts among the embodiments may be referred to each other, and each embodiment focuses on the differences from the other embodiments. In particular, for the system embodiment, since it is substantially similar to the method embodiment, the description is simple, and for the relevant points, reference may be made to the partial description of the method embodiment.
The above description is only for the preferred embodiment of the present invention, and is not intended to limit the scope of the present invention. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention shall fall within the protection scope of the present invention.

Claims (11)

1. A title generation method, comprising:
acquiring a target video;
respectively detecting areas with title characteristics from a plurality of image frames of the target video to obtain area positions;
determining the position of a title candidate region in the image frames according to the region positions corresponding to the image frames;
and performing text recognition on the title candidate areas in the plurality of image frames to obtain a target title of the target video.
2. The method according to claim 1, wherein the detecting the region having the title characteristic from the plurality of image frames of the target video respectively, and obtaining the region position comprises at least one of:
obtaining the area position of which the difference value of pixel values between different image frames is smaller than a preset threshold value by comparing different image frames;
respectively detecting areas containing texts in the image frames to obtain the positions of the areas containing the texts in the image frames;
and respectively carrying out edge detection on the plurality of image frames to obtain the area positions of the areas surrounded by the edges in the plurality of image frames.
3. The method according to claim 1 or 2, wherein the determining the position of the title candidate region in the image frame according to the region positions corresponding to the plurality of image frames comprises:
counting the area positions corresponding to the plurality of image frames to generate frequency distribution data of the area positions;
performing gradient operation on the frequency distribution data to obtain a gradient operation result;
and determining the position of the title candidate region in the image frame according to the gradient operation result.
4. The method of claim 1, wherein before the text identifying the title candidate regions in the plurality of image frames to obtain the target title of the target video, the method further comprises:
respectively detecting whether the image change rate of the title candidate area is smaller than a preset threshold value or not for each image frame;
and eliminating the title candidate area with the image change rate larger than a preset threshold value.
5. The method of claim 1, wherein the performing text recognition on the title candidate regions in the image frames to obtain the target title of the target video comprises:
respectively carrying out text recognition on the title candidate regions in the plurality of image frames to obtain candidate texts;
and selecting a target title of the target video according to the candidate text.
6. The method of claim 4, wherein the performing text recognition on the candidate regions of the titles in the image frames respectively to obtain candidate texts comprises:
when the title candidate area is subjected to text recognition, generating the appearance duration and/or time distribution of the candidate text in the target video according to the image frame of the candidate text;
before the selecting a target title of the target video according to the candidate text, the method further includes:
and eliminating the candidate texts of which the occurrence durations and/or time distributions do not accord with the preset time condition.
7. The method of claim 4, wherein before said selecting a target title of said target video according to said candidate text, said method further comprises:
detecting at least one of attribute information of text elements in the candidate text, text length of the candidate text and character categories contained in the candidate text;
and eliminating at least one candidate text which does not accord with a preset rule in the attribute information, the text length and the character category.
8. The method of claim 4, wherein selecting the target title of the target video according to the candidate text comprises:
inputting the candidate texts into a title detection network; the title detection network is used for detecting whether the text can be used as a title or not, and is obtained by adopting a title text sample and a non-title text sample for training;
detecting whether the candidate text can be used as a title or not by the title detection network, and outputting a title confidence coefficient;
and selecting the candidate text with the highest title confidence as a target title.
9. A title generation apparatus, comprising:
the video acquisition module is used for acquiring a target video;
the position detection module is used for respectively detecting areas with title characteristics from a plurality of image frames of the target video to obtain area positions;
the area determining module is used for determining the position of a title candidate area in the image frames according to the area positions corresponding to the image frames;
and the title generation module is used for performing text recognition on the title candidate areas in the image frames to obtain the target title of the target video.
10. An electronic device is characterized by comprising a processor, a communication interface, a memory and a communication bus, wherein the processor and the communication interface are used for realizing mutual communication by the memory through the communication bus;
a memory for storing a computer program;
a processor for implementing the method steps of any of claims 1 to 8 when executing a program stored in the memory.
11. A computer-readable storage medium, on which a computer program is stored which, when being executed by a processor, carries out the method according to any one of claims 1 to 8.
CN202110114237.2A 2021-01-27 2021-01-27 Title generation method, device, electronic equipment and storage medium Active CN112818984B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110114237.2A CN112818984B (en) 2021-01-27 2021-01-27 Title generation method, device, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110114237.2A CN112818984B (en) 2021-01-27 2021-01-27 Title generation method, device, electronic equipment and storage medium

Publications (2)

Publication Number Publication Date
CN112818984A true CN112818984A (en) 2021-05-18
CN112818984B CN112818984B (en) 2023-10-24

Family

ID=75860043

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110114237.2A Active CN112818984B (en) 2021-01-27 2021-01-27 Title generation method, device, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN112818984B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113807085A (en) * 2021-11-19 2021-12-17 成都索贝数码科技股份有限公司 Method for extracting title and subtitle aiming at news scene

Citations (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101467145A (en) * 2006-07-24 2009-06-24 谷歌公司 Method and apparatus for automatically annotating images
KR20160027651A (en) * 2014-09-02 2016-03-10 에스케이텔레콤 주식회사 Method and apparatus for indexing moving picture
CN108171235A (en) * 2018-01-08 2018-06-15 北京奇艺世纪科技有限公司 Title area detection method and system
CN108229476A (en) * 2018-01-08 2018-06-29 北京奇艺世纪科技有限公司 Title area detection method and system
CN108288060A (en) * 2018-02-23 2018-07-17 北京奇艺世纪科技有限公司 Title detection method, device and electronic equipment in a kind of video
CN108446603A (en) * 2018-02-28 2018-08-24 北京奇艺世纪科技有限公司 A kind of headline detection method and device
CN108495185A (en) * 2018-03-14 2018-09-04 北京奇艺世纪科技有限公司 A kind of video title generation method and device
CN108769776A (en) * 2018-05-31 2018-11-06 北京奇艺世纪科技有限公司 Main title detection method, device and electronic equipment
US20190095529A1 (en) * 2017-09-28 2019-03-28 Electronics And Telecommunications Research Institute Method and apparatus for generating title and keyframe of video
EP3499900A2 (en) * 2018-05-31 2019-06-19 Beijing Baidu Netcom Science and Technology Co., Ltd. Video processing method, apparatus and device
CN110263214A (en) * 2019-06-21 2019-09-20 北京百度网讯科技有限公司 Generation method, device, server and the storage medium of video title
CN110399526A (en) * 2019-07-26 2019-11-01 腾讯科技(深圳)有限公司 Generation method, device and the computer readable storage medium of video title
CN111626049A (en) * 2020-05-27 2020-09-04 腾讯科技(深圳)有限公司 Title correction method and device for multimedia information, electronic equipment and storage medium
US20200320307A1 (en) * 2019-04-08 2020-10-08 Baidu Usa Llc Method and apparatus for generating video
CN111984824A (en) * 2020-07-31 2020-11-24 河海大学 Multi-mode-based video recommendation method
WO2020253657A1 (en) * 2019-06-17 2020-12-24 腾讯科技(深圳)有限公司 Video clip positioning method and apparatus, computer device, and storage medium
BE1027349A1 (en) * 2020-04-01 2021-01-12 Yu Jian Trading Company Ltd A method, an apparatus, a storage medium and a terminal for generating a video title picture

Patent Citations (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101467145A (en) * 2006-07-24 2009-06-24 谷歌公司 Method and apparatus for automatically annotating images
KR20160027651A (en) * 2014-09-02 2016-03-10 에스케이텔레콤 주식회사 Method and apparatus for indexing moving picture
US20190095529A1 (en) * 2017-09-28 2019-03-28 Electronics And Telecommunications Research Institute Method and apparatus for generating title and keyframe of video
CN108171235A (en) * 2018-01-08 2018-06-15 北京奇艺世纪科技有限公司 Title area detection method and system
CN108229476A (en) * 2018-01-08 2018-06-29 北京奇艺世纪科技有限公司 Title area detection method and system
CN108288060A (en) * 2018-02-23 2018-07-17 北京奇艺世纪科技有限公司 Title detection method, device and electronic equipment in a kind of video
CN108446603A (en) * 2018-02-28 2018-08-24 北京奇艺世纪科技有限公司 A kind of headline detection method and device
CN108495185A (en) * 2018-03-14 2018-09-04 北京奇艺世纪科技有限公司 A kind of video title generation method and device
CN108769776A (en) * 2018-05-31 2018-11-06 北京奇艺世纪科技有限公司 Main title detection method, device and electronic equipment
EP3499900A2 (en) * 2018-05-31 2019-06-19 Beijing Baidu Netcom Science and Technology Co., Ltd. Video processing method, apparatus and device
US20200320307A1 (en) * 2019-04-08 2020-10-08 Baidu Usa Llc Method and apparatus for generating video
WO2020253657A1 (en) * 2019-06-17 2020-12-24 腾讯科技(深圳)有限公司 Video clip positioning method and apparatus, computer device, and storage medium
CN110263214A (en) * 2019-06-21 2019-09-20 北京百度网讯科技有限公司 Generation method, device, server and the storage medium of video title
CN110399526A (en) * 2019-07-26 2019-11-01 腾讯科技(深圳)有限公司 Generation method, device and the computer readable storage medium of video title
BE1027349A1 (en) * 2020-04-01 2021-01-12 Yu Jian Trading Company Ltd A method, an apparatus, a storage medium and a terminal for generating a video title picture
CN111626049A (en) * 2020-05-27 2020-09-04 腾讯科技(深圳)有限公司 Title correction method and device for multimedia information, electronic equipment and storage medium
CN111984824A (en) * 2020-07-31 2020-11-24 河海大学 Multi-mode-based video recommendation method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
汤鹏杰等: "从视频到语言:视频标题生成与描述研究综述", 《自动化学报》 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113807085A (en) * 2021-11-19 2021-12-17 成都索贝数码科技股份有限公司 Method for extracting title and subtitle aiming at news scene

Also Published As

Publication number Publication date
CN112818984B (en) 2023-10-24

Similar Documents

Publication Publication Date Title
CN106649316B (en) Video pushing method and device
CN111274442B (en) Method for determining video tag, server and storage medium
EP2471025B1 (en) A method and system for preprocessing the region of video containing text
CN111062259A (en) Form recognition method and device
WO2020155750A1 (en) Artificial intelligence-based corpus collecting method, apparatus, device, and storage medium
CN111708909B (en) Video tag adding method and device, electronic equipment and computer readable storage medium
CN111767713A (en) Keyword extraction method and device, electronic equipment and storage medium
US20190258629A1 (en) Data mining method based on mixed-type data
CN111460355A (en) Page parsing method and device
EP4273737A1 (en) Language labeling method and apparatus, and computer device and storage medium
CN111191591A (en) Watermark detection method, video processing method and related equipment
US9355099B2 (en) System and method for detecting explicit multimedia content
CN113076961B (en) Image feature library updating method, image detection method and device
CN112925905B (en) Method, device, electronic equipment and storage medium for extracting video subtitles
CN112818984B (en) Title generation method, device, electronic equipment and storage medium
CN113435438A (en) Video screen board extraction and video segmentation method for image and subtitle fusion
US11728914B2 (en) Detection device, detection method, and program
CN112163415A (en) User intention identification method and device for feedback content and electronic equipment
US20140307968A1 (en) Method and apparatus for automatic genre identification and classification
KR102028356B1 (en) Advertisement recommendation apparatus and method based on comments
CN115879002A (en) Training sample generation method, model training method and device
CN114220057A (en) Video trailer identification method and device, electronic equipment and readable storage medium
JP2010026923A (en) Method, device and program for document classification, and computer-readable recording medium
CN114140782A (en) Text recognition method and device, electronic equipment and storage medium
CN114048740A (en) Sensitive word detection method and device and computer readable storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant