CN113033190B - Subtitle generation method, subtitle generation device, medium and electronic equipment - Google Patents
Subtitle generation method, subtitle generation device, medium and electronic equipment Download PDFInfo
- Publication number
- CN113033190B CN113033190B CN202110420704.4A CN202110420704A CN113033190B CN 113033190 B CN113033190 B CN 113033190B CN 202110420704 A CN202110420704 A CN 202110420704A CN 113033190 B CN113033190 B CN 113033190B
- Authority
- CN
- China
- Prior art keywords
- information
- description information
- candidate
- original
- phrase
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 48
- 238000012549 training Methods 0.000 claims description 27
- 238000000605 extraction Methods 0.000 claims description 26
- 238000004590 computer program Methods 0.000 claims description 17
- 238000001914 filtration Methods 0.000 claims description 17
- 239000003607 modifier Substances 0.000 claims description 17
- 238000012545 processing Methods 0.000 claims description 12
- 230000011218 segmentation Effects 0.000 claims description 11
- 238000007670 refining Methods 0.000 abstract description 6
- 239000002023 wood Substances 0.000 description 21
- 239000007787 solid Substances 0.000 description 15
- 230000015654 memory Effects 0.000 description 9
- 238000010586 diagram Methods 0.000 description 7
- 230000006870 function Effects 0.000 description 7
- 238000004891 communication Methods 0.000 description 6
- 230000003287 optical effect Effects 0.000 description 5
- 239000013598 vector Substances 0.000 description 4
- 239000000463 material Substances 0.000 description 3
- 230000000717 retained effect Effects 0.000 description 3
- 241000238631 Hexapoda Species 0.000 description 2
- 239000000428 dust Substances 0.000 description 2
- 239000013307 optical fiber Substances 0.000 description 2
- 230000000644 propagated effect Effects 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 2
- 102100033814 Alanine aminotransferase 2 Human genes 0.000 description 1
- 101000779415 Homo sapiens Alanine aminotransferase 2 Proteins 0.000 description 1
- 238000012512 characterization method Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 239000002245 particle Substances 0.000 description 1
- 230000006403 short-term memory Effects 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 238000012800 visualization Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/258—Heading extraction; Automatic titling; Numbering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/279—Recognition of textual entities
- G06F40/289—Phrasal analysis, e.g. finite state techniques or chunking
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Studio Circuits (AREA)
Abstract
The disclosure relates to a subtitle generating method, a subtitle generating device, a subtitle generating medium and electronic equipment. The method comprises the following steps: acquiring original object description information of a target object to be introduced in a video, wherein the original object description information is used for describing original inherent characteristics of the target object; content refining is carried out on the original object description information, and brief description information of the target object is obtained; generating an object characteristic description phrase for describing the characteristics of the target object according to the original object description information; and generating caption information corresponding to the target object in the video according to the brief description information and the object characteristic description phrase. Therefore, the subtitle information related to the target object to be introduced in the video can be automatically generated, the readability and the accuracy of the subtitle information can be ensured, and a user can quickly know the target object through the video.
Description
Technical Field
The disclosure relates to the technical field of computers, and in particular relates to a subtitle generating method, a subtitle generating device, a subtitle generating medium and electronic equipment.
Background
With the rapid development of the internet, the amount of information in the internet has exploded, and users are increasingly accustomed to acquiring various required information from the internet. In particular, in daily life, for anything that is not known, a user can quickly refer to what he wants to know by searching the internet for the relevant content of the thing.
At present, information in the internet generally exists in the form of text, images, audios and videos and the like. Among them, information in the form of video is often used as a main way for users to understand things with the advantages of rich content, visualization, intuitiveness, etc. However, in the video production process, the related introduction text of the object is usually directly used as a subtitle, or the related introduction text is slightly colored and then used as a subtitle, but the related introduction text usually has more content and poor readability, so that a user can not conveniently and quickly know the object through the video.
Disclosure of Invention
This summary is provided to introduce a selection of concepts in a simplified form that are further described below in the detailed description. This summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter.
In a first aspect, the present disclosure provides a subtitle generating method, including:
acquiring original object description information of a target object to be introduced in a video, wherein the original object description information is used for describing original inherent characteristics of the target object;
content refining is carried out on the original object description information, and brief description information of the target object is obtained;
Generating an object characteristic description phrase for describing the characteristics of the target object according to the original object description information;
And generating caption information corresponding to the target object in the video according to the brief description information and the object characteristic description phrase.
In a second aspect, the present disclosure provides a subtitle generating apparatus, including:
The acquisition module is used for acquiring original object description information of a target object to be introduced in the video, wherein the original object description information is used for describing original inherent characteristics of the target object;
The extraction module is used for extracting the content of the original object description information acquired by the acquisition module to acquire brief description information of the target object;
The phrase generating module is used for generating an object characteristic description phrase for describing the characteristics of the target object according to the original object description information acquired by the acquiring module;
And the caption generating module is used for generating caption information corresponding to the target object in the video according to the brief description information obtained by the extracting module and the object characteristic description phrase generated by the phrase generating module.
In a third aspect, the present disclosure provides a computer readable medium having stored thereon a computer program which when executed by a processing device implements the steps of the method provided by the first aspect of the present disclosure.
In a fourth aspect, the present disclosure provides an electronic device comprising:
A storage device having a computer program stored thereon;
processing means for executing said computer program in said storage means to carry out the steps of the method provided by the first aspect of the present disclosure.
In the above technical solution, after original object description information for describing original intrinsic characteristics of a target object to be introduced in a video is obtained, brief description information of the target object and an object characteristic description phrase for describing characteristics of the target object are determined according to the original object description information; and then, generating subtitle information corresponding to the target object in the video according to the subtitle information and the target object. Therefore, the subtitle information related to the target object to be introduced in the video can be automatically generated, the readability and the accuracy of the subtitle information can be ensured, and a user can quickly know the target object through the video.
Additional features and advantages of the present disclosure will be set forth in the detailed description which follows.
Drawings
The above and other features, advantages, and aspects of embodiments of the present disclosure will become more apparent by reference to the following detailed description when taken in conjunction with the accompanying drawings. The same or similar reference numbers will be used throughout the drawings to refer to the same or like elements. It should be understood that the figures are schematic and that elements and components are not necessarily drawn to scale. In the drawings:
fig. 1 is a flowchart illustrating a subtitle generating method according to an exemplary embodiment.
Fig. 2 is a flowchart illustrating a method of generating subtitle information corresponding to a target object in a video according to brief description information and an object property description phrase according to an exemplary embodiment.
Fig. 3 is a flowchart illustrating a method of generating subtitle information corresponding to a target object in a video according to brief description information and an object property description phrase according to another exemplary embodiment.
Fig. 4 is a block diagram illustrating a subtitle generating apparatus according to an exemplary embodiment.
Fig. 5 is a block diagram of an electronic device, according to an example embodiment.
Detailed Description
Embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. While certain embodiments of the present disclosure have been shown in the accompanying drawings, it is to be understood that the present disclosure may be embodied in various forms and should not be construed as limited to the embodiments set forth herein, but are provided to provide a more thorough and complete understanding of the present disclosure. It should be understood that the drawings and embodiments of the present disclosure are for illustration purposes only and are not intended to limit the scope of the present disclosure.
It should be understood that the various steps recited in the method embodiments of the present disclosure may be performed in a different order and/or performed in parallel. Furthermore, method embodiments may include additional steps and/or omit performing the illustrated steps. The scope of the present disclosure is not limited in this respect.
The term "including" and variations thereof as used herein are intended to be open-ended, i.e., including, but not limited to. The term "based on" is based at least in part on. The term "one embodiment" means "at least one embodiment"; the term "another embodiment" means "at least one additional embodiment"; the term "some embodiments" means "at least some embodiments. Related definitions of other terms will be given in the description below.
It should be noted that the terms "first," "second," and the like in this disclosure are merely used to distinguish between different devices, modules, or units and are not used to define an order or interdependence of functions performed by the devices, modules, or units.
It should be noted that references to "one", "a plurality" and "a plurality" in this disclosure are intended to be illustrative rather than limiting, and those of ordinary skill in the art will appreciate that "one or more" is intended to be understood as "one or more" unless the context clearly indicates otherwise.
The names of messages or information interacted between the various devices in the embodiments of the present disclosure are for illustrative purposes only and are not intended to limit the scope of such messages or information.
Fig. 1 is a flowchart illustrating a subtitle generating method according to an exemplary embodiment. As shown in fig. 1, the method may include S101 to S104.
In S101, original object description information of a target object to be introduced in a video is acquired.
In the present disclosure, the target object may be, for example, flowers, birds, daily necessities, furniture, or the like. The original object description information is used to describe original inherent characteristics of the target object, such as style, specification, name, material, etc. Illustratively, the target object is a wardrobe, and its original object description information is: the high-end bedroom combined clothes cabinet is simple, modern and European style, is assembled by solid wood particle boards, and can be assembled in a sliding manner; and the capacity is large, and the cabinet can be used as a storage cabinet, the color matching is white in the atmosphere and oak, and the height is 1.8 meters, and the cabinet is provided with a top cabinet.
In S102, the original object description information is content-refined to obtain brief description information of the target object.
Illustratively, based on the original object description information of the target object "wardrobe" in the above example, the generated brief description information is: sliding and sliding assembled combined clothes cabinet.
In S103, an object property description phrase for describing the property of the target object is generated from the original object description information.
In the present disclosure, the object property description phrase can highlight a feature unique to the target object, and the number of generated object property description phrases may be one or more, which is not specifically limited in the present disclosure.
Illustratively, based on the original object description information of the target object "wardrobe" in the above example, 6 object property description phrases for describing properties of the wardrobe are generated as follows: ① European style, simple atmosphere; ② Constructing solid wood and high-end atmosphere; ③ Multifunctional solid wood, high capacity and foldability; ④ Simple atmosphere, water resistance, insect resistance and dust resistance; ⑤ European wood grain, simple but not simple; ⑥ Simple and modern, european style large capacity.
In S104, caption information corresponding to the target object in the video is generated from the brief description information and the object property description phrase.
In the above technical solution, after original object description information for describing original intrinsic characteristics of a target object to be introduced in a video is obtained, brief description information of the target object and an object characteristic description phrase for describing characteristics of the target object are determined according to the original object description information; and then, generating subtitle information corresponding to the target object in the video according to the subtitle information and the target object. Therefore, the subtitle information related to the target object to be introduced in the video can be automatically generated, the readability and the accuracy of the subtitle information can be ensured, and a user can quickly know the target object through the video.
The following describes in detail the embodiment of the brief description information of the target object obtained by content-refining the description information of the original object in S102.
Specifically, according to the original object description information of the target object, the brief description information of the target object can be obtained in various modes, in one embodiment, keywords can be extracted from the original object description information in a statistical mode, for example, words, which occur frequently exceeding a preset frequency threshold, in the original object description information are determined to be keywords; and then, splicing the keywords to obtain the brief description information.
In another embodiment, the original object description information may be input into a pre-trained content extraction model to perform content extraction on the original object description information to obtain the brief description information of the target object. Thus, the brief description information of the target object can be automatically extracted from the original object description information through the content extraction model, and the method is convenient and quick.
In the present disclosure, the content refinement model may be, for example, a transducer model, a Long Short-Term Memory (LSTM) model, or the like.
The content extraction model can be obtained through training in the following steps (1) to (7):
(1) First reference object description information of a first reference object is acquired, wherein the first reference object description information is used for describing original inherent characteristics of the first reference object.
(2) And segmenting the first reference object description information to obtain a first word sequence, wherein the words in the first word sequence are sequentially arranged according to the sequence of the words in the first reference object description information.
(3) And deleting each word except the first reference object in the first word sequence to obtain a new first word sequence, and sequentially connecting each word in the new first word sequence to obtain candidate brief description information of the first reference object.
(4) It is determined whether there are other words in any new first word sequence than the first reference object.
If there are other words in any new first word sequence except the first reference object, executing the following step (5); if no other words than the first reference object are present in any new first word sequence, the following step (6) is performed.
(5) And (3) taking the new first word sequence as the first word sequence for each new first word sequence, and returning to the step (3).
(6) The candidate brief description information with the highest fluency is taken as reference brief description information.
In the present disclosure, the fluency of candidate profile information may be determined by a pre-trained language model (e.g., a GPT2 model).
(7) Model training is performed by taking the first reference object description information as an input of the content refinement model in such a manner that the reference brief description information is taken as a target output of the content refinement model to obtain the content refinement model.
In the training process of the content extraction model, the reference brief description information can be directly determined based on the first reference object description information, so that the problem that the model training performance is affected due to the fact that training samples are limited can be avoided, the reference brief description information can be smooth as much as possible, and the content extraction performance of the content extraction model can be improved.
The following describes in detail a specific embodiment of generating an object property description phrase for describing the property of the target object from the original object description information in S103 described above.
Specifically, according to the original object description information of the target object, the object characteristic description phrase for describing the characteristics of the target object may be obtained in various manners, in one embodiment, the description words (typically adjectives) related to the characteristics of the target object may be extracted from the original object description information, and then, the description words are combined and spliced to obtain the object characteristic description phrase for describing the characteristics of the target object.
In another embodiment, descriptors related to the characteristics of the target object may be extracted from the original object description information; meanwhile, according to the corresponding relation between the pre-stored object and the characteristic descriptor, determining the target characteristic descriptor corresponding to the target object, wherein the corresponding relation can be formed according to a corpus; and combining and splicing the extracted description words related to the characteristics of the target object and the target characteristic description words determined according to the corresponding relation to obtain the object characteristic description phrase for describing the characteristics of the target object.
In yet another embodiment, the original object description information may be input into a pre-trained phrase generation model to obtain an object property description phrase for describing properties of the target object. Thus, object characteristic description phrases for describing characteristics of the target object can be automatically generated through the phrase generation model, and convenience and rapidness are realized. The phrase generation model may be, for example, a transducer model, an LSTM model, or the like.
The phrase generating model can be obtained through training in the following steps 1) to 5):
1) And acquiring second reference object description information and third reference object description information of a second reference object, wherein the second reference object description information and the third reference object description information are used for describing original inherent characteristics of the second reference object, and the similarity between the second reference object description information and the third reference object description information is larger than a preset similarity threshold.
2) And performing word segmentation on the second reference object description information to obtain a second word sequence, and performing word segmentation on the third reference object description information to obtain a third word sequence.
3) Extracting a phrase containing each modifier from the second reference object description information as a first candidate object characteristic description phrase aiming at each modifier for the second reference object in the second word sequence; for each modifier in the third word sequence for describing the second reference object, extracting a phrase containing the modifier from the third reference object description information as a second candidate object characteristic description phrase.
4) Each first candidate property description phrase and each second candidate property description phrase is determined to be a reference object property description phrase.
5) Model training is performed by taking the second reference object description information as input of a phrase generating model and taking the reference object characteristic description phrase as target output of the phrase generating model to obtain the phrase generating model.
In the training process of the phrase generation model, the reference object characteristic description phrase corresponding to the second reference object description information is determined based on the second reference object description information and third reference object description information similar to the second reference object description information, so that the reference object characteristic description phrase can embody the characteristic of the second reference object as far as possible, the phrase generation performance of the phrase generation model is improved, the generated object characteristic description phrase can fully embody the characteristic of the target object, and the accuracy of the subtitle information is improved. In addition, the phrase of the characteristic description of the reference object corresponding to the second reference object description information is automatically generated, so that the problem that the training performance of the model is influenced due to limited training samples can be avoided, and the phrase generating performance of the phrase generating model is further improved.
The following describes in detail the specific embodiment of generating the subtitle information corresponding to the target object in the video according to the brief description information and the object property description phrase in S104.
In one embodiment, the object property description phrase generated in S103 and used for describing the property of the target object is one, at this time, the brief description information generated in S102 and the object property description phrase generated in S103 and used for describing the property of the target object may be directly combined, so as to obtain the subtitle information corresponding to the target object in the video.
Illustratively, based on the original object description information of the target object "wardrobe" in the above example, the generated brief description information is: the sliding and sliding assembled combined clothes cabinet generates object characteristic description phrase for describing characteristics of the clothes cabinet, wherein the object characteristic description phrase is as follows: the method comprises the steps of constructing solid wood, constructing the high-end atmosphere, and combining the solid wood and the high-end atmosphere to obtain subtitle information 'push-pull sliding assembly combined clothes cabinet' corresponding to a target object 'wardrobe' in a video.
In another embodiment, the object property description phrase generated in S103 is plural and used for describing the property of the target object, and at this time, subtitle information corresponding to the target object in the video may be generated through S1041 and S1042 shown in fig. 2.
In S1041, the brief description information is respectively combined with each object property description phrase to obtain a plurality of candidate subtitle information.
In S1042, target subtitle information is determined from among the plurality of candidate subtitle information as subtitle information corresponding to the target object in the video.
Specifically, for each candidate subtitle information, a similarity between the candidate subtitle information and the original object description information may be determined; then, the candidate subtitle information having the highest similarity with the original object description information is determined as target subtitle information. The similarity may be, for example, cosine similarity, euclidean distance similarity, or the like.
In addition, the similarity between the candidate subtitle information and the original object description information may be determined by: firstly, determining semantic information corresponding to candidate subtitle information and semantic information corresponding to original object description information; then, the similarity between the semantic information corresponding to the candidate caption information and the semantic information corresponding to the original object description information is calculated, and the calculated similarity is used as the similarity between the candidate caption information and the original object description information. Specifically, for each candidate subtitle, the semantic vector (i.e., semantic information) corresponding to the generated candidate subtitle information and the semantic vector corresponding to the original object description information are respectively generated through the title vector characterization service, and then cosine similarity or euclidean distance similarity between the two semantic vectors is calculated and is used as the similarity between the candidate subtitle information and the original object description information.
Illustratively, based on the original object description information of the target object "wardrobe" in the above example, the generated brief description information is: the sliding and sliding assembled combined clothes cabinet generates object characteristic description phrase for describing characteristics of the clothes cabinet, which comprises the following steps: ① European style, simple atmosphere; ② Constructing solid wood and high-end atmosphere; ③ Multifunctional solid wood, high capacity and foldability; ④ Simple atmosphere, water resistance, insect resistance and dust resistance; ⑤ European wood grain, simple but not simple; ⑥ Simple and modern, european style large capacity; the brief description information 'push-pull sliding assembled combined clothes cabinet' is respectively combined with each object characteristic description phrase in the 6 object characteristic description phrases to obtain the following 6 candidate caption information:
(1) Push-pull sliding assembled combined clothes cabinet, european style and simple atmosphere;
(2) Sliding and sliding assembled combined clothes cabinet, solid wood is manufactured, and the high end is air;
(3) Push-pull sliding assembled combined clothes cabinet, multifunctional solid wood and high-capacity foldable;
(4) The sliding assembly combined clothes cabinet is simple in atmosphere, waterproof, insect-proof and dust-proof;
(5) The sliding assembly combined clothes cabinet is simple and not easy, and European wood grain is reduced;
(6) The sliding assembly combined clothes cabinet is simple and modern, and has European style and large capacity.
The candidate caption information with the highest similarity with the original object description information is the candidate caption information (2), so that the candidate caption information (2) can be determined as target caption information, namely, the caption information corresponding to a target object wardrobe in the video is a sliding-sliding assembled combined wardrobe, and the combined wardrobe is manufactured by solid wood and is high-end in atmosphere.
In addition, the object property description phrase generated in S103 for describing the property of the target object is not all derived from the original object description information, so that the subtitle information corresponding to the target object in the generated video may contain information in which the original object description information does not appear, based on the object property description phrase obtained in S103 for describing the property of the target object, so that the accuracy of the generated subtitle information cannot be guaranteed. For this reason, before determining the target subtitle information from the plurality of candidate subtitle information in S1042, the plurality of candidate subtitle information obtained by combining S1041 needs to be filtered to ensure accuracy and efficiency of the subsequent subtitle information generation. Specifically, as shown in fig. 3, S104 further includes S1043.
In S1043, for each candidate subtitle information, at least one object property description word for describing the property of the target object is extracted from the candidate subtitle information, and if there is an object property description word that does not appear in the original object description information, the candidate subtitle information is filtered out.
In the present disclosure, the object property description word may be a modifier describing a material, a class, a style, or the like of the target object. After the filtering operation is performed on the plurality of candidate subtitles in S1043, the above S1042 may determine the target subtitle information from the candidate subtitle information remaining after the filtering operation. Specifically, for each piece of candidate caption information remaining after the filtering operation, a similarity between the candidate caption information and the original object description information may be determined; then, the candidate subtitle information having the highest similarity with the original object description information is determined as target subtitle information.
A detailed description will be given below of a specific embodiment of extracting at least one object property description word for describing a property of a target object from the candidate subtitle information in S1043 described above. Specifically, named entity Recognition (NAMED ENTITY Recognizion, NER) may be performed on the candidate caption information to obtain at least one named entity, and the at least one named entity is used as at least one object property description term for describing properties of the target object.
Wherein, NER is used to identify an object property description entity, such as a material, a category, a style, etc., for describing a property of a target object in the candidate subtitle information.
For example, based on the target object "wardrobe" in the above example, 6 pieces of candidate subtitle information are obtained by combining in S1041, specifically as follows:
(1) Push-pull sliding assembled combined clothes cabinet, european style and simple atmosphere;
(2) Sliding and sliding assembled combined clothes cabinet, solid wood is manufactured, and the high end is air;
(3) Push-pull sliding assembled combined clothes cabinet, multifunctional solid wood and high-capacity foldable;
(4) The sliding assembly combined clothes cabinet is simple in atmosphere, waterproof, insect-proof and dust-proof;
(5) The sliding assembly combined clothes cabinet is simple and not easy, and European wood grain is reduced;
(6) The sliding assembly combined clothes cabinet is simple and modern, and has European style and large capacity.
Wherein, for the candidate subtitle information (1), the object characteristic description word extracted therefrom for describing the characteristic of the target object "wardrobe" includes: "European", "brief", "atmospheric", it can be seen that these three words all appear in the original object description information, and therefore, candidate subtitle information (1) is retained;
For the candidate subtitle information (2), object characteristic description words extracted therefrom for describing characteristics of the target object "wardrobe" include: "solid wood", "high-end", "atmosphere", it can be seen that these three words all appear in the original object description information, thus, the candidate subtitle information (2) is retained;
For the candidate subtitle information (3), object characteristic description words extracted therefrom for describing characteristics of the target object "wardrobe" include: "multifunction", "solid wood", "large capacity", "foldable", wherein "foldable" does not appear in the original object description information, i.e., there are object property description words in the candidate subtitle information (3) that do not appear in the original object description information, and therefore, the candidate subtitle information (3) is filtered out from the above 6 candidate subtitles;
for the candidate subtitle information (4), object characteristic description words extracted therefrom for describing characteristics of the target object "wardrobe" include: "brief", "atmosphere", "waterproof", "insect-proof", "dust-proof", wherein "waterproof", "insect-proof", "dust-proof" do not appear in the original object description information, namely there are object characteristic description words that do not appear in the original object description information in the candidate subtitle information (4), therefore, filter the candidate subtitle information (4) from above-mentioned 6 candidate subtitles;
For the candidate subtitle information (5), object characteristic description words extracted therefrom for describing characteristics of the target object "wardrobe" include: "European", "wood grain", "simple", "not simple", wherein "wood grain", "not simple" do not appear in the original object description information, namely there are object characteristic description words that do not appear in the original object description information in the candidate subtitle information (5), therefore, filter the candidate subtitle information (5) from above-mentioned 6 candidate subtitles;
For the candidate subtitle information (6), object characteristic description words extracted therefrom for describing characteristics of the target object "wardrobe" include: "brief", "modern", "European", "large capacity", it can be seen that the four words all appear in the original object description information, and thus, candidate subtitle information (6) is retained.
Thus, the candidate caption information remaining after the filtering operation includes candidate caption information (1), candidate caption information (2), and candidate caption information (6), that is: push-pull sliding assembled combined clothes cabinet, european style and simple atmosphere; sliding and sliding assembled combined clothes cabinet, solid wood is manufactured, and the high end is air; the sliding assembly combined clothes cabinet is simple and modern, and has European style and large capacity. After that, S1043 may determine a target subtitle from among the three candidate subtitles.
Based on the same inventive concept, the present disclosure also provides a subtitle generating apparatus. As shown in fig. 4, the apparatus 400 includes: an obtaining module 401, configured to obtain original object description information of a target object to be introduced in a video, where the original object description information is used to describe original intrinsic characteristics of the target object; a refinement module 402, configured to perform content refinement on the original object description information acquired by the acquisition module 401, so as to obtain brief description information of the target object; a phrase generating module 403, configured to generate an object characteristic description phrase for describing a characteristic of the target object according to the original object description information acquired by the acquiring module 402; and a caption generating module 404, configured to generate caption information corresponding to the target object in the video according to the brief description information obtained by the extracting module 402 and the object characteristic description phrase generated by the phrase generating module 403.
In the above technical solution, after original object description information for describing original intrinsic characteristics of a target object to be introduced in a video is obtained, brief description information of the target object and an object characteristic description phrase for describing characteristics of the target object are determined according to the original object description information; and then, generating subtitle information corresponding to the target object in the video according to the subtitle information and the target object. Therefore, the subtitle information related to the target object to be introduced in the video can be automatically generated, the readability and the accuracy of the subtitle information can be ensured, and a user can quickly know the target object through the video.
In one embodiment, the refinement module 402 is configured to statistically extract keywords from the original object description information; and then, splicing the keywords to obtain the brief description information.
In another embodiment, the refinement module 402 is configured to input the original object description information into a pre-trained content refinement model to perform content refinement on the original object description information to obtain the brief description information of the target object. Thus, the brief description information of the target object can be automatically extracted from the original object description information through the content extraction model, and the method is convenient and quick.
In one embodiment, the phrase generating module 403 is configured to extract, from the original object description information, a description word related to the characteristics of the target object, and then combine and splice the description words to obtain an object characteristic description phrase for describing the characteristics of the target object.
In another embodiment, the phrase generation module 403 is configured to: extracting descriptive words related to target object characteristics from the original object descriptive information; meanwhile, according to the corresponding relation between the pre-stored object and the characteristic descriptor, determining the target characteristic descriptor corresponding to the target object, wherein the corresponding relation can be formed according to a corpus; and combining and splicing the extracted description words related to the characteristics of the target object and the target characteristic description words determined according to the corresponding relation to obtain the object characteristic description phrase for describing the characteristics of the target object.
In yet another embodiment, the phrase generating module 403 is configured to input the original object description information into a pre-trained phrase generating model to obtain an object property description phrase for describing a property of the target object. Thus, object characteristic description phrases for describing characteristics of the target object can be automatically generated through the phrase generation model, and convenience and rapidness are realized.
Optionally, the object property description phrase is one; the caption generating module 404 is configured to directly combine the brief description information and the object characteristic description phrase to obtain caption information corresponding to a target object in the video.
Optionally, the object property description phrase is a plurality of; the subtitle generating module 404 includes: a combination sub-module, configured to combine the brief description information with each object characteristic description phrase to obtain multiple candidate subtitle information; and the determining submodule is used for determining target subtitle information from the plurality of candidate subtitle information as subtitle information corresponding to the target object in the video.
Optionally, the subtitle generating module 404 further includes: and a filtering sub-module for: extracting, for each of the candidate subtitle information, at least one object property description word for describing a property of the target object from the candidate subtitle information; if the object characteristic description words which do not appear in the original object description information exist, filtering the candidate subtitle information; the determining submodule is used for determining the target subtitle information from candidate subtitle information remained after filtering operation.
Optionally, the determining submodule includes: a similarity determining submodule, configured to determine, for each piece of candidate subtitle information remaining after the filtering operation, a similarity between the candidate subtitle information and the original object description information; and the subtitle determining sub-module is used for determining the candidate subtitle information with the highest similarity with the original object description information as the target subtitle information.
Optionally, the filtering sub-module is configured to identify the named entity of the candidate subtitle information, obtain at least one named entity, and use the at least one named entity as at least one object characteristic description term for describing characteristics of the target object.
Optionally, the determining submodule includes: a similarity determining sub-module, configured to determine, for each piece of candidate subtitle information, a similarity between the candidate subtitle information and the original object description information; and the subtitle determining sub-module is used for determining the candidate subtitle information with the highest similarity with the original object description information as the target subtitle information.
Optionally, the similarity determining submodule includes: the semantic information determining submodule is used for determining semantic information corresponding to the candidate subtitle information and semantic information corresponding to the original object description information; and the calculating sub-module is used for calculating the similarity between the semantic information corresponding to the candidate subtitle information and the semantic information corresponding to the original object description information, and taking the similarity as the similarity between the candidate subtitle information and the original object description information.
Optionally, the refinement module 402 is configured to input the original object description information into a pre-trained content refinement model, so as to refine the content of the original object description information, and obtain the brief description information of the target object; wherein the content refinement model is trained by a first model training device, wherein the first model training device comprises: the first descriptive information acquisition module is used for acquiring first reference object descriptive information of a first reference object, wherein the first reference object descriptive information is used for describing original inherent characteristics of the first reference object; the first word segmentation module is used for segmenting the first reference object description information to obtain a first word sequence; the first determining module is used for deleting each word except the first reference object in the first word sequence to obtain a new first word sequence, and sequentially connecting the words in the new first word sequence to obtain candidate brief description information of the first reference object; the judging module is used for judging whether other words except the first reference object exist in any new first word sequence; the triggering module is used for triggering the first determining module to delete each word except for the first reference object in the first word sequence to obtain a new first word sequence, and sequentially connecting the words in the new first word sequence to obtain candidate brief description information for each word except for the first reference object in the first word sequence; the second determining module is configured to use the candidate brief description information with the highest fluency as reference brief description information if no other word except the first reference object exists in any new first word sequence; and the first model training module is used for carrying out model training in a mode that the first reference object description information is used as the input of the content extraction model and the reference brief description information is used as the target output of the content extraction model so as to obtain the content extraction model.
Optionally, the phrase generating module 403 is configured to input the original object description information into a pre-trained phrase generating model, to obtain an object feature description phrase for describing a feature of the target object; the phrase generating model is trained by a second model training device, wherein the second model training device comprises: the second description information acquisition module is used for acquiring second reference object description information and third reference object description information of a second reference object, wherein the second reference object description information and the third reference object description information are used for describing original inherent characteristics of the second reference object, and the similarity between the second reference object description information and the third reference object description information is larger than a preset similarity threshold; the second word segmentation module is used for segmenting the second reference object description information to obtain a second word sequence, and segmenting the third reference object description information to obtain a third word sequence; a third determining module, configured to: extracting a phrase containing each modifier from the second reference object description information as a first candidate object characteristic description phrase aiming at each modifier in the second word sequence for describing the second reference object; extracting a phrase containing each modifier from the third reference object description information as a second candidate object characteristic description phrase aiming at each modifier in the third word sequence for describing the second reference object; a fourth determining module configured to determine each of the first candidate object property description phrases and each of the second candidate object property description phrases as reference object property description phrases; and the second model training module is used for carrying out model training in a mode that the second reference object description information is used as the input of the phrase generation model and the reference object characteristic description phrase is used as the target output of the phrase generation model so as to obtain the phrase generation model.
Note that the first model training device may be independent of the subtitle generating device 400, may be integrated into the subtitle generating device 400, and the second model training device may be independent of the subtitle generating device 400, may be integrated into the subtitle generating device 400, and is not particularly limited in this respect.
The present disclosure also provides a computer-readable medium having stored thereon a computer program which, when executed by a processing device, implements the steps of the above-described subtitle generating method provided by the present disclosure.
Referring now to fig. 5, a schematic diagram of an electronic device (e.g., a terminal device or server) 500 suitable for use in implementing embodiments of the present disclosure is shown. The terminal devices in the embodiments of the present disclosure may include, but are not limited to, mobile terminals such as mobile phones, notebook computers, digital broadcast receivers, PDAs (personal digital assistants), PADs (tablet computers), PMPs (portable multimedia players), in-vehicle terminals (e.g., in-vehicle navigation terminals), and the like, and stationary terminals such as digital TVs, desktop computers, and the like. The electronic device shown in fig. 5 is merely an example and should not be construed to limit the functionality and scope of use of the disclosed embodiments.
As shown in fig. 5, the electronic device 500 may include a processing means (e.g., a central processing unit, a graphics processor, etc.) 501, which may perform various appropriate actions and processes according to a program stored in a Read Only Memory (ROM) 502 or a program loaded from a storage means 508 into a Random Access Memory (RAM) 503. In the RAM 503, various programs and data required for the operation of the electronic apparatus 500 are also stored. The processing device 501, the ROM 502, and the RAM 503 are connected to each other via a bus 504. An input/output (I/O) interface 505 is also connected to bus 504.
In general, the following devices may be connected to the I/O interface 505: input devices 505 including, for example, a touch screen, touchpad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, and the like; an output device 507 including, for example, a Liquid Crystal Display (LCD), a speaker, a vibrator, and the like; storage 508 including, for example, magnetic tape, hard disk, etc.; and communication means 509. The communication means 509 may allow the electronic device 500 to communicate with other devices wirelessly or by wire to exchange data. While fig. 5 shows an electronic device 500 having various means, it is to be understood that not all of the illustrated means are required to be implemented or provided. More or fewer devices may be implemented or provided instead.
In particular, according to embodiments of the present disclosure, the processes described above with reference to flowcharts may be implemented as computer software programs. For example, embodiments of the present disclosure include a computer program product comprising a computer program embodied on a non-transitory computer readable medium, the computer program comprising program code for performing the method shown in the flow chart. In such an embodiment, the computer program may be downloaded and installed from a network via the communication means 509, or from the storage means 508, or from the ROM 502. The above-described functions defined in the methods of the embodiments of the present disclosure are performed when the computer program is executed by the processing device 501.
It should be noted that the computer readable medium described in the present disclosure may be a computer readable signal medium or a computer readable storage medium, or any combination of the two. The computer readable storage medium can be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or a combination of any of the foregoing. More specific examples of the computer-readable storage medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this disclosure, a computer-readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. In the present disclosure, however, the computer-readable signal medium may include a data signal propagated in baseband or as part of a carrier wave, with the computer-readable program code embodied therein. Such a propagated data signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination of the foregoing. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: electrical wires, fiber optic cables, RF (radio frequency), and the like, or any suitable combination of the foregoing.
In some embodiments, the clients, servers may communicate using any currently known or future developed network protocol, such as HTTP (HyperText Transfer Protocol ), and may be interconnected with any form or medium of digital data communication (e.g., a communication network). Examples of communication networks include a local area network ("LAN"), a wide area network ("WAN"), the internet (e.g., the internet), and peer-to-peer networks (e.g., ad hoc peer-to-peer networks), as well as any currently known or future developed networks.
The computer readable medium may be contained in the electronic device; or may exist alone without being incorporated into the electronic device.
The computer readable medium carries one or more programs which, when executed by the electronic device, cause the electronic device to: acquiring original object description information of a target object to be introduced in a video, wherein the original object description information is used for describing original inherent characteristics of the target object; content refining is carried out on the original object description information, and brief description information of the target object is obtained; generating an object characteristic description phrase for describing the characteristics of the target object according to the original object description information; and generating caption information corresponding to the target object in the video according to the brief description information and the object characteristic description phrase.
Computer program code for carrying out operations of the present disclosure may be written in one or more programming languages, including, but not limited to, an object oriented programming language such as Java, smalltalk, C ++ and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the case of a remote computer, the remote computer may be connected to the user's computer through any kind of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or may be connected to an external computer (for example, through the Internet using an Internet service provider).
The flowcharts and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present disclosure. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The modules described in the embodiments of the present disclosure may be implemented in software or hardware. The name of a module is not limited to the module itself in some cases, and for example, the acquisition module may also be described as "a module that acquires original object description information of a target object to be introduced in a video".
The functions described above herein may be performed, at least in part, by one or more hardware logic components. For example, without limitation, exemplary types of hardware logic components that may be used include: a Field Programmable Gate Array (FPGA), an Application Specific Integrated Circuit (ASIC), an Application Specific Standard Product (ASSP), a system on a chip (SOC), a Complex Programmable Logic Device (CPLD), and the like.
In the context of this disclosure, a machine-readable medium may be a tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. The machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium. The machine-readable medium may include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples of a machine-readable storage medium would include an electrical connection based on one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
According to one or more embodiments of the present disclosure, example 1 provides a subtitle generating method, including: acquiring original object description information of a target object to be introduced in a video, wherein the original object description information is used for describing original inherent characteristics of the target object; content refining is carried out on the original object description information, and brief description information of the target object is obtained; generating an object characteristic description phrase for describing the characteristics of the target object according to the original object description information; and generating caption information corresponding to the target object in the video according to the brief description information and the object characteristic description phrase.
Example 2 provides the method of example 1, the object property description phrase being a plurality of according to one or more embodiments of the present disclosure; and generating subtitle information corresponding to the target object in the video according to the brief description information and the object characteristic description phrase, wherein the subtitle information comprises the following components: combining the brief description information with each object characteristic description phrase respectively to obtain a plurality of candidate subtitle information; and determining target subtitle information from the plurality of candidate subtitle information as subtitle information corresponding to the target object in the video.
According to one or more embodiments of the present disclosure, example 3 provides the method of example 2, wherein generating subtitle information corresponding to the target object in the video according to the brief description information and the object property description phrase further includes: extracting, for each of the candidate subtitle information, at least one object property description word for describing a property of the target object from the candidate subtitle information; if the object characteristic description words which do not appear in the original object description information exist, filtering the candidate subtitle information; the determining target subtitle information from the plurality of candidate subtitle information includes: the target subtitle information is determined from the candidate subtitle information remaining after the filtering operation.
In accordance with one or more embodiments of the present disclosure, example 4 provides the method of example 3, the determining the target subtitle information from the candidate subtitle information remaining after the filtering operation, comprising: for each piece of candidate caption information left after filtering operation, determining the similarity between the candidate caption information and the original object description information; and determining the candidate caption information with the highest similarity with the original object description information as the target caption information.
According to one or more embodiments of the present disclosure, example 5 provides the method of example 3, the extracting at least one object property description word for describing a property of the target object from the candidate subtitle information, including: and carrying out named entity recognition on the candidate subtitle information to obtain at least one named entity, and taking the at least one named entity as at least one object characteristic description word for describing the characteristics of the target object.
In accordance with one or more embodiments of the present disclosure, example 6 provides the method of example 2, the determining target subtitle information from the plurality of candidate subtitle information, comprising: for each piece of candidate subtitle information, determining the similarity between the candidate subtitle information and the original object description information; and determining the candidate caption information with the highest similarity with the original object description information as the target caption information.
According to one or more embodiments of the present disclosure, example 7 provides the method of example 4 or 6, the determining a similarity between the candidate subtitle information and the original object description information, comprising: determining semantic information corresponding to the candidate subtitle information and semantic information corresponding to the original object description information; and calculating the similarity between the semantic information corresponding to the candidate caption information and the semantic information corresponding to the original object description information, and taking the similarity as the similarity between the candidate caption information and the original object description information.
In accordance with one or more embodiments of the present disclosure, example 8 provides the method of example 1, the content refining of the original object description information to obtain the brief description information of the target object, including: inputting the original object description information into a pre-trained content extraction model to extract the content of the original object description information so as to obtain the brief description information of the target object; wherein the content refinement model is trained by: acquiring first reference object description information of a first reference object, wherein the first reference object description information is used for describing original inherent characteristics of the first reference object; word segmentation is carried out on the first reference object description information to obtain a first word sequence; deleting each word except the first reference object in the first word sequence to obtain a new first word sequence, and sequentially connecting each word in the new first word sequence to obtain candidate brief description information of the first reference object; judging whether other words except the first reference object exist in any new first word sequence; if any word except the first reference object exists in any new first word sequence, repeating the step of deleting the word from the first word sequence to the step of judging whether any word except the first reference object exists in any new first word sequence by taking the new first word sequence as the first word sequence; if no other words except the first reference object exist in any new first word sequence, the candidate brief description information with the highest fluency is used as reference brief description information; model training is performed by taking the first reference object description information as an input of the content extraction model and taking the reference brief description information as a target output of the content extraction model so as to obtain the content extraction model.
According to one or more embodiments of the present disclosure, example 9 provides the method of example 1 or 8, the generating an object property description phrase for describing a property of the target object from the original object description information, comprising: inputting the original object description information into a pre-trained phrase generation model to obtain an object characteristic description phrase for describing the characteristics of the target object; the phrase generation model is trained by the following modes: acquiring second reference object description information and third reference object description information of a second reference object, wherein the second reference object description information and the third reference object description information are used for describing original inherent characteristics of the second reference object, and the similarity between the second reference object description information and the third reference object description information is larger than a preset similarity threshold; performing word segmentation on the second reference object description information to obtain a second word sequence, and performing word segmentation on the third reference object description information to obtain a third word sequence; extracting a phrase containing each modifier from the second reference object description information as a first candidate object characteristic description phrase aiming at each modifier in the second word sequence for describing the second reference object; extracting a phrase containing each modifier from the third reference object description information as a second candidate object characteristic description phrase aiming at each modifier in the third word sequence for describing the second reference object; determining each of the first candidate object property description phrases and each of the second candidate object property description phrases as reference object property description phrases; and performing model training in a mode that the second reference object description information is used as the input of the phrase generation model and the reference object characteristic description phrase is used as the target output of the phrase generation model so as to obtain the phrase generation model.
According to one or more embodiments of the present disclosure, example 10 provides a subtitle generating apparatus, including: the acquisition module is used for acquiring original object description information of a target object to be introduced in the video, wherein the original object description information is used for describing original inherent characteristics of the target object; the extraction module is used for extracting the content of the original object description information acquired by the acquisition module to acquire brief description information of the target object; the phrase generating module is used for generating an object characteristic description phrase for describing the characteristics of the target object according to the original object description information acquired by the acquiring module; and the caption generating module is used for generating caption information corresponding to the target object in the video according to the brief description information obtained by the extracting module and the object characteristic description phrase generated by the phrase generating module.
According to one or more embodiments of the present disclosure, example 11 provides a computer-readable medium having stored thereon a computer program which, when executed by a processing device, implements the steps of the method of any of examples 1-9.
Example 12 provides an electronic device according to one or more embodiments of the present disclosure, comprising: a storage device having a computer program stored thereon; processing means for executing the computer program in the storage means to implement the steps of the method of any one of examples 1-9.
The foregoing description is only of the preferred embodiments of the present disclosure and description of the principles of the technology being employed. It will be appreciated by persons skilled in the art that the scope of the disclosure referred to in this disclosure is not limited to the specific combinations of features described above, but also covers other embodiments which may be formed by any combination of features described above or equivalents thereof without departing from the spirit of the disclosure. Such as those described above, are mutually substituted with the technical features having similar functions disclosed in the present disclosure (but not limited thereto).
Moreover, although operations are depicted in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order. In certain circumstances, multitasking and parallel processing may be advantageous. Likewise, while several specific implementation details are included in the above discussion, these should not be construed as limiting the scope of the present disclosure. Certain features that are described in the context of separate embodiments can also be implemented in combination in a single embodiment. Conversely, various features that are described in the context of a single embodiment can also be implemented in multiple embodiments separately or in any suitable subcombination.
Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are example forms of implementing the claims. The specific manner in which the various modules perform the operations in the apparatus of the above embodiments have been described in detail in connection with the embodiments of the method, and will not be described in detail herein.
Claims (11)
1. A subtitle generating method, comprising:
acquiring original object description information of a target object to be introduced in a video, wherein the original object description information is used for describing original inherent characteristics of the target object;
inputting the original object description information into a pre-trained content extraction model to extract the content of the original object description information so as to obtain the brief description information of the target object;
Generating an object characteristic description phrase for describing the characteristics of the target object according to the original object description information;
generating caption information corresponding to the target object in the video according to the brief description information and the object characteristic description phrase;
Wherein the content refinement model is trained by:
Acquiring first reference object description information of a first reference object, wherein the first reference object description information is used for describing original inherent characteristics of the first reference object;
Word segmentation is carried out on the first reference object description information to obtain a first word sequence;
Deleting each word except the first reference object in the first word sequence to obtain a new first word sequence, and sequentially connecting each word in the new first word sequence to obtain candidate brief description information of the first reference object;
judging whether other words except the first reference object exist in any new first word sequence;
If any word except the first reference object exists in any new first word sequence, repeating the step of deleting the word from the first word sequence to the step of judging whether any word except the first reference object exists in any new first word sequence by taking the new first word sequence as the first word sequence;
If no other words except the first reference object exist in any new first word sequence, the candidate brief description information with the highest fluency is used as reference brief description information;
Model training is performed by taking the first reference object description information as an input of the content extraction model and taking the reference brief description information as a target output of the content extraction model so as to obtain the content extraction model.
2. The method of claim 1, wherein the object property description phrase is a plurality of;
and generating subtitle information corresponding to the target object in the video according to the brief description information and the object characteristic description phrase, wherein the subtitle information comprises the following components:
combining the brief description information with each object characteristic description phrase respectively to obtain a plurality of candidate subtitle information;
And determining target subtitle information from the plurality of candidate subtitle information as subtitle information corresponding to the target object in the video.
3. The method of claim 2, wherein generating subtitle information corresponding to the target object in the video from the brief description information and the object property description phrase further comprises:
Extracting, for each of the candidate subtitle information, at least one object property description word for describing a property of the target object from the candidate subtitle information; if the object characteristic description words which do not appear in the original object description information exist, filtering the candidate subtitle information;
the determining target subtitle information from the plurality of candidate subtitle information includes:
the target subtitle information is determined from the candidate subtitle information remaining after the filtering operation.
4. The method of claim 3, wherein the determining the target subtitle information from the candidate subtitle information remaining after the filtering operation comprises:
for each piece of candidate caption information left after filtering operation, determining the similarity between the candidate caption information and the original object description information;
and determining the candidate caption information with the highest similarity with the original object description information as the target caption information.
5. The method of claim 3, wherein the extracting at least one object property description word for describing a property of the target object from the candidate subtitle information comprises:
And carrying out named entity recognition on the candidate subtitle information to obtain at least one named entity, and taking the at least one named entity as at least one object characteristic description word for describing the characteristics of the target object.
6. The method of claim 2, wherein the determining target caption information from the plurality of candidate caption information comprises:
for each piece of candidate subtitle information, determining the similarity between the candidate subtitle information and the original object description information;
and determining the candidate caption information with the highest similarity with the original object description information as the target caption information.
7. The method according to claim 4 or 6, wherein the determining of the similarity between the candidate subtitle information and the original object description information includes:
determining semantic information corresponding to the candidate subtitle information and semantic information corresponding to the original object description information;
And calculating the similarity between the semantic information corresponding to the candidate caption information and the semantic information corresponding to the original object description information, and taking the similarity as the similarity between the candidate caption information and the original object description information.
8. The method of claim 1, wherein generating an object property description phrase for describing properties of the target object from the original object description information comprises:
inputting the original object description information into a pre-trained phrase generation model to obtain an object characteristic description phrase for describing the characteristics of the target object;
the phrase generation model is trained by the following modes:
Acquiring second reference object description information and third reference object description information of a second reference object, wherein the second reference object description information and the third reference object description information are used for describing original inherent characteristics of the second reference object, and the similarity between the second reference object description information and the third reference object description information is larger than a preset similarity threshold;
Performing word segmentation on the second reference object description information to obtain a second word sequence, and performing word segmentation on the third reference object description information to obtain a third word sequence;
Extracting a phrase containing each modifier from the second reference object description information as a first candidate object characteristic description phrase aiming at each modifier in the second word sequence for describing the second reference object; extracting a phrase containing each modifier from the third reference object description information as a second candidate object characteristic description phrase aiming at each modifier in the third word sequence for describing the second reference object;
Determining each of the first candidate object property description phrases and each of the second candidate object property description phrases as reference object property description phrases;
And performing model training in a mode that the second reference object description information is used as the input of the phrase generation model and the reference object characteristic description phrase is used as the target output of the phrase generation model so as to obtain the phrase generation model.
9. A subtitle generating apparatus, comprising:
The acquisition module is used for acquiring original object description information of a target object to be introduced in the video, wherein the original object description information is used for describing original inherent characteristics of the target object;
The extraction module is used for inputting the original object description information into a pre-trained content extraction model so as to extract the content of the original object description information acquired by the acquisition module and obtain the brief description information of the target object;
The phrase generating module is used for generating an object characteristic description phrase for describing the characteristics of the target object according to the original object description information acquired by the acquiring module;
The subtitle generating module is used for generating subtitle information corresponding to the target object in the video according to the brief description information obtained by the extracting module and the object characteristic description phrase generated by the phrase generating module;
Wherein the content refinement model is trained by a first model training device comprising:
the first descriptive information acquisition module is used for acquiring first reference object descriptive information of a first reference object, wherein the first reference object descriptive information is used for describing original inherent characteristics of the first reference object;
the first word segmentation module is used for segmenting the first reference object description information to obtain a first word sequence;
The first determining module is used for deleting each word except the first reference object in the first word sequence to obtain a new first word sequence, and sequentially connecting the words in the new first word sequence to obtain candidate brief description information of the first reference object;
The judging module is used for judging whether other words except the first reference object exist in any new first word sequence;
The triggering module is used for triggering the first determining module to delete each word except for the first reference object in the first word sequence to obtain a new first word sequence, and sequentially connecting the words in the new first word sequence to obtain candidate brief description information for each word except for the first reference object in the first word sequence;
The second determining module is configured to use the candidate brief description information with the highest fluency as reference brief description information if no other word except the first reference object exists in any new first word sequence;
And the first model training module is used for carrying out model training in a mode that the first reference object description information is used as the input of the content extraction model and the reference brief description information is used as the target output of the content extraction model so as to obtain the content extraction model.
10. A computer readable medium on which a computer program is stored, characterized in that the program, when being executed by a processing device, carries out the steps of the method according to any one of claims 1-8.
11. An electronic device, comprising:
A storage device having a computer program stored thereon;
Processing means for executing said computer program in said storage means to carry out the steps of the method according to any one of claims 1-8.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110420704.4A CN113033190B (en) | 2021-04-19 | 2021-04-19 | Subtitle generation method, subtitle generation device, medium and electronic equipment |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110420704.4A CN113033190B (en) | 2021-04-19 | 2021-04-19 | Subtitle generation method, subtitle generation device, medium and electronic equipment |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113033190A CN113033190A (en) | 2021-06-25 |
CN113033190B true CN113033190B (en) | 2024-05-17 |
Family
ID=76456858
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110420704.4A Active CN113033190B (en) | 2021-04-19 | 2021-04-19 | Subtitle generation method, subtitle generation device, medium and electronic equipment |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113033190B (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114222193B (en) * | 2021-12-03 | 2024-01-05 | 北京影谱科技股份有限公司 | Video subtitle time alignment model training method and system |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107577663A (en) * | 2017-08-24 | 2018-01-12 | 北京奇艺世纪科技有限公司 | A kind of key-phrase extraction method and apparatus |
CN110765244A (en) * | 2019-09-18 | 2020-02-07 | 平安科技(深圳)有限公司 | Method and device for acquiring answering, computer equipment and storage medium |
CN110928981A (en) * | 2019-11-18 | 2020-03-27 | 佰聆数据股份有限公司 | Method, system and storage medium for establishing and perfecting iteration of text label system |
CN111723566A (en) * | 2019-03-21 | 2020-09-29 | 阿里巴巴集团控股有限公司 | Method and device for reconstructing product information |
CN111859940A (en) * | 2019-04-23 | 2020-10-30 | 北京嘀嘀无限科技发展有限公司 | Keyword extraction method and device, electronic equipment and storage medium |
CN111859930A (en) * | 2020-07-27 | 2020-10-30 | 北京字节跳动网络技术有限公司 | Title generation method and device, electronic equipment and storage medium |
CN112446208A (en) * | 2020-12-09 | 2021-03-05 | 北京有竹居网络技术有限公司 | Method, device and equipment for generating advertisement title and storage medium |
-
2021
- 2021-04-19 CN CN202110420704.4A patent/CN113033190B/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107577663A (en) * | 2017-08-24 | 2018-01-12 | 北京奇艺世纪科技有限公司 | A kind of key-phrase extraction method and apparatus |
CN111723566A (en) * | 2019-03-21 | 2020-09-29 | 阿里巴巴集团控股有限公司 | Method and device for reconstructing product information |
CN111859940A (en) * | 2019-04-23 | 2020-10-30 | 北京嘀嘀无限科技发展有限公司 | Keyword extraction method and device, electronic equipment and storage medium |
CN110765244A (en) * | 2019-09-18 | 2020-02-07 | 平安科技(深圳)有限公司 | Method and device for acquiring answering, computer equipment and storage medium |
CN110928981A (en) * | 2019-11-18 | 2020-03-27 | 佰聆数据股份有限公司 | Method, system and storage medium for establishing and perfecting iteration of text label system |
CN111859930A (en) * | 2020-07-27 | 2020-10-30 | 北京字节跳动网络技术有限公司 | Title generation method and device, electronic equipment and storage medium |
CN112446208A (en) * | 2020-12-09 | 2021-03-05 | 北京有竹居网络技术有限公司 | Method, device and equipment for generating advertisement title and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN113033190A (en) | 2021-06-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109886326B (en) | Cross-modal information retrieval method and device and storage medium | |
CN111368185B (en) | Data display method and device, storage medium and electronic equipment | |
CN107527619B (en) | Method and device for positioning voice control service | |
CN112037792B (en) | Voice recognition method and device, electronic equipment and storage medium | |
CN109948073B (en) | Content retrieval method, terminal, server, electronic device, and storage medium | |
CN109036397B (en) | Method and apparatus for presenting content | |
WO2020244487A1 (en) | Easter egg presentation method and apparatus, electronic device, and computer readable storage medium | |
JP7140913B2 (en) | Video distribution statute of limitations determination method and device | |
CN111178056A (en) | Deep learning based file generation method and device and electronic equipment | |
CN109815448B (en) | Slide generation method and device | |
US20230112005A1 (en) | Image special effect configuration method, image recognition method, apparatus and electronic device | |
CN113033190B (en) | Subtitle generation method, subtitle generation device, medium and electronic equipment | |
CN112954453B (en) | Video dubbing method and device, storage medium and electronic equipment | |
JP2024506495A (en) | Methods, devices, equipment and media for processing minutes | |
CN110232181B (en) | Comment analysis method and device | |
CN113971400B (en) | Text detection method and device, electronic equipment and storage medium | |
CN113014853B (en) | Interactive information processing method and device, electronic equipment and storage medium | |
CN110377842A (en) | Voice remark display methods, system, medium and electronic equipment | |
CN116186545A (en) | Training and application methods and devices of pre-training model, electronic equipment and medium | |
CN111737572B (en) | Search statement generation method and device and electronic equipment | |
CN111782895B (en) | Retrieval processing method and device, readable medium and electronic equipment | |
US9484033B2 (en) | Processing and cross reference of realtime natural language dialog for live annotations | |
CN112651221A (en) | Data processing method and device and data processing device | |
CN115050349B (en) | Method, apparatus, device and medium for text-to-audio conversion | |
CN114997120B (en) | Method, device, terminal and storage medium for generating document tag |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |