CN113688283B - Method and device for determining video subtitle matching degree and electronic equipment - Google Patents

Method and device for determining video subtitle matching degree and electronic equipment Download PDF

Info

Publication number
CN113688283B
CN113688283B CN202110997692.1A CN202110997692A CN113688283B CN 113688283 B CN113688283 B CN 113688283B CN 202110997692 A CN202110997692 A CN 202110997692A CN 113688283 B CN113688283 B CN 113688283B
Authority
CN
China
Prior art keywords
foreign language
video
chinese
speech
language speech
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110997692.1A
Other languages
Chinese (zh)
Other versions
CN113688283A (en
Inventor
牟晋勇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing QIYI Century Science and Technology Co Ltd
Original Assignee
Beijing QIYI Century Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing QIYI Century Science and Technology Co Ltd filed Critical Beijing QIYI Century Science and Technology Co Ltd
Priority to CN202110997692.1A priority Critical patent/CN113688283B/en
Publication of CN113688283A publication Critical patent/CN113688283A/en
Application granted granted Critical
Publication of CN113688283B publication Critical patent/CN113688283B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/783Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/7844Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using original textual content or text extracted from visual content or transcript of audio data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/63Querying
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/68Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/683Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/685Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using automatically derived transcript of audio data, e.g. lyrics
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/73Querying
    • G06F16/735Filtering based on additional data, e.g. user or group profiles
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/40Processing or translation of natural language
    • G06F40/42Data-driven translation
    • G06F40/44Statistical methods, e.g. probability models

Abstract

The embodiment of the invention provides a method and a device for determining the matching degree of video subtitles and electronic equipment, wherein the method comprises the following steps: obtaining a foreign language speech of a target foreign language video and a corresponding Chinese caption; inquiring phonetic symbols corresponding to words included in foreign language speech from a preset word phonetic symbol library, and determining pronunciation change times corresponding to each word based on the phonetic symbols; and determining the matching degree of the foreign language speech of the target foreign language video and the Chinese caption according to the pronunciation change times and the word number of the Chinese caption. By adopting the method, the foreign language video with high user preference degree, namely the foreign language speech and the Chinese subtitle, can be recommended to the user according to the determined matching degree of the foreign language speech of the target foreign language video and the Chinese subtitle, and the recommending effect of recommending the foreign language video to the user is improved.

Description

Method and device for determining video subtitle matching degree and electronic equipment
Technical Field
The present invention relates to the field of computer technologies, and in particular, to a method and an apparatus for determining a video subtitle matching degree, and an electronic device.
Background
In daily movie watching life, a user can watch a large amount of videos such as foreign language movies, television shows, variety and the like, such as English videos such as hollywood movies, disney cartoons and the like. In general, characters in a foreign language video are expressed in a foreign language, and simultaneously, a corresponding translated Chinese subtitle is displayed below the video, so that a viewer can know the meaning of the video expression.
However, in the actual playing process of the foreign language video, the user can feel that the character expression in the foreign language video is unnatural due to the difference between the foreign language pronunciation mouth shape change of the character in the foreign language video and the translated Chinese subtitle, so that the watching experience of the user is affected, and the preference degree of the user on the foreign language video is further reduced.
Therefore, the degree of correspondence between the mouth shape change of the character expression foreign language words in the foreign language video and the Chinese subtitle can influence the preference degree of the user on the foreign language video, and the higher the degree of correspondence between the mouth shape change and the Chinese subtitle is, the higher the preference degree of the user on the foreign language video is. Based on the above, if the video software recommends the foreign language video with the mouth shape of the character expressing the foreign language word and the very low degree of correspondence with the Chinese caption to the user, the use experience of the user is likely to be affected. Therefore, how to determine the correspondence between the mouth shape variation of the characters expressing the foreign language words in the foreign language video and the Chinese subtitle, so as to improve the recommendation effect of the foreign language video becomes a problem to be solved urgently.
Disclosure of Invention
The embodiment of the invention aims to provide a method and a device for determining the matching degree of video captions and electronic equipment, so as to determine the corresponding degree of the mouth shape change of characters expressing foreign language words in foreign language videos and Chinese captions.
In a first aspect of the present invention, there is provided a method for determining a matching degree of video subtitles, including:
obtaining a foreign language speech of a target foreign language video and a corresponding Chinese caption;
inquiring phonetic symbols corresponding to words included in the foreign language speech from a preset word phonetic symbol library, and determining pronunciation change times corresponding to each word based on the phonetic symbols;
and determining the matching degree of the foreign language speech of the target foreign language video and the Chinese caption according to the pronunciation change times and the word number of the Chinese caption.
Optionally, the determining, based on the phonetic symbols, the number of pronunciation changes corresponding to each word includes:
determining the number of vowels in the phonetic symbol corresponding to each word;
and determining the number of the vowels as the pronunciation change times corresponding to the word.
Optionally, the determining, according to the number of pronunciation changes and the number of words of the chinese subtitle, the matching degree of the foreign language speech of the target foreign language video and the chinese subtitle includes:
Aiming at each foreign language speech of the target foreign language video, determining the sum of pronunciation change times corresponding to words included in the foreign language speech as the pronunciation change times of the foreign language speech;
determining the word number of Chinese subtitles corresponding to each foreign language line of the target foreign language video;
traversing each foreign language speech of the target foreign language video, and determining the difference rate between the foreign language speech and the corresponding Chinese subtitle based on the pronunciation change times of the foreign language speech and the number of Chinese subtitle characters corresponding to the foreign language speech;
and determining the matching degree of the foreign language speech lines of the target foreign language video and the Chinese subtitles according to the difference rate between each sentence of foreign language speech lines of the target foreign language video and the corresponding Chinese subtitles.
Optionally, the determining the difference rate between the foreign language line and the corresponding chinese subtitle based on the pronunciation change times of the foreign language line and the number of words of the chinese subtitle corresponding to the foreign language line includes:
if the pronunciation change times of the foreign language speech is larger than the number of Chinese characters of the Chinese characters corresponding to the foreign language speech, determining the ratio of the number of the Chinese characters corresponding to the foreign language speech to the pronunciation change times of the foreign language speech as the difference rate between the foreign language speech and the corresponding Chinese characters;
If the pronunciation change times of the foreign language speech is less than or equal to the number of Chinese characters of the Chinese characters corresponding to the foreign language speech, determining the ratio of the pronunciation change times of the foreign language speech to the number of Chinese characters of the Chinese characters corresponding to the foreign language speech as the difference rate between the foreign language speech and the corresponding Chinese characters.
Optionally, the determining, according to the difference rate between each foreign language speech of the target foreign language video and the corresponding chinese subtitle, the matching degree of the foreign language speech of the target foreign language video and the chinese subtitle includes:
calculating an average value of the difference rates between the foreign language speech of the target foreign language video and the corresponding Chinese subtitles, and taking the average value as an average difference rate;
calculating a standard deviation of the difference rate between the foreign language speech of the target foreign language video and the corresponding Chinese caption based on the average difference rate and the difference rate between each foreign language speech of the target foreign language video and the corresponding Chinese caption;
if the average difference rate is greater than or equal to a preset difference rate threshold value and the standard deviation is less than or equal to a preset fluctuation threshold value, determining that the matching degree of the foreign language speech lines of the target foreign language video and the Chinese subtitles is a first matching degree, otherwise, determining that the matching degree of the foreign language speech lines of the target foreign language video and the Chinese subtitles is a second matching degree, wherein the first matching degree is greater than the second matching degree.
Optionally, after the determining the matching degree of the foreign language speech line of the target foreign language video and the chinese subtitle, the method further includes:
if the target foreign language video is the video to be recommended, recommending the target foreign language video to a user based on the matching degree of the foreign language speech of the target foreign language video and the Chinese subtitle.
In a second aspect of the present invention, there is also provided a device for determining a matching degree of video subtitles, including:
the speech acquisition module is used for acquiring the foreign speech of the target foreign language video and the corresponding Chinese captions;
the change frequency determining module is used for inquiring phonetic symbols corresponding to words included in the foreign language speech from a preset word phonetic symbol library and determining the pronunciation change frequency corresponding to each word based on the phonetic symbols;
and the matching degree determining module is used for determining the matching degree of the foreign language speech of the target foreign language video and the Chinese subtitle according to the pronunciation change times and the word number of the Chinese subtitle.
Optionally, the change number determining module is specifically configured to determine the number of vowels in a phonetic symbol corresponding to each word; and determining the number of the vowels as the pronunciation change times corresponding to the word.
Optionally, the matching degree determining module includes:
the frequency determining submodule is used for determining the sum of pronunciation change times corresponding to words included in each foreign language speech word as the pronunciation change times of the foreign language speech word aiming at each foreign language speech word of the target foreign language video;
the word number determining submodule is used for determining the word number of the Chinese subtitle corresponding to each foreign language line of the target foreign language video;
the difference rate determination submodule is used for traversing each foreign language speech of the target foreign language video and determining the difference rate between the foreign language speech and the corresponding Chinese subtitle based on the pronunciation change times of the foreign language speech and the number of Chinese subtitle words corresponding to the foreign language speech;
and the matching degree determining sub-module is used for determining the matching degree of the foreign language speech lines of the target foreign language video and the Chinese subtitles according to the difference rate between each sentence of the foreign language speech lines of the target foreign language video and the corresponding Chinese subtitles.
Optionally, the difference rate determining submodule is specifically configured to determine, as a difference rate between the foreign language speech word and the corresponding chinese subtitle, a ratio of a number of words of the chinese subtitle corresponding to the foreign language speech word to a number of pronunciation changes of the foreign language speech word if the number of pronunciation changes of the foreign language speech word is greater than the number of words of the chinese subtitle corresponding to the foreign language speech word; if the pronunciation change times of the foreign language speech is less than or equal to the number of Chinese characters of the Chinese characters corresponding to the foreign language speech, determining the ratio of the pronunciation change times of the foreign language speech to the number of Chinese characters of the Chinese characters corresponding to the foreign language speech as the difference rate between the foreign language speech and the corresponding Chinese characters.
Optionally, the matching degree determining submodule is specifically configured to calculate an average value of difference rates between the foreign language speech of the target foreign language video and the corresponding chinese subtitle, as an average difference rate; calculating a standard deviation of the difference rate between the foreign language speech of the target foreign language video and the corresponding Chinese caption based on the average difference rate and the difference rate between each foreign language speech of the target foreign language video and the corresponding Chinese caption; if the average difference rate is greater than or equal to a preset difference rate threshold value and the standard deviation is less than or equal to a preset fluctuation threshold value, determining that the matching degree of the foreign language speech lines of the target foreign language video and the Chinese subtitles is a first matching degree, otherwise, determining that the matching degree of the foreign language speech lines of the target foreign language video and the Chinese subtitles is a second matching degree, wherein the first matching degree is greater than the second matching degree.
Optionally, the device further includes a video recommendation module, configured to recommend the target foreign language video to a user based on a matching degree of a foreign language speech of the target foreign language video and a chinese subtitle if the target foreign language video is a video to be recommended.
In yet another aspect of the present invention, there is also provided an electronic device including a processor, a communication interface, a memory, and a communication bus, wherein the processor, the communication interface, and the memory perform communication with each other through the communication bus;
A memory for storing a computer program;
and the processor is used for realizing any of the steps of the method for determining the video subtitle matching degree when executing the program stored in the memory.
In still another aspect of the implementation of the present invention, there is further provided a computer readable storage medium, where a computer program is stored, where the computer program is executed by a processor to implement the method for determining a matching degree of video subtitles according to any of the above.
In yet another aspect of the implementation of the present invention, there is also provided a computer program product containing instructions that, when executed on a computer, cause the computer to perform the method for determining the matching degree of video subtitles as described in any of the above.
By adopting the method provided by the embodiment of the invention, the foreign language speech of the target foreign language video and the corresponding Chinese captions can be obtained; inquiring phonetic symbols corresponding to words included in foreign language speech from a preset word phonetic symbol library, and determining pronunciation change times corresponding to each word based on the phonetic symbols; and determining the matching degree of the foreign language speech of the target foreign language video and the Chinese caption according to the pronunciation change times and the word number of the Chinese caption. Furthermore, according to the determined matching degree of the foreign language speech of the target foreign language video and the Chinese subtitle, the foreign language video with high user preference degree, namely the foreign language speech with high matching degree of the foreign language speech and the Chinese subtitle, can be recommended to the user, and the recommending effect of the user on the foreign language video is improved.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below.
Fig. 1 is a flowchart of a method for determining a matching degree of a video subtitle according to an embodiment of the present invention;
FIG. 2 is a flowchart for determining matching degree of a foreign language speech line and a Chinese caption of a target foreign language video according to an embodiment of the present invention;
fig. 3 is a schematic structural diagram of a device for determining matching degree of video subtitles according to an embodiment of the present invention;
fig. 4 is a schematic diagram of another structure of a device for determining matching degree of video subtitles according to an embodiment of the present invention;
fig. 5 is a schematic structural diagram of a device for determining matching degree of video subtitles according to an embodiment of the present invention;
fig. 6 is a schematic structural diagram of an electronic device according to an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be described below with reference to the accompanying drawings in the embodiments of the present invention.
The method for determining the matching degree of the video subtitles is provided in the embodiment of the invention, so as to determine the corresponding degree of the mouth shape change of the character expression foreign language word in the foreign language video and the Chinese subtitles, and further improve the recommending effect of the foreign language video.
Fig. 1 is a flowchart of a method for determining a matching degree of a video subtitle according to an embodiment of the present invention, where, as shown in fig. 1, the method includes:
step 101, obtaining foreign language speech of the target foreign language video and corresponding Chinese captions.
Step 102, inquiring phonetic symbols corresponding to words included in the foreign language speech from a preset word phonetic symbol library, and determining the pronunciation change times corresponding to each word based on the phonetic symbols.
And step 103, determining the matching degree of the foreign language speech of the target foreign language video and the Chinese caption according to the pronunciation change times and the word number of the Chinese caption.
By adopting the method provided by the embodiment of the invention, the foreign language speech of the target foreign language video and the corresponding Chinese captions can be obtained; inquiring phonetic symbols corresponding to words included in foreign language speech from a preset word phonetic symbol library, and determining pronunciation change times corresponding to each word based on the phonetic symbols; and determining the matching degree of the foreign language speech of the target foreign language video and the Chinese caption according to the pronunciation change times and the word number of the Chinese caption. Furthermore, according to the determined matching degree of the foreign language speech of the target foreign language video and the Chinese subtitle, the foreign language video with high user preference degree, namely the foreign language speech with high matching degree of the foreign language speech and the Chinese subtitle, can be recommended to the user, and the recommending effect of the user on the foreign language video is improved.
In the embodiment of the invention, the foreign language video can be English video, french video or German video, and the like. Specifically, the foreign language video may be a foreign language movie video, a foreign language variety video, a foreign language news video, and the like. The target foreign language video may be a complete foreign language movie video, a complete foreign language variety video or a complete foreign language news video, and the target foreign language video may also be a partial segment of a complete foreign language movie video, a partial segment of a complete foreign language variety video or a partial segment of a complete foreign language news video, which is not specifically limited herein.
In the embodiment of the invention, the foreign language video can be stored in the database of the server. The server side can directly determine the matching degree of the foreign language lines and the Chinese subtitles aiming at the target foreign language video needing to determine the matching degree of the video subtitles. In the embodiment of the invention, the matching degree of the foreign language speech and the Chinese caption can be determined according to the target foreign language video played at the client: when the matching degree of the video subtitles needs to be determined, the client can send the foreign language speech of the target foreign language video and the corresponding Chinese subtitles to the server, and then the server determines the matching degree of the foreign language speech and the Chinese subtitles by adopting the method provided by the embodiment of the invention.
In the embodiment of the invention, a corresponding relation library of the foreign language words and the phonetic symbols thereof, namely a preset word phonetic symbol library, can be maintained at the server side. Each foreign language word and the phonetic symbol thereof are stored in a preset word phonetic symbol library, and each foreign language word and the phonetic symbol thereof in the preset word phonetic symbol library can be correspondingly stored. For example, a preset word phonetic symbol library storing each english word and its phonetic symbol may be maintained at the server, and then the phonetic symbol of each english word may be found in the preset word phonetic symbol library, for example, the english phonetic symbol of the english word banana may be foundAnd American phonetic symbol->
In one possible implementation manner, the determining the number of pronunciation change times corresponding to each word based on the phonetic symbols may specifically include steps A1-A2:
step A1: the number of vowels in the phonetic symbol corresponding to each word is determined.
Step A2: the number of vowels is determined as the number of pronunciation changes corresponding to the word.
If the target foreign language video is an English video, the server can maintain the English word and the phonetic symbol thereof in a preset English word phonetic symbol library. The preset english word phonetic symbol library stores vowels, consonants and the like in the english phonetic symbols, for example, the vocabulary of the stored vowels may include the following single vowels and double vowels:
Single vowels: [ i:]、[i]、[u:]、[u]、/>[ɑ:]、[Λ]、[e]、/>
diphthong: [ ei ]]、[ai]、[au]、/>
In this embodiment, for each word of the target foreign language video, a phonetic symbol of the target foreign language video is searched in the preset phonetic symbol library of english words, the number of vowels in the phonetic symbol is determined, and the number of vowels in the phonetic symbol is determined as the corresponding pronunciation change times. If the target foreign language video includes the english word banana, the phonetic symbols of the banana may be searched in the preset english word phonetic symbol library: english phonetic symbolAnd American phonetic symbol->The english phonetic symbols +.>The vowels of (a) include->[ɑ:]And->American phonetic symbol of banana->The vowels of (a) includeAnd->Therefore, the number of vowels in the phonetic symbol of the banana is 3, either in the American phonetic symbol or in the English phonetic symbol, and the number of pronunciation changes corresponding to the English word banana can be determined to be 3. The number of pronunciation changes corresponding to a word may reflect the number of times the character is changed in shape when reading the word, and the number of times the character is changed in shape when reading the word is consistent with the number of pronunciation changes corresponding to the word, e.g., the character is changed 3 times in shape when reading the word banana, and the number of times of change in shape is consistent with the number of pronunciation changes corresponding to banana.
In a possible implementation manner, fig. 2 is a flowchart of determining a matching degree of a foreign language speech line of a target foreign language video and a chinese subtitle according to an embodiment of the present invention, as shown in fig. 2, the step of determining the matching degree of the foreign language speech line of the target foreign language video and the chinese subtitle according to the number of pronunciation change times and the number of words of the chinese subtitle may specifically include:
Step 201, for each foreign language speech of the target foreign language video, determining the sum of pronunciation change times corresponding to the words included in the foreign language speech as the pronunciation change times of the foreign language speech.
In the embodiment of the invention, each foreign language speech of the target foreign language video corresponds to a Chinese caption, and each foreign language speech of the target foreign language video and the corresponding Chinese caption are stored with the identification of the target foreign language video. Each foreign language line and the corresponding Chinese caption have self marks which can be the foreign language lineNumbering of the order of the speech and its corresponding chinese captions in all sentence speech of the target video, e.g. the target foreign language video a is identified as V id The identification of the third foreign language speech in the target foreign language video A is the number E of the sequence of the third foreign language speech in all the speech in the target foreign language video A 3 And the identification of Chinese captions corresponding to the third foreign language speech is also the serial number C of the sequence of the Chinese captions in all the speech of the target foreign language video A 3 In addition, the third foreign language speech in the target foreign language video A and the Chinese subtitle corresponding to the third foreign language speech are stored with the identifier V of the target foreign language video A id
Each foreign language word may include at least one word. In this step, the sum of the number of pronunciation changes corresponding to all the words included in each foreign language speech line may be determined as the number of pronunciation changes of the foreign language speech line.
For example, the target foreign language video is an english video S, where the first sentence in the english video S is "a yellow kana", and the sentence includes a word: a. yellow and banana. The english phonetic symbols of word a can be searched in the preset word phonetic symbol libraryAnd American phonetic symbol->Further, the number of vowels in the phonetic symbol of the word a can be determined to be 1, namely the pronunciation change frequency of the word a is 1; english phonetic symbols of the word yellow can be found +.>And American phonetic symbol->Further, the number of vowels in the phonetic symbol of the word yellow can be determined to be 2, namely the pronunciation change times of the word yellow is 2; english phonetic symbols of the word banana can be found +.>And American phonetic symbol->Further, it is possible to determine that the number of vowels in the phonetic symbol of the word banana is 3, i.e., the number of pronunciation changes of the word yellow is 3. Therefore, the sum "1+2+3=6" of the pronunciation change times corresponding to all the words "a", "yellow" and "banna" in the first sentence in "a yellow banna" can be determined as the pronunciation change times of the first sentence in the english video S for "a yellow banna".
Step 202, determining the number of words of the Chinese subtitle corresponding to each foreign language line of the target foreign language video.
For example, if the first sentence in the english video S is "a yellow banana" and the corresponding chinese subtitle is "a yellow banana", the number of words of the first sentence in the english video S corresponding to "a yellow banana" may be determined to be 7.
Step 203, traversing each foreign language speech of the target foreign language video, and determining the difference rate between the foreign language speech and the corresponding Chinese subtitle based on the pronunciation change times of the foreign language speech and the number of Chinese subtitle words corresponding to the foreign language speech.
In the step, if the pronunciation change times of the foreign language speech is larger than the number of Chinese characters corresponding to the foreign language speech, determining the ratio of the number of Chinese characters corresponding to the foreign language speech to the pronunciation change times of the foreign language speech as the difference rate between the foreign language speech and the corresponding Chinese characters; if the pronunciation change times of the foreign language speech is less than or equal to the number of Chinese characters of the Chinese characters corresponding to the foreign language speech, determining the ratio of the pronunciation change times of the foreign language speech to the number of Chinese characters of the Chinese characters corresponding to the foreign language speech as the difference rate between the foreign language speech and the corresponding Chinese characters.
For example, if the english video S includes N english speech, the english speech set of all the sentences is S English :[E 1 ,E 2 ,…,E N ]. Can traverse English in each sentenceSpeech line E i I is more than or equal to 1 and less than or equal to N, and each English word E is processed i Word u of each Word in (a) X Searching corresponding phonetic symbols SoundMark/u from a preset word phonetic symbol library X Calculating the phonetic symbol SoundMark/u X Number of middle vowels Vcount/u x So each sentence of speech E of English video S can be obtained i The number of vowels corresponding to each word is set Vc: [ Vcount/u ] 1 ,Vcount_ 2 ,…,Vcount_ X ]. Furthermore, all English speech sets S in the English video S can be obtained English Vowel total number set SVc corresponding to each English speech: [ Vc ] _all_1 ,Vc _all_2 ,…,Vc _all_N ]I.e. all English speech sets S in the English video S English Pronunciation change times SVc corresponding to each English speech: [ Vc ] _all_1 ,Vc _all_2 ,…,Vc _all_N ]。
If the target foreign language video is English video S, the Chinese subtitle set corresponding to each English speech in English video S is S chinese :[C 1 ,C 2 ,…,C N ]Traversing each sentence of Chinese caption C i I is more than or equal to 1 and less than or equal to N, and each sentence of Chinese caption C i Number of Chinese characters Ccount _x Furthermore, a Chinese caption set S corresponding to each English speech in the English video S can be obtained Chinese A set Cc of the number of words of each sentence of chinese subtitles: [ Ccount _1 ,Ccount _2 ,…,Ccount _X ]. And each sentence of Chinese caption C i For English speech E i Corresponding to the above.
With the first English speech E in English video S 1 Chinese caption C corresponding to the same 1 For example, if the sentence is English speech E 1 The number of pronunciation variations Vc _all_1 English speech E larger than the sentence 1 Word number Ccount of corresponding Chinese caption _1 English speech E 1 Word number Ccount of corresponding Chinese caption _1 The pronunciation change times Vc of English speech _all_1 Is determined as the difference rate P between the foreign language speech line and the corresponding Chinese caption 1 =Ccount _1 /Vc _all_1 . If the processSentence English speech E 1 The number of pronunciation variations Vc _all_1 The word number Ccount of Chinese subtitles corresponding to English lines of the sentence is smaller than or equal to _1 The pronunciation change times Vc of English speech of the sentence are changed _all_1 Word number Ccount of Chinese caption corresponding to English speech of the sentence _1 Is determined as the difference rate P between the English speech of the sentence and the corresponding Chinese subtitle 1 =Vc _all_1 /Ccount _1
Traversing each sentence of the English video S, the difference rate between each sentence of the English video S and the corresponding Chinese subtitle can be obtained, and the difference rate between each sentence of the English video S and the corresponding Chinese subtitle can form a set SP: [ P ] 1 ,…,P N ]。P i The difference rate between the i-th sentence in the English video S and the corresponding Chinese subtitle is equal to or more than 1 and equal to or less than N.
Step 204, determining the matching degree of the foreign language speech lines of the target foreign language video and the Chinese subtitles according to the difference rate between each sentence of foreign language speech lines of the target foreign language video and the corresponding Chinese subtitles.
In this step, determining the matching degree of the foreign language speech line and the chinese subtitle may specifically include the following steps B1 to B3:
step B1: and calculating an average value of the difference rates between the foreign language speech of the target foreign language video and the corresponding Chinese subtitles as an average difference rate.
For example, if the target foreign language video is an english video S, and the difference rate set SP between each sentence in the english video S and the corresponding chinese subtitle: [ P ] 1 ,…,P N ]The average value P of the difference rates between all the lines of the English video S and the corresponding Chinese subtitles can be calculated avr =(P 1 +…+P N ) N, then P can be avr As the average difference rate.
Step B2: based on the average difference rate and the difference rate between each foreign language speech of the target foreign language video and the corresponding Chinese caption, calculating the standard deviation of the difference rate between the foreign language speech of the target foreign language video and the corresponding Chinese caption.
For example, if the target foreign language video is an english video S, and the difference rate set SP between each sentence in the english video S and the corresponding chinese subtitle: [ P ] 1 ,…,P N ]And average difference rate P of english video S avr =(P 1 +…+P N ) and/N, calculating the standard deviation sigma of the difference rate between all the lines of the English video S and the corresponding Chinese subtitles:
step B3: if the average difference rate is greater than or equal to a preset difference rate threshold value and the standard deviation is less than or equal to a preset fluctuation threshold value, determining that the matching degree of the foreign language speech of the target foreign language video and the Chinese subtitle is a first matching degree, otherwise, determining that the matching degree of the foreign language speech of the target foreign language video and the Chinese subtitle is a second matching degree.
Wherein the first degree of matching is greater than the second degree of matching.
In the embodiment of the invention, a difference rate threshold value P is preset valid Is smaller than 1, and can be adjusted on the premise of smaller than 1 according to practical application conditions, and is not particularly limited herein. The preset fluctuation threshold may be set to σ valid =P avr M, M may be an empirical reference value, and M may be appropriately adjusted according to practical application conditions, which is not particularly limited herein.
In the embodiment of the invention, the practical application meaning of the standard deviation reflection of the difference rate between the foreign language speech of the target foreign language video and the corresponding Chinese caption can be as follows: when the data distribution is relatively scattered (i.e., the data fluctuates more around the average difference rate), the sum of squares of differences between each data and the average difference rate is larger, and the standard deviation after the arithmetic square root is larger; when the data distribution is relatively concentrated, the sum of squares of differences between individual data and the average difference rate is small. Thus the larger the standard deviation σ, the larger the fluctuation of the data; the smaller the standard deviation σ, the smaller the fluctuation of the data.
In the embodiment of the invention, the average difference rate is larger than or equal to the preset difference rate threshold value, and the difference rate between the actual foreign language pronunciation mouth shape change of the character in the target foreign language video and the word number of the Chinese subtitle is smaller, namely the matching degree of the foreign language speech of the target foreign language video and the Chinese subtitle is higher, so that the method is acceptable to users; and if the standard variance is smaller than or equal to a preset fluctuation threshold, the standard variance indicates that the actual pronunciation mouth shape change of each sentence of the character in the target foreign language video and the quantity difference rate of Chinese subtitles are not greatly fluctuated, and the difference rate is relatively stable and continuous. Therefore, if the average difference rate is greater than or equal to the preset difference rate threshold and the standard deviation is less than or equal to the preset fluctuation threshold, determining that the matching degree of the foreign language speech of the target foreign language video and the Chinese subtitle is the first matching degree, wherein the matching degree of the foreign language speech of the target foreign language video and the Chinese subtitle is higher, otherwise, the matching degree of the foreign language speech of the target foreign language video and the Chinese subtitle is lower.
By adopting the method provided by the embodiment of the invention, the difference rate between each foreign language word and the corresponding Chinese subtitle can be determined through the pronunciation change number of each foreign language word of the target foreign language video and the word number of the corresponding Chinese subtitle, so that the foreign language video with high user preference, namely, the foreign language video with high matching degree of the foreign language word and the Chinese subtitle, is recommended to the user according to the matching degree between the foreign language word of the target foreign language video and the corresponding Chinese subtitle, and the recommending effect of the foreign language video recommended to the user is improved.
In one possible implementation manner, the step of determining the matching degree of the foreign language speech of the target foreign language video and the chinese subtitle according to the number of pronunciation change times and the number of words of the chinese subtitle may specifically further include the following steps C1-C3:
step C1: and determining the sum of pronunciation change times corresponding to all words of the target foreign language video as the total pronunciation change times.
Step C2: and determining the total word number of Chinese subtitles corresponding to all the foreign language lines of the target foreign language video.
Step C3: and calculating the absolute value of the ratio of the total word number to the total pronunciation change number, and taking the absolute value as the matching degree of the foreign language speech of the target foreign language video and the Chinese subtitle.
If the matching degree is larger than the preset matching degree threshold, the difference rate between the actual foreign language pronunciation mouth shape change of the character in the target foreign language video and the word number of the Chinese subtitle is smaller, namely the matching degree of the foreign language speech of the target foreign language video and the Chinese subtitle is higher. If the matching degree is not greater than the preset matching degree threshold, the difference rate between the actual foreign language pronunciation mouth shape change of the character in the target foreign language video and the word number of the Chinese subtitle is larger, namely the matching degree of the foreign language speech of the target foreign language video and the Chinese subtitle is lower. The preset matching degree threshold may be specifically set according to practical applications, and is not specifically limited herein.
In one possible implementation manner, after determining the matching degree of the foreign language speech of the target foreign language video and the chinese subtitle, if the target foreign language video is the video to be recommended, the target foreign language video may be recommended to the user based on the matching degree of the foreign language speech of the target foreign language video and the chinese subtitle. Specifically, if the target foreign language video is a video to be recommended, and the determined matching degree of the foreign language speech of the target foreign language video and the Chinese subtitle is the first matching degree, the matching degree of the foreign language speech of the target foreign language video and the Chinese subtitle is higher, namely, the difference between the mouth shape change of the person speaking the foreign language in the target foreign language video and the Chinese subtitle is smaller, the target foreign language video looks more natural, the interestingness of the user is higher and is more loved by the user, so that the matching degree of the foreign language speech of the target foreign language video and the Chinese subtitle can be used as a characteristic dimension of recommending the video to the user, namely, the matching degree of the foreign speech of the target foreign language video and the Chinese subtitle can be considered when determining whether to recommend the target foreign language video to the user or not.
By adopting the method provided by the embodiment of the invention, the matching degree of the foreign language speech of the target foreign language video and the Chinese caption can be determined by analyzing the difference rate between the mouth shape change of the person speaking the foreign language in the target foreign language video and the Chinese caption number, and the matching degree can be extracted as one characteristic dimension of the target foreign language video to be used as one dimension of a video recommendation system, so that the recommendation accuracy of the video recommendation system is improved, and the recommendation effect of recommending the foreign language video to the user is improved.
Based on the same inventive concept, according to the method for determining the matching degree of video subtitles provided in the foregoing embodiment of the present invention, correspondingly, another embodiment of the present invention further provides a device for determining the matching degree of video subtitles, where a schematic structural diagram of the device is shown in fig. 3, and the device specifically includes:
the speech acquisition module 301 is configured to acquire a foreign speech of a target foreign speech video and a corresponding chinese caption;
the change number determining module 302 is configured to query a preset word phonetic symbol library for phonetic symbols corresponding to words included in the foreign language speech, and determine a pronunciation change number corresponding to each word based on the phonetic symbols;
and the matching degree determining module 303 is configured to determine the matching degree of the foreign language speech of the target foreign language video and the chinese subtitle according to the pronunciation change times and the word number of the chinese subtitle.
By adopting the device provided by the embodiment of the invention, the foreign language speech of the target foreign language video and the corresponding Chinese captions can be obtained; inquiring phonetic symbols corresponding to words included in foreign language speech from a preset word phonetic symbol library, and determining pronunciation change times corresponding to each word based on the phonetic symbols; and determining the matching degree of the foreign language speech of the target foreign language video and the Chinese caption according to the pronunciation change times and the word number of the Chinese caption. Furthermore, according to the determined matching degree of the foreign language speech of the target foreign language video and the Chinese subtitle, the foreign language video with high user preference degree, namely the foreign language speech with high matching degree of the foreign language speech and the Chinese subtitle, can be recommended to the user, and the recommending effect of the user on the foreign language video is improved.
Optionally, the number of changes determining module 302 is specifically configured to determine the number of vowels in the phonetic symbol corresponding to each word; and determining the number of the vowels as the pronunciation change times corresponding to the word.
Optionally, referring to fig. 4, the matching degree determining module 303 includes:
a number determining submodule 401, configured to determine, for each foreign language speech of the target foreign language video, a sum of pronunciation change numbers corresponding to words included in the foreign language speech as a pronunciation change number of the foreign language speech;
a word number determining sub-module 402, configured to determine the number of words of the chinese subtitle corresponding to each foreign language line of the target foreign language video;
a difference rate determining submodule 403, configured to traverse each foreign language speech of the target foreign language video, and determine a difference rate between the foreign language speech and the corresponding chinese subtitle based on the number of pronunciation changes of the foreign language speech and the number of words of the chinese subtitle corresponding to the foreign language speech;
and the matching degree determining sub-module 404 is configured to determine the matching degree of the foreign language speech lines of the target foreign language video and the chinese subtitles according to the difference rate between each sentence of the foreign language speech lines of the target foreign language video and the corresponding chinese subtitles.
Optionally, the difference rate determining submodule 403 is specifically configured to determine, as the difference rate between the foreign language speech word and the corresponding chinese subtitle, a ratio of the number of words of the chinese subtitle corresponding to the foreign language speech word to the number of pronunciation changes of the foreign language speech word if the number of pronunciation changes of the foreign language speech word is greater than the number of words of the chinese subtitle corresponding to the foreign language speech word; if the pronunciation change times of the foreign language speech is less than or equal to the number of Chinese characters of the Chinese characters corresponding to the foreign language speech, determining the ratio of the pronunciation change times of the foreign language speech to the number of Chinese characters of the Chinese characters corresponding to the foreign language speech as the difference rate between the foreign language speech and the corresponding Chinese characters.
Optionally, the matching degree determining submodule 404 is specifically configured to calculate an average value of difference rates between the foreign language speech of the target foreign language video and the corresponding chinese subtitle, as an average difference rate; calculating a standard deviation of the difference rate between the foreign language speech of the target foreign language video and the corresponding Chinese caption based on the average difference rate and the difference rate between each foreign language speech of the target foreign language video and the corresponding Chinese caption; if the average difference rate is greater than or equal to a preset difference rate threshold value and the standard deviation is less than or equal to a preset fluctuation threshold value, determining that the matching degree of the foreign language speech lines of the target foreign language video and the Chinese subtitles is a first matching degree, otherwise, determining that the matching degree of the foreign language speech lines of the target foreign language video and the Chinese subtitles is a second matching degree, wherein the first matching degree is greater than the second matching degree.
Optionally, referring to fig. 5, the apparatus further includes a video recommendation module 501, configured to recommend the target foreign language video to the user based on the matching degree of the foreign language speech of the target foreign language video and the chinese subtitle if the target foreign language video is the video to be recommended.
By adopting the device provided by the embodiment of the invention, the matching degree of the foreign language speech of the target foreign language video and the Chinese caption can be determined by analyzing the difference rate between the mouth shape change of the person speaking the foreign language in the target foreign language video and the Chinese caption number, and the matching degree can be extracted as one characteristic dimension of the target foreign language video to be used as one dimension of a video recommendation system, so that the recommendation accuracy of the video recommendation system is improved, and the recommendation effect of recommending the foreign language video to a user is improved.
The embodiment of the invention also provides an electronic device, as shown in fig. 6, which comprises a processor 601, a communication interface 602, a memory 603 and a communication bus 604, wherein the processor 601, the communication interface 602 and the memory 603 complete communication with each other through the communication bus 604,
a memory 603 for storing a computer program;
the processor 601 is configured to execute the program stored in the memory 603, and implement the following steps:
Obtaining a foreign language speech of a target foreign language video and a corresponding Chinese caption;
inquiring phonetic symbols corresponding to words included in the foreign language speech from a preset word phonetic symbol library, and determining pronunciation change times corresponding to each word based on the phonetic symbols;
and determining the matching degree of the foreign language speech of the target foreign language video and the Chinese caption according to the pronunciation change times and the word number of the Chinese caption.
The communication bus mentioned by the above terminal may be a peripheral component interconnect standard (Peripheral Component Interconnect, abbreviated as PCI) bus or an extended industry standard architecture (Extended Industry Standard Architecture, abbreviated as EISA) bus, etc. The communication bus may be classified as an address bus, a data bus, a control bus, or the like. For ease of illustration, the figures are shown with only one bold line, but not with only one bus or one type of bus.
The communication interface is used for communication between the terminal and other devices.
The memory may include random access memory (Random Access Memory, RAM) or non-volatile memory (non-volatile memory), such as at least one disk memory. Optionally, the memory may also be at least one memory device located remotely from the aforementioned processor.
The processor may be a general-purpose processor, including a central processing unit (Central Processing Unit, CPU for short), a network processor (Network Processor, NP for short), etc.; but also digital signal processors (Digital Signal Processor, DSP for short), application specific integrated circuits (Application Specific Integrated Circuit, ASIC for short), field-programmable gate arrays (Field-Programmable Gate Array, FPGA for short) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components.
In yet another embodiment of the present invention, a computer readable storage medium is provided, where a computer program is stored, where the computer program is executed by a processor to implement the method for determining a matching degree of video subtitles according to any of the foregoing embodiments.
In yet another embodiment of the present invention, a computer program product containing instructions that, when executed on a computer, cause the computer to perform the method for determining a matching degree of video subtitles according to any of the above embodiments is also provided.
In the above embodiments, it may be implemented in whole or in part by software, hardware, firmware, or any combination thereof. When implemented in software, may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer instructions. When loaded and executed on a computer, produces a flow or function in accordance with embodiments of the present invention, in whole or in part. The computer may be a general purpose computer, a special purpose computer, a computer network, or other programmable apparatus. The computer instructions may be stored in or transmitted from one computer-readable storage medium to another, for example, by wired (e.g., coaxial cable, optical fiber, digital Subscriber Line (DSL)), or wireless (e.g., infrared, wireless, microwave, etc.). The computer readable storage medium may be any available medium that can be accessed by a computer or a data storage device such as a server, data center, etc. that contains an integration of one or more available media. The usable medium may be a magnetic medium (e.g., floppy Disk, hard Disk, magnetic tape), an optical medium (e.g., DVD), or a semiconductor medium (e.g., solid State Disk (SSD)), etc.
It is noted that relational terms such as first and second, and the like are used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Moreover, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.
In this specification, each embodiment is described in a related manner, and identical and similar parts of each embodiment are all referred to each other, and each embodiment mainly describes differences from other embodiments. In particular, for the apparatus, the electronic device and the storage medium, since they are substantially similar to the method embodiments, the description is relatively simple, and the relevant points are referred to in the description of the method embodiments.
The foregoing description is only of the preferred embodiments of the present invention and is not intended to limit the scope of the present invention. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present invention are included in the protection scope of the present invention.

Claims (7)

1. A method for determining a matching degree of video subtitles, comprising:
obtaining a foreign language speech of a target foreign language video and a corresponding Chinese caption;
inquiring phonetic symbols corresponding to words included in the foreign language speech lines from a preset word phonetic symbol library, determining the number of vowels in the phonetic symbols corresponding to each word, and determining the number of vowels as the pronunciation change times corresponding to the word;
aiming at each foreign language speech of the target foreign language video, determining the sum of pronunciation change times corresponding to words included in the foreign language speech as the pronunciation change times of the foreign language speech;
determining the word number of Chinese subtitles corresponding to each foreign language line of the target foreign language video;
traversing each foreign language speech of the target foreign language video, and determining the difference rate between the foreign language speech and the corresponding Chinese subtitle based on the pronunciation change times of the foreign language speech and the number of Chinese subtitle characters corresponding to the foreign language speech;
And determining the matching degree of the foreign language speech lines of the target foreign language video and the Chinese subtitles according to the difference rate between each sentence of foreign language speech lines of the target foreign language video and the corresponding Chinese subtitles.
2. The method of claim 1, wherein determining the difference rate between the foreign language speech word and the corresponding chinese subtitle based on the number of pronunciation variations of the foreign language speech word and the number of words of the chinese subtitle corresponding to the foreign language speech word comprises:
if the pronunciation change times of the foreign language speech is larger than the number of Chinese characters of the Chinese characters corresponding to the foreign language speech, determining the ratio of the number of the Chinese characters corresponding to the foreign language speech to the pronunciation change times of the foreign language speech as the difference rate between the foreign language speech and the corresponding Chinese characters;
if the pronunciation change times of the foreign language speech is less than or equal to the number of Chinese characters of the Chinese characters corresponding to the foreign language speech, determining the ratio of the pronunciation change times of the foreign language speech to the number of Chinese characters of the Chinese characters corresponding to the foreign language speech as the difference rate between the foreign language speech and the corresponding Chinese characters.
3. The method according to claim 1, wherein determining the matching degree of the foreign language speech lines of the target foreign language video with the chinese subtitles according to the difference rate between each sentence of the foreign language speech lines of the target foreign language video and the corresponding chinese subtitles comprises:
Calculating an average value of the difference rates between the foreign language speech of the target foreign language video and the corresponding Chinese subtitles, and taking the average value as an average difference rate;
calculating a standard deviation of the difference rate between the foreign language speech of the target foreign language video and the corresponding Chinese caption based on the average difference rate and the difference rate between each foreign language speech of the target foreign language video and the corresponding Chinese caption;
if the average difference rate is greater than or equal to a preset difference rate threshold value and the standard deviation is less than or equal to a preset fluctuation threshold value, determining that the matching degree of the foreign language speech lines of the target foreign language video and the Chinese subtitles is a first matching degree, otherwise, determining that the matching degree of the foreign language speech lines of the target foreign language video and the Chinese subtitles is a second matching degree, wherein the first matching degree is greater than the second matching degree.
4. A method according to any one of claims 1-3, wherein after said determining the degree of matching of the foreign language speech of the target foreign language video to the chinese subtitle, the method further comprises:
if the target foreign language video is the video to be recommended, recommending the target foreign language video to a user based on the matching degree of the foreign language speech of the target foreign language video and the Chinese subtitle.
5. A video subtitle matching degree determining apparatus, comprising:
the speech acquisition module is used for acquiring the foreign speech of the target foreign language video and the corresponding Chinese captions;
the change number determining module is used for inquiring phonetic symbols corresponding to words included in the foreign language speech from a preset word phonetic symbol library, determining the number of vowels in the phonetic symbols corresponding to each word and determining the number of vowels as the pronunciation change number corresponding to the word;
the matching degree determining module is used for determining the sum of pronunciation change times corresponding to words included in each foreign language speech word as the pronunciation change times of the foreign language speech word for each foreign language speech word of the target foreign language video; determining the word number of Chinese subtitles corresponding to each foreign language line of the target foreign language video; traversing each foreign language speech of the target foreign language video, and determining the difference rate between the foreign language speech and the corresponding Chinese subtitle based on the pronunciation change times of the foreign language speech and the number of Chinese subtitle characters corresponding to the foreign language speech; and determining the matching degree of the foreign language speech lines of the target foreign language video and the Chinese subtitles according to the difference rate between each sentence of foreign language speech lines of the target foreign language video and the corresponding Chinese subtitles.
6. The electronic equipment is characterized by comprising a processor, a communication interface, a memory and a communication bus, wherein the processor, the communication interface and the memory are communicated with each other through the communication bus;
a memory for storing a computer program;
a processor for carrying out the method steps of any one of claims 1-4 when executing a program stored on a memory.
7. A computer-readable storage medium, characterized in that the computer-readable storage medium has stored therein a computer program which, when executed by a processor, implements the method steps of any of claims 1-4.
CN202110997692.1A 2021-08-27 2021-08-27 Method and device for determining video subtitle matching degree and electronic equipment Active CN113688283B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110997692.1A CN113688283B (en) 2021-08-27 2021-08-27 Method and device for determining video subtitle matching degree and electronic equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110997692.1A CN113688283B (en) 2021-08-27 2021-08-27 Method and device for determining video subtitle matching degree and electronic equipment

Publications (2)

Publication Number Publication Date
CN113688283A CN113688283A (en) 2021-11-23
CN113688283B true CN113688283B (en) 2023-09-05

Family

ID=78583504

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110997692.1A Active CN113688283B (en) 2021-08-27 2021-08-27 Method and device for determining video subtitle matching degree and electronic equipment

Country Status (1)

Country Link
CN (1) CN113688283B (en)

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20110110539A (en) * 2010-04-01 2011-10-07 티제이커뮤니케이션 주식회사 Caption data structure and caption player for synchronizing syllables between a sound source and caption data
WO2013097429A1 (en) * 2011-12-30 2013-07-04 Lg Electronics (China) R & D Center Co., Ltd Method and apparatus for recognizing video captions
CN103854648A (en) * 2012-12-08 2014-06-11 上海能感物联网有限公司 Chinese and foreign language voiced image data bidirectional reversible voice converting and subtitle labeling method
CN103984772A (en) * 2014-06-04 2014-08-13 百度在线网络技术(北京)有限公司 Method and device for generating text retrieval subtitle library and video retrieval method and device
CN106126619A (en) * 2016-06-20 2016-11-16 中山大学 A kind of video retrieval method based on video content and system
CN110096715A (en) * 2019-05-06 2019-08-06 北京理工大学 A kind of fusion pronunciation character Chinese-Vietnamese statistical machine translation method
CN110430448A (en) * 2019-07-31 2019-11-08 北京奇艺世纪科技有限公司 A kind of barrage processing method, device and electronic equipment
CN110798635A (en) * 2019-10-16 2020-02-14 重庆爱奇艺智能科技有限公司 Method and device for matching subtitle files for video

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN100483399C (en) * 2005-10-09 2009-04-29 株式会社东芝 Training transliteration model, segmentation statistic model and automatic transliterating method and device
CN112825561A (en) * 2019-11-21 2021-05-21 上海幻电信息科技有限公司 Subtitle display method, system, computer device and readable storage medium

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20110110539A (en) * 2010-04-01 2011-10-07 티제이커뮤니케이션 주식회사 Caption data structure and caption player for synchronizing syllables between a sound source and caption data
WO2013097429A1 (en) * 2011-12-30 2013-07-04 Lg Electronics (China) R & D Center Co., Ltd Method and apparatus for recognizing video captions
CN103854648A (en) * 2012-12-08 2014-06-11 上海能感物联网有限公司 Chinese and foreign language voiced image data bidirectional reversible voice converting and subtitle labeling method
CN103984772A (en) * 2014-06-04 2014-08-13 百度在线网络技术(北京)有限公司 Method and device for generating text retrieval subtitle library and video retrieval method and device
CN106126619A (en) * 2016-06-20 2016-11-16 中山大学 A kind of video retrieval method based on video content and system
CN110096715A (en) * 2019-05-06 2019-08-06 北京理工大学 A kind of fusion pronunciation character Chinese-Vietnamese statistical machine translation method
CN110430448A (en) * 2019-07-31 2019-11-08 北京奇艺世纪科技有限公司 A kind of barrage processing method, device and electronic equipment
CN110798635A (en) * 2019-10-16 2020-02-14 重庆爱奇艺智能科技有限公司 Method and device for matching subtitle files for video

Also Published As

Publication number Publication date
CN113688283A (en) 2021-11-23

Similar Documents

Publication Publication Date Title
CN106331778B (en) Video recommendation method and device
US11017178B2 (en) Methods, devices, and systems for constructing intelligent knowledge base
JP4580885B2 (en) Scene information extraction method, scene extraction method, and extraction apparatus
CN111767461B (en) Data processing method and device
CN107885852B (en) APP recommendation method and system based on APP usage record
CN112328906A (en) Content item recommendation method, device, equipment and storage medium
CN111372141B (en) Expression image generation method and device and electronic equipment
CN108021619B (en) Event description object recommendation method and device
EP4310695A1 (en) Data processing method and apparatus, computer device, and storage medium
CN111914564B (en) Text keyword determination method and device
CN116227474B (en) Method and device for generating countermeasure text, storage medium and electronic equipment
CN110019948B (en) Method and apparatus for outputting information
CN107506459A (en) A kind of film recommendation method based on film similarity
CN112487300A (en) Video recommendation method and device, electronic equipment and storage medium
CN109190116B (en) Semantic analysis method, system, electronic device and storage medium
CN113688283B (en) Method and device for determining video subtitle matching degree and electronic equipment
CN113672793A (en) Information recall method and device, electronic equipment and storage medium
US11475080B2 (en) Natural language-based search and discovery of content service
CN111787409A (en) Movie and television comment data processing method and device
CN110750708A (en) Keyword recommendation method and device and electronic equipment
US11531811B2 (en) Method and system for extracting keywords from text
CN114254634A (en) Multimedia data mining method, device, storage medium and equipment
CN110659419B (en) Method and related device for determining target user
CN106815288A (en) A kind of video related information generation method and its device
CN117252215A (en) Translation model training and translating method and device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant