CN110516266A - Video caption automatic translating method, device, storage medium and computer equipment - Google Patents

Video caption automatic translating method, device, storage medium and computer equipment Download PDF

Info

Publication number
CN110516266A
CN110516266A CN201910894066.2A CN201910894066A CN110516266A CN 110516266 A CN110516266 A CN 110516266A CN 201910894066 A CN201910894066 A CN 201910894066A CN 110516266 A CN110516266 A CN 110516266A
Authority
CN
China
Prior art keywords
subtitle file
video
file
original
caption
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910894066.2A
Other languages
Chinese (zh)
Inventor
张启
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to CN201910894066.2A priority Critical patent/CN110516266A/en
Publication of CN110516266A publication Critical patent/CN110516266A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/435Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/488Data services, e.g. news ticker
    • H04N21/4884Data services, e.g. news ticker for displaying subtitles
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8547Content authoring involving timestamps for synchronizing content

Abstract

The invention belongs to technical field of data processing, a kind of video caption automatic translating method, device, storage medium and computer equipment are disclosed.Device in the present invention, including data obtaining module, preprocessing module, translation module and memory module.The present invention is by translating and exporting to the subtitle of video after obtaining the caption translating request of viewer, realize the function of video caption automatic translation, avoid problem at high cost caused by human translation and limited languages, the Production Time of video caption translation is saved, it can be according to the actual demand free switching languages of viewer, the user experience of viewer is improved, there is good practical value.

Description

Video caption automatic translating method, device, storage medium and computer equipment
Technical field
The invention belongs to technical field of data processing, and in particular to a kind of video caption automatic translating method, device, storage Medium and computer equipment.
Background technique
As a large amount of video is made, issues and is propagated using Digital Media mode, the viewings such as computer, home theater view The equipment of frequency is also gradually popularized, and people can be contacted fast, on a large scale from country variant, different regions, different language The propagation of video, video is more convenient and quick;But the following main problem is, for the video of different language, Most of video viewers understand that there is also difficulties even with audio and subtitle;As propagated most wide film and electricity in video Depending on play, viewer understands the most of dialogue of personage or monologue in video to the plot of film and TV play, therefore sees If there are obstacles in language understanding by the person of seeing, substantially fail to understand film and TV play;Therefore, video production side is come It says, the difficulty of language understanding means to lose potentially large number of spectators, and for viewer, then it is a large amount of to lose appreciation The chance of outstanding video work.
In existing video production process, the caption translating production of the video of different language, which mainly relies on, to be accomplished manually, Problem on obstacle, due to cost of labor etc., human translation word are watched caused by the languages of some subtitles although can solve Curtain usually only translates several commonly used languages, it is difficult to meet the demand that different viewers obtains the subtitle of its suitable languages, thus Influence the Experience Degree of viewer.
Summary of the invention
In order to solve the above problems existing in the present technology, it is an object of that present invention to provide a kind of video caption automatic translations Method, apparatus, storage medium and computer equipment, present invention saves the Production Times of video caption translation, and can root According to the actual demand free switching languages of viewer, there is good practical value.
The technical scheme adopted by the invention is as follows:
A kind of video caption automatic translating method, comprising the following steps:
Whether real-time detection has caption translating request, if so, then obtaining caption translating request, and is requested according to caption translating Determine sensing word, wherein sensing word includes source languages, target language and video location;
The corresponding video file in current video position is extracted, the corresponding original subtitle text of current video file is then obtained Part;
Currently whether the corresponding languages of original subtitle file are source languages for judgement, if so, then to current original subtitle file It is split and generates multiple phrases, if not, output translation request error message;
The language model of target language is extracted, and multiple phrases after segmentation are sequentially input to the language model of target language It is interior, obtain target subtitle file;
Current goal subtitle file is corresponding with original subtitle file, and current goal subtitle file is exported to working as forward sight Information is completed in frequency file, output translation.
Preferably, caption translating request is voice request and/or text request;When caption translating request is voice request When, determining sensing word, specific step is as follows:
Current speech is requested to carry out pretreatment operation, wherein pretreatment operation includes noise elimination and grammer segmentation;
It is converted to by pretreated voice request, obtains text request, then determined by identification text request Sensing word.
Preferably, when obtaining the corresponding original subtitle file of current video file, the specific steps are as follows:
Judge whether to prestore the corresponding original subtitle file of current video file;
If so, then continuing the judgement currently corresponding languages of original subtitle file;
If not, obtaining the video frame of current video file, identification shape is successively then carried out to the subtitle of all video frames At original document, all original documents are arranged according to the sequence of corresponding video frame, form original subtitle file.
Preferably, by current goal subtitle file and original subtitle file to it is corresponding when
A1. current goal subtitle file is traversed, is judged with the presence or absence of sensitive word in current goal subtitle file, if so, then Sensitive word is replaced operation and/or delete operation, then current goal subtitle file is input to the grammer of target language The verification of smoothness degree is carried out in model, is carried out if not, current goal subtitle file is input in the syntactic model of target language Clear and coherent degree verification;
A2. judge whether the clear and coherent rate of smoothness degree verification is higher than threshold value, if so, final goal subtitle file is then exported, and Final goal subtitle file is corresponding with original subtitle file, finally current final goal subtitle file is exported to current video Target subtitle file after mark if not, the phrase by clear and coherent rate lower than threshold value is labeled, and is re-entered mesh by file Verification object subtitle file is obtained after the language model of poster kind, and step A1 then is repeated to verification object subtitle file.
Preferably, by final goal subtitle file and original subtitle file to it is corresponding when by final goal subtitle file Timestamp is corresponding with the timestamp of original subtitle file.
Preferably, original subtitle file includes that original captioned test and scene corresponding with each original captioned test are believed Breath;When being split to current original subtitle file, multiple phrases are generated after dividing original captioned test, each phrase is corresponding There is a scene information;Extract target language language model after, by after segmentation multiple phrases and corresponding scene information successively It inputs in the language model of target language, obtains target subtitle file.
A kind of device based on above-mentioned video caption automatic translating method, including data obtaining module, preprocessing module, turn over Translate module and memory module;
Data obtaining module, for obtaining caption translating request, video file and original subtitle corresponding with video file File is also used to be requested to determine sensing word according to caption translating, wherein sensing word includes source languages, target language and video position It sets;
Preprocessing module is also used to for being split processing to original subtitle file to the original subtitle text after segmentation Part carries out languages judgement;
Memory module, for storing the language model of video file, original subtitle file and multiple languages;
Then translation module exports target subtitle text for calling the corresponding language model of the target language in sensing word Part.
Preferably, above-mentioned video caption automatic translation device further includes correction verification module;Memory module is also used to store The syntactic model of multiple languages;
Correction verification module, for calling the corresponding syntactic model of the target language in sensing word, then to target subtitle file Carry out the verification of smoothness degree, and the clear and coherent rate result of output verification.
A kind of computer readable storage medium, it is characterised in that: be stored with computer program;The calculation machine program is located When managing device execution, above-mentioned video caption automatic translating method is realized.
A kind of computer equipment, including one or more processors further include memory, further include one or more calculating Machine program;One or more computer programs are stored in memory, and are configured as by one or more of Processor executes;One or more computer programs are configured for executing above-mentioned video caption automatic translation side Method.
The invention has the benefit that
By the way that the subtitle of video is translated and exported after obtaining the caption translating request of viewer, video words are realized The function of curtain automatic translation, avoids problem at high cost caused by human translation and limited languages, has saved video caption translation Production Time can improve the user experience of viewer according to the actual demand free switching languages of viewer, have very Good practical value.
Other beneficial effects of the invention will be described in detail in a specific embodiment.
Detailed description of the invention
In order to more clearly explain the embodiment of the invention or the technical proposal in the existing technology, to embodiment or will show below There is attached drawing needed in technical description to be briefly described, it should be apparent that, the accompanying drawings in the following description is only this Some embodiments of invention for those of ordinary skill in the art without creative efforts, can be with It obtains other drawings based on these drawings.
Fig. 1 is the flow diagram of video caption automatic translating method in embodiment 1.
Specific embodiment
With reference to the accompanying drawing and specific embodiment come the present invention is further elaborated.It should be noted that for Although the explanation of these way of example is to be used to help understand the present invention, but and do not constitute a limitation of the invention.The present invention Disclosed function detail is only used for description example embodiments of the present invention.However, this hair can be embodied with many alternative forms It is bright, and be not construed as limiting the invention in the embodiment that the present invention illustrates.
It should be appreciated that terminology used in the present invention is only used for description specific embodiment, it is not intended to limit of the invention show Example embodiment.If term " includes ", " including ", "comprising" and/or " containing " are in the present invention by use, specified institute's sound Bright feature, integer, step, operation, unit and/or component existence, and be not excluded for one or more other features, number Amount, step, operation, unit, component and/or their combination existence or increase.
It should be appreciated that it will be further noted that the function action occurred may go out with attached drawing in some alternative embodiments Existing sequence is different.For example, depending on related function action, can actually substantially be executed concurrently, or sometimes Two figures continuously shown can be executed in reverse order.
It should be appreciated that providing specific details, in the following description in order to which example embodiment is understood completely. However, those of ordinary skill in the art are it is to be understood that implementation example embodiment without these specific details. Such as system can be shown in block diagrams, to avoid with unnecessary details come so that example is unclear.In other instances, may be used Or not show well-known process, structure and technology unnecessary details, to avoid making example embodiment unclear.
Embodiment 1:
As shown in Figure 1, the present embodiment provides a kind of video caption automatic translating methods, comprising the following steps:
Whether real-time detection has caption translating request, if so, then obtaining caption translating request, and is requested according to caption translating Determine sensing word, wherein sensing word includes source languages, target language and video location;
The corresponding video file in current video position is extracted, the corresponding original subtitle text of current video file is then obtained Part;
Currently whether the corresponding languages of original subtitle file are source languages for judgement, if so, then to current original subtitle file It is split and generates multiple phrases, if not, output translation request error message;
The language model of target language is extracted, and multiple phrases after segmentation are sequentially input to the language model of target language It is interior, obtain target subtitle file;
Current goal subtitle file is corresponding with original subtitle file, and current goal subtitle file is exported to working as forward sight Information is completed in frequency file, output translation.
In the present embodiment, by the way that the subtitle of video is translated and is exported after obtaining the caption translating request of viewer, The function of realizing video caption automatic translation avoids problem at high cost caused by human translation and limited languages, has saved view The Production Time of frequency caption translating can improve the user of viewer according to the actual demand free switching languages of viewer Experience Degree has good practical value.
For example, user can issue caption translating request, caption translating by oneself oral account or by man-machine interface May include in request " English to Chinese ", " Chinese to English ", " Ying Yide ", " Ying Yi ", " Ying Yihan ", " Ying Yi Russia ", " Han Yi Russia ", The sensing word of two languages such as " Chinese translates Korea Spro ", " Chinese translates day ", " Yi Russia " translation, thus the user convenient for different language habit exists Independently switch the languages for oneself needing to translate when watching video, the subtitle after then translating can occur simultaneously with original subtitle Subtitle in video pictures or after translation can substitute original subtitle and appear in video pictures, it is possible thereby to promote each languages Information exchange and cultural exchanges.
As one of preferred embodiment, the language model of each languages is based on CNN (convolutional neural networks) Or training forms RNN (Recognition with Recurrent Neural Network) etc. in advance, so that the data processings such as caption translating are more efficient, and then realizes The automatic real time translation of video caption.
Embodiment 2
The present embodiment is the further improvement made on the basis of embodiment 1, and the difference of the present embodiment and embodiment 1 exists In:
In the present embodiment, caption translating request is voice request and/or text request;When caption translating requests to ask for voice When asking, determining sensing word, specific step is as follows:
Current speech is requested to carry out pretreatment operation, wherein pretreatment operation includes noise elimination and grammer segmentation;Its In, grammer segmentation includes that punctuation mark prejudges, spoken language repeats screening out, segmenting for words.
It is converted to by pretreated voice request, obtains text request, then determined by identification text request Sensing word.
It is real using Python third party library jieba when voice request is Chinese as one of preferred embodiment Now the text request of current speech request is segmented, recognition speed is fast, so that the recognition accuracy of voice request is bright It is aobvious to improve.
Embodiment 3
The present embodiment is the further improvement made on the basis of embodiment 1 and any embodiment 2, the present embodiment and reality It applies example 1 and any difference of embodiment 2 is:
In the present embodiment, when obtaining the corresponding original subtitle file of current video file, the specific steps are as follows:
Judge whether to prestore the corresponding original subtitle file of current video file;
If so, then continuing the judgement currently corresponding languages of original subtitle file;
If not, obtaining the video frame of current video file, identification shape is successively then carried out to the subtitle of all video frames At original document, all original documents are arranged according to the sequence of corresponding video frame, form original subtitle file.
As one of preferred embodiment, OCR recognizer is used when identifying to the subtitle of video frame, by This, which can be convenient, is rapidly converted to text information for image information.
Alternatively preferred embodiment, the lower section for presetting picture in video frame is caption area, when progress word When curtain identification, the caption area of each video frame is only identified;Thus avoid when the texts such as billboard occur in video pictures also into Subtitle mistake caused by row identification.
Embodiment 4
The present embodiment is the further improvement made on the basis of embodiment 1-3 is any, the present embodiment and embodiment 1-3 Any difference is:
In the present embodiment, by current goal subtitle file and original subtitle file to it is corresponding when
A1. current goal subtitle file is traversed, is judged with the presence or absence of sensitive word in current goal subtitle file, if so, then Sensitive word is replaced operation and/or delete operation, then current goal subtitle file is input to the grammer of target language The verification of smoothness degree is carried out in model, is carried out if not, current goal subtitle file is input in the syntactic model of target language Clear and coherent degree verification;
A2. judge whether the clear and coherent rate of smoothness degree verification is higher than threshold value, if so, final goal subtitle file is then exported, and Final goal subtitle file is corresponding with original subtitle file, finally current final goal subtitle file is exported to current video Target subtitle file after mark if not, the phrase by clear and coherent rate lower than threshold value is labeled, and is re-entered mesh by file Verification object subtitle file is obtained after the language model of poster kind, and step A1 then is repeated to verification object subtitle file.
Sensitive word detection and the verification of smoothness degree make the subtitle accuracy rate after translation higher, and more adduction rule, avoid pair Video itself causes deleterious effect.
Embodiment 5
The present embodiment is the further improvement made on the basis of embodiment 4, and the difference of the present embodiment and embodiment 4 exists In:
In the present embodiment, by final goal subtitle file and original subtitle file to it is corresponding when by final goal subtitle file Timestamp it is corresponding with the timestamp of original subtitle file;Thus it is watched caused by avoiding subtitle from misplacing uncomfortable.
As one of preferred embodiment, when the timestamp of original subtitle file is the broadcasting of subtitle in video Between, during caption translating, each subtitle, which is bound, unique corresponding play time.
Embodiment 6
The present embodiment is the further improvement made on the basis of embodiment 1-5 is any, the present embodiment and embodiment 1-5 Any difference is:
In the present embodiment, original subtitle file includes original captioned test and scene corresponding with each original captioned test Information;When being split to current original subtitle file, multiple phrases are generated after dividing original captioned test, each phrase is right Ying Youyi scene information;Extract target language language model after, by after segmentation multiple phrases and corresponding scene information according to In the language model of secondary input target language, target subtitle file is obtained.
In the present embodiment, by presetting scene information, so that caption translating is more accurate;Meanwhile language model is in training Scene information training can be added in the process, so that language model can be according to current subtitle during caption Scene translated.
Embodiment 7
The present embodiment is made on the basis of embodiment 1-6 is any, and the present embodiment provides one kind to be appointed based on embodiment 1-6 The device of the one video caption automatic translating method, including data obtaining module, preprocessing module, translation module and storage mould Block;
Data obtaining module, for obtaining caption translating request, video file and original subtitle corresponding with video file File is also used to be requested to determine sensing word according to caption translating, wherein sensing word includes source languages, target language and video position It sets;
Preprocessing module is also used to for being split processing to original subtitle file to the original subtitle text after segmentation Part carries out languages judgement;
Memory module, for storing the language model of video file, original subtitle file and multiple languages;
Then translation module exports target subtitle text for calling the corresponding language model of the target language in sensing word Part.
Embodiment 8
The present embodiment is the further improvement made on the basis of embodiment 7, and the difference of the present embodiment and embodiment 7 exists In:
In the present embodiment, above-mentioned video caption automatic translation device further includes correction verification module;Memory module is also used to deposit Store up the syntactic model of multiple languages;
Correction verification module, for calling the corresponding syntactic model of the target language in sensing word, then to target subtitle file Carry out the verification of smoothness degree, and the clear and coherent rate result of output verification.
Embodiment 9
The present embodiment is made on the basis of embodiment 1-6 is any, and the present embodiment provides a kind of computer-readable storages Medium, it is characterised in that: be stored with computer program;When the calculation machine program is executed by processor, above-mentioned video is realized Subtitle automatic translating method.
Wherein, computer readable storage medium described in the present embodiment includes but is not limited to that any kind of disk is (including soft Disk, hard disk, CD, CD-ROM and magneto-optic disk), ROM (Read-Only Memory, read-only memory), RAM (RandomAcceSS Memory, immediately memory), EPROM (EraSable Programmable Read-Only Memory, Erarable Programmable Read only Memory), EEPROM (Electrically EraSable Programmable Read-Only Memory, Electrically Erasable Programmable Read-Only Memory), flash memory, magnetic card or light card.It is, storage Equipment includes that can be by equipment (for example, computer, mobile phone) with any medium for the form storage or transmission information that can be read Read-only memory, disk or CD etc., memory disclosed in the present embodiment are only used as citing rather than as restrictions.
Any video words of embodiment 1-7 may be implemented in computer readable storage medium provided in an embodiment of the present invention Curtain automatic translating method, concrete function realize refer to embodiment 1-7 it is any in explanation, details are not described herein.
Embodiment 10
The present embodiment is made on the basis of embodiment any one of 1-6, and the present embodiment provides a kind of computer equipment, packets One or more processors are included, further include memory, further include one or more computer programs;One or more meters Calculation machine program is stored in memory, and is configured as being executed by one or more of processors;Described one or more A computer program is configured for executing above-mentioned video caption automatic translating method.
Wherein, computer equipment described in the present embodiment can be server, personal computer and network equipment etc. and set It is standby.It, can be with it will be understood by those skilled in the art that device structure described in the present embodiment does not constitute the restriction to all devices Including than illustrating more or fewer components, or the certain components of combination.It includes that video caption is automatic that memory, which can be used for storing, The computer program of interpretation method and each functional module, processor runs the computer program for being stored in memory, to hold The various function application and data processing of row equipment.
Embodiments described above is only schematical, can if being related to unit as illustrated by the separation member It is physically separated with being or may not be;If being related to component shown as a unit, can be or can also be with It is not physical unit, it can it is in one place, or may be distributed over multiple network units.It can be according to actual Need that some or all of the units may be selected to achieve the purpose of the solution of this embodiment.Those of ordinary skill in the art are not In the case where paying creative labor, it can understand and implement.
The above embodiments are merely illustrative of the technical solutions of the present invention, rather than its limitations;Although with reference to the foregoing embodiments Invention is explained in detail, those skilled in the art should understand that: it still can be to aforementioned each implementation Technical solution documented by example is modified or equivalent replacement of some of the technical features.And these modification or Replacement, the spirit and scope for technical solution of various embodiments of the present invention that it does not separate the essence of the corresponding technical solution.
The present invention is not limited to above-mentioned optional embodiment, anyone can show that other are each under the inspiration of the present invention The product of kind form.Above-mentioned specific embodiment should not be understood the limitation of pairs of protection scope of the present invention, protection of the invention Range should be subject to be defined in claims, and specification can be used for interpreting the claims.

Claims (10)

1. a kind of video caption automatic translating method, it is characterised in that: the following steps are included:
Whether real-time detection has caption translating request, if so, then obtaining caption translating request, and requests to determine according to caption translating Sensing word, wherein sensing word includes source languages, target language and video location;
The corresponding video file in current video position is extracted, the corresponding original subtitle file of current video file is then obtained;
Currently whether the corresponding languages of original subtitle file are source languages for judgement, if so, then carrying out to current original subtitle file Segmentation generates multiple phrases, if not, output translation request error message;
The language model of target language is extracted, and multiple phrases after segmentation are sequentially input in the language model of target language, Obtain target subtitle file;
Current goal subtitle file is corresponding with original subtitle file, and current goal subtitle file is exported to current video text Information is completed in part, output translation.
2. video caption automatic translating method according to claim 1, it is characterised in that: caption translating request is asked for voice It asks and/or text request;When caption translating request is voice request, determining sensing word, specific step is as follows:
Current speech is requested to carry out pretreatment operation, wherein pretreatment operation includes noise elimination and grammer segmentation;
It is converted to by pretreated voice request, obtains text request, then determined and be directed toward by identification text request Word.
3. video caption automatic translating method according to claim 2, it is characterised in that: it is corresponding to obtain current video file Original subtitle file when, the specific steps are as follows:
Judge whether to prestore the corresponding original subtitle file of current video file;
If so, then continuing the judgement currently corresponding languages of original subtitle file;
If not, obtaining the video frame of current video file, identification is successively then carried out to the subtitle of all video frames and is formed just All original documents are arranged according to the sequence of corresponding video frame, form original subtitle file by beginning file.
4. video caption automatic translating method according to claim 3, it is characterised in that: by current goal subtitle file with When original subtitle file is to corresponding to
A1. current goal subtitle file is traversed, is judged with the presence or absence of sensitive word in current goal subtitle file, if so, then will be quick Sense word is replaced operation and/or delete operation, then current goal subtitle file is input to the syntactic model of target language Middle progress smoothness degree verification, carries out smoothness if not, current goal subtitle file is input in the syntactic model of target language Degree verification;
A2. judge whether the clear and coherent rate of smoothness degree verification is higher than threshold value, if so, then exporting final goal subtitle file, and will most Whole target subtitle file is corresponding with original subtitle file, finally exports current final goal subtitle file to current video text Target subtitle file after mark if not, the phrase by clear and coherent rate lower than threshold value is labeled, and is re-entered target by part Verification object subtitle file is obtained after the language model of languages, and step A1 then is repeated to verification object subtitle file.
5. video caption automatic translating method according to claim 4, it is characterised in that: by final goal subtitle file with Original subtitle file is to corresponding with the timestamp of original subtitle file by the timestamp of final goal subtitle file when corresponding to.
6. -5 any video caption automatic translating method according to claim 1, it is characterised in that: original subtitle file packet Include original captioned test and scene information corresponding with each original captioned test;Current original subtitle file is split When, multiple phrases are generated after dividing original captioned test, each phrase is corresponding with a scene information;Extract the language of target language Say model after, by after segmentation multiple phrases and corresponding scene information sequentially input in the language model of target language, obtain Target subtitle file.
7. a kind of video caption automatic translation device, it is characterised in that: including data obtaining module, preprocessing module, translation mould Block and memory module;
Data obtaining module, for obtaining caption translating request, video file and original subtitle file corresponding with video file, It is also used to be requested to determine sensing word according to caption translating, wherein sensing word includes source languages, target language and video location;
Preprocessing module, for being split processing to original subtitle file, be also used to the original subtitle file after segmentation into The judgement of row languages;
Memory module, for storing the language model of video file, original subtitle file and multiple languages;
Then translation module exports target subtitle file for calling the corresponding language model of the target language in sensing word.
8. video caption automatic translation device according to claim 7, it is characterised in that: further include correction verification module;Storage Module is also used to store the syntactic model of multiple languages;
Then correction verification module carries out target subtitle file for calling the corresponding syntactic model of the target language in sensing word Clear and coherent degree verification, and the clear and coherent rate result of output verification.
9. a kind of computer readable storage medium, it is characterised in that: be stored with computer program;The calculation machine program is processed When device executes, any video caption automatic translating method of claim 1 to 6 is realized.
10. a kind of computer equipment, it is characterised in that: further include memory including one or more processors, further include one Or multiple computer programs;One or more computer programs are stored in memory, and are configured as by described One or more processors execute;It is any that one or more computer programs are configured for perform claim requirement 1 to 6 The video caption automatic translating method.
CN201910894066.2A 2019-09-20 2019-09-20 Video caption automatic translating method, device, storage medium and computer equipment Pending CN110516266A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910894066.2A CN110516266A (en) 2019-09-20 2019-09-20 Video caption automatic translating method, device, storage medium and computer equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910894066.2A CN110516266A (en) 2019-09-20 2019-09-20 Video caption automatic translating method, device, storage medium and computer equipment

Publications (1)

Publication Number Publication Date
CN110516266A true CN110516266A (en) 2019-11-29

Family

ID=68633107

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910894066.2A Pending CN110516266A (en) 2019-09-20 2019-09-20 Video caption automatic translating method, device, storage medium and computer equipment

Country Status (1)

Country Link
CN (1) CN110516266A (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111368557A (en) * 2020-03-06 2020-07-03 北京字节跳动网络技术有限公司 Video content translation method, device, equipment and computer readable medium
CN111901675A (en) * 2020-07-13 2020-11-06 腾讯科技(深圳)有限公司 Multimedia data playing method and device, computer equipment and storage medium
CN112511910A (en) * 2020-11-23 2021-03-16 浪潮天元通信信息系统有限公司 Real-time subtitle processing method and device
CN112995749A (en) * 2021-02-07 2021-06-18 北京字节跳动网络技术有限公司 Method, device and equipment for processing video subtitles and storage medium
CN113038184A (en) * 2021-03-01 2021-06-25 北京百度网讯科技有限公司 Data processing method, device, equipment and storage medium
CN113438542A (en) * 2021-05-28 2021-09-24 北京智慧星光信息技术有限公司 Subtitle real-time translation method, system, electronic equipment and storage medium
CN113660432A (en) * 2021-08-17 2021-11-16 安徽听见科技有限公司 Translation subtitle production method and device, electronic equipment and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101090461A (en) * 2006-06-13 2007-12-19 中国科学院计算技术研究所 Automatic translation method for digital video captions
CN102104741A (en) * 2009-12-16 2011-06-22 新奥特(北京)视频技术有限公司 Method and device for arranging multi-language captions
CN104391839A (en) * 2014-11-13 2015-03-04 百度在线网络技术(北京)有限公司 Method and device for machine translation
CN110134973A (en) * 2019-04-12 2019-08-16 深圳壹账通智能科技有限公司 Video caption real time translating method, medium and equipment based on artificial intelligence

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101090461A (en) * 2006-06-13 2007-12-19 中国科学院计算技术研究所 Automatic translation method for digital video captions
CN102104741A (en) * 2009-12-16 2011-06-22 新奥特(北京)视频技术有限公司 Method and device for arranging multi-language captions
CN104391839A (en) * 2014-11-13 2015-03-04 百度在线网络技术(北京)有限公司 Method and device for machine translation
CN110134973A (en) * 2019-04-12 2019-08-16 深圳壹账通智能科技有限公司 Video caption real time translating method, medium and equipment based on artificial intelligence

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111368557A (en) * 2020-03-06 2020-07-03 北京字节跳动网络技术有限公司 Video content translation method, device, equipment and computer readable medium
CN111901675A (en) * 2020-07-13 2020-11-06 腾讯科技(深圳)有限公司 Multimedia data playing method and device, computer equipment and storage medium
CN111901675B (en) * 2020-07-13 2021-09-21 腾讯科技(深圳)有限公司 Multimedia data playing method and device, computer equipment and storage medium
CN112511910A (en) * 2020-11-23 2021-03-16 浪潮天元通信信息系统有限公司 Real-time subtitle processing method and device
CN112995749A (en) * 2021-02-07 2021-06-18 北京字节跳动网络技术有限公司 Method, device and equipment for processing video subtitles and storage medium
CN112995749B (en) * 2021-02-07 2023-05-26 北京字节跳动网络技术有限公司 Video subtitle processing method, device, equipment and storage medium
CN113038184A (en) * 2021-03-01 2021-06-25 北京百度网讯科技有限公司 Data processing method, device, equipment and storage medium
CN113438542A (en) * 2021-05-28 2021-09-24 北京智慧星光信息技术有限公司 Subtitle real-time translation method, system, electronic equipment and storage medium
CN113438542B (en) * 2021-05-28 2022-11-08 北京智慧星光信息技术有限公司 Subtitle real-time translation method, system, electronic equipment and storage medium
CN113660432A (en) * 2021-08-17 2021-11-16 安徽听见科技有限公司 Translation subtitle production method and device, electronic equipment and storage medium

Similar Documents

Publication Publication Date Title
CN110516266A (en) Video caption automatic translating method, device, storage medium and computer equipment
KR101990023B1 (en) Method for chunk-unit separation rule and display automated key word to develop foreign language studying, and system thereof
CN110517689B (en) Voice data processing method, device and storage medium
EP4099709A1 (en) Data processing method and apparatus, device, and readable storage medium
CN114465737B (en) Data processing method and device, computer equipment and storage medium
CN109256133A (en) A kind of voice interactive method, device, equipment and storage medium
CN111488487B (en) Advertisement detection method and detection system for all-media data
CN112399258A (en) Live playback video generation playing method and device, storage medium and electronic equipment
CN113392273A (en) Video playing method and device, computer equipment and storage medium
CN113035199A (en) Audio processing method, device, equipment and readable storage medium
CN111526405A (en) Media material processing method, device, equipment, server and storage medium
CN113703579B (en) Data processing method, device, electronic equipment and storage medium
CN111460094A (en) Method and device for optimizing audio splicing based on TTS (text to speech)
Ham et al. Ksl-guide: A large-scale korean sign language dataset including interrogative sentences for guiding the deaf and hard-of-hearing
WO2021169825A1 (en) Speech synthesis method and apparatus, device and storage medium
CN116567351B (en) Video processing method, device, equipment and medium
US20230326369A1 (en) Method and apparatus for generating sign language video, computer device, and storage medium
CN115690280B (en) Three-dimensional image pronunciation mouth shape simulation method
CN116366872A (en) Live broadcast method, device and system based on man and artificial intelligence
CN109905756A (en) TV subtitling dynamic creation method and relevant device based on artificial intelligence
CN113033357B (en) Subtitle adjusting method and device based on mouth shape characteristics
CN113761865A (en) Sound and text realignment and information presentation method and device, electronic equipment and storage medium
Xu et al. Gabor based lipreading with a new audiovisual mandarin corpus
Virkkunen Automatic speech recognition for the hearing impaired in an augmented reality application
CN111160051A (en) Data processing method and device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination