CN110648653A - Subtitle realization method, device and system based on intelligent voice mouse and storage medium - Google Patents
Subtitle realization method, device and system based on intelligent voice mouse and storage medium Download PDFInfo
- Publication number
- CN110648653A CN110648653A CN201910923592.7A CN201910923592A CN110648653A CN 110648653 A CN110648653 A CN 110648653A CN 201910923592 A CN201910923592 A CN 201910923592A CN 110648653 A CN110648653 A CN 110648653A
- Authority
- CN
- China
- Prior art keywords
- voice
- mouse
- subtitle
- model
- user
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 29
- 238000003860 storage Methods 0.000 title claims abstract description 9
- 238000012549 training Methods 0.000 claims abstract description 15
- 238000007781 pre-processing Methods 0.000 claims abstract description 10
- 238000004590 computer program Methods 0.000 claims description 9
- 238000000605 extraction Methods 0.000 claims description 9
- 230000006870 function Effects 0.000 claims description 6
- 238000013519 translation Methods 0.000 claims description 5
- 238000012937 correction Methods 0.000 claims description 4
- 238000012986 modification Methods 0.000 claims description 4
- 230000004048 modification Effects 0.000 claims description 4
- 238000003058 natural language processing Methods 0.000 claims description 4
- 238000004458 analytical method Methods 0.000 claims description 3
- 230000005540 biological transmission Effects 0.000 claims description 3
- 230000015572 biosynthetic process Effects 0.000 claims description 3
- 238000003786 synthesis reaction Methods 0.000 claims description 3
- 238000005520 cutting process Methods 0.000 claims description 2
- 238000012545 processing Methods 0.000 abstract description 2
- 238000004422 calculation algorithm Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 230000011218 segmentation Effects 0.000 description 2
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000009826 distribution Methods 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000003825 pressing Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L13/00—Speech synthesis; Text to speech systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0354—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 2D relative movements between the device, or an operating part thereof, and a plane or surface, e.g. 2D mice, trackballs, pens or pucks
- G06F3/03543—Mice or pucks
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/02—Feature extraction for speech recognition; Selection of recognition unit
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/06—Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
- G10L15/063—Training
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/14—Speech classification or search using statistical models, e.g. Hidden Markov Models [HMMs]
- G10L15/142—Hidden Markov Models [HMMs]
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/18—Speech classification or search using natural language modelling
- G10L15/1822—Parsing for meaning understanding
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/225—Feedback of the input speech
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Acoustics & Sound (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Multimedia (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- Probability & Statistics with Applications (AREA)
- Computer Vision & Pattern Recognition (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
The invention relates to the field of voice signal processing, in particular to a caption realization method, a device, a system and a storage medium based on an intelligent voice mouse, wherein the method comprises the following steps: the method comprises the steps of realizing voice acquisition at an intelligent voice mouse end, then preprocessing acquired voice files, storing and managing the preprocessed files, then training the obtained data through a model to obtain user intentions, freely sharing the data into a local area network through a plurality of mobile ends, interconnecting a plurality of devices in the local area network in real time, receiving the data of the plurality of devices, and finally displaying the voice content of a user through subtitles.
Description
Technical Field
The invention relates to the field of voice signal processing, in particular to a subtitle realization method, a device and a system based on an intelligent voice mouse and a storage medium.
Background
At present, when people are in a meeting, most of the contents depend on the ppt of a speaker and the explanation contents of the speaker, most of the contents are stated by the speaker, but in many cases, the participants cannot timely and accurately understand the meaning of the speaker. By utilizing machine learning techniques for deep understanding of natural language, the field of speech recognition will quickly enable commercial deployment, which has been the focus of industrial and academic interest. In various fields of artificial intelligence, natural language processing is the most mature technology, and therefore various large enterprises are introduced to carry out military distribution. In the next 3 years, the mature voice product can be rapidly deployed commercially through the cloud platform and the intelligent hardware platform, and the prospect is very wide.
The invention provides a subtitle realization method, a device and a system based on an intelligent voice mouse and a storage medium, which realize data sharing among different devices, establish application-level connection, realize real-time recording, synchronously share the results of recording, voice recognition and voice translation to computer devices in a local area network, and display subtitle contents on a display terminal, so that people can understand conference contents more clearly.
Disclosure of Invention
In order to solve the above problems in the prior art, the present invention provides a method for implementing subtitles based on an intelligent voice mouse, which comprises the following steps:
step S1: starting recording and finishing recording at the intelligent voice mouse end through an appointed key to realize voice acquisition;
step S2: preprocessing the collected voice file, completing voice recognition and automatic result correction, synchronously completing voice translation and voice synthesis, and storing and managing the preprocessed file;
step S3: training the data obtained in the step S2 through a model to obtain the user intention;
step S4: a plurality of mobile terminals freely share data in a local area network;
step S5: and the real-time interconnection of the multiple devices in the local area network receives the data of the multiple devices in real time and displays the voice content of the user through subtitles.
Preferably, the preprocessing procedure in step S2 includes:
step S21: dividing words by using jieba and Hanlp as open source Chinese word dividing tools, and dividing a Chinese character sequence into separate words;
step S22: using MITIE as a tool for feature extraction and entity identification to identify an entity contained in a text sequence;
step S23: providing an intention judgment service in a mode of combining a plurality of schemes, and marking the category of a sentence by using sklern as an intention judgment tool;
preferably, the model training in step S3 includes the following steps:
step S31: performing feature extraction by using an HMM model, an average perceptron and CRF + +;
step S32: training the prepared corpus;
step S33: cutting the model;
step S34: and storing the trained model.
Preferably, in step S5, the subtitle is modified through a personalized modification function of a computer-side subtitle interface.
Preferably, the multiple devices perform respective management for data statistics and analysis, and the multiple devices include a function for updating software of multiple clients.
In order to achieve the above object, the present invention further provides a caption implementing device based on the intelligent voice mouse, which comprises
The pickup module is used for acquiring a user voice command through the mobile terminal, collecting and sorting the user voice command and transferring the user voice command to the next module;
the preprocessing module is used for carrying out natural language processing on the collected data so as to facilitate the next module to judge the sentence content of the user;
the model training module is used for performing model training by taking MITIE as feature extraction and entity recognition and using sklern as intention judgment to obtain the user intention;
the transmission module is used for freely sharing data to a computer in the local area network through a plurality of mobile terminals;
and the caption display module displays the text content of the input voice through the caption.
In order to achieve the above object, the present invention further provides a caption implementing system based on an intelligent voice mouse, which includes an intelligent voice mouse terminal, a memory, a processor, a display terminal and a computer program stored in the memory and capable of running on the processor, wherein the processor implements the steps of the method when executing the computer program.
To achieve the above object, the present invention also provides a computer-readable storage medium having stored thereon a computer program, which when executed by a processor, performs the steps of the above method.
The invention has the beneficial effects that:
the invention utilizes the voice recognition technology and combines hardware equipment to realize data sharing, real-time recording and real-time voice subtitle display among different equipment, and in a conference, participants can display the speech content of a speaker on a computer screen of the participants by pressing a voice key of a mouse to pick up the speech of the speaker, so that the conference of people is more intelligent and convenient, and the situation that the meaning of the participants cannot be correctly understood due to mishearing of the participants because the speech of the conference speaker is not clear is avoided.
Drawings
Fig. 1 is an overall flowchart of a subtitle implementation method based on an intelligent voice mouse according to embodiment 1 of the present invention.
Fig. 2 is a block diagram of a subtitle implementation apparatus based on an intelligent voice mouse according to embodiment 2 of the present invention.
Fig. 3 is a specific flowchart of a subtitle implementation method based on an intelligent voice mouse in embodiment 1 of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings of the present invention, and it is obvious that the described embodiments are only some embodiments of the present invention, and not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Example 1
Fig. 1 is an overall flowchart of a subtitle implementation method based on an intelligent voice mouse according to an embodiment 1 of the present invention. As shown in fig. 1, a subtitle implementation method based on an intelligent voice mouse includes the following steps:
step S1: and starting and ending the recording at the intelligent voice mouse end through the designated key to realize voice acquisition.
Step S2: and preprocessing the collected voice file, completing voice recognition and automatic result correction, synchronously completing voice translation and voice synthesis, and storing and managing the preprocessed file.
In the step, in the preprocessing process, firstly, jieba and HanLp are used as open-source Chinese word segmentation tools for word segmentation, a Chinese character sequence is segmented into individual words, then, MITIE is used as a tool for feature extraction and entity identification to identify entities contained in a text sequence, next, an intention judgment service is provided by adopting a mode of combining various schemes, and sklern is used as a tool for intention judgment to label the category of a sentence.
Step S3: and training the data obtained in the step S2 through a model to obtain the user intention.
In this step, the model training is to perform feature extraction using an HMM model, an average perceptron and CRF + +, then train the prepared corpus, then cut the model, and finally store the trained model.
Step S4: and a plurality of mobile terminals freely share data into the local area network.
Step S5: and the real-time interconnection of the multiple devices in the local area network receives the data of the multiple devices in real time and displays the voice content of the user through subtitles.
In this step, the multiple devices perform respective management for data statistics and analysis, and the multiple devices include a function of updating the software of the multiple clients; and the subtitle is modified through a personalized modification function of a subtitle interface at the computer end.
Once the user opens the caption mode (by clicking the interface button, the caption works), 3 relatively independent but interdependent functional modules are started simultaneously: the recording module, the voice recognition module and the subtitle display module, wherein the specific dependency relationship refers to the attached figure 3;
when performing voice recognition through a tuning flight voice recognition interface (dynamic correction), two points need to be paid attention:
(1) after 30 seconds, detecting silence, actively disconnecting the message flight voice recognition interface, and reconnecting;
(2) the message flight voice recognition interface is overtime, passively disconnected and reconnected,
when acquiring a speech recognition interface, three main points are needed:
(1) once the identification result is obtained, immediately sending the identification result to a caption display module;
(2) no recognition results are available for 1.5 seconds, and punctuation symbols "," or "are sent in an analog manner. ";
(3) when the correct voice recognition result is determined, translating the sentence;
the caption display module realizes specific algorithm, core thought and attention.
1. Initialization: calculating the positions of each control and layout of a caption window, creating a virtual notepad with a determined width, determining Chinese characters and English characters, determining a row of at least 32 Chinese characters and calculating the size of the characters (theoretically, the number of the displayed characters in a row can be determined to be between 32 and 33, and a certain error can be caused by calculation under different resolutions);
2. writing into the virtual notepad every time the latest recognition result is received, determining the content of the last line of the virtual notepad, and only displaying the last line of the virtual notepad on the caption;
3. if the number of lines of the last line of the notepad displayed last time is less than the number of lines of the last line of the notepad displayed this time, the line is being changed, and at this time, the subtitle has a dynamic effect of 'two lines turning up pages';
the first character of each line of the "virtual notepad" cannot be a punctuation mark;
the last N words of the last line of the virtual notepad are set as color, and the other words are white, wherein N is the minimum value between the number of dynamically corrected words and 10;
6. the exact positions of the color font and the white font need to be calculated;
7. the display of the Chinese original text and the English translation is modularized, independent and not interfered with each other;
8. after setting for 5 seconds, the subtitle interface is emptied, the virtual notepad is emptied, and the recognition result is obtained from the beginning next time;
9. after the caption is finished, the recognized text is stored into a local txt file according to the requirement;
10. the subtitle can be selected with or without background; when the background exists, the color and the transparency of the background can be modified;
11. simple font shadows are realized by using a simple Gaussian fuzzy algorithm, so that the subtitles can be displayed under all desktop backgrounds when the subtitles have no background (the effect is slightly poor under the background of some whistles);
example 2
Fig. 2 is a block diagram of a subtitle implementation apparatus based on an intelligent voice mouse according to an embodiment 2 of the present invention. As shown in fig. 2, the embodiment provides a caption implementing device based on an intelligent voice mouse, which includes
The pickup module is used for acquiring a user voice command through the mobile terminal, collecting and sorting the user voice command and transferring the user voice command to the next module;
the preprocessing module is used for carrying out natural language processing on the collected data so as to facilitate the next module to judge the sentence content of the user;
the model training module is used for performing model training by taking MITIE as feature extraction and entity recognition and using sklern as intention judgment to obtain the user intention;
the transmission module is used for freely sharing data to a computer in the local area network through a plurality of mobile terminals;
and the caption display module displays the text content of the input voice through the caption.
Example 3
The embodiment provides a caption implementation system based on an intelligent voice mouse, which comprises an intelligent voice mouse end, a memory, a processor, a display terminal and a computer program which is stored on the memory and can run on the processor, wherein the processor implements the steps of the method when executing the computer program.
Example 4
The present embodiment provides a computer-readable storage medium, on which a computer program is stored, which program, when being executed by a processor, carries out the steps of the above-mentioned method.
In summary, the method, the apparatus, the system and the storage medium for realizing subtitles based on the intelligent voice mouse disclosed in the above embodiments of the present invention can make the conference more intelligent and convenient, and avoid the situation that the meaning of the conference presenter cannot be correctly understood due to the fact that the conference presenter mislistens because the speech of the conference presenter is unclear.
The above description is only for the specific embodiments of the present invention, but the scope of the present invention is not limited thereto, and any person skilled in the art can understand that the changes or modifications within the technical scope of the present invention are included in the scope of the present invention, and therefore, the scope of the present invention should be subject to the protection scope of the claims.
Claims (8)
1. A caption realization method based on an intelligent voice mouse is characterized by comprising the following steps:
step S1: starting recording and finishing recording at the intelligent voice mouse end through an appointed key to realize voice acquisition;
step S2: preprocessing the collected voice file, completing voice recognition and automatic result correction, synchronously completing voice translation and voice synthesis, and storing and managing the preprocessed file;
step S3: training the data obtained in the step S2 through a model to obtain the user intention;
step S4: a plurality of mobile terminals freely share data in a local area network;
step S5: and the real-time interconnection of the multiple devices in the local area network receives the data of the multiple devices in real time and displays the voice content of the user through subtitles.
2. The subtitle implementing method based on the intelligent voice mouse of claim 1, wherein: the preprocessing process in step S2 includes:
step S21: dividing words by using jieba and Hanlp as open source Chinese word dividing tools, and dividing a Chinese character sequence into separate words;
step S22: using MITIE as a tool for feature extraction and entity identification to identify an entity contained in a text sequence;
step S23: the method provides an intention judgment service in a mode of combining various schemes, and labels the category of the sentence by using sklern as an intention judgment tool.
3. The subtitle implementing method based on the intelligent voice mouse of claim 1, wherein: the model training in step S3 includes the following steps:
step S31: performing feature extraction by using an HMM model, an average perceptron and CRF + +;
step S32: training the prepared corpus;
step S33: cutting the model;
step S34: and storing the trained model.
4. The subtitle implementing method based on the intelligent voice mouse of claim 1, wherein: and in the step S5, modifying the subtitles through a personalized modification function of a computer-side subtitle interface.
5. The subtitle implementing method based on the intelligent voice mouse of claim 1, wherein: the multiple devices perform respective management for data statistics and analysis, and the multiple devices include a function for updating multiple client software.
6. The utility model provides a subtitle realization device based on intelligence voice mouse which characterized in that: comprises that
The pickup module is used for acquiring a user voice command through the mobile terminal, collecting and sorting the user voice command and transferring the user voice command to the next module;
the preprocessing module is used for carrying out natural language processing on the collected data so as to facilitate the next module to judge the sentence content of the user;
the model training module is used for performing model training by taking MITIE as feature extraction and entity recognition and using sklern as intention judgment to obtain the user intention;
the transmission module is used for freely sharing data to a computer in the local area network through a plurality of mobile terminals;
and the caption display module displays the text content of the input voice through the caption.
7. The utility model provides a caption implementation system based on intelligence pronunciation mouse, includes intelligence pronunciation mouse end, memory, treater, display terminal and store on the memory and can be at the computer program of treater operation which characterized in that: the processor, when executing the computer program, realizes the steps of the method of any of the preceding claims 1 to 5.
8. A computer-readable storage medium having stored thereon a computer program, characterized in that: the program when executed by a processor implements the steps of the method of any of claims 1 to 5.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910923592.7A CN110648653A (en) | 2019-09-27 | 2019-09-27 | Subtitle realization method, device and system based on intelligent voice mouse and storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910923592.7A CN110648653A (en) | 2019-09-27 | 2019-09-27 | Subtitle realization method, device and system based on intelligent voice mouse and storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
CN110648653A true CN110648653A (en) | 2020-01-03 |
Family
ID=68992837
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910923592.7A Pending CN110648653A (en) | 2019-09-27 | 2019-09-27 | Subtitle realization method, device and system based on intelligent voice mouse and storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110648653A (en) |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102170553A (en) * | 2010-02-26 | 2011-08-31 | 夏普株式会社 | Conference system, information processor, conference supporting method and information processing method |
CN105446159A (en) * | 2016-01-08 | 2016-03-30 | 北京光年无限科技有限公司 | Intelligent household system and data processing method thereof |
US20160140966A1 (en) * | 2014-11-14 | 2016-05-19 | Sylvia Ann Mines | Portable speech transcription, interpreter and calculation device |
CN106371801A (en) * | 2016-09-23 | 2017-02-01 | 安徽声讯信息技术有限公司 | Voice mouse system based on voice recognition technology |
US20170092274A1 (en) * | 2015-09-24 | 2017-03-30 | Otojoy LLC | Captioning system and/or method |
CN108304466A (en) * | 2017-12-27 | 2018-07-20 | 中国银联股份有限公司 | A kind of user view recognition methods and user view identifying system |
CN108519828A (en) * | 2018-03-21 | 2018-09-11 | 安徽咪鼠科技有限公司 | A kind of intelligent wireless mouse for realizing speech transcription based on speech recognition |
CN109379641A (en) * | 2018-11-14 | 2019-02-22 | 腾讯科技(深圳)有限公司 | A kind of method for generating captions and device |
CN109857327A (en) * | 2017-03-27 | 2019-06-07 | 三角兽(北京)科技有限公司 | Information processing unit, information processing method and storage medium |
-
2019
- 2019-09-27 CN CN201910923592.7A patent/CN110648653A/en active Pending
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102170553A (en) * | 2010-02-26 | 2011-08-31 | 夏普株式会社 | Conference system, information processor, conference supporting method and information processing method |
US20160140966A1 (en) * | 2014-11-14 | 2016-05-19 | Sylvia Ann Mines | Portable speech transcription, interpreter and calculation device |
US20170092274A1 (en) * | 2015-09-24 | 2017-03-30 | Otojoy LLC | Captioning system and/or method |
CN105446159A (en) * | 2016-01-08 | 2016-03-30 | 北京光年无限科技有限公司 | Intelligent household system and data processing method thereof |
CN106371801A (en) * | 2016-09-23 | 2017-02-01 | 安徽声讯信息技术有限公司 | Voice mouse system based on voice recognition technology |
CN109857327A (en) * | 2017-03-27 | 2019-06-07 | 三角兽(北京)科技有限公司 | Information processing unit, information processing method and storage medium |
CN108304466A (en) * | 2017-12-27 | 2018-07-20 | 中国银联股份有限公司 | A kind of user view recognition methods and user view identifying system |
CN108519828A (en) * | 2018-03-21 | 2018-09-11 | 安徽咪鼠科技有限公司 | A kind of intelligent wireless mouse for realizing speech transcription based on speech recognition |
CN109379641A (en) * | 2018-11-14 | 2019-02-22 | 腾讯科技(深圳)有限公司 | A kind of method for generating captions and device |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN104050160B (en) | Interpreter's method and apparatus that a kind of machine is blended with human translation | |
CN111582241B (en) | Video subtitle recognition method, device, equipment and storage medium | |
KR20210038449A (en) | Question and answer processing, language model training method, device, equipment and storage medium | |
CN107680019A (en) | A kind of implementation method of Examination Scheme, device, equipment and storage medium | |
CN110149265B (en) | Message display method and device and computer equipment | |
CN110335592B (en) | Speech phoneme recognition method and device, storage medium and electronic device | |
CN113450759A (en) | Voice generation method, device, electronic equipment and storage medium | |
CN111046148A (en) | Intelligent interaction system and intelligent customer service robot | |
KR20190120847A (en) | Ar-based writing practice method and program | |
CN115760500A (en) | Method, device, equipment and storage medium for optimizing teacher reading and amending operation | |
CN114528840A (en) | Chinese entity identification method, terminal and storage medium fusing context information | |
CN113301382A (en) | Video processing method, device, medium, and program product | |
CN112466277A (en) | Rhythm model training method and device, electronic equipment and storage medium | |
CN113963306B (en) | Courseware title making method and device based on artificial intelligence | |
CN110648653A (en) | Subtitle realization method, device and system based on intelligent voice mouse and storage medium | |
CN112289321B (en) | Explanation synchronization video highlight processing method and device, computer equipment and medium | |
CN115273057A (en) | Text recognition method and device, dictation correction method and device and electronic equipment | |
CN114331932A (en) | Target image generation method and device, computing equipment and computer storage medium | |
CN114490967A (en) | Training method of dialogue model, dialogue method and device of dialogue robot and electronic equipment | |
CN113408290A (en) | Intelligent marking method and system for Chinese text | |
CN116009682A (en) | Interactive display method and device, electronic equipment and readable medium | |
CN113569112A (en) | Tutoring strategy providing method, system, device and medium based on question | |
CN111582281A (en) | Picture display optimization method and device, electronic equipment and storage medium | |
US20240086452A1 (en) | Tracking concepts within content in content management systems and adaptive learning systems | |
CN113762223B (en) | Question splitting model training method, question splitting method and related device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20200103 |
|
RJ01 | Rejection of invention patent application after publication |