CN110265027A - A kind of audio frequency transmission method for meeting shorthand system - Google Patents
A kind of audio frequency transmission method for meeting shorthand system Download PDFInfo
- Publication number
- CN110265027A CN110265027A CN201910532574.6A CN201910532574A CN110265027A CN 110265027 A CN110265027 A CN 110265027A CN 201910532574 A CN201910532574 A CN 201910532574A CN 110265027 A CN110265027 A CN 110265027A
- Authority
- CN
- China
- Prior art keywords
- audio
- meeting
- server
- terminal
- section
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 25
- 230000005540 biological transmission Effects 0.000 title claims abstract description 24
- 238000005520 cutting process Methods 0.000 claims abstract description 13
- 230000009466 transformation Effects 0.000 claims abstract description 4
- 238000012546 transfer Methods 0.000 abstract description 4
- 238000003058 natural language processing Methods 0.000 description 16
- 230000008569 process Effects 0.000 description 8
- 238000006243 chemical reaction Methods 0.000 description 5
- 238000005516 engineering process Methods 0.000 description 3
- 230000008859 change Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012795 verification Methods 0.000 description 2
- 230000004888 barrier function Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000010485 coping Effects 0.000 description 1
- 230000001934 delay Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 230000002035 prolonged effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 239000000725 suspension Substances 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/04—Segmentation; Word boundary detection
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/28—Constructional details of speech recognition systems
- G10L15/30—Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/50—Network services
- H04L67/56—Provisioning of proxy services
- H04L67/568—Storing data temporarily at an intermediate stage, e.g. caching
Landscapes
- Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Telephonic Communication Services (AREA)
- Document Processing Apparatus (AREA)
Abstract
The invention discloses a kind of audio frequency transmission methods for meeting shorthand system, when meeting carries out, meeting shorthand terminal includes conference audio and audio stream is cut and numbered according to natural sentences, and the audio section after cutting is respectively sent to ASR server and collaborative editing server;ASR server is used for audio section Content Transformation Cheng Yici text, and collaborative editing server is positioned for audio.Meeting shorthand terminal cuts audio stream according to natural sentences, reduces the bandwidth of accounting during audio transmission, keeps its transmission quicker, the text return speed of ASR server is also faster;After a segment of audio section and its corresponding File Transfer to human-edited's terminal, its corresponding text can be modified according to the audio section, to realize the real-time amendment to the minutes of dynamic generation.
Description
Technical field
The present invention relates to voice stenography technical field, especially a kind of audio frequency transmission method for meeting shorthand system.
Background technique
In conference process, the hoc scenario and particular content of meeting are recorded by record personnel, are formed meeting
View record.Most traditional form is to arrange verification meeting by record personnel on site shorthand and according to session recording after meeting adjourned
View record.
With the development of speech recognition technology (ASR) and natural language processing technique (NLP), the audio energy that is generated in meeting
It is enough to be directly converted into text in real time at meeting scene and generate minutes, considerably reduce the workload of record personnel.
Speech recognition technology be by the vocabulary Content Transformation in human speech be computer-readable input, such as key,
Binary coding or character string;Natural language processing technique research how is realized between people and computer with nature language
Speech carries out efficient communication;The two combines, it will be able to which human speech is converted to the wirtiting form of human language --- text
This.But this conversion process cannot be guaranteed very precisely, particularly with some terms without in input system, personage
Name etc., system have no idea to judge should be specifically what word.Such as input voice " Zhang Ziyi ", system is for this star's
Name can be recognized and converted into correct text;It inputs voice " Zhang Erlei ", for this strange phrase, system is only
Can word for word transliteration and select system be arranged default option, as system default " zhang " preferentially " chapter " when, voice " Zhang Erlei " can
It can will be converted into text " Zhang Erlei ", which results in the presence of mistake.Certainly, actual mistake is not limited only to this.
The accuracy rate of existing meeting shorthand system is substantially in 90-95% or so, and for mistake present in text, having must
It is modified.Currently, the correcting mode used, after mainly still meeting adjourned, records personnel according to session recording to meeting
View record carry out arrangement verification so that minutes at original text generate there are certain time delays, there are certain inconveniences.
It is readily apparent that therewith, optimal correcting mode, real time modifying certainly is carried out to text made of audio conversion, but is existed
Technology barrier be how to realize one side audio just in typing, while one side text is generating, to text carry out in time,
It rapidly corrects, that is, how the text progress just in dynamic generation is corrected in time, rapidly.
Summary of the invention
In view of the above-mentioned problems, the present invention provide it is a kind of facilitate meeting shorthand system realize to just in the text of dynamic generation
Carry out in time, rapidly modified audio frequency transmission method.
A kind of audio frequency transmission method for meeting shorthand system, when meeting carries out, meeting shorthand terminal includes meeting sound
Frequency simultaneously audio stream is cut and is numbered according to natural sentences, and by the audio section after cutting be respectively sent to ASR server and
Collaborative editing server;ASR server is used for audio section Content Transformation Cheng Yici text, and collaborative editing server is used for audio
Positioning.
Further, meeting take down in short-hand terminal when cutting audio stream, to each audio section at the beginning of, the end time,
Audiocode is recorded and generates journal file;Meeting takes down in short-hand terminal and believes one of journal file or a variety of audio frequency characteristics
Breath connection audio section is sent to ASR server together.
Further, cutting the time interval between audio section is 0.00001ms;Audio section duration is limited within 60s.
Further, it while meeting shorthand terminal cutting audio stream, replicates audio stream and is sent to collaborative editing service
Device.
Further, when meeting shorthand terminal detects network interruption, stop sending data to ASR server, and will
Data are temporarily deposited in memory, and when network is again coupled to, data are orderly sent to ASR server by memory.
Beneficial effects of the present invention: 1. meetings shorthand terminal cuts audio stream according to natural sentences, reduces audio
The bandwidth of accounting in transmission process keeps its transmission quicker, and the text return speed of ASR server is also faster;A segment of audio
After section and its corresponding File Transfer to human-edited's terminal, its corresponding text can be modified according to the audio section, from
And realize the real-time amendment to the minutes of dynamic generation;2. audio feature information is sent to ASR together with audio section
Server corresponds audio section and text convenient for collaborative editing server;3. coping with treatment mechanism when suspension, energy
Audio and text after enough well solving network reconnection send problem.
Detailed description of the invention
Fig. 1 is that system block diagram is taken down in short-hand in meeting;
Fig. 2 is audio volume control schematic diagram.
Specific embodiment
The present invention will be further described in detail below with reference to the accompanying drawings and specific embodiments.The embodiment of the present invention is
It is provided for the sake of example and description, and is not exhaustively or to limit the invention to disclosed form.Very much
Modifications and variations are obvious for the ordinary skill in the art.Selection and description embodiment are in order to more preferable
Illustrate the principle of the present invention and practical application, and makes those skilled in the art it will be appreciated that the present invention is suitable to design
In the various embodiments with various modifications of special-purpose.
Embodiment 1
The invention discloses a kind of audio frequency transmission method for meeting shorthand system, system such as Fig. 1 institute is taken down in short-hand in the meeting referred to
Show, terminal is mainly taken down in short-hand by the meeting for including conference audio, the ASR server that speech-recognition services are provided, provides natural language
It handles the NLP server of service, the collaborative editing server of back-office support and the human-edited for correcting minutes is provided
Terminal is constituted, and the meeting shorthand terminal is respectively at the ASR server, the NLP server, the collaborative editing server
It is bi-directionally connected, the collaborative editing server is bi-directionally connected with human-edited's terminal.
Meeting shorthand terminal is disposed on meeting scene, is included and pretreated autonomous device to conference audio;People
Work editor terminal is the equipment such as the desktop computer for being mounted with specific software, notebook, and the specific software refers to can be realized it
The software of necessary functions.
Human-edited's terminal and meeting shorthand terminal can be located at different location, such as meeting in Beijing,
Record personnel carry out the amendment of minutes in Shanghai.
Meeting shorthand terminal, ASR server, NLP server, the connection type between human-edited's terminal can use but
It is not limited to cable network, WiFi network, 4G network.
Audio frequency transmission method disclosed in the present embodiment is to take down in short-hand terminal when meeting is carried out by meeting and include conference audio
And audio stream is cut and numbered according to natural sentences, the time interval cut between audio section is 0.00001ms, then will be cut
Audio section after cutting is respectively sent to ASR server and collaborative editing server;Meeting takes down in short-hand terminal when cutting audio stream, right
At the beginning of each audio section, end time, Audiocode recorded and generate journal file;Meeting takes down in short-hand terminal for day
One of will file or a variety of audio feature informations connection audio section are sent to ASR server together.
Here ASR server and NLP server is existing third-party server.ASR server will be in audio section
Appearance is converted into a text, is mechanical conversion in this conversion process, wherein that there are wrong words quite a lot is (mostly same
Sound character error);NLP server is automatically corrected a text according to natural language, this conversion process is namely based on people
The habit of natural language carries out the process of automatic error-correcting to a text.NLP server is back to the two of meeting shorthand terminal
Secondary text, accuracy is up to 90-95%, but there are still certain error rates.
Here collaborative editing server is for corresponding audio section and secondary text according to journal file.
Audio stream is cut according to natural sentences, this is one of main points of the present embodiment.
People has pause when normally speaking, and the natural sentences in the present embodiment refer to this sentence between adjacent pause
Words, such as " thick mad sound my that the Yellow River " in Fig. 2, " not only ringing in the mansion of the United Nations ".It is carried out according to natural sentences
Audio stream cutting, first is that with can guaranteeing audio-frequency information integrality, the case where preventing audio data from losing;Second is that reducing sound
The bandwidth occupied in frequency transmission process quickly reaches speech text change server convenient for audio, reduces because network traffic congestion causes
Audio is jammed in the road for being sent to speech text change server, this like on the road of a congestion, bicycle,
Battery truck, especially pedestrian can shuttle from automobile gap, and network transmission is similarly.
It fluctuates when detecting in a period of time without audio, just audio stream is cut, it is then subsequent in 0.00001ms
It is continuous to start to process.It will be set to 0.00001ms between audio section, is loss and mistake in order to reduce audio as far as possible
Position.For example, be averaged if being divided into 0.1ms between audio section comprising an audio section interval among 5s audio, 1h audio meeting
72ms deviation is generated, the deviation that 4h audio generates reaches 288ms;If being divided into 0.00001ms between audio section, it is averaged, 1h sound
Frequency only generates 0.0072ms deviation, and the deviation that 4h audio generates is also only 0.0288ms.
If all not detecting pause prolonged enough in 60s, audio stream is cut by force, is avoided
Audio section is too long, influences the transmission speed of audio section and the response speed of ASR server and NLP server.
When audio stream is cut to form audio section, it is with the audio stream that is generating with regard to independent, it is meant that this section
The end of audio, this section audio can be played back by also implying that, convenient for being modified to its corresponding text.
One of journal file or a variety of audio feature informations connection audio section are sent to ASR server together, this
It is the two of the main points of the present embodiment.
At the beginning of audio section, the end time is subject to Beijing time.At the beginning of audio section, the end time,
And its corresponding Audiocode is the information that meeting shorthand terminal can obtain in audio cutting process, but audio section pair
The text answered is the secondary text that NLP server returns.
Ideally, a segment of audio corresponds to passage, carries out corresponding in sequence, but there may be one section
Audio does not correspond to a possibility that text, such as situations such as live play song.What this related to how to return to NLP server
The problem of secondary text and audio section correspond.In the present embodiment, solution to this problem is, if audio section not with
Corresponding text, meeting shorthand terminal is marked in journal file, and collaborative editing server is according to journal file by sound
Frequency range and secondary text are corresponded, if encountering some audio section has label, are just skipped, in order to avoid there is text
The problem of mistake corresponding with audio section occurs.How meeting shorthand terminal knows which section audio section does not have corresponding text, this
It is the data judgement returned by ASR server, such as time started, end time, audio is numbered into one such information
Or much information carries out fusion and forms characteristic information connection audio section sending jointly to ASR server, ASR server, which returns, to be carried
Text of this feature information, meeting shorthand terminal can know that this audio section is sended over either with or without corresponding text.
Audio section is numbered, this is the third key point of the present embodiment.
In the transmission process of audio section and text, audio section is big and text is small, therefore text is often earlier than audio section
Ground is transferred to collaborative editing server, i.e., simultaneously nonsimultaneous transmission takes to collaborative editing server, collaborative editing for audio section and text
How business device knows which section text will be corresponding to which section audio.In the present embodiment, terminal is taken down in short-hand to each section by meeting
Audio and text are numbered to solve the problems, such as this.
Since meeting shorthand terminal, ASR server, NLP server, collaborative editing server, human-edited's terminal are all
By network connection, during meeting carries out, it may occur however that the case where network interruption.When meeting shorthand terminal detects in network
When disconnected, stop sending data to ASR server/NLP server, and data are temporarily deposited in memory, when network connects again
When connecing, data are orderly sent to ASR server/NLP server by memory, after avoiding network reconnection, ASR server/NLP
Server centered receives audio data, is mistakenly considered by attack, and closes meeting shorthand terminal and the connection between it.For
Prevent offline condition occur between meeting shorthand terminal and collaborative editing server, collaborative editing server memory has the meeting of backup
Discuss audio.The conference audio of backup can be used for after the conference is over, and human-edited's terminal transfers conference audio to minutes again
It is modified, rather than minutes must be modified in conference process;It is also possible to prevent meeting from taking down in short-hand terminal
There is the problem of when transmitting obstacle, human-edited's terminal can not obtain audio-frequency information generation between collaborative editing server.
There is offline condition, collaborative editing server memory to prevent meeting from taking down in short-hand between terminal and collaborative editing server
There is the conference audio of backup.The conference audio of backup can be used for after the conference is over, and human-edited's terminal transfers conference audio again
Minutes are modified, rather than minutes must be modified in conference process;It is also possible to prevent meeting
When there is transmission obstacle between view shorthand terminal and collaborative editing server, human-edited's terminal can not obtain asking for audio-frequency information
Topic occurs.
Obviously, described embodiment is only a part of the embodiments of the present invention, instead of all the embodiments.It is based on
Embodiment in the present invention, this field and those of ordinary skill in the related art institute without creative labor
The every other embodiment obtained, all should belong to the scope of protection of the invention.
Claims (6)
1. a kind of audio frequency transmission method for meeting shorthand system, which is characterized in that when meeting carries out, meeting is taken down in short-hand terminal and received
Record conference audio is simultaneously cut and is numbered to audio stream according to natural sentences, and the audio section after cutting is respectively sent to ASR
Server and collaborative editing server;ASR server is used for audio section Content Transformation Cheng Yici text, collaborative editing server
It is positioned for audio.
2. audio frequency transmission method according to claim 1, which is characterized in that meeting takes down in short-hand terminal when cutting audio stream,
At the beginning of to each audio section, end time, Audiocode recorded and generate journal file;Terminal is taken down in short-hand in meeting will
One of journal file or a variety of audio feature informations connection audio section are sent to ASR server together.
3. audio frequency transmission method according to claim 2, which is characterized in that cutting audio section between time interval be
0.00001ms。
4. audio frequency transmission method according to claim 1-3, which is characterized in that audio section duration is limited in 60s
Within.
5. audio frequency transmission method according to claim 4, which is characterized in that the same of terminal cutting audio stream is taken down in short-hand in meeting
When, it replicates audio stream and is sent to collaborative editing server.
6. audio frequency transmission method according to claim 4, which is characterized in that when meeting shorthand terminal detects network interruption
When, stop sending data to ASR server, and data are temporarily deposited in memory, when network is again coupled to, passes through memory
Data are orderly sent to ASR server.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910532574.6A CN110265027B (en) | 2019-06-19 | 2019-06-19 | Audio transmission method for conference shorthand system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910532574.6A CN110265027B (en) | 2019-06-19 | 2019-06-19 | Audio transmission method for conference shorthand system |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110265027A true CN110265027A (en) | 2019-09-20 |
CN110265027B CN110265027B (en) | 2021-07-27 |
Family
ID=67919465
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910532574.6A Active CN110265027B (en) | 2019-06-19 | 2019-06-19 | Audio transmission method for conference shorthand system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110265027B (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113536744A (en) * | 2021-07-15 | 2021-10-22 | 维沃移动通信(杭州)有限公司 | Conference recording method and device |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101101590A (en) * | 2006-07-04 | 2008-01-09 | 王建波 | Sound and character correspondence relation table generation method and positioning method |
WO2010024426A1 (en) * | 2008-08-29 | 2010-03-04 | ヤマハ株式会社 | Sound recording device |
KR20100065495A (en) * | 2008-12-08 | 2010-06-17 | (주) 소리포스 | Integration system and method for shorthand of multilateral video conference |
CN103067793A (en) * | 2011-10-21 | 2013-04-24 | 上海博泰悦臻网络技术服务有限公司 | Intercom device between server and vehicles and off-line communication method and system between vehicles |
CN103336773A (en) * | 2012-05-18 | 2013-10-02 | 徐信 | System and method for audio and video speech processing and retrieval |
CN105845129A (en) * | 2016-03-25 | 2016-08-10 | 乐视控股(北京)有限公司 | Method and system for dividing sentences in audio and automatic caption generation method and system for video files |
CN106057193A (en) * | 2016-07-13 | 2016-10-26 | 深圳市沃特沃德股份有限公司 | Conference record generation method based on telephone conference and device |
US20170277784A1 (en) * | 2016-03-22 | 2017-09-28 | International Business Machines Corporation | Audio summarization of meetings driven by user participation |
CN108074570A (en) * | 2017-12-26 | 2018-05-25 | 安徽声讯信息技术有限公司 | Surface trimming, transmission, the audio recognition method preserved |
CN109147791A (en) * | 2017-06-16 | 2019-01-04 | 深圳市轻生活科技有限公司 | A kind of shorthand system and method |
-
2019
- 2019-06-19 CN CN201910532574.6A patent/CN110265027B/en active Active
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101101590A (en) * | 2006-07-04 | 2008-01-09 | 王建波 | Sound and character correspondence relation table generation method and positioning method |
WO2010024426A1 (en) * | 2008-08-29 | 2010-03-04 | ヤマハ株式会社 | Sound recording device |
KR20100065495A (en) * | 2008-12-08 | 2010-06-17 | (주) 소리포스 | Integration system and method for shorthand of multilateral video conference |
CN103067793A (en) * | 2011-10-21 | 2013-04-24 | 上海博泰悦臻网络技术服务有限公司 | Intercom device between server and vehicles and off-line communication method and system between vehicles |
CN103336773A (en) * | 2012-05-18 | 2013-10-02 | 徐信 | System and method for audio and video speech processing and retrieval |
US20170277784A1 (en) * | 2016-03-22 | 2017-09-28 | International Business Machines Corporation | Audio summarization of meetings driven by user participation |
CN105845129A (en) * | 2016-03-25 | 2016-08-10 | 乐视控股(北京)有限公司 | Method and system for dividing sentences in audio and automatic caption generation method and system for video files |
CN106057193A (en) * | 2016-07-13 | 2016-10-26 | 深圳市沃特沃德股份有限公司 | Conference record generation method based on telephone conference and device |
CN109147791A (en) * | 2017-06-16 | 2019-01-04 | 深圳市轻生活科技有限公司 | A kind of shorthand system and method |
CN108074570A (en) * | 2017-12-26 | 2018-05-25 | 安徽声讯信息技术有限公司 | Surface trimming, transmission, the audio recognition method preserved |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113536744A (en) * | 2021-07-15 | 2021-10-22 | 维沃移动通信(杭州)有限公司 | Conference recording method and device |
Also Published As
Publication number | Publication date |
---|---|
CN110265027B (en) | 2021-07-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10917519B2 (en) | Semiautomated relay method and apparatus | |
US9336689B2 (en) | Methods and apparatuses related to text caption error correction | |
US9710819B2 (en) | Real-time transcription system utilizing divided audio chunks | |
US10680995B1 (en) | Continuous multimodal communication and recording system with automatic transmutation of audio and textual content | |
DE60223131T2 (en) | METHOD AND DEVICE FOR CODING AND DECODING PAUSE INFORMATION | |
CN110557451B (en) | Dialogue interaction processing method and device, electronic equipment and storage medium | |
US20150106091A1 (en) | Conference transcription system and method | |
CN109147779A (en) | Voice data processing method and device | |
US11356492B2 (en) | Preventing audio dropout | |
JP2007189671A (en) | System and method for enabling speaker (WHO-IS-SPEAKING) (WIS) signal application | |
US10192554B1 (en) | Transcription of communications using multiple speech recognition systems | |
CN109417583B (en) | System and method for transcribing audio signal into text in real time | |
EP0847178A3 (en) | Multimedia conferencing using parallel networks | |
GB2390722A (en) | Spoken language interface | |
CN110265026A (en) | A kind of meeting shorthand system and meeting stenography method | |
US20120259924A1 (en) | Method and apparatus for providing summary information in a live media session | |
EP3729770B1 (en) | Managing streamed audio communication sessions | |
US11580954B2 (en) | Systems and methods of handling speech audio stream interruptions | |
CN110265027A (en) | A kind of audio frequency transmission method for meeting shorthand system | |
US8195457B1 (en) | System and method for automatically sending text of spoken messages in voice conversations with voice over IP software | |
US20210020181A1 (en) | Automated Audio-to-Text Transcription in Multi-Device Teleconferences | |
CN110263313A (en) | A kind of man-machine coordination edit methods for meeting shorthand | |
CN112825551A (en) | Method and system for prompting important contents of video conference and transferring and storing important contents | |
CN110264998A (en) | A kind of audio localization method for meeting shorthand system | |
JP4509590B2 (en) | Speech recognition system and program thereof |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |