CN113810653A - Audio and video based method and system for talkback tracking of multi-party network conference - Google Patents

Audio and video based method and system for talkback tracking of multi-party network conference Download PDF

Info

Publication number
CN113810653A
CN113810653A CN202111094320.4A CN202111094320A CN113810653A CN 113810653 A CN113810653 A CN 113810653A CN 202111094320 A CN202111094320 A CN 202111094320A CN 113810653 A CN113810653 A CN 113810653A
Authority
CN
China
Prior art keywords
audio
data
talkback
video data
video
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202111094320.4A
Other languages
Chinese (zh)
Inventor
张力
张瑞峰
朱庆祥
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangzhou Ketianshichang Information Technology Co ltd
Original Assignee
Guangzhou Ketianshichang Information Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangzhou Ketianshichang Information Technology Co ltd filed Critical Guangzhou Ketianshichang Information Technology Co ltd
Priority to CN202111094320.4A priority Critical patent/CN113810653A/en
Publication of CN113810653A publication Critical patent/CN113810653A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/1066Session management
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/1066Session management
    • H04L65/1083In-session procedures
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/40Support for services or applications
    • H04L65/403Arrangements for multi-party communication, e.g. for conferences
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/60Network streaming of media packets
    • H04L65/75Media network packet handling
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems
    • H04N7/152Multipoint control units therefor

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Business, Economics & Management (AREA)
  • General Business, Economics & Management (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Telephonic Communication Services (AREA)

Abstract

The invention discloses a method and a system for a talkback tracking multi-party network conference based on audio and video, which comprises the following steps of establishing communication connection between a conference center of a conference initiator and terminals of a plurality of participants: recording the conference center as a main speaker; acquiring audio and video data of a main speaker and transmitting the audio and video data to each terminal; identifying audio and video data, judging whether the talkback switching indication information exists, and if so, executing talkback switching processing; the talkback switching process includes: identifying other audio and video data after the talkback switching indication information, and comparing the audio and video data with the participant identity information base to obtain the identity data of the next speaker; and searching and determining a terminal of the new speaker according to the participant identity data pre-associated with the participant terminal, and acquiring audio and video data of the new speaker. The method and the device have the effect of improving the fluency of the conference process.

Description

Audio and video based method and system for talkback tracking of multi-party network conference
Technical Field
The application relates to the technical field of online conferences, in particular to a method and a system for a talkback tracking multi-party network conference based on audio and video.
Background
The development of the digital technology leads the audio and video technology to be integrated with communication and further live and work. The technology is particularly popularized in the aspect of enterprise collaborative office, and the user can conveniently work at home, remotely communicate and the like.
Patent publication No. CN112689115A discloses a control method of a multiparty conferencing system. The method comprises the following steps: the conference center and the terminal are coupled to a communication network; the conference center authenticates the terminal; the conference center designates a certain terminal as a video access terminal Cv, processes first video data and a plurality of audio data sent by the video access terminal Cv to generate first mixed audio data, and sends the first mixed audio data to the terminal. And converting the mixed audio data into a conference summary text, storing and outputting the conference summary text.
The above provides a multiparty conference system, which can improve the rate of terminal environment configuration in the conference process, improve the rate of audio/video code conversion, and improve the fluency of audio/video of multiparty conference, but has the following defects:
for a large conference, besides part of people on the line, in a host place, a lot of participants exist under the line, so the conference is usually displayed in the form of an electronic large screen, however, with the switching of the speaker, the content released on the large screen depends on manual work, and is relatively rigid, and the switching fluency is not good, so a new technical scheme is provided in the application.
Disclosure of Invention
In order to improve the fluency of the conference process, the application provides a method and a system for tracking the multi-party network conference by the talkback based on the audio and video.
In a first aspect, the present application provides a method for tracking a multiparty netmeeting by a talkback based on audio and video, which adopts the following technical scheme:
a method for tracking a multi-party network conference by a talkback based on audio and video comprises the following steps of establishing communication connection between a conference center of a conference initiator and terminals of a plurality of participants, and further comprises the following steps:
recording the conference center as a main speaker;
acquiring audio and video data of a main speaker and transmitting the audio and video data to each terminal; and the number of the first and second groups,
identifying audio and video data, judging whether talkback switching indication information exists or not, and executing talkback switching processing if the talkback switching indication information exists;
the talkback switching process includes:
identifying other audio and video data after the talkback switching indication information, and comparing the audio and video data with the participant identity information base to obtain the identity data of the next speaker; and the number of the first and second groups,
and searching and determining a terminal of the new talkback according to the participant identity data pre-associated with the participant terminal, and acquiring the audio and video data of the new talkback.
Optionally, the recognizing the audio/video data includes:
separating audio and video data to obtain audio data and video data;
performing audio translation of the text on the audio data to identify text information; and/or the presence of a gas in the gas,
carrying out image recognition on the video data, and recognizing human behavior;
the talkback switching indication information comprises pre-selected character information and/or human body behavior information.
Optionally, the method further includes:
acquiring manuscript data of a main speaker and switching setting data of chapters, sections or pages of the manuscript; wherein the chapter, section or page switching setting data includes time-consuming data of each chapter, section or page; and the number of the first and second groups,
according to the data set by the chapter, section or page switching, calculating the total time consumption of the manuscript and generating a dynamic progress bar based on the total time consumption of the manuscript;
and the progress bar is sent to the conference center and/or the terminals of the participants.
Optionally, the generation manner of the chapter, section, or page switching setting data includes:
setting time consumption for each chapter, section or page respectively; or the like, or, alternatively,
setting time consumption for a plurality of chapters, sections or pages; or the like, or, alternatively,
the time consumption of setting each chapter, section or page is the same, and the setting is carried out once.
Optionally, the method further includes:
acquiring chapters, sections or pages generated by the pre-arranged interaction activities as interaction nodes; and
determining a section of the progress bar, which is pre-generated on the progress bar, as an interactive section according to the interactive node;
and the identification audio and video data is executed in the interactive section.
Optionally, the method further includes:
performing silence detection on the audio data; or the like, or, alternatively,
the acquired audio and video data is data after silence detection;
the special encoded frames resulting from silence detection are removed prior to executing the audio translated text.
Optionally, the method further includes:
acquiring feedback information of a participant terminal;
judging whether a hand-lifting behavior exists, if so, sending a hand-lifting prompt to a conference center;
and judging whether the feedback information of the conference center agrees to behavior, if so, transmitting the audio and video data in the feedback information of the participant terminal to the conference center.
In a second aspect, the present application provides an audio/video-based lecture tracking multi-party netmeeting system, which adopts the following technical scheme:
an audio-video based walkie-talkie tracking multiparty network conferencing system comprising a memory and a processor, the memory having stored thereon a computer program which can be loaded by the processor and which performs any of the methods described above.
In summary, the present application includes at least one of the following beneficial technical effects:
1. through audio and video analysis and processing of a speaker, a speaker switching instruction is recognized, the identity of the speaker is further recognized according to the speaker switching instruction, the speaker and equipment are actively tracked and switched, and the audio and video of a new speaker are transmitted to a conference center and terminals of other participants, so that seamless switching of the speaker of the network conference is realized;
2. according to the time consumption of each chapter, section or page of the manuscript, a progress bar is generated based on the time consumption to prompt the progress and time of the speaker to control the PPT, the lecture and the like.
Drawings
FIG. 1 is a schematic diagram of a communication architecture of the present application;
FIG. 2 is a schematic overall flow diagram of the present application;
FIG. 3 is a schematic diagram of an interface of an existing conference, as referred to in the embodiments of the present application;
fig. 4 is a schematic diagram of a progress bar of the present application.
Detailed Description
The present application is described in further detail below with reference to figures 1-4.
The embodiment of the application discloses a method for tracking a multi-party network conference by a talkback based on audio and video.
Referring to fig. 1 and 2, an audio-video based method for a lecture tracking multiparty network conference includes:
s1, establishing communication connection between the conference center of the conference initiator and the terminals of a plurality of participants; and the number of the first and second groups,
and S2, interacting the multi-party audio and video data.
The conference center comprises a control computer, a multimedia large screen and a field camera which are distributed in a conference place; the computer is used for connecting and controlling the multimedia large screen and the field camera; the multimedia large screen is used for displaying manuscripts and the like of a main speech on site selected by a conference initiator, and the site camera is used for collecting a video image on site.
The terminals of the participants comprise computers, mobile phones or flat panels, and the terminals are connected with the computers through the Internet so that users can remotely communicate with each other through audio and video.
The interactive multiparty audio and video data comprises:
101. recording the conference center as a main speaker;
102. acquiring audio and video data of a main speaker and transmitting the audio and video data to each terminal; and the number of the first and second groups,
103. and identifying audio and video data, judging whether the talkback switching indication information exists, and if so, executing talkback switching processing.
For the audio and video data of the main speaker, the audio and video acquisition device of the main speaker (here, the conference center) acquires feedback.
Because the talkback switching information is set to: including pre-selected textual information and/or human behavior information, identifying audio-visual data includes: firstly, analyzing audio and video data to obtain audio data and video data; subsequently, performing audio translation text on the audio data to identify text information; and/or performing image recognition on the video data to recognize human behavior.
Specifically, the voice translation of the text information is obtained by using a commercially available voice translation platform (such as a flight), and the text information content is as "please speak XX now"; that is, after the speaker utters the content of the sentence, the content is translated by the audio translation platform, and then the text is recognized, and when the speaker switching instruction is recognized, the speaker switching process is executed.
Similarly, the human behavior information, such as the speaker standing on the front side of the multimedia large screen, makes an action of lifting one arm obliquely upwards at 135 degrees; at this time, the presence of the call-to-speak switching instruction is recognized through the motion recognition technology. In order to improve the accuracy, the two modes can be used in a matched mode, namely the difference between the character information and the action is within +/-3S, namely the talkback switching indication exists in the judgment.
In the above stage, except for the conference initiator, other participants have no speaking right temporarily, and the conference center does not show the audios and videos of other participants, so that the problem that the participants forget to turn off the phone call and an embarrassing event occurs in the existing part of the conference system is solved.
Regarding the talkback handover process, it includes:
201. identifying other audio and video data after the talkback switching indication information, and comparing the audio and video data with the participant identity information base to obtain the identity data of the next speaker; and the number of the first and second groups,
202. and searching and determining a terminal of the new talkback according to the participant identity data pre-associated with the participant terminal, and acquiring the audio and video data of the new talkback.
Before the meeting participant identity information base is taken into the meeting, the meeting participants actively enter or the meeting initiator actively enters; and in the identity input stage, automatically binding the equipment identification code and the network address to be used for obtaining the audio and video data of the main speaker in a matching manner.
According to the content, after the method is applied, the audio and video of the main speaker are monitored in the conference process, and after the switching indication is made, the follow-up audio and video data are actively identified, tracked and switched to the next main speaker selected by the main speaker, so that the audio and video data of the new main speaker are played through the conference center, the intelligent seamless switching of the large-screen content of the conference center is realized, meanwhile, the audio and video data transmitted to other participants are switched, and the conference fluency and experience are improved.
The method also comprises the following steps:
301. acquiring manuscript data of a main speaker and switching setting data of chapters, sections or pages of the manuscript; and the number of the first and second groups,
302. and calculating total time consumption of the manuscripts according to the switching setting data of the chapters, the sections or the pages, and generating a dynamic progress bar based on the total time consumption of the manuscripts.
Wherein the chapter, section or page switching setting data includes time-consuming data of each chapter, section or page; the progress bar is sent to the conference center and/or the terminals of the participants.
Reference is made to fig. 4, which is a schematic diagram of a progress bar, and the technology thereof can be referred to clock. The numbers in the figure represent the time consumed per PPT page, and the total time consumed for this PPT or lecture document; through the rightward pushing of the progress bar, the time consumption without PPT can be accurately mastered so as to help the speaker adjust and control the time and progress of the speech.
The method is provided with the following contents:
referring to fig. 3, in the existing network conference, many display a clock counter (such as a conference) on the right or left of the conference, accumulate the duration of the conference, and have no other role besides; however, if people who have training requirements, people who have specific requirements for speech time or some speaker groups in conferences can help them to control training or speech time, the quality of the related conferences can be effectively improved.
The generation manner of the switching setting data for a chapter, section, or page includes:
setting time consumption for each chapter, section or page respectively; or the like, or, alternatively,
setting time consumption for a plurality of chapters, sections or pages; or the like, or, alternatively,
the time consumption of setting each chapter, section or page is the same, and the setting is carried out once.
Specifically, for example: the PPT file (manuscript) has 25 pages, and the time spent by each page of PPT can be independently set, such as 1 minute on the first page and 2 minutes on the second page; or setting time for the page needing time consuming control; or making a uniform time-consuming setting, such as: PPT5 minutes per page.
For this application, above-mentioned progress bar and relevant setting thereof not only assists user to talkback progress etc. and controls, and it still is applied to the aforesaid talkback of cooperateing and switches, specifically includes:
401. acquiring chapters, sections or pages generated by the pre-arranged interaction activities as interaction nodes; and the number of the first and second groups,
402. and determining a section pre-generated on the progress bar according to the interactive nodes as an interactive section.
Wherein, the prearranged interaction activities are determined by related personnel at the manuscript uploading stage; or more conveniently, interactive marks are made in the manuscripts, and the identification marks are automatically determined; and subsequently, identifying audio and video data and executing the audio and video data in the interactive section.
The advantages of the above arrangement are: when the conference is unnecessary, such as the conference is not required to be picked and remembered, the audio and video data of the whole conference process are not identified, so that the talkback switching is more targeted, the resource waste is reduced, and the like.
In view of the above, the method further comprises:
performing silence detection on the audio data; or the acquired audio and video data is data after silence detection;
subsequently, the special encoded frames from the silence detection are removed before the audio translation is performed.
Specifically, silence detection may be integrated into an encoding module that processes the captured speech (audio); the silence detection algorithm is matched with the noise suppression algorithm to identify whether voice input exists at present, and if no voice input exists, a special coding frame (for example, the length is 0) is coded and output; subsequently, by eliminating the encoded frames, more accurate "audio" data can be obtained, and translation errors, translation costs, and the like can be reduced.
The above-mentioned single line interaction that is the main one, and in the course of meeting, etc., there is a demand that many parties interact at the same time, therefore this method also includes:
501. acquiring feedback information of a participant terminal;
502. judging whether a hand-lifting behavior exists, if so, sending a hand-lifting prompt to a conference center;
503. and judging whether the feedback information of the conference center agrees to behavior, if so, transmitting the audio and video data in the feedback information of the participant terminal to the conference center.
The hand-lifting behavior comprises hand-lifting trigger information generated by man-machine interaction between the terminal and the participants, and the hand-lifting trigger information is displayed on an interactive interface of a large screen and a computer after being received by the conference center so as to inform a main speaker and related personnel.
According to the content, the method is also suitable for simultaneous online interaction of multiple users, and can meet the requirements of more users.
The embodiment of the application also discloses a master-speaking tracking multi-party network conference system based on the audio and video.
An audio-video based lecture tracking multiparty netmeeting system comprises a memory having stored thereon a computer program which can be loaded by a processor and which performs any of the methods described above.
The above embodiments are preferred embodiments of the present application, and the protection scope of the present application is not limited by the above embodiments, so: all equivalent changes made according to the structure, shape and principle of the present application shall be covered by the protection scope of the present application.

Claims (8)

1. A method for tracking a multi-party network conference by a talkback based on audio and video comprises the steps of establishing communication connection between a conference center of a conference initiator and terminals of a plurality of participants, and is characterized by further comprising the following steps:
recording the conference center as a main speaker;
acquiring audio and video data of a main speaker and transmitting the audio and video data to each terminal; and the number of the first and second groups,
identifying audio and video data, judging whether talkback switching indication information exists or not, and executing talkback switching processing if the talkback switching indication information exists;
the talkback switching process includes:
identifying other audio and video data after the talkback switching indication information, and comparing the audio and video data with the participant identity information base to obtain the identity data of the next speaker; and the number of the first and second groups,
and searching and determining a terminal of the new talkback according to the participant identity data pre-associated with the participant terminal, and acquiring the audio and video data of the new talkback.
2. The audio-video based intercom tracking multi-party netmeeting method of claim 1, wherein said identifying audio-video data comprises:
separating audio and video data to obtain audio data and video data;
performing audio translation of the text on the audio data to identify text information; and/or the presence of a gas in the gas,
carrying out image recognition on the video data, and recognizing human behavior;
the talkback switching indication information comprises pre-selected character information and/or human body behavior information.
3. The audio-video based intercom tracking multi-party netmeeting method of claim 1, further comprising:
acquiring manuscript data of a main speaker and switching setting data of chapters, sections or pages of the manuscript; wherein the chapter, section or page switching setting data includes time-consuming data of each chapter, section or page; and the number of the first and second groups,
according to the data set by the chapter, section or page switching, calculating the total time consumption of the manuscript and generating a dynamic progress bar based on the total time consumption of the manuscript;
and the progress bar is sent to the conference center and/or the terminals of the participants.
4. The audio-video based lecture tracking multiparty netmeeting method of claim 3, wherein the chapter, section or page switch setting data is generated in a manner comprising:
setting time consumption for each chapter, section or page respectively; or the like, or, alternatively,
setting time consumption for a plurality of chapters, sections or pages; or the like, or, alternatively,
the time consumption of setting each chapter, section or page is the same, and the setting is carried out once.
5. The audio-video based intercom tracking multi-party netmeeting method of claim 4, further comprising:
acquiring chapters, sections or pages generated by the pre-arranged interaction activities as interaction nodes; and
determining a section of the progress bar, which is pre-generated on the progress bar, as an interactive section according to the interactive node;
and the identification audio and video data is executed in the interactive section.
6. The audio-video based intercom tracking multi-party netmeeting method of claim 5, further comprising:
performing silence detection on the audio data; or the like, or, alternatively,
the acquired audio and video data is data after silence detection;
the special encoded frames resulting from silence detection are removed prior to executing the audio translated text.
7. The audio-video based intercom tracking multi-party netmeeting method of claim 1, further comprising:
acquiring feedback information of a participant terminal;
judging whether a hand-lifting behavior exists, if so, sending a hand-lifting prompt to a conference center;
and judging whether the feedback information of the conference center agrees to behavior, if so, transmitting the audio and video data in the feedback information of the participant terminal to the conference center.
8. An audio-video based walkie-talkie tracking multiparty network conferencing system comprising a memory and a processor, the memory having stored thereon a computer program which can be loaded by the processor and which carries out the method of any one of claims 1 to 7.
CN202111094320.4A 2021-09-17 2021-09-17 Audio and video based method and system for talkback tracking of multi-party network conference Pending CN113810653A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111094320.4A CN113810653A (en) 2021-09-17 2021-09-17 Audio and video based method and system for talkback tracking of multi-party network conference

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111094320.4A CN113810653A (en) 2021-09-17 2021-09-17 Audio and video based method and system for talkback tracking of multi-party network conference

Publications (1)

Publication Number Publication Date
CN113810653A true CN113810653A (en) 2021-12-17

Family

ID=78939742

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111094320.4A Pending CN113810653A (en) 2021-09-17 2021-09-17 Audio and video based method and system for talkback tracking of multi-party network conference

Country Status (1)

Country Link
CN (1) CN113810653A (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030122863A1 (en) * 2001-12-28 2003-07-03 International Business Machines Corporation Navigation tool for slide presentations
CN107507474A (en) * 2016-06-14 2017-12-22 广东紫旭科技有限公司 A kind of long-distance interactive recording and broadcasting system video switching method
CN111405232A (en) * 2020-03-05 2020-07-10 深圳震有科技股份有限公司 Video conference speaker picture switching processing method and device, equipment and medium
CN112445556A (en) * 2019-08-29 2021-03-05 珠海金山办公软件有限公司 Presentation play time reminding method and device

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030122863A1 (en) * 2001-12-28 2003-07-03 International Business Machines Corporation Navigation tool for slide presentations
CN107507474A (en) * 2016-06-14 2017-12-22 广东紫旭科技有限公司 A kind of long-distance interactive recording and broadcasting system video switching method
CN112445556A (en) * 2019-08-29 2021-03-05 珠海金山办公软件有限公司 Presentation play time reminding method and device
CN111405232A (en) * 2020-03-05 2020-07-10 深圳震有科技股份有限公司 Video conference speaker picture switching processing method and device, equipment and medium

Similar Documents

Publication Publication Date Title
CN108615527B (en) Data processing method, device and storage medium based on simultaneous interpretation
US10885318B2 (en) Performing artificial intelligence sign language translation services in a video relay service environment
US8630854B2 (en) System and method for generating videoconference transcriptions
US20070285505A1 (en) Method and apparatus for video conferencing having dynamic layout based on keyword detection
US20100283829A1 (en) System and method for translating communications between participants in a conferencing environment
US20100253689A1 (en) Providing descriptions of non-verbal communications to video telephony participants who are not video-enabled
EP1486949A1 (en) Audio video conversion apparatus and method, and audio video conversion program
CN112653902B (en) Speaker recognition method and device and electronic equipment
CN111405234A (en) Video conference information system and method with integration of cloud computing and edge computing
CN102984496B (en) The processing method of the audiovisual information in video conference, Apparatus and system
US20120259924A1 (en) Method and apparatus for providing summary information in a live media session
US10726247B2 (en) System and method for monitoring qualities of teaching and learning
CN111554280A (en) Real-time interpretation service system for mixing interpretation contents using artificial intelligence and interpretation contents of interpretation experts
CN116524791A (en) Lip language learning auxiliary training system based on meta universe and application thereof
US20210312143A1 (en) Real-time call translation system and method
CN116366800B (en) Online conference method and device, storage medium and electronic equipment
JP2006229903A (en) Conference supporting system, method and computer program
CN113810653A (en) Audio and video based method and system for talkback tracking of multi-party network conference
US11848026B2 (en) Performing artificial intelligence sign language translation services in a video relay service environment
US20220308825A1 (en) Automatic toggling of a mute setting during a communication session
CN111028837B (en) Voice conversation method, voice recognition system and computer storage medium
CN113225521B (en) Video conference control method and device and electronic equipment
CN110996036A (en) Remote online conference management system based on AI intelligent technology
CN112351238B (en) Video conference all-in-one machine and image transmission detection system thereof
KR102041730B1 (en) System and Method for Providing Both Way Simultaneous Interpretation System

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20211217

RJ01 Rejection of invention patent application after publication