CN106782610A - A kind of acoustical testing method of audio conferencing - Google Patents
A kind of acoustical testing method of audio conferencing Download PDFInfo
- Publication number
- CN106782610A CN106782610A CN201611004655.1A CN201611004655A CN106782610A CN 106782610 A CN106782610 A CN 106782610A CN 201611004655 A CN201611004655 A CN 201611004655A CN 106782610 A CN106782610 A CN 106782610A
- Authority
- CN
- China
- Prior art keywords
- audio
- source
- local
- composite tone
- tonequality
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000012360 testing method Methods 0.000 title claims abstract description 28
- 239000002131 composite material Substances 0.000 claims abstract description 27
- 238000000034 method Methods 0.000 claims abstract description 12
- 230000015572 biosynthetic process Effects 0.000 claims description 28
- 238000003786 synthesis reaction Methods 0.000 claims description 28
- 241001269238 Data Species 0.000 claims description 6
- 230000001360 synchronised effect Effects 0.000 claims description 5
- 238000010586 diagram Methods 0.000 description 4
- 230000005540 biological transmission Effects 0.000 description 3
- 238000011156 evaluation Methods 0.000 description 2
- 101000783776 Homo sapiens Lysosomal cobalamin transporter ABCD4 Proteins 0.000 description 1
- 102100020978 Lysosomal cobalamin transporter ABCD4 Human genes 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000001303 quality assessment method Methods 0.000 description 1
- 238000010998 test method Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/60—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for measuring the quality of voice signals
Landscapes
- Engineering & Computer Science (AREA)
- Quality & Reliability (AREA)
- Computational Linguistics (AREA)
- Signal Processing (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Telephonic Communication Services (AREA)
Abstract
The present invention provides a kind of acoustical testing method of audio conferencing, and methods described is:Obtain the corresponding source audio of each member for participating in audio conferencing;Each member constructs the Composite tone close with target audio by audio system based on source audio, after i.e. each member deducts the source audio of itself, Composite tone is formed after the source audio of other members is mixed, the Composite tone is the corresponding Composite tone of each member itself;Using ITU T, P.862 recommendation carries out PESQ and calculates to obtain current real-time tonequality the target audio data that each member sends itself corresponding Composite tone and audio conference server, and the records of values of tonequality is got off to complete the test of tonequality.The method test is simple, and test and comparison is accurate.
Description
Technical field
The present invention relates to technical field of communication equipment, more particularly to a kind of acoustical testing method of audio conferencing.
Background technology
In existing acoustical testing model, that the most commonly used is PESQ (Perceptual evaluation of
Speech quality) be:The method of testing of subjective speech quality assessment.The method is that P.862 recommendation is provided ITU-T
Objective MOS (Mean Opinion Score, mean subjective opinion point) value evaluation method.Use the method for testing of PESQ, it is necessary to
Active audio frequency and target audio.Judge the front and rear difference of tonequality by calculating source audio with the PESQ of target audio, enter
And draw MOS values.MOS values are exactly the quantized result of current correspondence tonequality.
In audio conferencing, each member is added in same meeting, can only hear the non-audio of oneself, that is, for
For audio conferencing, what each member heard is the audio mixing result of remaining member.This also just brings up the sound that each member hears
Frequency is special.Everyone is different.
Because can parliament the audio of each participant is synthesized, and it is remaining and meeting that each participant hears
The audio mixing result of person, that is, the audio that everyone hears is not quite similar.Therefore the tonequality of testing audio meeting can be held very much
The easy source audio for getting each participant, but target audio has been the data being mixed, generally with source
Audio has had very big difference, therefore is not easy based on ITU-T that P.862 recommendation carries out the tonequality assessment of PESQ, if base
In ITU-T, P.862 recommendation carries out tone frequency estimation, and tonequality is relatively poor certainly.
For the situation of the above, can be easy to get the source audio of each participant, but comparing is difficult to get mesh
Mark with phonetic symbols frequently, or can so think, easily get target audio, but source audio can not be corresponding with target audio,
Therefore it is accomplished by being configured to the source audio for meeting target audio.
The content of the invention
The technical problem to be solved in the present invention, is to provide a kind of acoustical testing method of audio conferencing, the method test
Simply, and test and comparison is accurate.
What the present invention was realized in:A kind of acoustical testing method of audio conferencing, methods described is:
Obtain the corresponding source audio of each member for participating in audio conferencing;
Each member constructs the Composite tone close with target audio by audio system based on source audio, i.e., each into
After member deducts the source audio of itself, Composite tone is formed after the source audio of other members is mixed, the Composite tone is every
The corresponding Composite tone of individual member itself;
Each member uses the target audio data that itself corresponding Composite tone sends with audio conference server
P.862 recommendation carries out PESQ and calculates to obtain current real-time tonequality ITU-T, and the records of values of tonequality is got off to complete
The test of tonequality.
Further, methods described is specially:After meeting starts, the corresponding source of each member for participating in audio conferencing is obtained
Audio;Audio conference server needs locally to synthesize to each member does a time synchronized, informs that local synthesis is needed from which
Individual time point starts synthesis;After informing, local synthesis starts locally to be synthesized according to the source audio of each member, now services
Device proceeds by the packet parsing sended over from local source audio, and carries out synthetic operation;When certain time point is synthesized to, no
Need to the member for carrying out synthetic operation do notice do other operate;But must be notified to the local synthesis of other members, its
Its member locally synthesizes after the operation that the member for carrying out synthetic operation is muted is received, it is necessary to timely respond, will
The source data for carrying out synthetic operation member is substituted for mute audio.
Further, data are notified because audio conference server also has a network delay to local, therefore, is also accused
Know that carry out the member of synthetic operation is muted period in which, then local synthetic model starts the member for carrying out synthetic operation
The data of silence period synchronize operation treatment.
Further, after audio conference server sends back the RTP messages of synthesis, local reception audio conference server
The RTP message datas beamed back, the RTP message datas are the target audio data of audio conference server transmission.
The invention has the advantages that:1st, using the characteristic of existing audio conferencing, go to construct the source sound of suitable target audio
Frequently.2nd, the tonequality situation of each participant is calculated based on ITU-TP.862 recommendations using the source audio of construction.3rd, it is right to add
Some details parameters such as network delay carry out synthesis adjustment to the source audio for constructing, and reach more accurate acoustical testing.
Brief description of the drawings
The present invention is further illustrated in conjunction with the embodiments with reference to the accompanying drawings.
Fig. 1 is the inventive method schematic flow sheet.
Fig. 2 is the frame diagram of audio conferencing sound intermediate frequency synthesis.
Fig. 3 is the frame diagram for constructing source audio and test assessment.
Fig. 4 is Conference control relevant control timing diagram.
Fig. 5 is the tonequality estimation flow schematic diagram of single user.
Specific embodiment
Refer to shown in Fig. 1, a kind of acoustical testing method of audio conferencing, methods described is:Obtain and participate in audio conferencing
The corresponding source audio of each member;
Each member constructs the Composite tone close with target audio by audio system based on source audio, i.e., each into
After member deducts the source audio of itself, Composite tone is formed after the source audio of other members is mixed, the Composite tone is every
The corresponding Composite tone of individual member itself;
Each member uses the target audio data that itself corresponding Composite tone sends with audio conference server
P.862 recommendation carries out PESQ and calculates to obtain current real-time tonequality ITU-T, and the records of values of tonequality is got off to complete
The test of tonequality.
Wherein, after audio conference server sends back the RTP messages of synthesis, local reception audio conference server is beamed back
RTP message datas, the RTP message datas be audio conference server transmission target audio data.
Methods described is specially:After meeting starts, the corresponding source audio of each member for participating in audio conferencing is obtained;Audio
Conference server needs locally to synthesize to each member and does a time synchronized, informs that local synthesis needs to be opened from which time point
Begin to synthesize;After informing, local synthesis starts locally to be synthesized according to the source audio of each member, and now server is proceeded by
From the packet parsing that local source audio is sended over, and carry out synthetic operation;When being synthesized to certain time point, it is not necessary to carrying out
The member of synthetic operation does notice and does other operations;But must notify that other members are local to the local synthesis of other members
Synthesis is after the operation that the member for carrying out synthetic operation is muted is received, it is necessary to timely responds, will carry out synthesis behaviour
The source data for making member is substituted for mute audio.Because audio conference server also has network delay to local, therefore notify
In data, also to inform that carry out the member of synthetic operation is muted period in which, then local synthetic model will be synthesized
The data that the member of operation starts silence period synchronize operation treatment.
With reference to an instantiation, the present invention will be further described:
1st, the framework of audio conferencing sound intermediate frequency synthesis
The resultant frame of the audio conferencing of main flow is illustrated as shown in Fig. 2 simply doing here:In Fig. 2 examples, a meeting
View has 4 participants, and each participant has a say in a meeting.The audio that user A can hear is user B, C, D speech
Audio after audio mixing;The audio that user B hears is the audio after user A, C, D speech audio mixing, and equally, what user C heard is to use
Audio after A, B, D the speech audio mixing of family;User D is then to hear the audio after user A, B, C speech audio mixing.And if carrying out audio
Meeting acoustical testing, usually uses the source audio that some are fixed.Such as party A-subscriber uses party B-subscriber with the source audio of party A-subscriber, party B-subscriber
Source audio, C user using C user source audio, D user then using D user source audio.
2nd, construct source audio and test the framework of assessment
Above, have been able to get the ABCD4 source audio of user, a core of Current patents application is exactly
The Composite tone close with target audio is constructed based on source audio, then after Composite tone and target audio are carried out into PESQ calculating,
Can obtain the tonequality situation of each participant.As shown in figure 3, under the framework, for synthesis or fairly simple, i.e.,:
The Composite tone of party A-subscriber is:The source audio of party B-subscriber, the source audio of C user, the source audio of D user are mixed
Audio;
The Composite tone of party B-subscriber is:The source audio of party A-subscriber, the source audio of C user, the source audio of D user are mixed
Audio;
The Composite tone of C user is:The source audio of party A-subscriber, the source audio of party B-subscriber, the source audio of D user are mixed
Audio;
The Composite tone of D user is:The source audio of party A-subscriber, the source audio of party B-subscriber, the source audio of C user are mixed
Audio;But network transmission time delay must also be related to, and correspond to some composite characters of audio conferencing:Such as certain into
The gain of member is enhanced, and the gain of some members is lowered and waits some specially treateds, therefore this is when be accomplished by testing, audio
Meeting is required to inform the source audio of local synthesis goes how to construct.
3rd, Conference control relevant control sequential
Such as the sequential of Fig. 4, the control situation occurred in one meeting of test frame of simple illustration.Meeting first starts
Afterwards, audio conference server needs locally to synthesize to each user and does a time synchronized, informs that local synthesis is needed from which
Time point starts synthesis.After informing, each local synthesis starts locally to be synthesized according to the source audio of each user, now sound
Frequency Conference server also begins to the packet parsing for sended over from local source audio, and is synthesized.When being synthesized to certain
At time point, after C user is muted, the sound of oneself can not be heard due to C user, therefore C need not notify it to C user
Do any operation.But must be notified to the local synthesis of party A-subscriber and party B-subscriber.Locally synthesis is receiving C use to AB user
After the operation that family is muted, it is necessary to timely respond, the source data of C user is substituted for mute audio.And due to audio
Conference server also has a network delay to local, therefore notifies data, also to inform that C user is muted period in which.
Then the data that C user starts silence period are synchronized operation treatment by local synthetic model.
4th, the tonequality assessment of single user
As shown in figure 5, be exactly treatment of the user for tonequality, the final assessment of tonequality is in user side (i.e. user
Locally) carry out calculating and evaluate what is come, after audio conference server sends back the RTP messages of synthesis, locally need audio
The RTP data that Conference server is beamed back, that is, the target data that audio conference server is beamed back.And with the source number of local synthesis
According to.PESQ calculating is carried out using recommendation P.862, it is possible to calculate current real-time tonequality.And give the numerical value of tonequality
Record.
Although the foregoing describing specific embodiment of the invention, those familiar with the art should manage
Solution, the specific embodiment described by us is merely exemplary, and rather than for the restriction to the scope of the present invention, is familiar with this
The technical staff in field should be covered of the invention in the equivalent modification and change made according to spirit of the invention
In scope of the claimed protection.
Claims (4)
1. a kind of acoustical testing method of audio conferencing, it is characterised in that:Methods described is:
Obtain the corresponding source audio of each member for participating in audio conferencing;
Each member constructs the Composite tone close with target audio by audio system based on source audio, i.e. each member button
After except the source audio of itself, Composite tone is formed after the source audio of other members is mixed, the Composite tone for each into
Member itself corresponding Composite tone;
The target audio data that itself corresponding Composite tone and audio conference server send are used ITU-T by each member
P.862 recommendation carries out PESQ and calculates to obtain current real-time tonequality, and the records of values of tonequality is got off to complete tonequality
Test.
2. the acoustical testing method of a kind of audio conferencing according to claim 1, it is characterised in that:Methods described is specific
For:After meeting starts, the corresponding source audio of each member for participating in audio conferencing is obtained;Audio conference server is needed to each
Locally a time synchronized is done in synthesis to member, informs that local synthesis needs to synthesize since which time point;It is local to close after informing
Into starting locally to be synthesized according to the source audio of each member, now server is proceeded by and sended over from local source audio
Packet parsing, and carry out synthetic operation;When being synthesized to certain time point, it is not necessary to notified to the member for carrying out synthetic operation
Do other operations;But must be notified to the local synthesis of other members, locally synthesis is synthesized receiving other members
After the operation that the member of operation is muted, it is necessary to timely respond, the source data for carrying out synthetic operation member is substituted for
Mute audio.
3. the acoustical testing method of a kind of audio conferencing according to claim 2, it is characterised in that:Due to audio conferencing clothes
Business device also has a network delay, therefore notice data to local, and which also to inform carries out the member of synthetic operation in period
It is muted, then local synthetic model synchronizes at operation the data that the member for carrying out synthetic operation starts silence period
Reason.
4. the acoustical testing method of a kind of audio conferencing according to claim 1, it is characterised in that:Audio conference server
After the RTP messages of synthesis are sent back, the RTP message datas that local reception audio conference server is beamed back, the RTP message datas
The target audio data that as audio conference server sends.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201611004655.1A CN106782610B (en) | 2016-11-15 | 2016-11-15 | A kind of acoustical testing method of audio conferencing |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201611004655.1A CN106782610B (en) | 2016-11-15 | 2016-11-15 | A kind of acoustical testing method of audio conferencing |
Publications (2)
Publication Number | Publication Date |
---|---|
CN106782610A true CN106782610A (en) | 2017-05-31 |
CN106782610B CN106782610B (en) | 2019-09-20 |
Family
ID=58968339
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201611004655.1A Active CN106782610B (en) | 2016-11-15 | 2016-11-15 | A kind of acoustical testing method of audio conferencing |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106782610B (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111405122A (en) * | 2020-03-18 | 2020-07-10 | 苏州科达科技股份有限公司 | Audio call testing method, device and storage medium |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040002852A1 (en) * | 2002-07-01 | 2004-01-01 | Kim Doh-Suk | Auditory-articulatory analysis for speech quality assessment |
CN101217759A (en) * | 2007-12-28 | 2008-07-09 | 中国移动通信集团浙江有限公司 | A ringtone quality detecting method of CRBT |
US20090204922A1 (en) * | 2008-02-13 | 2009-08-13 | Microsoft Corporation | Techniques to manage communications resources for a multimedia conference event |
CN103067217A (en) * | 2012-12-14 | 2013-04-24 | 北京思特奇信息技术股份有限公司 | Indicating system and method of communication network service quality |
CN103581934A (en) * | 2012-07-19 | 2014-02-12 | 中兴通讯股份有限公司 | Terminal voice quality evaluation method and terminal |
CN103841275A (en) * | 2013-07-24 | 2014-06-04 | 同济大学 | Interactive audio experience quality evaluation platform and method based on QoS |
US20160028881A1 (en) * | 2014-07-22 | 2016-01-28 | International Business Machines Corporation | MONITORING VOICE OVER INTERNET PROTOCOL (VoIP) QUALITY DURING AN ONGOING CALL |
CN105788611A (en) * | 2016-02-25 | 2016-07-20 | 成都普创通信技术股份有限公司 | Audio quality online monitoring system |
-
2016
- 2016-11-15 CN CN201611004655.1A patent/CN106782610B/en active Active
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040002852A1 (en) * | 2002-07-01 | 2004-01-01 | Kim Doh-Suk | Auditory-articulatory analysis for speech quality assessment |
CN101217759A (en) * | 2007-12-28 | 2008-07-09 | 中国移动通信集团浙江有限公司 | A ringtone quality detecting method of CRBT |
US20090204922A1 (en) * | 2008-02-13 | 2009-08-13 | Microsoft Corporation | Techniques to manage communications resources for a multimedia conference event |
CN103581934A (en) * | 2012-07-19 | 2014-02-12 | 中兴通讯股份有限公司 | Terminal voice quality evaluation method and terminal |
CN103067217A (en) * | 2012-12-14 | 2013-04-24 | 北京思特奇信息技术股份有限公司 | Indicating system and method of communication network service quality |
CN103841275A (en) * | 2013-07-24 | 2014-06-04 | 同济大学 | Interactive audio experience quality evaluation platform and method based on QoS |
US20160028881A1 (en) * | 2014-07-22 | 2016-01-28 | International Business Machines Corporation | MONITORING VOICE OVER INTERNET PROTOCOL (VoIP) QUALITY DURING AN ONGOING CALL |
CN105788611A (en) * | 2016-02-25 | 2016-07-20 | 成都普创通信技术股份有限公司 | Audio quality online monitoring system |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111405122A (en) * | 2020-03-18 | 2020-07-10 | 苏州科达科技股份有限公司 | Audio call testing method, device and storage medium |
CN111405122B (en) * | 2020-03-18 | 2021-09-24 | 苏州科达科技股份有限公司 | Audio call testing method, device and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN106782610B (en) | 2019-09-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Jelassi et al. | Quality of experience of VoIP service: A survey of assessment approaches and open issues | |
US10044854B2 (en) | Method and system for providing captioned telephone service with automated speech recognition | |
Möller | Assessment and prediction of speech quality in telecommunications | |
Takahashi et al. | Perceptual QoS assessment technologies for VoIP | |
CN101689365B (en) | Method of controlling a video conference | |
CN100546322C (en) | Chat and tele-conferencing system with the translation of Text To Speech and speech-to-text | |
CN1946107B (en) | Interactive telephony trainer and exerciser | |
US8358599B2 (en) | System for providing audio highlighting of conference participant playout | |
US8208645B2 (en) | System and method for harmonizing calibration of audio between networked conference rooms | |
US20120259924A1 (en) | Method and apparatus for providing summary information in a live media session | |
Woods et al. | Predicting the effect of hearing loss and audibility on amplified speech reception in a multi-talker listening scenario | |
EP2973559B1 (en) | Audio transmission channel quality assessment | |
CN105991964A (en) | Method and apparatus for broadcasting dynamic information in multimedia conference | |
CN104954727B (en) | Audio-visual synchronization control device and method | |
US20110313765A1 (en) | Conversational Subjective Quality Test Tool | |
CN105991854A (en) | System and method for visualizing VOIP teleconference on intelligent terminal | |
Daengsi et al. | QoE modeling for voice over IP: simplified E-model enhancement utilizing the subjective MOS prediction model: a case of G. 729 and Thai users | |
Gallardo et al. | Human speaker identification of known voices transmitted through different user interfaces and transmission channels | |
Skowronek et al. | Assessment of cognitive load, speech communication quality and quality of experience for spatial and non-spatial audio conferencing calls | |
KR101511795B1 (en) | Notification of dropped audio in a teleconference call | |
JP2015092347A (en) | Emotion-expressing animation face display system, method and program | |
CN106782610B (en) | A kind of acoustical testing method of audio conferencing | |
Berndtsson et al. | Methods for human-centered evaluation of MediaSync in real-time communication | |
KR20150042628A (en) | Method and apparatus of providing a distributed telepresense service | |
JP2013219495A (en) | Emotion-expressing animation face display system, method, and program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
CP03 | Change of name, title or address |
Address after: Software Park Siming District of Xiamen city in Fujian province 361000 two sunrise Road No. 56 unit 1001 Patentee after: Fujian Xingwang Intelligent Technology Co., Ltd Address before: Software Park Siming District of Xiamen city in Fujian province 350000 two sunrise Road No. 56 unit 1001 Patentee before: FUJIAN STAR-NET WISDOM TECHNOLOGY Co.,Ltd. |
|
CP03 | Change of name, title or address |