CN106782610A - A kind of acoustical testing method of audio conferencing - Google Patents

A kind of acoustical testing method of audio conferencing Download PDF

Info

Publication number
CN106782610A
CN106782610A CN201611004655.1A CN201611004655A CN106782610A CN 106782610 A CN106782610 A CN 106782610A CN 201611004655 A CN201611004655 A CN 201611004655A CN 106782610 A CN106782610 A CN 106782610A
Authority
CN
China
Prior art keywords
audio
source
local
composite tone
tonequality
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201611004655.1A
Other languages
Chinese (zh)
Other versions
CN106782610B (en
Inventor
洪剑平
陈锦辉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fujian Xingwang Intelligent Technology Co., Ltd
Original Assignee
Fujian Star-Net Wisdom Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fujian Star-Net Wisdom Technology Co Ltd filed Critical Fujian Star-Net Wisdom Technology Co Ltd
Priority to CN201611004655.1A priority Critical patent/CN106782610B/en
Publication of CN106782610A publication Critical patent/CN106782610A/en
Application granted granted Critical
Publication of CN106782610B publication Critical patent/CN106782610B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/60Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for measuring the quality of voice signals

Landscapes

  • Engineering & Computer Science (AREA)
  • Quality & Reliability (AREA)
  • Computational Linguistics (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Telephonic Communication Services (AREA)

Abstract

The present invention provides a kind of acoustical testing method of audio conferencing, and methods described is:Obtain the corresponding source audio of each member for participating in audio conferencing;Each member constructs the Composite tone close with target audio by audio system based on source audio, after i.e. each member deducts the source audio of itself, Composite tone is formed after the source audio of other members is mixed, the Composite tone is the corresponding Composite tone of each member itself;Using ITU T, P.862 recommendation carries out PESQ and calculates to obtain current real-time tonequality the target audio data that each member sends itself corresponding Composite tone and audio conference server, and the records of values of tonequality is got off to complete the test of tonequality.The method test is simple, and test and comparison is accurate.

Description

A kind of acoustical testing method of audio conferencing
Technical field
The present invention relates to technical field of communication equipment, more particularly to a kind of acoustical testing method of audio conferencing.
Background technology
In existing acoustical testing model, that the most commonly used is PESQ (Perceptual evaluation of Speech quality) be:The method of testing of subjective speech quality assessment.The method is that P.862 recommendation is provided ITU-T Objective MOS (Mean Opinion Score, mean subjective opinion point) value evaluation method.Use the method for testing of PESQ, it is necessary to Active audio frequency and target audio.Judge the front and rear difference of tonequality by calculating source audio with the PESQ of target audio, enter And draw MOS values.MOS values are exactly the quantized result of current correspondence tonequality.
In audio conferencing, each member is added in same meeting, can only hear the non-audio of oneself, that is, for For audio conferencing, what each member heard is the audio mixing result of remaining member.This also just brings up the sound that each member hears Frequency is special.Everyone is different.
Because can parliament the audio of each participant is synthesized, and it is remaining and meeting that each participant hears The audio mixing result of person, that is, the audio that everyone hears is not quite similar.Therefore the tonequality of testing audio meeting can be held very much The easy source audio for getting each participant, but target audio has been the data being mixed, generally with source Audio has had very big difference, therefore is not easy based on ITU-T that P.862 recommendation carries out the tonequality assessment of PESQ, if base In ITU-T, P.862 recommendation carries out tone frequency estimation, and tonequality is relatively poor certainly.
For the situation of the above, can be easy to get the source audio of each participant, but comparing is difficult to get mesh Mark with phonetic symbols frequently, or can so think, easily get target audio, but source audio can not be corresponding with target audio, Therefore it is accomplished by being configured to the source audio for meeting target audio.
The content of the invention
The technical problem to be solved in the present invention, is to provide a kind of acoustical testing method of audio conferencing, the method test Simply, and test and comparison is accurate.
What the present invention was realized in:A kind of acoustical testing method of audio conferencing, methods described is:
Obtain the corresponding source audio of each member for participating in audio conferencing;
Each member constructs the Composite tone close with target audio by audio system based on source audio, i.e., each into After member deducts the source audio of itself, Composite tone is formed after the source audio of other members is mixed, the Composite tone is every The corresponding Composite tone of individual member itself;
Each member uses the target audio data that itself corresponding Composite tone sends with audio conference server P.862 recommendation carries out PESQ and calculates to obtain current real-time tonequality ITU-T, and the records of values of tonequality is got off to complete The test of tonequality.
Further, methods described is specially:After meeting starts, the corresponding source of each member for participating in audio conferencing is obtained Audio;Audio conference server needs locally to synthesize to each member does a time synchronized, informs that local synthesis is needed from which Individual time point starts synthesis;After informing, local synthesis starts locally to be synthesized according to the source audio of each member, now services Device proceeds by the packet parsing sended over from local source audio, and carries out synthetic operation;When certain time point is synthesized to, no Need to the member for carrying out synthetic operation do notice do other operate;But must be notified to the local synthesis of other members, its Its member locally synthesizes after the operation that the member for carrying out synthetic operation is muted is received, it is necessary to timely respond, will The source data for carrying out synthetic operation member is substituted for mute audio.
Further, data are notified because audio conference server also has a network delay to local, therefore, is also accused Know that carry out the member of synthetic operation is muted period in which, then local synthetic model starts the member for carrying out synthetic operation The data of silence period synchronize operation treatment.
Further, after audio conference server sends back the RTP messages of synthesis, local reception audio conference server The RTP message datas beamed back, the RTP message datas are the target audio data of audio conference server transmission.
The invention has the advantages that:1st, using the characteristic of existing audio conferencing, go to construct the source sound of suitable target audio Frequently.2nd, the tonequality situation of each participant is calculated based on ITU-TP.862 recommendations using the source audio of construction.3rd, it is right to add Some details parameters such as network delay carry out synthesis adjustment to the source audio for constructing, and reach more accurate acoustical testing.
Brief description of the drawings
The present invention is further illustrated in conjunction with the embodiments with reference to the accompanying drawings.
Fig. 1 is the inventive method schematic flow sheet.
Fig. 2 is the frame diagram of audio conferencing sound intermediate frequency synthesis.
Fig. 3 is the frame diagram for constructing source audio and test assessment.
Fig. 4 is Conference control relevant control timing diagram.
Fig. 5 is the tonequality estimation flow schematic diagram of single user.
Specific embodiment
Refer to shown in Fig. 1, a kind of acoustical testing method of audio conferencing, methods described is:Obtain and participate in audio conferencing The corresponding source audio of each member;
Each member constructs the Composite tone close with target audio by audio system based on source audio, i.e., each into After member deducts the source audio of itself, Composite tone is formed after the source audio of other members is mixed, the Composite tone is every The corresponding Composite tone of individual member itself;
Each member uses the target audio data that itself corresponding Composite tone sends with audio conference server P.862 recommendation carries out PESQ and calculates to obtain current real-time tonequality ITU-T, and the records of values of tonequality is got off to complete The test of tonequality.
Wherein, after audio conference server sends back the RTP messages of synthesis, local reception audio conference server is beamed back RTP message datas, the RTP message datas be audio conference server transmission target audio data.
Methods described is specially:After meeting starts, the corresponding source audio of each member for participating in audio conferencing is obtained;Audio Conference server needs locally to synthesize to each member and does a time synchronized, informs that local synthesis needs to be opened from which time point Begin to synthesize;After informing, local synthesis starts locally to be synthesized according to the source audio of each member, and now server is proceeded by From the packet parsing that local source audio is sended over, and carry out synthetic operation;When being synthesized to certain time point, it is not necessary to carrying out The member of synthetic operation does notice and does other operations;But must notify that other members are local to the local synthesis of other members Synthesis is after the operation that the member for carrying out synthetic operation is muted is received, it is necessary to timely responds, will carry out synthesis behaviour The source data for making member is substituted for mute audio.Because audio conference server also has network delay to local, therefore notify In data, also to inform that carry out the member of synthetic operation is muted period in which, then local synthetic model will be synthesized The data that the member of operation starts silence period synchronize operation treatment.
With reference to an instantiation, the present invention will be further described:
1st, the framework of audio conferencing sound intermediate frequency synthesis
The resultant frame of the audio conferencing of main flow is illustrated as shown in Fig. 2 simply doing here:In Fig. 2 examples, a meeting View has 4 participants, and each participant has a say in a meeting.The audio that user A can hear is user B, C, D speech Audio after audio mixing;The audio that user B hears is the audio after user A, C, D speech audio mixing, and equally, what user C heard is to use Audio after A, B, D the speech audio mixing of family;User D is then to hear the audio after user A, B, C speech audio mixing.And if carrying out audio Meeting acoustical testing, usually uses the source audio that some are fixed.Such as party A-subscriber uses party B-subscriber with the source audio of party A-subscriber, party B-subscriber Source audio, C user using C user source audio, D user then using D user source audio.
2nd, construct source audio and test the framework of assessment
Above, have been able to get the ABCD4 source audio of user, a core of Current patents application is exactly The Composite tone close with target audio is constructed based on source audio, then after Composite tone and target audio are carried out into PESQ calculating, Can obtain the tonequality situation of each participant.As shown in figure 3, under the framework, for synthesis or fairly simple, i.e.,:
The Composite tone of party A-subscriber is:The source audio of party B-subscriber, the source audio of C user, the source audio of D user are mixed Audio;
The Composite tone of party B-subscriber is:The source audio of party A-subscriber, the source audio of C user, the source audio of D user are mixed Audio;
The Composite tone of C user is:The source audio of party A-subscriber, the source audio of party B-subscriber, the source audio of D user are mixed Audio;
The Composite tone of D user is:The source audio of party A-subscriber, the source audio of party B-subscriber, the source audio of C user are mixed Audio;But network transmission time delay must also be related to, and correspond to some composite characters of audio conferencing:Such as certain into The gain of member is enhanced, and the gain of some members is lowered and waits some specially treateds, therefore this is when be accomplished by testing, audio Meeting is required to inform the source audio of local synthesis goes how to construct.
3rd, Conference control relevant control sequential
Such as the sequential of Fig. 4, the control situation occurred in one meeting of test frame of simple illustration.Meeting first starts Afterwards, audio conference server needs locally to synthesize to each user and does a time synchronized, informs that local synthesis is needed from which Time point starts synthesis.After informing, each local synthesis starts locally to be synthesized according to the source audio of each user, now sound Frequency Conference server also begins to the packet parsing for sended over from local source audio, and is synthesized.When being synthesized to certain At time point, after C user is muted, the sound of oneself can not be heard due to C user, therefore C need not notify it to C user Do any operation.But must be notified to the local synthesis of party A-subscriber and party B-subscriber.Locally synthesis is receiving C use to AB user After the operation that family is muted, it is necessary to timely respond, the source data of C user is substituted for mute audio.And due to audio Conference server also has a network delay to local, therefore notifies data, also to inform that C user is muted period in which. Then the data that C user starts silence period are synchronized operation treatment by local synthetic model.
4th, the tonequality assessment of single user
As shown in figure 5, be exactly treatment of the user for tonequality, the final assessment of tonequality is in user side (i.e. user Locally) carry out calculating and evaluate what is come, after audio conference server sends back the RTP messages of synthesis, locally need audio The RTP data that Conference server is beamed back, that is, the target data that audio conference server is beamed back.And with the source number of local synthesis According to.PESQ calculating is carried out using recommendation P.862, it is possible to calculate current real-time tonequality.And give the numerical value of tonequality Record.
Although the foregoing describing specific embodiment of the invention, those familiar with the art should manage Solution, the specific embodiment described by us is merely exemplary, and rather than for the restriction to the scope of the present invention, is familiar with this The technical staff in field should be covered of the invention in the equivalent modification and change made according to spirit of the invention In scope of the claimed protection.

Claims (4)

1. a kind of acoustical testing method of audio conferencing, it is characterised in that:Methods described is:
Obtain the corresponding source audio of each member for participating in audio conferencing;
Each member constructs the Composite tone close with target audio by audio system based on source audio, i.e. each member button After except the source audio of itself, Composite tone is formed after the source audio of other members is mixed, the Composite tone for each into Member itself corresponding Composite tone;
The target audio data that itself corresponding Composite tone and audio conference server send are used ITU-T by each member P.862 recommendation carries out PESQ and calculates to obtain current real-time tonequality, and the records of values of tonequality is got off to complete tonequality Test.
2. the acoustical testing method of a kind of audio conferencing according to claim 1, it is characterised in that:Methods described is specific For:After meeting starts, the corresponding source audio of each member for participating in audio conferencing is obtained;Audio conference server is needed to each Locally a time synchronized is done in synthesis to member, informs that local synthesis needs to synthesize since which time point;It is local to close after informing Into starting locally to be synthesized according to the source audio of each member, now server is proceeded by and sended over from local source audio Packet parsing, and carry out synthetic operation;When being synthesized to certain time point, it is not necessary to notified to the member for carrying out synthetic operation Do other operations;But must be notified to the local synthesis of other members, locally synthesis is synthesized receiving other members After the operation that the member of operation is muted, it is necessary to timely respond, the source data for carrying out synthetic operation member is substituted for Mute audio.
3. the acoustical testing method of a kind of audio conferencing according to claim 2, it is characterised in that:Due to audio conferencing clothes Business device also has a network delay, therefore notice data to local, and which also to inform carries out the member of synthetic operation in period It is muted, then local synthetic model synchronizes at operation the data that the member for carrying out synthetic operation starts silence period Reason.
4. the acoustical testing method of a kind of audio conferencing according to claim 1, it is characterised in that:Audio conference server After the RTP messages of synthesis are sent back, the RTP message datas that local reception audio conference server is beamed back, the RTP message datas The target audio data that as audio conference server sends.
CN201611004655.1A 2016-11-15 2016-11-15 A kind of acoustical testing method of audio conferencing Active CN106782610B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201611004655.1A CN106782610B (en) 2016-11-15 2016-11-15 A kind of acoustical testing method of audio conferencing

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201611004655.1A CN106782610B (en) 2016-11-15 2016-11-15 A kind of acoustical testing method of audio conferencing

Publications (2)

Publication Number Publication Date
CN106782610A true CN106782610A (en) 2017-05-31
CN106782610B CN106782610B (en) 2019-09-20

Family

ID=58968339

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201611004655.1A Active CN106782610B (en) 2016-11-15 2016-11-15 A kind of acoustical testing method of audio conferencing

Country Status (1)

Country Link
CN (1) CN106782610B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111405122A (en) * 2020-03-18 2020-07-10 苏州科达科技股份有限公司 Audio call testing method, device and storage medium

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040002852A1 (en) * 2002-07-01 2004-01-01 Kim Doh-Suk Auditory-articulatory analysis for speech quality assessment
CN101217759A (en) * 2007-12-28 2008-07-09 中国移动通信集团浙江有限公司 A ringtone quality detecting method of CRBT
US20090204922A1 (en) * 2008-02-13 2009-08-13 Microsoft Corporation Techniques to manage communications resources for a multimedia conference event
CN103067217A (en) * 2012-12-14 2013-04-24 北京思特奇信息技术股份有限公司 Indicating system and method of communication network service quality
CN103581934A (en) * 2012-07-19 2014-02-12 中兴通讯股份有限公司 Terminal voice quality evaluation method and terminal
CN103841275A (en) * 2013-07-24 2014-06-04 同济大学 Interactive audio experience quality evaluation platform and method based on QoS
US20160028881A1 (en) * 2014-07-22 2016-01-28 International Business Machines Corporation MONITORING VOICE OVER INTERNET PROTOCOL (VoIP) QUALITY DURING AN ONGOING CALL
CN105788611A (en) * 2016-02-25 2016-07-20 成都普创通信技术股份有限公司 Audio quality online monitoring system

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040002852A1 (en) * 2002-07-01 2004-01-01 Kim Doh-Suk Auditory-articulatory analysis for speech quality assessment
CN101217759A (en) * 2007-12-28 2008-07-09 中国移动通信集团浙江有限公司 A ringtone quality detecting method of CRBT
US20090204922A1 (en) * 2008-02-13 2009-08-13 Microsoft Corporation Techniques to manage communications resources for a multimedia conference event
CN103581934A (en) * 2012-07-19 2014-02-12 中兴通讯股份有限公司 Terminal voice quality evaluation method and terminal
CN103067217A (en) * 2012-12-14 2013-04-24 北京思特奇信息技术股份有限公司 Indicating system and method of communication network service quality
CN103841275A (en) * 2013-07-24 2014-06-04 同济大学 Interactive audio experience quality evaluation platform and method based on QoS
US20160028881A1 (en) * 2014-07-22 2016-01-28 International Business Machines Corporation MONITORING VOICE OVER INTERNET PROTOCOL (VoIP) QUALITY DURING AN ONGOING CALL
CN105788611A (en) * 2016-02-25 2016-07-20 成都普创通信技术股份有限公司 Audio quality online monitoring system

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111405122A (en) * 2020-03-18 2020-07-10 苏州科达科技股份有限公司 Audio call testing method, device and storage medium
CN111405122B (en) * 2020-03-18 2021-09-24 苏州科达科技股份有限公司 Audio call testing method, device and storage medium

Also Published As

Publication number Publication date
CN106782610B (en) 2019-09-20

Similar Documents

Publication Publication Date Title
Jelassi et al. Quality of experience of VoIP service: A survey of assessment approaches and open issues
US10044854B2 (en) Method and system for providing captioned telephone service with automated speech recognition
Möller Assessment and prediction of speech quality in telecommunications
Takahashi et al. Perceptual QoS assessment technologies for VoIP
CN101689365B (en) Method of controlling a video conference
CN100546322C (en) Chat and tele-conferencing system with the translation of Text To Speech and speech-to-text
CN1946107B (en) Interactive telephony trainer and exerciser
US8358599B2 (en) System for providing audio highlighting of conference participant playout
US8208645B2 (en) System and method for harmonizing calibration of audio between networked conference rooms
US20120259924A1 (en) Method and apparatus for providing summary information in a live media session
Woods et al. Predicting the effect of hearing loss and audibility on amplified speech reception in a multi-talker listening scenario
EP2973559B1 (en) Audio transmission channel quality assessment
CN105991964A (en) Method and apparatus for broadcasting dynamic information in multimedia conference
CN104954727B (en) Audio-visual synchronization control device and method
US20110313765A1 (en) Conversational Subjective Quality Test Tool
CN105991854A (en) System and method for visualizing VOIP teleconference on intelligent terminal
Daengsi et al. QoE modeling for voice over IP: simplified E-model enhancement utilizing the subjective MOS prediction model: a case of G. 729 and Thai users
Gallardo et al. Human speaker identification of known voices transmitted through different user interfaces and transmission channels
Skowronek et al. Assessment of cognitive load, speech communication quality and quality of experience for spatial and non-spatial audio conferencing calls
KR101511795B1 (en) Notification of dropped audio in a teleconference call
JP2015092347A (en) Emotion-expressing animation face display system, method and program
CN106782610B (en) A kind of acoustical testing method of audio conferencing
Berndtsson et al. Methods for human-centered evaluation of MediaSync in real-time communication
KR20150042628A (en) Method and apparatus of providing a distributed telepresense service
JP2013219495A (en) Emotion-expressing animation face display system, method, and program

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CP03 Change of name, title or address

Address after: Software Park Siming District of Xiamen city in Fujian province 361000 two sunrise Road No. 56 unit 1001

Patentee after: Fujian Xingwang Intelligent Technology Co., Ltd

Address before: Software Park Siming District of Xiamen city in Fujian province 350000 two sunrise Road No. 56 unit 1001

Patentee before: FUJIAN STAR-NET WISDOM TECHNOLOGY Co.,Ltd.

CP03 Change of name, title or address