CN106782610B - A kind of acoustical testing method of audio conferencing - Google Patents

A kind of acoustical testing method of audio conferencing Download PDF

Info

Publication number
CN106782610B
CN106782610B CN201611004655.1A CN201611004655A CN106782610B CN 106782610 B CN106782610 B CN 106782610B CN 201611004655 A CN201611004655 A CN 201611004655A CN 106782610 B CN106782610 B CN 106782610B
Authority
CN
China
Prior art keywords
audio
source
local
synthesis
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201611004655.1A
Other languages
Chinese (zh)
Other versions
CN106782610A (en
Inventor
洪剑平
陈锦辉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fujian Xingwang Intelligent Technology Co., Ltd
Original Assignee
Fujian Star-Net Wisdom Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fujian Star-Net Wisdom Technology Co Ltd filed Critical Fujian Star-Net Wisdom Technology Co Ltd
Priority to CN201611004655.1A priority Critical patent/CN106782610B/en
Publication of CN106782610A publication Critical patent/CN106782610A/en
Application granted granted Critical
Publication of CN106782610B publication Critical patent/CN106782610B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/60Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for measuring the quality of voice signals

Landscapes

  • Engineering & Computer Science (AREA)
  • Quality & Reliability (AREA)
  • Computational Linguistics (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Telephonic Communication Services (AREA)

Abstract

The present invention provides a kind of acoustical testing method of audio conferencing, the method are as follows: obtains the corresponding source audio of each member for participating in audio conferencing;Each member is based on source audio by audio system and constructs and Composite tone similar in target audio, after i.e. each member deducts the source audio of itself, Composite tone is formed after the source audio of other members is mixed, which is the corresponding Composite tone of each member itself;By itself corresponding Composite tone and the target audio data of audio conference server transmission, using ITU-T, P.862 recommendation progress PESQ calculating obtains real-time sound quality currently to each member, and the records of values of sound quality is got off, to complete the test of sound quality.This method test is simple, and test and comparison is accurate.

Description

A kind of acoustical testing method of audio conferencing
Technical field
The present invention relates to technical field of communication equipment more particularly to a kind of acoustical testing methods of audio conferencing.
Background technique
In existing acoustical testing model, that the most commonly used is PESQ (Perceptual evaluation of Speech quality) i.e.: the test method of subjective speech quality assessment.This method is that P.862 recommendation provides ITU-T Objective MOS (Mean Opinion Score, mean subjective opinion point) value evaluation method.Use the test method of PESQ, it is necessary to Active audio frequency and target audio.By calculating the PESQ of source audio and target audio, to judge the front and back difference of sound quality, into And obtain MOS value.MOS value is exactly the quantized result of current corresponding sound quality.
In audio conferencing, each member is added in the same meeting, can only hear oneself non-audio, that is, for For audio conferencing, what each member heard is the audio mixing result of remaining member.This also just brings up the sound that each member hears Frequency is special.Everyone is different.
Because can parliament the audio of each participant is synthesized, and what each participant heard is remaining and meeting The audio mixing of person is as a result, the audio that namely everyone hears is not quite similar.Therefore the sound quality of testing audio meeting can be held very much The easy source audio for getting each participant, but target audio has been the data being mixed, under normal circumstances with source Audio has had very big difference, therefore is not easy to assess based on the ITU-T sound quality that P.862 recommendation carries out PESQ, if base In ITU-T, P.862 recommendation carries out tone frequency estimation, and sound quality is relatively poor certainly.
It for above situation, can be easy to get the source audio of each participant, but more difficult get mesh Mark with phonetic symbols frequency, or can so think, it is easy to get target audio, but source audio can not be corresponding with target audio, Therefore it is configured to meet the source audio of target audio with regard to needs.
Summary of the invention
The technical problem to be solved in the present invention is to provide a kind of acoustical testing method of audio conferencing, this method test Simply, and test and comparison is accurate.
The present invention is implemented as follows: a kind of acoustical testing method of audio conferencing, the method are as follows:
Obtain the corresponding source audio of each member for participating in audio conferencing;
Each member by audio system be based on source audio construct with Composite tone similar in target audio, i.e., it is each at After member deducts the source audio of itself, Composite tone is formed after the source audio of other members is mixed, which is every The corresponding Composite tone of a member itself;
Each member uses the target audio data that itself corresponding Composite tone and audio conference server are sent P.862 recommendation carries out the current real-time sound quality of PESQ calculating acquisition to ITU-T, and the records of values of sound quality is got off, to complete The test of sound quality.
Further, the method specifically: after meeting starts, obtain the corresponding source of each member for participating in audio conferencing Audio;Audio conference server, which needs locally to synthesize to each member, does a time synchronization, informs that local synthesis is needed from which A time point starts to synthesize;After informing, local synthesis starts to carry out local synthesis according to the source audio of each member, services at this time Device starts the packet parsing send over from local source audio, and carries out synthetic operation;When being synthesized to some time point, no It needs to do notice to the member for carrying out synthetic operation and does other operations;But it must be notified to the local synthesis of other members, Its member locally synthesizes after receiving the operation that the member for carrying out synthetic operation is muted, it is necessary to timely responds, it will The source data for carrying out synthetic operation member is substituted for mute audio.
Further, network delay is also had to local due to audio conference server, in notification data, also accused Know that the member for carrying out synthetic operation is muted period in which, then local synthetic model starts the member for carrying out synthetic operation The data of silence period synchronize operation processing.
Further, after audio conference server sends back the RTP message of synthesis, local reception audio conference server The RTP message data sent back to, the RTP message data are the target audio data that audio conference server is sent.
The present invention has the advantage that 1, using the characteristic of existing audio conferencing, remove the source sound for constructing suitable target audio Frequently.2, the sound quality situation of each participant is calculated based on ITU-TP.862 recommendation using the source audio of construction.3, addition pair Some details parameters such as network delay carry out synthesis adjustment to the source audio of construction, reach more accurate acoustical testing.
Detailed description of the invention
The present invention is further illustrated in conjunction with the embodiments with reference to the accompanying drawings.
Fig. 1 is the method for the present invention flow diagram.
Fig. 2 is the frame diagram of audio conferencing sound intermediate frequency synthesis.
Fig. 3 is the frame diagram for constructing source audio and carrying out test assessment.
Fig. 4 is Conference control relevant control timing diagram.
Fig. 5 is the sound quality estimation flow schematic diagram of single user.
Specific embodiment
Refering to Figure 1, a kind of acoustical testing method of audio conferencing, the method are as follows: obtain and participate in audio conferencing The corresponding source audio of each member;
Each member by audio system be based on source audio construct with Composite tone similar in target audio, i.e., it is each at After member deducts the source audio of itself, Composite tone is formed after the source audio of other members is mixed, which is every The corresponding Composite tone of a member itself;
Each member uses the target audio data that itself corresponding Composite tone and audio conference server are sent P.862 recommendation carries out the current real-time sound quality of PESQ calculating acquisition to ITU-T, and the records of values of sound quality is got off, to complete The test of sound quality.
Wherein, after audio conference server sends back the RTP message of synthesis, local reception audio conference server is sent back to RTP message data, the RTP message data be audio conference server send target audio data.
The method specifically: after meeting starts, obtain the corresponding source audio of each member for participating in audio conferencing;Audio Conference server, which needs locally to synthesize to each member, does a time synchronization, informs that local synthesis needs to open from which time point Begin to synthesize;After informing, local synthesis starts to carry out local synthesis according to the source audio of each member, and server starts to carry out at this time The packet parsing sended over from local source audio, and carry out synthetic operation;When being synthesized to some time point, do not need to progress The member of synthetic operation does notice and does other operations;But it must notify that other members are local to the local synthesis of other members It synthesizes after receiving the operation that the member for carrying out synthetic operation is muted, it is necessary to timely respond, synthesis behaviour will be carried out The source data for making member is substituted for mute audio.Since audio conference server also has network delay to local, notify In data, also to inform that the member for carrying out synthetic operation is muted period in which, then local synthetic model will synthesize The data that the member of operation starts silence period synchronize operation processing.
Below with reference to a specific example, the present invention will be further described:
1, the frame of audio conferencing sound intermediate frequency synthesis
The resultant frame of the audio conferencing of mainstream illustrates as shown in Fig. 2, simply doing here: in Fig. 2 example, a meeting View has 4 participants, and each participant has a say in a meeting.The audio that user A can be heard is user B, C, D speech Audio after audio mixing;The audio that user B is heard is the audio after user A, C, D speech audio mixing, and equally, what user C was heard is to use Audio after family A, B, D speech audio mixing;User D is then the audio heard after user A, B, C speech audio mixing.And if carrying out audio Meeting acoustical testing usually uses the source audio of some fixations.For example party A-subscriber's source audio of party A-subscriber, party B-subscriber use party B-subscriber Source audio, C user uses the source audio of C user, and D user then uses the source audio of D user.
2, it constructs source audio and carries out the frame of test assessment
Above, have been able to get the source audio of ABCD4 user, a core of Current patents application is exactly Based on source audio construction and Composite tone similar in target audio, then after Composite tone and target audio are carried out PESQ calculating, The sound quality situation of each participant can be obtained.As shown in figure 3, under the frame, for synthesis or fairly simple, it may be assumed that
The Composite tone of party A-subscriber are as follows: the source audio of party B-subscriber, the source audio of C user, the source audio of D user mix Audio;
The Composite tone of party B-subscriber are as follows: the source audio of party A-subscriber, the source audio of C user, the source audio of D user mix Audio;
The Composite tone of C user are as follows: the source audio of party A-subscriber, the source audio of party B-subscriber, the source audio of D user mix Audio;
The Composite tone of D user are as follows: the source audio of party A-subscriber, the source audio of party B-subscriber, the source audio of C user mix Audio;But network transmission delay, and some composite characters of corresponding audio conferencing must also be related to: such as some at The gain of member is enhanced, some specially treateds such as the gain of certain members is lowered, therefore when this just needs to test, audio Meeting is required to inform that the source audio locally synthesized goes how to construct.
3, Conference control relevant control timing
Such as the timing of Fig. 4, the control situation that occurs in the meeting of one test frame of simple illustration.Meeting starts first Afterwards, audio conference server, which needs locally to synthesize to each user, does a time synchronization, informs that local synthesis is needed from which Time point starts to synthesize.After informing, each local synthesis starts to carry out local synthesis according to the source audio of each user, at this time sound Frequency Conference server also starts the packet parsing send over from local source audio, and is synthesized.When being synthesized to some Time point, after C user is muted, since C user can not hear that the sound of oneself, C do not need to notify it to C user Do any operation.But it must be notified to the local synthesis of party A-subscriber and party B-subscriber.Locally synthesis is receiving C use to AB user After the operation that family is muted, it is necessary to timely respond, the source data of C user is substituted for mute audio.And due to audio Conference server also has network delay to local, therefore in notification data, also to inform that C user is muted period in which. Then the data that C user starts silence period are synchronized operation processing by local synthetic model.
4, the sound quality assessment of single user
As shown in figure 5, being exactly processing of the user for sound quality, the final assessment of sound quality is in user side (i.e. user Local) carry out what calculating evaluated, after audio conference server sends back the RTP message of synthesis, locally need audio The RTP data that Conference server is sent back to, that is, the target data that audio conference server is sent back to.And with the source number that locally synthesizes According to.PESQ calculating is carried out using recommendation P.862, so that it may calculate current real-time sound quality.And the numerical value of sound quality is given It records.
Although specific embodiments of the present invention have been described above, those familiar with the art should be managed Solution, we are merely exemplary described specific embodiment, rather than for the restriction to the scope of the present invention, it is familiar with this The technical staff in field should be covered of the invention according to modification and variation equivalent made by spirit of the invention In scope of the claimed protection.

Claims (3)

1. a kind of acoustical testing method of audio conferencing, it is characterised in that: the method are as follows:
Obtain the corresponding source audio of each member for participating in audio conferencing;
Each member is based on source audio by audio system and constructs and Composite tone similar in target audio, i.e., each member's button Except forming Composite tone after the source audio of other members is mixed after itself source audio, the Composite tone be it is each at Itself corresponding Composite tone of member;
The target audio data that itself corresponding Composite tone and audio conference server are sent are used ITU-T by each member P.862 recommendation carries out PESQ and calculates the current real-time sound quality of acquisition, and the records of values of sound quality is got off, to complete sound quality Test;
After meeting starts, the corresponding source audio of each member for participating in audio conferencing is obtained;Audio conference server is needed to every Locally a time synchronization is done in synthesis to a member, informs that local synthesis needs to synthesize since which time point;It is local after informing Synthesis starts to carry out local synthesis according to the source audio of each member, and server starts be transmitted across from local source audio at this time The packet parsing come, and carry out synthetic operation;When being synthesized to some time point, do not need to do to the member for carrying out synthetic operation logical Know and does other operations;But it must be notified to the local synthesis of other members, other members locally close receiving by synthesis After the operation being muted at the member of operation, it is necessary to timely respond, the source data for carrying out synthetic operation member is replaced At mute audio.
2. a kind of acoustical testing method of audio conferencing according to claim 1, it is characterised in that: since audio conferencing takes Business device also has network delay to local, therefore in notification data, also to inform the member of progress synthetic operation in which period It is muted, then local synthetic model synchronizes the data that the member for carrying out synthetic operation starts silence period at operation Reason.
3. a kind of acoustical testing method of audio conferencing according to claim 1, it is characterised in that: audio conference server After the RTP message of synthesis is sent back, the RTP message data that local reception audio conference server is sent back to, the RTP message data The as target audio data of audio conference server transmission.
CN201611004655.1A 2016-11-15 2016-11-15 A kind of acoustical testing method of audio conferencing Active CN106782610B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201611004655.1A CN106782610B (en) 2016-11-15 2016-11-15 A kind of acoustical testing method of audio conferencing

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201611004655.1A CN106782610B (en) 2016-11-15 2016-11-15 A kind of acoustical testing method of audio conferencing

Publications (2)

Publication Number Publication Date
CN106782610A CN106782610A (en) 2017-05-31
CN106782610B true CN106782610B (en) 2019-09-20

Family

ID=58968339

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201611004655.1A Active CN106782610B (en) 2016-11-15 2016-11-15 A kind of acoustical testing method of audio conferencing

Country Status (1)

Country Link
CN (1) CN106782610B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111405122B (en) * 2020-03-18 2021-09-24 苏州科达科技股份有限公司 Audio call testing method, device and storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101217759A (en) * 2007-12-28 2008-07-09 中国移动通信集团浙江有限公司 A ringtone quality detecting method of CRBT
CN103067217A (en) * 2012-12-14 2013-04-24 北京思特奇信息技术股份有限公司 Indicating system and method of communication network service quality
CN103581934A (en) * 2012-07-19 2014-02-12 中兴通讯股份有限公司 Terminal voice quality evaluation method and terminal
CN103841275A (en) * 2013-07-24 2014-06-04 同济大学 Interactive audio experience quality evaluation platform and method based on QoS
CN105788611A (en) * 2016-02-25 2016-07-20 成都普创通信技术股份有限公司 Audio quality online monitoring system

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7165025B2 (en) * 2002-07-01 2007-01-16 Lucent Technologies Inc. Auditory-articulatory analysis for speech quality assessment
US8713440B2 (en) * 2008-02-13 2014-04-29 Microsoft Corporation Techniques to manage communications resources for a multimedia conference event
US9325838B2 (en) * 2014-07-22 2016-04-26 International Business Machines Corporation Monitoring voice over internet protocol (VoIP) quality during an ongoing call

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101217759A (en) * 2007-12-28 2008-07-09 中国移动通信集团浙江有限公司 A ringtone quality detecting method of CRBT
CN103581934A (en) * 2012-07-19 2014-02-12 中兴通讯股份有限公司 Terminal voice quality evaluation method and terminal
CN103067217A (en) * 2012-12-14 2013-04-24 北京思特奇信息技术股份有限公司 Indicating system and method of communication network service quality
CN103841275A (en) * 2013-07-24 2014-06-04 同济大学 Interactive audio experience quality evaluation platform and method based on QoS
CN105788611A (en) * 2016-02-25 2016-07-20 成都普创通信技术股份有限公司 Audio quality online monitoring system

Also Published As

Publication number Publication date
CN106782610A (en) 2017-05-31

Similar Documents

Publication Publication Date Title
Jelassi et al. Quality of experience of VoIP service: A survey of assessment approaches and open issues
EP2901669B1 (en) Near-end indication that the end of speech is received by the far end in an audio or video conference
Takahashi et al. Perceptual QoS assessment technologies for VoIP
US7664231B2 (en) Method and device for quality evaluation of an audio signal and device and method for obtaining a quality evaluation result
TW200828867A (en) Communication system
EP2973559B1 (en) Audio transmission channel quality assessment
CN105791738A (en) Method and device for adjusting video window in video conference
Daengsi et al. QoE modeling for voice over IP: simplified E-model enhancement utilizing the subjective MOS prediction model: a case of G. 729 and Thai users
CN104575521A (en) Method for evaluating voice quality of LTE communication system
CN106782610B (en) A kind of acoustical testing method of audio conferencing
DE60132196T2 (en) TEST SIGNAL
DE60118922T2 (en) MEASURE THE TRUE LANGUAGE QUALITY DURING OPERATION BY MEASURING OBJECTIVE ERROR PARAMETER
KR101511795B1 (en) Notification of dropped audio in a teleconference call
EP2194525A1 (en) Conversational subjective quality test tool
Möller et al. Telephone speech quality prediction: towards network planning and monitoring models for modern network scenarios
DE60307737T2 (en) Notification for waiting calls
CN100516788C (en) Acoustic automatic gaining effect testing method and system
CN109246664A (en) A kind of tone testing method and device
Skowronek et al. Quality assessment of asymmetric multiparty telephone conferences: a systematic method from technical degradations to perceived impairments.
Daengsi VoIP quality measurement: recommendation of MOS and enhanced objective measurement method for standard Thai spoken language
Daengsi et al. Speech quality assessment of VoIP: G. 711 VS G. 722 based on interview tests with Thai users
Wuttidittachotti et al. Subjective MOS model and simplified E-model enhancement for Skype associated with packet loss effects: a case using conversation-like tests with Thai users
Gallardo et al. Variable voice likability affecting subjective speech quality assessments
DE602004004824T2 (en) Automatic treatment of conversation groups
CN108809915B (en) Session management method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CP03 Change of name, title or address

Address after: Software Park Siming District of Xiamen city in Fujian province 361000 two sunrise Road No. 56 unit 1001

Patentee after: Fujian Xingwang Intelligent Technology Co., Ltd

Address before: Software Park Siming District of Xiamen city in Fujian province 350000 two sunrise Road No. 56 unit 1001

Patentee before: FUJIAN STAR-NET WISDOM TECHNOLOGY Co.,Ltd.

CP03 Change of name, title or address