CN117115719A - Method for determining video conference quality - Google Patents

Method for determining video conference quality Download PDF

Info

Publication number
CN117115719A
CN117115719A CN202311367912.8A CN202311367912A CN117115719A CN 117115719 A CN117115719 A CN 117115719A CN 202311367912 A CN202311367912 A CN 202311367912A CN 117115719 A CN117115719 A CN 117115719A
Authority
CN
China
Prior art keywords
target video
data analysis
analysis unit
qualified
quality
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202311367912.8A
Other languages
Chinese (zh)
Inventor
石金川
朱正辉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong Baolun Electronics Co ltd
Original Assignee
Guangdong Baolun Electronics Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong Baolun Electronics Co ltd filed Critical Guangdong Baolun Electronics Co ltd
Priority to CN202311367912.8A priority Critical patent/CN117115719A/en
Publication of CN117115719A publication Critical patent/CN117115719A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/761Proximity, similarity or dissimilarity measures
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/57Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for processing of video signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems

Abstract

The invention relates to the field of data processing, in particular to a method for determining video conference quality, which comprises the following steps: the data acquisition unit acquires target quality data aiming at the target video; the data acquisition unit performs multi-frame acquisition on the target video to acquire a target video frame and judges whether the target video frame is qualified or not according to the similarity between the target video frame and the reference video frame; the data analysis unit uniformly divides the target video frame into four sub-areas and sequentially detects the similarity between each sub-area and the corresponding part of the reference video frame; the data analysis unit counts the number of the qualified target video frames and calculates the ratio of the number of the qualified target video frames to the total number of the target video frames, and marks the ratio as a reference ratio, and the data analysis unit determines whether the quality of the target video image is qualified according to the reference ratio; the data analysis unit compares the voice definition of the target video with a preset voice definition to judge whether the voice definition of the target video is qualified or not; the invention improves the determination accuracy of the video conference quality.

Description

Method for determining video conference quality
Technical Field
The invention relates to the field of data processing, in particular to a method for determining video conference quality.
Background
The video conference quality refers to the performance of the video and audio in terms of definition, stability, real-time performance and the like in the video conference process. The importance of determining video conference quality is becoming increasingly important in modern tele-working and tele-collaboration environments. Video conferencing is an important tool for real-time communication and collaboration. Good video conference quality can ensure clear and accurate voice and image transmission between participants, thereby facilitating efficient communication and understanding. How to make efficient analytical decisions regarding video quality and audio quality of video conferences is a problem that the following technicians are urgent to address.
Chinese patent publication No. CN115174842a discloses a method, related apparatus and system for determining quality of a video conference, where the method includes inputting N sets of flow characteristic data obtained in N sampling periods into a quality judgment model, to obtain a quality judgment result of the video conference in a target period. According to the technical scheme, the quality of the video conference can be judged according to the flow characteristic data in the network. In this way, even without a conference screen, the video conference screen quality during the video conference can be determined, thereby facilitating the video conference provider to optimize the video conference service. It can be seen that the above technical solution has the following problems: the determined data volume is large, and the quality determination method cannot be adjusted according to the actual detection result, so that the method for determining the quality of the video conference has poor flexibility.
Disclosure of Invention
Therefore, the invention provides a method for determining the quality of a video conference, which is used for overcoming the defect that the flexibility of the method for determining the quality of the video conference is poor because the quality judging method cannot be adjusted according to the actual detection result in the prior art.
To achieve the above object, the present invention provides a method for determining video conference quality, comprising:
the data acquisition unit acquires target quality data aiming at the target video;
the data acquisition unit performs multi-frame acquisition on the target video to acquire a target video frame and judges whether the target video frame is qualified or not according to the similarity between the target video frame and the reference video frame;
the data analysis unit performs partition processing on the target video frame under the first data analysis condition, uniformly divides the target video frame into four sub-areas and sequentially performs similarity detection on the sub-areas and the corresponding parts of the reference video frame;
the data analysis unit counts the number of qualified target video frames under the second data analysis condition, calculates the ratio of the number of the qualified target video frames to the total number of the target video frames, marks the ratio as a reference ratio, and determines whether the quality of the target video image is qualified or not according to the reference ratio;
the data analysis unit compares the voice definition of the target video with a preset voice definition under a fourth data analysis condition to judge whether the voice definition of the target video is qualified or not;
the first data analysis condition is that the similarity is in a second preset similarity range; the second data analysis condition is that whether the target video frame is qualified or not is determined to be finished; and the fourth data analysis condition is that whether the target video image quality is qualified or not is judged to be finished.
Further, the data acquisition unit acquires a target video frame by aiming at the target video in a multi-frame manner, and the data analysis unit judges whether the target video frame is qualified or not according to the similarity between the target video frame and the reference video frame;
if the similarity is in a first preset similarity range, the data analysis unit judges that the target video frame is unqualified;
if the similarity is in a second preset similarity range, the data analysis unit judges that the target video frame is possibly qualified and performs partition comparison;
and if the similarity is in a third preset similarity range, the data analysis unit judges that the target video frame is qualified.
Further, the data analysis unit performs partition processing on the target video frame under the first data analysis condition, uniformly divides the target video frame into four sub-areas, and sequentially performs similarity detection on the sub-areas and the corresponding parts of the reference video frame;
if the number of the qualified sub-areas is larger than the number of the preset sub-areas, the data analysis unit judges that the target video frame is qualified;
and if the number of the qualified sub-areas is smaller than or equal to the number of the preset sub-areas, the data analysis unit judges that the target video frame is unqualified.
Further, the data analysis unit counts the number of qualified target video frames under the second data analysis condition, calculates the ratio of the number of the qualified target video frames to the total number of the target video frames, marks the ratio as a reference ratio, and determines whether the quality of the target video image is qualified or not according to the reference ratio;
if the reference ratio is in a first preset ratio range, the data analysis unit judges that the quality of the target video image is qualified;
if the reference ratio is in the second preset ratio range, the data analysis unit judges that the quality of the target video image is possibly disqualified;
if the reference ratio is in a third preset ratio range, the data analysis unit judges that the quality of the target video image is unqualified;
and determining whether the target video frame is qualified or not according to the second data analysis condition.
Further, the data analysis unit judges the re-extraction quantity of the target video frames according to the reference ratio of the target video under a third data analysis condition;
the extraction quantity of the target video frames and the reference ratio of the target video are in a negative correlation relationship;
and the third data analysis condition is that if the reference ratio is in a second preset ratio range, the data analysis unit judges that the target video image quality is possibly disqualified.
Further, the data analysis unit compares the voice definition of the target video with a preset voice definition under a fourth data analysis condition to judge whether the voice definition of the target video is qualified or not;
if the voice definition of the target video is smaller than the first preset voice definition, the data analysis unit judges that the voice definition of the target video is unqualified;
if the voice definition of the target video is greater than or equal to the first preset voice definition and less than the second preset voice definition, the data analysis unit judges that the voice definition of the target video is possibly qualified;
if the voice definition of the target video is greater than or equal to the second preset voice definition, the data analysis unit judges that the voice definition of the target video is qualified;
and the fourth data analysis condition is that whether the target video image quality is qualified or not is judged to be finished.
Further, the calculation formula of the voice definition is as follows:
K=(N1×α1+N2×α2)/N
wherein N1 is a paragraph duration of sound blur in the target video, N2 is a paragraph duration of sound too large or too small in the target video, N is a total duration of the target video, α1 is a first weight coefficient, α2 is a second weight coefficient, 0 < α1 < α2 < 1, α1+α2=1.
Further, the data analysis unit is provided with a sound allowable threshold range, if the sound decibel frequency in the target video is larger than the sound allowable threshold range, the data analysis unit judges that the current sound of the target video is too large, and if the sound decibel frequency in the target video is smaller than the sound allowable threshold range, the data analysis unit judges that the current sound of the target video is too small.
Further, the display unit transmits the judging information of the data analysis unit to the user side under the fifth data analysis condition;
and the fifth data analysis condition is that the sound definition and video quality of the target video are judged to be finished.
Further, the display unit is connected with the data analysis unit through 5g remote communication.
Compared with the prior art, the method has the advantages that the data acquisition unit acquires the target video frames by multi-frame acquisition aiming at the target video, the data analysis unit judges whether the target video frames are qualified or not according to the similarity between the target video frames and the reference video frames, and does not perform quality analysis of the full video section, so that the data processing amount is reduced, the data analysis unit carries out partition processing on the target video frames under the first data analysis condition, uniformly divides the target video frames into four sub-areas and sequentially carries out similarity detection on the sub-areas and the corresponding parts of the reference video frames, misjudgment of a system when the similarity is in an intermediate threshold value is avoided, the judgment precision of the system is further improved, the data analysis unit judges the re-extraction amount of the target video frames according to the reference ratio of the target video under the third data analysis condition, the problem of poor data characterization caused by excessively small extracted data amount is avoided, and the judgment precision of the method is further improved.
Drawings
Fig. 1 is a schematic diagram of a method for determining video conference quality according to an embodiment of the present invention.
Detailed Description
In order that the objects and advantages of the invention will become more apparent, the invention will be further described with reference to the following examples; it should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention.
Preferred embodiments of the present invention are described below with reference to the accompanying drawings. It should be understood by those skilled in the art that these embodiments are merely for explaining the technical principles of the present invention, and are not intended to limit the scope of the present invention.
It should be noted that, in the description of the present invention, terms such as "upper," "lower," "left," "right," "inner," "outer," and the like indicate directions or positional relationships based on the directions or positional relationships shown in the drawings, which are merely for convenience of description, and do not indicate or imply that the apparatus or elements must have a specific orientation, be constructed and operated in a specific orientation, and thus should not be construed as limiting the present invention.
Furthermore, it should be noted that, in the description of the present invention, unless explicitly specified and limited otherwise, the terms "mounted," "connected," and "connected" are to be construed broadly, and may be either fixedly connected, detachably connected, or integrally connected, for example; can be mechanically or electrically connected; can be directly connected or indirectly connected through an intermediate medium, and can be communication between two elements. The specific meaning of the above terms in the present invention can be understood by those skilled in the art according to the specific circumstances.
Referring to fig. 1, which is a schematic diagram of a method for determining quality of a video conference according to an embodiment of the present invention, the present invention provides a method for determining quality of a video conference, including:
the data acquisition unit acquires target quality data aiming at the target video;
the data acquisition unit performs multi-frame acquisition on the target video to acquire a target video frame and judges whether the target video frame is qualified or not according to the similarity between the target video frame and the reference video frame;
the data analysis unit performs partition processing on the target video frame under the first data analysis condition, uniformly divides the target video frame into four sub-areas and sequentially performs similarity detection on the sub-areas and the corresponding parts of the reference video frame;
the data analysis unit counts the number of qualified target video frames under the second data analysis condition, calculates the ratio of the number of the qualified target video frames to the total number of the target video frames, marks the ratio as a reference ratio, and determines whether the quality of the target video image is qualified or not according to the reference ratio;
the data analysis unit compares the voice definition of the target video with a preset voice definition under a fourth data analysis condition to judge whether the voice definition of the target video is qualified or not;
the first data analysis condition is that the similarity is in a second preset similarity range; the second data analysis condition is that whether the target video frame is qualified or not is determined to be finished; and the fourth data analysis condition is that whether the target video image quality is qualified or not is judged to be finished.
Specifically, the data acquisition unit performs multi-frame acquisition on the target video to acquire a target video frame, and the data analysis unit judges whether the target video frame is qualified or not according to the similarity between the target video frame and the reference video frame;
if the similarity is in a first preset similarity range, the data analysis unit judges that the target video frame is unqualified;
if the similarity is in a second preset similarity range, the data analysis unit judges that the target video frame is possibly qualified and performs partition comparison;
and if the similarity is in a third preset similarity range, the data analysis unit judges that the target video frame is qualified.
Specifically, the data analysis unit performs partition processing on a target video frame under a first data analysis condition, uniformly divides the target video frame into four sub-areas, and sequentially performs similarity detection on the sub-areas and corresponding parts of a reference video frame;
if the number of the qualified sub-areas is larger than the number of the preset sub-areas, the data analysis unit judges that the target video frame is qualified;
and if the number of the qualified sub-areas is smaller than or equal to the number of the preset sub-areas, the data analysis unit judges that the target video frame is unqualified.
Specifically, the data analysis unit counts the number of qualified target video frames under the second data analysis condition, calculates the ratio of the number of the qualified target video frames to the total number of the target video frames, marks the ratio as a reference ratio, and determines whether the quality of the target video image is qualified or not according to the reference ratio;
if the reference ratio is in a first preset ratio range, the data analysis unit judges that the quality of the target video image is qualified;
if the reference ratio is in the second preset ratio range, the data analysis unit judges that the quality of the target video image is possibly disqualified;
if the reference ratio is in a third preset ratio range, the data analysis unit judges that the quality of the target video image is unqualified;
and determining whether the target video frame is qualified or not according to the second data analysis condition.
Specifically, the data analysis unit judges the re-extraction quantity of the target video frames according to the reference ratio of the target video under a third data analysis condition;
the extraction quantity of the target video frames and the reference ratio of the target video are in a negative correlation relationship;
and the third data analysis condition is that if the reference ratio is in a second preset ratio range, the data analysis unit judges that the target video image quality is possibly disqualified.
Specifically, the data analysis unit compares the voice definition of the target video with a preset voice definition under a fourth data analysis condition to determine whether the voice definition of the target video is qualified;
if the voice definition of the target video is smaller than the first preset voice definition, the data analysis unit judges that the voice definition of the target video is unqualified;
if the voice definition of the target video is greater than or equal to the first preset voice definition and less than the second preset voice definition, the data analysis unit judges that the voice definition of the target video is possibly qualified;
if the voice definition of the target video is greater than or equal to the second preset voice definition, the data analysis unit judges that the voice definition of the target video is qualified;
and the fourth data analysis condition is that whether the target video image quality is qualified or not is judged to be finished.
Specifically, the calculation formula of the speech intelligibility is:
K=(N1×α1+N2×α2)/N
wherein N1 is a paragraph duration of sound blur in the target video, N2 is a paragraph duration of sound too large or too small in the target video, N is a total duration of the target video, α1 is a first weight coefficient, α2 is a second weight coefficient, 0 < α1 < α2 < 1, α1+α2=1.
Specifically, the data analysis unit is provided with a sound allowable threshold range, if the sound decibel frequency in the target video is larger than the sound allowable threshold range, the data analysis unit determines that the current sound of the target video is too large, and if the sound decibel frequency in the target video is smaller than the sound allowable threshold range, the data analysis unit determines that the current sound of the target video is too small.
Specifically, the display unit transmits the judgment information of the data analysis unit to the user side under the fifth data analysis condition;
and the fifth data analysis condition is that the sound definition and video quality of the target video are judged to be finished.
Specifically, the display unit is connected with the data analysis unit through 5g remote communication.
Thus far, the technical solution of the present invention has been described in connection with the preferred embodiments shown in the drawings, but it is easily understood by those skilled in the art that the scope of protection of the present invention is not limited to these specific embodiments. Equivalent modifications and substitutions for related technical features may be made by those skilled in the art without departing from the principles of the present invention, and such modifications and substitutions will be within the scope of the present invention.
The foregoing description is only of the preferred embodiments of the invention and is not intended to limit the invention; various modifications and variations of the present invention will be apparent to those skilled in the art. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (10)

1. A method of determining video conference quality, comprising:
the data acquisition unit acquires target quality data aiming at the target video;
the data acquisition unit performs multi-frame acquisition on the target video to acquire a target video frame and judges whether the target video frame is qualified or not according to the similarity between the target video frame and the reference video frame;
the data analysis unit performs partition processing on the target video frame under the first data analysis condition, uniformly divides the target video frame into four sub-areas and sequentially performs similarity detection on the sub-areas and the corresponding parts of the reference video frame;
the data analysis unit counts the number of qualified target video frames under the second data analysis condition, calculates the ratio of the number of the qualified target video frames to the total number of the target video frames, marks the ratio as a reference ratio, and determines whether the quality of the target video image is qualified or not according to the reference ratio;
the data analysis unit compares the voice definition of the target video with a preset voice definition under a fourth data analysis condition to judge whether the voice definition of the target video is qualified or not;
the first data analysis condition is that the similarity is in a second preset similarity range; the second data analysis condition is that whether the target video frame is qualified or not is determined to be finished; and the fourth data analysis condition is that whether the target video image quality is qualified or not is judged to be finished.
2. The method for determining the quality of a video conference according to claim 1, wherein the data acquisition unit performs multi-frame acquisition on the target video to acquire a target video frame, and the data analysis unit determines whether the target video frame is qualified according to the similarity between the target video frame and the reference video frame;
if the similarity is in a first preset similarity range, the data analysis unit judges that the target video frame is unqualified;
if the similarity is in a second preset similarity range, the data analysis unit judges that the target video frame is possibly qualified and performs partition comparison;
and if the similarity is in a third preset similarity range, the data analysis unit judges that the target video frame is qualified.
3. The method for determining the quality of a video conference according to claim 2, wherein the data analysis unit performs a partition process on the target video frame under the first data analysis condition, uniformly divides the target video frame into four sub-regions, and sequentially performs a similarity detection on each sub-region and a corresponding portion of the reference video frame;
if the number of the qualified sub-areas is larger than the number of the preset sub-areas, the data analysis unit judges that the target video frame is qualified;
and if the number of the qualified sub-areas is smaller than or equal to the number of the preset sub-areas, the data analysis unit judges that the target video frame is unqualified.
4. A method of determining the quality of a video conference according to claim 3, wherein the data analysis unit counts the number of acceptable target video frames under the second data analysis condition and calculates the ratio of the number of acceptable target video frames to the total number of target video frames, and marks the ratio as a reference ratio, and the data analysis unit determines whether the quality of the target video image is acceptable according to the reference ratio;
if the reference ratio is in a first preset ratio range, the data analysis unit judges that the quality of the target video image is qualified;
if the reference ratio is in the second preset ratio range, the data analysis unit judges that the quality of the target video image is possibly disqualified;
if the reference ratio is in a third preset ratio range, the data analysis unit judges that the quality of the target video image is unqualified;
and determining whether the target video frame is qualified or not according to the second data analysis condition.
5. The method for determining the quality of a video conference according to claim 4, wherein said data analysis unit determines the re-extraction number of the target video frames based on the reference ratio of the target video under the third data analysis condition;
the extraction quantity of the target video frames and the reference ratio of the target video are in a negative correlation relationship;
and the third data analysis condition is that if the reference ratio is in a second preset ratio range, the data analysis unit judges that the target video image quality is possibly disqualified.
6. The method for determining the quality of a video conference according to claim 5, wherein the data analysis unit compares the speech intelligibility of the target video with a preset speech intelligibility under a fourth data analysis condition to determine whether the speech intelligibility of the target video is acceptable;
if the voice definition of the target video is smaller than the first preset voice definition, the data analysis unit judges that the voice definition of the target video is unqualified;
if the voice definition of the target video is greater than or equal to the first preset voice definition and less than the second preset voice definition, the data analysis unit judges that the voice definition of the target video is possibly qualified;
if the voice definition of the target video is greater than or equal to the second preset voice definition, the data analysis unit judges that the voice definition of the target video is qualified;
and the fourth data analysis condition is that whether the target video image quality is qualified or not is judged to be finished.
7. The method of determining video conference quality according to claim 6, wherein the speech intelligibility is calculated by the formula:
K=(N1×α1+N2×α2)/N
wherein N1 is a paragraph duration of sound blur in the target video, N2 is a paragraph duration of sound too large or too small in the target video, N is a total duration of the target video, α1 is a first weight coefficient, α2 is a second weight coefficient, 0 < α1 < α2 < 1, α1+α2=1.
8. The method according to claim 7, wherein the data analysis unit is provided with a sound permission threshold range, and the data analysis unit determines that the current sound of the target video is too loud if the sound decibel frequency in the target video is greater than the sound permission threshold range, and determines that the current sound of the target video is too loud if the sound decibel frequency in the target video is less than the sound permission threshold range.
9. The method for determining the quality of a video conference according to claim 8, wherein the display unit transmits the determination information of the data analysis unit to the user side under a fifth data analysis condition;
and the fifth data analysis condition is that the sound definition and video quality of the target video are judged to be finished.
10. The method of determining the quality of a videoconference according to claim 9, wherein the display unit is connected to the data analysis unit by 5g telecommunications.
CN202311367912.8A 2023-10-23 2023-10-23 Method for determining video conference quality Pending CN117115719A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311367912.8A CN117115719A (en) 2023-10-23 2023-10-23 Method for determining video conference quality

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311367912.8A CN117115719A (en) 2023-10-23 2023-10-23 Method for determining video conference quality

Publications (1)

Publication Number Publication Date
CN117115719A true CN117115719A (en) 2023-11-24

Family

ID=88809454

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311367912.8A Pending CN117115719A (en) 2023-10-23 2023-10-23 Method for determining video conference quality

Country Status (1)

Country Link
CN (1) CN117115719A (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140278399A1 (en) * 2013-03-14 2014-09-18 Polycom, Inc. Speech fragment detection for management of interaction in a remote conference
CN105959684A (en) * 2016-05-26 2016-09-21 天津大学 Stereo image quality evaluation method based on binocular fusion
CN107809632A (en) * 2017-11-01 2018-03-16 深圳创维数字技术有限公司 Audio frequency and video method of testing, device, test system and readable storage medium storing program for executing
CN113873226A (en) * 2021-09-13 2021-12-31 深圳市欢太科技有限公司 Encoding and decoding quality testing method and device, computer equipment and storage medium
CN115221340A (en) * 2021-12-23 2022-10-21 飞虎互动科技(北京)有限公司 Audio and video session quality inspection method and device
CN116385745A (en) * 2023-02-22 2023-07-04 深圳绿米联创科技有限公司 Image recognition method, device, electronic equipment and storage medium

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140278399A1 (en) * 2013-03-14 2014-09-18 Polycom, Inc. Speech fragment detection for management of interaction in a remote conference
CN105959684A (en) * 2016-05-26 2016-09-21 天津大学 Stereo image quality evaluation method based on binocular fusion
CN107809632A (en) * 2017-11-01 2018-03-16 深圳创维数字技术有限公司 Audio frequency and video method of testing, device, test system and readable storage medium storing program for executing
CN113873226A (en) * 2021-09-13 2021-12-31 深圳市欢太科技有限公司 Encoding and decoding quality testing method and device, computer equipment and storage medium
CN115221340A (en) * 2021-12-23 2022-10-21 飞虎互动科技(北京)有限公司 Audio and video session quality inspection method and device
CN116385745A (en) * 2023-02-22 2023-07-04 深圳绿米联创科技有限公司 Image recognition method, device, electronic equipment and storage medium

Similar Documents

Publication Publication Date Title
EP2381738A1 (en) Adaptive volume adjustment method, device and communication terminal
USRE40054E1 (en) Video-assisted audio signal processing system and method
US9392221B2 (en) Videoconferencing endpoint having multiple voice-tracking cameras
CN112380972B (en) Volume adjusting method applied to television scene
EP1622349A1 (en) Teleconference volume level monitoring and feedback on the volume level
WO2018209879A1 (en) Method and device for automatically selecting camera image, and audio and video system
US9094523B2 (en) Metric for meeting commencement in a voice conferencing system
US20110285809A1 (en) Automatic Camera Framing for Videoconferencing
CN103841357A (en) Microphone array sound source positioning method, device and system based on video tracking
KR20160145730A (en) Volume adjusting method, apparatus and terminal of user terminal
CN105979197A (en) Remote conference control method and device based on automatic recognition of howling sound
US20140118472A1 (en) Active Speaker Indicator for Conference Participants
CN107820037A (en) The methods, devices and systems of audio signal, image procossing
US9165182B2 (en) Method and apparatus for using face detection information to improve speaker segmentation
CN105933557A (en) Volume intelligent adjusting method for conference participants in synchronous voice conference and volume intelligent adjusting system thereof
CN114333853A (en) Audio data processing method, equipment and system
CN117115719A (en) Method for determining video conference quality
US9602758B2 (en) Communication apparatus, conference system, computer-readable recording medium, and display control method
CN112788280A (en) Intelligent conference media management cloud platform based on cloud computing and voice feature analysis
CN112543302B (en) Intelligent noise reduction method and equipment in multi-person teleconference
CN111083627B (en) Fault microphone detection method and device
US10867609B2 (en) Transcription generation technique selection
CN116916111A (en) Video processing method suitable for cloud video conference
CN116980522B (en) System and method for notifying customer image based on intelligent quality inspection
CN113923395A (en) Method, equipment and storage medium for improving conference quality

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination