CN117115719A - Method for determining video conference quality - Google Patents
Method for determining video conference quality Download PDFInfo
- Publication number
- CN117115719A CN117115719A CN202311367912.8A CN202311367912A CN117115719A CN 117115719 A CN117115719 A CN 117115719A CN 202311367912 A CN202311367912 A CN 202311367912A CN 117115719 A CN117115719 A CN 117115719A
- Authority
- CN
- China
- Prior art keywords
- target video
- data analysis
- analysis unit
- qualified
- quality
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 29
- 238000007405 data analysis Methods 0.000 claims abstract description 129
- 238000005192 partition Methods 0.000 claims description 10
- 238000001514 detection method Methods 0.000 claims description 9
- 238000000605 extraction Methods 0.000 claims description 7
- 238000004891 communication Methods 0.000 description 5
- 230000004048 modification Effects 0.000 description 4
- 238000012986 modification Methods 0.000 description 4
- 238000010586 diagram Methods 0.000 description 2
- 238000006467 substitution reaction Methods 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000012512 characterization method Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/74—Image or video pattern matching; Proximity measures in feature spaces
- G06V10/761—Proximity, similarity or dissimilarity measures
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/57—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for processing of video signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/14—Systems for two-way working
- H04N7/15—Conference systems
Abstract
The invention relates to the field of data processing, in particular to a method for determining video conference quality, which comprises the following steps: the data acquisition unit acquires target quality data aiming at the target video; the data acquisition unit performs multi-frame acquisition on the target video to acquire a target video frame and judges whether the target video frame is qualified or not according to the similarity between the target video frame and the reference video frame; the data analysis unit uniformly divides the target video frame into four sub-areas and sequentially detects the similarity between each sub-area and the corresponding part of the reference video frame; the data analysis unit counts the number of the qualified target video frames and calculates the ratio of the number of the qualified target video frames to the total number of the target video frames, and marks the ratio as a reference ratio, and the data analysis unit determines whether the quality of the target video image is qualified according to the reference ratio; the data analysis unit compares the voice definition of the target video with a preset voice definition to judge whether the voice definition of the target video is qualified or not; the invention improves the determination accuracy of the video conference quality.
Description
Technical Field
The invention relates to the field of data processing, in particular to a method for determining video conference quality.
Background
The video conference quality refers to the performance of the video and audio in terms of definition, stability, real-time performance and the like in the video conference process. The importance of determining video conference quality is becoming increasingly important in modern tele-working and tele-collaboration environments. Video conferencing is an important tool for real-time communication and collaboration. Good video conference quality can ensure clear and accurate voice and image transmission between participants, thereby facilitating efficient communication and understanding. How to make efficient analytical decisions regarding video quality and audio quality of video conferences is a problem that the following technicians are urgent to address.
Chinese patent publication No. CN115174842a discloses a method, related apparatus and system for determining quality of a video conference, where the method includes inputting N sets of flow characteristic data obtained in N sampling periods into a quality judgment model, to obtain a quality judgment result of the video conference in a target period. According to the technical scheme, the quality of the video conference can be judged according to the flow characteristic data in the network. In this way, even without a conference screen, the video conference screen quality during the video conference can be determined, thereby facilitating the video conference provider to optimize the video conference service. It can be seen that the above technical solution has the following problems: the determined data volume is large, and the quality determination method cannot be adjusted according to the actual detection result, so that the method for determining the quality of the video conference has poor flexibility.
Disclosure of Invention
Therefore, the invention provides a method for determining the quality of a video conference, which is used for overcoming the defect that the flexibility of the method for determining the quality of the video conference is poor because the quality judging method cannot be adjusted according to the actual detection result in the prior art.
To achieve the above object, the present invention provides a method for determining video conference quality, comprising:
the data acquisition unit acquires target quality data aiming at the target video;
the data acquisition unit performs multi-frame acquisition on the target video to acquire a target video frame and judges whether the target video frame is qualified or not according to the similarity between the target video frame and the reference video frame;
the data analysis unit performs partition processing on the target video frame under the first data analysis condition, uniformly divides the target video frame into four sub-areas and sequentially performs similarity detection on the sub-areas and the corresponding parts of the reference video frame;
the data analysis unit counts the number of qualified target video frames under the second data analysis condition, calculates the ratio of the number of the qualified target video frames to the total number of the target video frames, marks the ratio as a reference ratio, and determines whether the quality of the target video image is qualified or not according to the reference ratio;
the data analysis unit compares the voice definition of the target video with a preset voice definition under a fourth data analysis condition to judge whether the voice definition of the target video is qualified or not;
the first data analysis condition is that the similarity is in a second preset similarity range; the second data analysis condition is that whether the target video frame is qualified or not is determined to be finished; and the fourth data analysis condition is that whether the target video image quality is qualified or not is judged to be finished.
Further, the data acquisition unit acquires a target video frame by aiming at the target video in a multi-frame manner, and the data analysis unit judges whether the target video frame is qualified or not according to the similarity between the target video frame and the reference video frame;
if the similarity is in a first preset similarity range, the data analysis unit judges that the target video frame is unqualified;
if the similarity is in a second preset similarity range, the data analysis unit judges that the target video frame is possibly qualified and performs partition comparison;
and if the similarity is in a third preset similarity range, the data analysis unit judges that the target video frame is qualified.
Further, the data analysis unit performs partition processing on the target video frame under the first data analysis condition, uniformly divides the target video frame into four sub-areas, and sequentially performs similarity detection on the sub-areas and the corresponding parts of the reference video frame;
if the number of the qualified sub-areas is larger than the number of the preset sub-areas, the data analysis unit judges that the target video frame is qualified;
and if the number of the qualified sub-areas is smaller than or equal to the number of the preset sub-areas, the data analysis unit judges that the target video frame is unqualified.
Further, the data analysis unit counts the number of qualified target video frames under the second data analysis condition, calculates the ratio of the number of the qualified target video frames to the total number of the target video frames, marks the ratio as a reference ratio, and determines whether the quality of the target video image is qualified or not according to the reference ratio;
if the reference ratio is in a first preset ratio range, the data analysis unit judges that the quality of the target video image is qualified;
if the reference ratio is in the second preset ratio range, the data analysis unit judges that the quality of the target video image is possibly disqualified;
if the reference ratio is in a third preset ratio range, the data analysis unit judges that the quality of the target video image is unqualified;
and determining whether the target video frame is qualified or not according to the second data analysis condition.
Further, the data analysis unit judges the re-extraction quantity of the target video frames according to the reference ratio of the target video under a third data analysis condition;
the extraction quantity of the target video frames and the reference ratio of the target video are in a negative correlation relationship;
and the third data analysis condition is that if the reference ratio is in a second preset ratio range, the data analysis unit judges that the target video image quality is possibly disqualified.
Further, the data analysis unit compares the voice definition of the target video with a preset voice definition under a fourth data analysis condition to judge whether the voice definition of the target video is qualified or not;
if the voice definition of the target video is smaller than the first preset voice definition, the data analysis unit judges that the voice definition of the target video is unqualified;
if the voice definition of the target video is greater than or equal to the first preset voice definition and less than the second preset voice definition, the data analysis unit judges that the voice definition of the target video is possibly qualified;
if the voice definition of the target video is greater than or equal to the second preset voice definition, the data analysis unit judges that the voice definition of the target video is qualified;
and the fourth data analysis condition is that whether the target video image quality is qualified or not is judged to be finished.
Further, the calculation formula of the voice definition is as follows:
K=(N1×α1+N2×α2)/N
wherein N1 is a paragraph duration of sound blur in the target video, N2 is a paragraph duration of sound too large or too small in the target video, N is a total duration of the target video, α1 is a first weight coefficient, α2 is a second weight coefficient, 0 < α1 < α2 < 1, α1+α2=1.
Further, the data analysis unit is provided with a sound allowable threshold range, if the sound decibel frequency in the target video is larger than the sound allowable threshold range, the data analysis unit judges that the current sound of the target video is too large, and if the sound decibel frequency in the target video is smaller than the sound allowable threshold range, the data analysis unit judges that the current sound of the target video is too small.
Further, the display unit transmits the judging information of the data analysis unit to the user side under the fifth data analysis condition;
and the fifth data analysis condition is that the sound definition and video quality of the target video are judged to be finished.
Further, the display unit is connected with the data analysis unit through 5g remote communication.
Compared with the prior art, the method has the advantages that the data acquisition unit acquires the target video frames by multi-frame acquisition aiming at the target video, the data analysis unit judges whether the target video frames are qualified or not according to the similarity between the target video frames and the reference video frames, and does not perform quality analysis of the full video section, so that the data processing amount is reduced, the data analysis unit carries out partition processing on the target video frames under the first data analysis condition, uniformly divides the target video frames into four sub-areas and sequentially carries out similarity detection on the sub-areas and the corresponding parts of the reference video frames, misjudgment of a system when the similarity is in an intermediate threshold value is avoided, the judgment precision of the system is further improved, the data analysis unit judges the re-extraction amount of the target video frames according to the reference ratio of the target video under the third data analysis condition, the problem of poor data characterization caused by excessively small extracted data amount is avoided, and the judgment precision of the method is further improved.
Drawings
Fig. 1 is a schematic diagram of a method for determining video conference quality according to an embodiment of the present invention.
Detailed Description
In order that the objects and advantages of the invention will become more apparent, the invention will be further described with reference to the following examples; it should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention.
Preferred embodiments of the present invention are described below with reference to the accompanying drawings. It should be understood by those skilled in the art that these embodiments are merely for explaining the technical principles of the present invention, and are not intended to limit the scope of the present invention.
It should be noted that, in the description of the present invention, terms such as "upper," "lower," "left," "right," "inner," "outer," and the like indicate directions or positional relationships based on the directions or positional relationships shown in the drawings, which are merely for convenience of description, and do not indicate or imply that the apparatus or elements must have a specific orientation, be constructed and operated in a specific orientation, and thus should not be construed as limiting the present invention.
Furthermore, it should be noted that, in the description of the present invention, unless explicitly specified and limited otherwise, the terms "mounted," "connected," and "connected" are to be construed broadly, and may be either fixedly connected, detachably connected, or integrally connected, for example; can be mechanically or electrically connected; can be directly connected or indirectly connected through an intermediate medium, and can be communication between two elements. The specific meaning of the above terms in the present invention can be understood by those skilled in the art according to the specific circumstances.
Referring to fig. 1, which is a schematic diagram of a method for determining quality of a video conference according to an embodiment of the present invention, the present invention provides a method for determining quality of a video conference, including:
the data acquisition unit acquires target quality data aiming at the target video;
the data acquisition unit performs multi-frame acquisition on the target video to acquire a target video frame and judges whether the target video frame is qualified or not according to the similarity between the target video frame and the reference video frame;
the data analysis unit performs partition processing on the target video frame under the first data analysis condition, uniformly divides the target video frame into four sub-areas and sequentially performs similarity detection on the sub-areas and the corresponding parts of the reference video frame;
the data analysis unit counts the number of qualified target video frames under the second data analysis condition, calculates the ratio of the number of the qualified target video frames to the total number of the target video frames, marks the ratio as a reference ratio, and determines whether the quality of the target video image is qualified or not according to the reference ratio;
the data analysis unit compares the voice definition of the target video with a preset voice definition under a fourth data analysis condition to judge whether the voice definition of the target video is qualified or not;
the first data analysis condition is that the similarity is in a second preset similarity range; the second data analysis condition is that whether the target video frame is qualified or not is determined to be finished; and the fourth data analysis condition is that whether the target video image quality is qualified or not is judged to be finished.
Specifically, the data acquisition unit performs multi-frame acquisition on the target video to acquire a target video frame, and the data analysis unit judges whether the target video frame is qualified or not according to the similarity between the target video frame and the reference video frame;
if the similarity is in a first preset similarity range, the data analysis unit judges that the target video frame is unqualified;
if the similarity is in a second preset similarity range, the data analysis unit judges that the target video frame is possibly qualified and performs partition comparison;
and if the similarity is in a third preset similarity range, the data analysis unit judges that the target video frame is qualified.
Specifically, the data analysis unit performs partition processing on a target video frame under a first data analysis condition, uniformly divides the target video frame into four sub-areas, and sequentially performs similarity detection on the sub-areas and corresponding parts of a reference video frame;
if the number of the qualified sub-areas is larger than the number of the preset sub-areas, the data analysis unit judges that the target video frame is qualified;
and if the number of the qualified sub-areas is smaller than or equal to the number of the preset sub-areas, the data analysis unit judges that the target video frame is unqualified.
Specifically, the data analysis unit counts the number of qualified target video frames under the second data analysis condition, calculates the ratio of the number of the qualified target video frames to the total number of the target video frames, marks the ratio as a reference ratio, and determines whether the quality of the target video image is qualified or not according to the reference ratio;
if the reference ratio is in a first preset ratio range, the data analysis unit judges that the quality of the target video image is qualified;
if the reference ratio is in the second preset ratio range, the data analysis unit judges that the quality of the target video image is possibly disqualified;
if the reference ratio is in a third preset ratio range, the data analysis unit judges that the quality of the target video image is unqualified;
and determining whether the target video frame is qualified or not according to the second data analysis condition.
Specifically, the data analysis unit judges the re-extraction quantity of the target video frames according to the reference ratio of the target video under a third data analysis condition;
the extraction quantity of the target video frames and the reference ratio of the target video are in a negative correlation relationship;
and the third data analysis condition is that if the reference ratio is in a second preset ratio range, the data analysis unit judges that the target video image quality is possibly disqualified.
Specifically, the data analysis unit compares the voice definition of the target video with a preset voice definition under a fourth data analysis condition to determine whether the voice definition of the target video is qualified;
if the voice definition of the target video is smaller than the first preset voice definition, the data analysis unit judges that the voice definition of the target video is unqualified;
if the voice definition of the target video is greater than or equal to the first preset voice definition and less than the second preset voice definition, the data analysis unit judges that the voice definition of the target video is possibly qualified;
if the voice definition of the target video is greater than or equal to the second preset voice definition, the data analysis unit judges that the voice definition of the target video is qualified;
and the fourth data analysis condition is that whether the target video image quality is qualified or not is judged to be finished.
Specifically, the calculation formula of the speech intelligibility is:
K=(N1×α1+N2×α2)/N
wherein N1 is a paragraph duration of sound blur in the target video, N2 is a paragraph duration of sound too large or too small in the target video, N is a total duration of the target video, α1 is a first weight coefficient, α2 is a second weight coefficient, 0 < α1 < α2 < 1, α1+α2=1.
Specifically, the data analysis unit is provided with a sound allowable threshold range, if the sound decibel frequency in the target video is larger than the sound allowable threshold range, the data analysis unit determines that the current sound of the target video is too large, and if the sound decibel frequency in the target video is smaller than the sound allowable threshold range, the data analysis unit determines that the current sound of the target video is too small.
Specifically, the display unit transmits the judgment information of the data analysis unit to the user side under the fifth data analysis condition;
and the fifth data analysis condition is that the sound definition and video quality of the target video are judged to be finished.
Specifically, the display unit is connected with the data analysis unit through 5g remote communication.
Thus far, the technical solution of the present invention has been described in connection with the preferred embodiments shown in the drawings, but it is easily understood by those skilled in the art that the scope of protection of the present invention is not limited to these specific embodiments. Equivalent modifications and substitutions for related technical features may be made by those skilled in the art without departing from the principles of the present invention, and such modifications and substitutions will be within the scope of the present invention.
The foregoing description is only of the preferred embodiments of the invention and is not intended to limit the invention; various modifications and variations of the present invention will be apparent to those skilled in the art. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present invention should be included in the protection scope of the present invention.
Claims (10)
1. A method of determining video conference quality, comprising:
the data acquisition unit acquires target quality data aiming at the target video;
the data acquisition unit performs multi-frame acquisition on the target video to acquire a target video frame and judges whether the target video frame is qualified or not according to the similarity between the target video frame and the reference video frame;
the data analysis unit performs partition processing on the target video frame under the first data analysis condition, uniformly divides the target video frame into four sub-areas and sequentially performs similarity detection on the sub-areas and the corresponding parts of the reference video frame;
the data analysis unit counts the number of qualified target video frames under the second data analysis condition, calculates the ratio of the number of the qualified target video frames to the total number of the target video frames, marks the ratio as a reference ratio, and determines whether the quality of the target video image is qualified or not according to the reference ratio;
the data analysis unit compares the voice definition of the target video with a preset voice definition under a fourth data analysis condition to judge whether the voice definition of the target video is qualified or not;
the first data analysis condition is that the similarity is in a second preset similarity range; the second data analysis condition is that whether the target video frame is qualified or not is determined to be finished; and the fourth data analysis condition is that whether the target video image quality is qualified or not is judged to be finished.
2. The method for determining the quality of a video conference according to claim 1, wherein the data acquisition unit performs multi-frame acquisition on the target video to acquire a target video frame, and the data analysis unit determines whether the target video frame is qualified according to the similarity between the target video frame and the reference video frame;
if the similarity is in a first preset similarity range, the data analysis unit judges that the target video frame is unqualified;
if the similarity is in a second preset similarity range, the data analysis unit judges that the target video frame is possibly qualified and performs partition comparison;
and if the similarity is in a third preset similarity range, the data analysis unit judges that the target video frame is qualified.
3. The method for determining the quality of a video conference according to claim 2, wherein the data analysis unit performs a partition process on the target video frame under the first data analysis condition, uniformly divides the target video frame into four sub-regions, and sequentially performs a similarity detection on each sub-region and a corresponding portion of the reference video frame;
if the number of the qualified sub-areas is larger than the number of the preset sub-areas, the data analysis unit judges that the target video frame is qualified;
and if the number of the qualified sub-areas is smaller than or equal to the number of the preset sub-areas, the data analysis unit judges that the target video frame is unqualified.
4. A method of determining the quality of a video conference according to claim 3, wherein the data analysis unit counts the number of acceptable target video frames under the second data analysis condition and calculates the ratio of the number of acceptable target video frames to the total number of target video frames, and marks the ratio as a reference ratio, and the data analysis unit determines whether the quality of the target video image is acceptable according to the reference ratio;
if the reference ratio is in a first preset ratio range, the data analysis unit judges that the quality of the target video image is qualified;
if the reference ratio is in the second preset ratio range, the data analysis unit judges that the quality of the target video image is possibly disqualified;
if the reference ratio is in a third preset ratio range, the data analysis unit judges that the quality of the target video image is unqualified;
and determining whether the target video frame is qualified or not according to the second data analysis condition.
5. The method for determining the quality of a video conference according to claim 4, wherein said data analysis unit determines the re-extraction number of the target video frames based on the reference ratio of the target video under the third data analysis condition;
the extraction quantity of the target video frames and the reference ratio of the target video are in a negative correlation relationship;
and the third data analysis condition is that if the reference ratio is in a second preset ratio range, the data analysis unit judges that the target video image quality is possibly disqualified.
6. The method for determining the quality of a video conference according to claim 5, wherein the data analysis unit compares the speech intelligibility of the target video with a preset speech intelligibility under a fourth data analysis condition to determine whether the speech intelligibility of the target video is acceptable;
if the voice definition of the target video is smaller than the first preset voice definition, the data analysis unit judges that the voice definition of the target video is unqualified;
if the voice definition of the target video is greater than or equal to the first preset voice definition and less than the second preset voice definition, the data analysis unit judges that the voice definition of the target video is possibly qualified;
if the voice definition of the target video is greater than or equal to the second preset voice definition, the data analysis unit judges that the voice definition of the target video is qualified;
and the fourth data analysis condition is that whether the target video image quality is qualified or not is judged to be finished.
7. The method of determining video conference quality according to claim 6, wherein the speech intelligibility is calculated by the formula:
K=(N1×α1+N2×α2)/N
wherein N1 is a paragraph duration of sound blur in the target video, N2 is a paragraph duration of sound too large or too small in the target video, N is a total duration of the target video, α1 is a first weight coefficient, α2 is a second weight coefficient, 0 < α1 < α2 < 1, α1+α2=1.
8. The method according to claim 7, wherein the data analysis unit is provided with a sound permission threshold range, and the data analysis unit determines that the current sound of the target video is too loud if the sound decibel frequency in the target video is greater than the sound permission threshold range, and determines that the current sound of the target video is too loud if the sound decibel frequency in the target video is less than the sound permission threshold range.
9. The method for determining the quality of a video conference according to claim 8, wherein the display unit transmits the determination information of the data analysis unit to the user side under a fifth data analysis condition;
and the fifth data analysis condition is that the sound definition and video quality of the target video are judged to be finished.
10. The method of determining the quality of a videoconference according to claim 9, wherein the display unit is connected to the data analysis unit by 5g telecommunications.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311367912.8A CN117115719A (en) | 2023-10-23 | 2023-10-23 | Method for determining video conference quality |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311367912.8A CN117115719A (en) | 2023-10-23 | 2023-10-23 | Method for determining video conference quality |
Publications (1)
Publication Number | Publication Date |
---|---|
CN117115719A true CN117115719A (en) | 2023-11-24 |
Family
ID=88809454
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202311367912.8A Pending CN117115719A (en) | 2023-10-23 | 2023-10-23 | Method for determining video conference quality |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117115719A (en) |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140278399A1 (en) * | 2013-03-14 | 2014-09-18 | Polycom, Inc. | Speech fragment detection for management of interaction in a remote conference |
CN105959684A (en) * | 2016-05-26 | 2016-09-21 | 天津大学 | Stereo image quality evaluation method based on binocular fusion |
CN107809632A (en) * | 2017-11-01 | 2018-03-16 | 深圳创维数字技术有限公司 | Audio frequency and video method of testing, device, test system and readable storage medium storing program for executing |
CN113873226A (en) * | 2021-09-13 | 2021-12-31 | 深圳市欢太科技有限公司 | Encoding and decoding quality testing method and device, computer equipment and storage medium |
CN115221340A (en) * | 2021-12-23 | 2022-10-21 | 飞虎互动科技(北京)有限公司 | Audio and video session quality inspection method and device |
CN116385745A (en) * | 2023-02-22 | 2023-07-04 | 深圳绿米联创科技有限公司 | Image recognition method, device, electronic equipment and storage medium |
-
2023
- 2023-10-23 CN CN202311367912.8A patent/CN117115719A/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140278399A1 (en) * | 2013-03-14 | 2014-09-18 | Polycom, Inc. | Speech fragment detection for management of interaction in a remote conference |
CN105959684A (en) * | 2016-05-26 | 2016-09-21 | 天津大学 | Stereo image quality evaluation method based on binocular fusion |
CN107809632A (en) * | 2017-11-01 | 2018-03-16 | 深圳创维数字技术有限公司 | Audio frequency and video method of testing, device, test system and readable storage medium storing program for executing |
CN113873226A (en) * | 2021-09-13 | 2021-12-31 | 深圳市欢太科技有限公司 | Encoding and decoding quality testing method and device, computer equipment and storage medium |
CN115221340A (en) * | 2021-12-23 | 2022-10-21 | 飞虎互动科技(北京)有限公司 | Audio and video session quality inspection method and device |
CN116385745A (en) * | 2023-02-22 | 2023-07-04 | 深圳绿米联创科技有限公司 | Image recognition method, device, electronic equipment and storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP2381738A1 (en) | Adaptive volume adjustment method, device and communication terminal | |
USRE40054E1 (en) | Video-assisted audio signal processing system and method | |
US9392221B2 (en) | Videoconferencing endpoint having multiple voice-tracking cameras | |
CN112380972B (en) | Volume adjusting method applied to television scene | |
EP1622349A1 (en) | Teleconference volume level monitoring and feedback on the volume level | |
WO2018209879A1 (en) | Method and device for automatically selecting camera image, and audio and video system | |
US9094523B2 (en) | Metric for meeting commencement in a voice conferencing system | |
US20110285809A1 (en) | Automatic Camera Framing for Videoconferencing | |
CN103841357A (en) | Microphone array sound source positioning method, device and system based on video tracking | |
KR20160145730A (en) | Volume adjusting method, apparatus and terminal of user terminal | |
CN105979197A (en) | Remote conference control method and device based on automatic recognition of howling sound | |
US20140118472A1 (en) | Active Speaker Indicator for Conference Participants | |
CN107820037A (en) | The methods, devices and systems of audio signal, image procossing | |
US9165182B2 (en) | Method and apparatus for using face detection information to improve speaker segmentation | |
CN105933557A (en) | Volume intelligent adjusting method for conference participants in synchronous voice conference and volume intelligent adjusting system thereof | |
CN114333853A (en) | Audio data processing method, equipment and system | |
CN117115719A (en) | Method for determining video conference quality | |
US9602758B2 (en) | Communication apparatus, conference system, computer-readable recording medium, and display control method | |
CN112788280A (en) | Intelligent conference media management cloud platform based on cloud computing and voice feature analysis | |
CN112543302B (en) | Intelligent noise reduction method and equipment in multi-person teleconference | |
CN111083627B (en) | Fault microphone detection method and device | |
US10867609B2 (en) | Transcription generation technique selection | |
CN116916111A (en) | Video processing method suitable for cloud video conference | |
CN116980522B (en) | System and method for notifying customer image based on intelligent quality inspection | |
CN113923395A (en) | Method, equipment and storage medium for improving conference quality |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |