CN111277823A - System and method for audio and video synchronization test - Google Patents

System and method for audio and video synchronization test Download PDF

Info

Publication number
CN111277823A
CN111277823A CN202010148288.2A CN202010148288A CN111277823A CN 111277823 A CN111277823 A CN 111277823A CN 202010148288 A CN202010148288 A CN 202010148288A CN 111277823 A CN111277823 A CN 111277823A
Authority
CN
China
Prior art keywords
audio
video
signal
processing module
sound
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010148288.2A
Other languages
Chinese (zh)
Inventor
戎玲
胡融
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Jinghui Electronic Equipment Minhang Co ltd
Third Research Institute of the Ministry of Public Security
Original Assignee
Shanghai Jinghui Electronic Equipment Minhang Co ltd
Third Research Institute of the Ministry of Public Security
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Jinghui Electronic Equipment Minhang Co ltd, Third Research Institute of the Ministry of Public Security filed Critical Shanghai Jinghui Electronic Equipment Minhang Co ltd
Priority to CN202010148288.2A priority Critical patent/CN111277823A/en
Publication of CN111277823A publication Critical patent/CN111277823A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N17/00Diagnosis, testing or measuring for television systems or their details
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/4302Content synchronisation processes, e.g. decoder synchronisation
    • H04N21/4307Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/141Systems for two-way working between two video terminals, e.g. videophone
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • H04N7/183Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a single remote source
    • H04N7/186Video door telephones

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Studio Devices (AREA)

Abstract

The invention relates to a system and a method for audio and video synchronous testing, which can accurately test audio and video, audio and audio, and video synchronously by building an image and sound closed loop system and sending synchronous graphic command signals and sound command signals by computer equipment, thereby achieving the purpose that people and people have face-to-face speaking experience though being remote. Moreover, the whole process is automatically tested, and the interference error of the external environment is reduced. In addition, by the technical scheme of the invention, the performance of the developed or to-be-popularized product can be evaluated in a preparation manner.

Description

System and method for audio and video synchronization test
Technical Field
The invention relates to an audio and video device testing technology, in particular to an audio and video synchronization testing system and method.
Background
In the current situation of high-speed development of new technologies, audio and video transmission is a very common tool, in a system for audio and video synchronous transmission, digital end-to-end equipment without standard interfaces or non-standard transmission protocols or adopting an encryption technology is adopted, but because delay generated in the current digital end-to-end equipment for audio and video data transmission is different, the defects such as lip sound asynchrony and the like in the video call process can be generated, and especially in the application with higher requirements on audio and video synchronous performance such as real-time remote audio and video communication, the user experience of the audio and video equipment can be greatly reduced; for example, in the current house safety technology precaution product building intercom system, when the visitor presses the room number at building gate visitor caller end, the owner of indoor receiver end hears the calling sound and answers, can in time watch the real-time image of visitor better, and then so that owner comparatively accurate differentiates who is the visitor, and in order to enable indoor receiver end to hear the appearance of the caller when the caller's sound is watched, just need to make equipment can realize lip sound synchronization, and once the lip sound is asynchronous, not only can reduce user experience, still can make owner can't differentiate visitor's identity simultaneously, and then bring the potential safety hazard.
At present, in order to improve the audio and video synchronization performance of a remote audio and video communication device to be researched and developed or tested to be popularized and applied, an electric signal-image signal testing method is generally adopted for testing (such as lip sound synchronization testing and the like), but because interfaces, video devices and the like adopted by the remote audio and video communication device are possibly nonstandard, the remote audio and video communication device can only adopt end-to-end whole-course image signal testing, the testing effect is not ideal, the audio and video synchronization performance of the current device to be tested cannot be accurately obtained, and therefore the performance of a product to be researched and developed or to be popularized and applied cannot be subjected to prepared evaluation.
Moreover, as the image quality and the voice quality of people are improved, the synchronization of the image and the voice is required, and meanwhile, people hope to be close to the face-to-face speech of people through the subjective feeling of video chat or video call, so that the synchronization of the audio and the video needs to be correspondingly tested.
Disclosure of Invention
Aiming at the defects of the prior art, the invention provides a system and a method for audio and video synchronization test of digital end-to-end equipment which has no standard interface or non-standard transmission protocol or adopts encryption technology.
A system for audio and video synchronization test is characterized in that:
the computing equipment is respectively electrically connected with a video signal generating device, an audio generating and collecting instrument and a camera;
the video signal generating device is electrically connected with the camera, and the audio generating and collecting instrument is also electrically connected with a microphone and a loudspeaker;
the loudspeaker and the video signal generating device are close to a first device to be tested;
the camera and the microphone are close to a second device to be tested;
the first equipment to be tested and the second equipment to be tested are connected through a network;
the computing equipment is used for generating a graphic command signal and sending the graphic command signal to the video signal generating device;
the video signal generating device is used for generating a graphic test signal according to the graphic command signal and outputting a synchronous trigger signal to the camera;
the first device to be tested acquires the graph test signal and sends the graph test signal to the second device to be tested through a network;
the second device to be tested is used for receiving and displaying the graphic test signal sent by the first device to be tested;
the camera is used for receiving the synchronous trigger signal, shooting the second device to be tested and sending an image stream obtained by shooting to the computing device;
the computing equipment is further used for computing video delay time from the first equipment to be tested to the second equipment to be tested to collect the graphic test signal according to the image stream.
The method is characterized in that: the computing equipment is also used for generating a sound command signal synchronous with the graphic command signal and sending the sound command signal to the audio generation and acquisition instrument;
the audio generation and acquisition instrument is used for receiving the sound command signal and generating a sound test signal, and the sound test signal is externally emitted through the loudspeaker;
the first device to be tested collects the sound test signal and sends the sound test signal to the second device to be tested through a network;
the second device to be tested is used for receiving and playing the sound test signal sent by the first device to be tested;
the audio generation and acquisition instrument is also used for acquiring an audio stream of the sound test signal played by the second device to be tested through a microphone and forwarding the audio stream to the computing device;
the computing equipment is further used for the audio stream to calculate the audio delay time from the collection of the sound test signal by the first device to be tested to the playing of the sound test signal by the second device to be tested.
The method is characterized in that: and the computing equipment is also used for computing the audio and video time difference between the time of playing the sound test signal and the time of playing the graphic test signal by the second equipment to be tested according to the video delay time and the audio delay time.
The method is characterized in that: the computing device includes:
the first processing module is respectively and electrically connected with a third processing module and the video signal generating device and is used for generating the synchronous graphic command signal and sending the synchronous graphic command signal to the video signal generating device and the third processing module;
the second processing module is respectively and electrically connected with a fourth processing module, the audio generation and acquisition instrument and the first processing module, and is used for generating the sound command signal synchronous with the graphic command signal and sending the sound command signal to the audio generation and acquisition instrument and the fourth processing module;
the third processing module is electrically connected with the camera and is used for receiving the image stream sent by the camera, calculating the video delay time according to the image stream and the graphics command signal, and sending the video delay time to a fifth processing module;
the fourth processing module is further electrically connected to the audio generation and acquisition instrument, and is configured to receive the audio stream sent by the audio generation and acquisition instrument, calculate the audio time difference according to the audio stream and the sound command signal, and send the audio delay time to a fifth processing module;
and the fifth processing module is respectively connected with the third processing module and the fourth processing module, and calculates the audio and video time difference between the time of playing the sound test signal and the time of playing the graphic test signal by the second device to be tested according to the video delay time and the audio delay time.
The video signal generation device is characterized in that the video signal generation device is an LED array, the LED array is composed of N × M LED lamps, N and M are positive integers, N represents N rows, M represents M columns, and the LED lamps are used for being sequentially lightened at a preset time interval according to the graphic command signal;
the LED array generates a synchronous trigger signal through the TTL module and outputs the synchronous trigger signal to the camera, when the LED lamp is lightened, the camera is triggered to start shooting, and when the LED lamp is extinguished, the camera is triggered to stop shooting.
The system is characterized in that the first processing module is used for presetting the pulse number W of the graphic command signals, controlling the number of the lighted LED lamps to be W, and sequentially and completely lighting the W lamps to form a one-time test period Ts;
wherein the value range of W is 2-N M;
the third processing module is configured to extract a last frame image, locate a lit LED lamp of the last frame image, and if the lit LED lamp of the last frame image is the nth one, the video delay time Δ ta is calculated according to the following formula:
Δta=(W-n)Ts/(W)。
wherein the fourth processing module is configured to calculate the audio delay time Δ tb according to a difference assumed to be p between the number of pulses of the audio command signal with frequency fb that has been transmitted within Ts time and the number of pulses of the received audio test signal, and the formula is as follows:
Δtb=p/fb。
the fifth processing module is configured to calculate the audio/video time difference Δ tab according to the video delay time Δ ta and the audio delay time Δ tb, where the calculation formula is as follows:
Δtab=Δta-Δtb。
a method for audio-video synchronization testing, characterized by using a system for audio-video synchronization testing according to claim 1, the method comprising the following steps:
step S11, the computing equipment generates a graphic command signal and sends the graphic command signal to a video signal generating device;
step S21, the video signal generating device generates a pattern test signal according to the pattern command signal and outputs a synchronous trigger signal to the camera;
step S31, the first device to be tested collects the graphic test signal and sends the graphic test signal to the second device to be tested through a network;
step S41, the second device to be tested receives and displays the graphic test signal sent by the first device to be tested;
step S51, the camera receives the synchronous trigger signal, shoots the second device to be tested, and sends the image stream obtained by shooting to the computing device;
step S61, the computing device calculates, according to the image stream, a video delay time from the acquisition of the graphics test signal by the first device under test to the display of the graphics test signal by the second device under test.
The method is characterized in that: step S12: the computing equipment generates a sound command signal synchronous with the graphic command signal and sends the sound command signal to the audio generation and acquisition instrument;
step S22, the audio frequency generation and acquisition instrument receives the sound command signal to generate a sound test signal, and the sound test signal is emitted through the loudspeaker;
step S32, the first device to be tested also collects the sound test signal and sends the sound test signal to the second device to be tested through the network;
in step S42, the second device under test receives and plays the sound test signal sent by the first device under test;
step S52, the audio generation and acquisition instrument obtains an audio stream of the sound test signal played by the second device under test via a microphone and forwards the audio stream to the computing device;
step S62: the audio stream of the computing equipment calculates the audio delay time from the collection of the sound test signal by the first device to be tested to the playing of the sound test signal by the second device to be tested;
wherein the steps S12-S62 are performed simultaneously with the steps S11-S61.
The method is characterized in that: includes step S7: and the computing equipment computes the audio-video time difference between the time of playing the sound test signal and the time of playing the graphic test signal by the second equipment to be tested according to the video delay time and the audio delay time.
The method is characterized in that: the step S11 specifically includes: the first processing module of the computing equipment generates the graphics command signal and sends the graphics command signal to the video signal generating device and the third processing module; the step S12 specifically includes: the second processing module of the computing equipment also generates a sound command signal synchronous with the graphic command signal and sends the sound command signal to the audio generation and acquisition instrument and a fourth processing module of the computing equipment;
the step S61 specifically includes: the third processing module receives the image stream sent by the camera, calculates the video delay time according to the image stream and the graphics command signal, and sends the video delay time to a fifth processing module of the computing device;
the step S62 specifically includes: the fourth processing module receives the audio stream sent by the audio generation and acquisition instrument, calculates the audio time difference according to the audio stream and the sound command signal, and sends the audio delay time to a fifth processing module;
the step S7 specifically includes: and the fifth processing module calculates the audio and video time difference between the time of playing the sound test signal and the time of playing the graphic test signal by the second device to be tested according to the video delay time and the audio delay time.
It is characterized in that the step S21 specifically includes: the video signal generating device is an LED array, the LED array is composed of N x M LED lamps, N and M are positive integers, N represents N rows, M represents M columns, the LED lamps are sequentially lightened through a preset time interval according to the graphic command signal, the LED array generates a synchronous triggering signal through a TTL module and outputs the synchronous triggering signal to the camera, when the LED lamps are lightened, the camera is triggered to start shooting, and when the LED lamps are extinguished, the camera is triggered to stop shooting.
It is characterized in that the step S11 specifically includes: the first processing module presets the pulse number W of the graphic command signals, controls the number of the lighted LED lamps to be W, and lights the W lamps in sequence to be lighted into a one-time test period Ts;
wherein the value range of W is 2-N M;
the step S61 specifically includes: the third processing module extracts a last frame image, positions the lighted LED lamp of the last frame image, and assumes that the lighted LED lamp of the last frame image is the nth one, the video delay time Δ ta is calculated according to the following formula:
Δta=(W-n)Ts/(W)。
characterized in that step S62 specifically includes: the fourth processing module calculates the audio delay time Δ tb according to the assumption that the difference between the number of pulses of the sound command signal with frequency fb that has been transmitted within Ts time and the number of pulses of the received sound test signal is p, and the formula is as follows:
Δtb=p/fb。
characterized in that step S7 specifically includes: the fifth processing module calculates the audio/video time difference Δ tab according to the video delay time Δ ta and the audio delay time Δ tb, and the calculation formula is as follows:
Δtab=Δta-Δtb。
drawings
Fig. 1-3 are schematic structural diagrams of a system for testing synchronization of a medium audio and video according to an embodiment of the present invention.
FIG. 4 is a diagram of the shape of the correlation signal according to the present invention.
Fig. 5-7 are flow charts of methods for testing the synchronization of the mediant video according to embodiments of the present invention.
The system comprises a first device to be tested, a second device to be tested, a loudspeaker, a video signal generating device, a first processing module, a second processing module, a third processing module, a fourth processing module, a fifth processing module, a microphone, a camera, a computing device, a microphone, a camera, a computing device, a microphone, a camera, a computing device, a microphone, a loudspeaker, a microphone, a computing device, a microphone, a.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
It should be noted that the embodiments and features of the embodiments may be combined with each other without conflict.
The invention is further described with reference to the following drawings and specific examples, which are not intended to be limiting.
Fig. 1 to 3 are schematic structural diagrams of an audio and video synchronization test system in an embodiment of the present invention. As shown in fig. 1-2, the embodiment provides an audio and video synchronization testing system, which can be applied to performance testing of devices to be tested (such as audio and video devices) for transmitting and playing audio and video, and in particular, can be used for audio and video synchronization performance testing of devices with high requirements on audio and video synchronization performance, such as real-time remote audio and video communication (for example, building video intercom devices in home security devices), and the like, and the system can include the following features.
The system for audio and video synchronization testing shown in fig. 2 comprises a first device to be tested 1, a loudspeaker 2, a video signal generating device 3, a second device to be tested 4, a microphone 5, a camera 6, a computing device 7 and an audio generation and acquisition instrument 8. The video signal generating device 3 is electrically connected with the camera 6 and the computing device 7, the audio generation and acquisition instrument 8 is electrically connected with the computing device 7, the loudspeaker 2 and the microphone 5, the computing device 8 is further electrically connected with the second device to be tested 4, and the first device to be tested 1 is connected with the second device to be tested 2 through a network. Specifically, the computing device 7, the audio generation and acquisition instrument 8, the video signal generation device 3, the camera 6, and the second device under test 4 may be electrically connected by USB interfaces, respectively.
In particular, the computing device 7 comprises at least one processing module for processing the relevant information.
The video signal generating device 3 can be arranged close to the first device to be tested 1 so that the first device to be tested 1 can obtain the graphic test signal of the video signal generating device 3; the video signal generating device 3 can generate a graphic test signal (generally, information can be displayed by displaying dynamic light and shadow), for example, the graphic test signal can be generated by a display screen or an LED lamp matrix. In a preferred embodiment, the video signal generating means 3 is a matrix of LED lamps. The video signal generating device 3 is electrically connected with the camera 6 and the computing device 7 respectively, the video signal generating device 3 is controlled by a processor in the computing device 7, receives a graphic command signal sent by a data transceiver module of the computing device 7 to generate a graphic test signal, and can output a synchronous trigger signal to be sent to the camera 6. This synchronous trigger signal may be generated using a TTL (transistor-transistor logic level) signal system. So that the video signal generating means 3 causes the computer to start shooting while displaying the pattern test signal.
The audio generation and collection instrument 8 is connected with the loudspeaker 2, receives the sound command signal from the data receiving module of the computing device 7, generates a sound test signal according to the sound command signal (generally, the sound test signal can be generated by simulating human sound with the loudspeaker), controls the loudspeaker 2 to play, and the loudspeaker 2 can be close to the first equipment to be tested 1, so that the first equipment to be tested 1 can collect audio information.
The first device to be tested 1 sends the acquired graphic test signal and the acquired sound test signal to the second device to be tested 4 through the network, and the second device to be tested 4 receives the graphic test signal and the sound test signal and plays the graphic test signal and the sound test signal.
Specifically, the processor of the computing device 7 synchronously generates a graphics command signal and a sound command signal, and the graphics command signal and the sound command signal are respectively sent to the video signal generating device 3 and the audio generation and acquisition instrument 8 through the data transceiver module, so that the graphics test signal generated by the video signal generating device 3 and the sound test signal played by the loudspeaker 2 are ensured to be simultaneously performed. The camera 6 is close to the second device to be tested 4, so that photographing processing of the second device to be tested 4 is facilitated, the video signal generating device 3 generates a graphic test signal and simultaneously outputs a synchronous trigger signal to start the camera 6 to shoot, the camera 6 shoots the displayed video information of the second device to be tested 4, and the obtained image is transmitted to the computing device 7. Specifically, the high-speed camera 6 has a fast continuous shooting speed compared with a common camera, and can reach tens of frames or even thousands of frames per second.
Specifically, a microphone 5 is further provided near the second device under test 4, the audio generation and acquisition instrument 8 acquires the sound graphic signals output by the second device under test 4 in real time through the microphone 5 and transmits the sound graphic signals to the data transceiver module of the computing device 7, and the data transceiver module transmits the sound graphic signals to the processor.
Specifically, the audio generator 8 includes an audio generator for generating a sound test signal according to the sound command signal sent by the computing device 7, and an audio collector for collecting the sound test signal.
The processor of the computing device 7 controls to simultaneously generate a graphic command signal and a sound command signal, and respectively sends the signals to the video signal generating device 3 and the audio generation and acquisition instrument 8 through interfaces, and receives the image acquired by the camera and the sound test information acquired by the audio generation and acquisition instrument 8, so as to analyze and calculate the delay information of the audio and the video. The data transceiver module of the computing device 7 is further configured to receive a graphic test signal and a sound test signal from the second device under test 4 in real time, an audio and video training model is installed in the computing device 7, the computing device trains the audio and video training model according to the delay information, the graphic test signal and the sound test signal, and sends the trained model to the device under test, such as a second device under test lamp, to be applied to the device under test, so that the first device under test 1 and the second device under test 2 perform audio and video synchronization, video and audio and video synchronization, and audio and video synchronization. The device has the advantages that the subjective feeling of face-to-face speech between people is achieved, when the device is used in occasions such as buildings, the device can know the limb information and the voice information of people outside the building in time, and the response information of people inside the building can be fed back to the people outside the building in time.
The computing device 7 contains at least one processing module to perform various operations.
Fig. 2 constitutes an image-sound ring ratio system, which achieves a test effect on audio-video delay.
Specifically, it takes a period of time for the first device under test 1 to collect the pattern test signal and transmit the pattern test signal to the second device under test 4 in a remote location. Therefore, when people outside the building do a movement or the lips change, people inside the building cannot see the movement in time, the people need to be placed for a period of time to see the movement, and if the response of the people inside the building is responded, the people outside the building can see the movement almost in the same time. As is the transmission of sound. In addition, the audio and video information such as lip sound is not synchronized because there is a time difference between the audio and video transmissions. When the delay time difference is larger, the remote communication is seriously influenced, and the experience is very poor.
In particular, the computing device 7 comprises a first processing module 71, a second processing module 72, a third processing module 73, a fourth processing module 74 and a fifth processing module 75.
The first processing module 71, which is electrically connected to a third processing module 73 and the video signal generating device 3, respectively, and is configured to generate a synchronous graphics command signal and send the synchronous graphics command signal to the video signal generating device 3 and the third processing module 73;
the second processing module 72 is electrically connected to a fourth processing module 74, the audio generation and collection instrument 8 and the first processing module 71, respectively, and is configured to generate a sound command signal synchronized with the graphics command signal and send the sound command signal to the audio generation and collection instrument 8 and the fourth processing module 74;
the third processing module 73, electrically connected to the camera 6, is configured to receive the image stream sent by the camera 6, calculate a video delay time according to the image stream and the graphics command signal, and send the video delay time to the fifth processing module 75;
the fourth processing module 74 is further electrically connected to the audio generation and acquisition instrument 8, and is configured to receive the audio stream sent by the audio generation and acquisition instrument 8, calculate an audio delay time according to the audio stream and the sound command signal, and send the audio delay time to the fifth processing module 75;
the fifth processing module 75 is connected to the third processing module 73 and the fourth processing module 74, respectively, and calculates an audio/video time difference between the time when the second device under test 4 plays the sound test signal and the time when the second device under test 4 plays the graphic test signal according to the video delay time and the audio delay time.
Referring to fig. 3, therefore, in the present invention, as an embodiment of the present invention, the computing device 7 simultaneously issues two command signals, i.e. a graphics command signal and a sound command signal, at t0, and controls to simultaneously generate a graphics test signal and a sound test signal in the vicinity of the first device under test 1, respectively, and the generating time of the graphics test signal and the sound test signal may be marked as t 0.
The camera 6 is triggered to shoot the second device under test 4 while the pattern test signal is generated, and the time when the camera 6 acquires the pattern test signal on the second device under test 4 is Ta, so that the difference between the time of the pattern test signal displayed by the second device under test 4 and the time when the pattern test signal is acquired by the first device under test 1 is Δ a-Ta-t 0, that is, the time of Ta-t0 is delayed. The time of the sound test signal played by the second device under test collected by the audio generation and collection instrument 8 is Tb, and then the difference between the time of the sound test signal displayed by the second device under test 4 and the time of the sound test signal collected by the first device under test 1 is Tb-t 0. Furthermore, the delay time for starting playing the graphic test signal and the sound test signal is Δ ab — Ta-Tb. That is, the audio/video delay time can be obtained by using the correlation rule, the difference signal obtained by comparing the pattern test signal obtained by the second device under test 4 with the pattern test signal collected by the first device under test in real time, the difference signal obtained by comparing the sound test signal obtained by the second device under test 4 with the sound test signal collected by the first device under test in real time, and the difference signal obtained by comparing the pattern test signal played by the second device under test 4 with the sound test signal played by the first device under test. The time difference calculation can be performed taking t0 as a reference. Ta, Tb and t0 are all the same in time unit. If the values of delta a, delta b and delta ab are zero, the synchronization of audio and video, audio and video is represented. Correspondingly, the output audio and video synchronization performance parameters are judged to be the optimal audio and video, audio and video synchronization performance parameters. If Δ a, Δ b, Δ ab are not zero, this indicates that the video delay Δ a is reached, and the audio delay Δ b is reached, which requires optimization of the synchronization performance parameters.
And audio and video synchronization performance judgment is carried out according to the time difference, and audio and video synchronization model training is further carried out, so that the synchronization of audio and video, video and video, and audio and video can be achieved, and people-to-people communication between people inside the building, people outside the building and the like can be carried out face-to-face communication experience.
Referring to fig. 3, the computing device 7 includes a first processing module 71, a second processing module 72, a third processing module 73, and a fourth processing module 74.
The first processing module 71 is used for generating a synchronous graphics command signal and a synchronous sound command signal, sending the graphics command signal to the video signal generating device 3 and the second processing module 72, and sending the sound command signal to the audio generation and acquisition instrument 8 and the third processing module 73;
the second processing module 72 is electrically connected to the first processing module 71 and the camera 6, respectively, and configured to receive the image stream sent by the camera 6, calculate a video delay time according to the image stream and the graphics command signal, and send the video delay time to the fourth processing module 74.
The third processing module 73 is electrically connected to the first processing module 71 and the audio generation and acquisition instrument 8, and is configured to receive the audio stream sent by the audio generation and acquisition instrument 8, calculate an audio time difference according to the audio stream and the sound command signal, and send the audio delay time to the fourth processing module 74;
the fourth processing module 74 is electrically connected to the second processing module 72 and the third processing module 73, respectively, and calculates an audio/video time difference between a time when the second device under test 4 plays the sound test signal and a time when the second device under test plays the graphic test signal according to the video delay time and the audio delay time.
Further, the first device to be tested comprises an image acquisition unit for acquiring the graphical test signal, for example, and an audio acquisition unit for acquiring the acoustic test signal.
As one embodiment of the present invention, preferably, the LED pattern generator may include at least one row of N LED lamps arranged along a straight line (e.g., an N × N LED lamp array, and a row of LED lamps may be selected for testing); and N LED lamps are sequentially lit at a frequency f along the same direction along which they extend, and the lighting time of each LED lamp may be 1/(N × f), or N × M LED lamps may be sequentially lit at a frequency f from left to right from top to bottom, and each lamp is lit for a time of 1/(N × f), each lamp being controlled by the processing means in the computing device 7.
The camera 6 preferably adopts a high-speed camera, and compared with a common camera, the high-speed camera has high image stability, high transmission capability and high anti-interference capability.
As a simpler embodiment of the invention, it is assumed that an N × M LED lamp array is selected as the test, half of the lamps are turned on by a high level command sent by the processing device, the lamp is turned off to indicate a low level, and the synchronous trigger signal triggers the camera 6 with a ttl level. The processing means sends a series of first square wave command signals with high and low levels (for example, high level is 5v, low level is 0) to the video signal generating means 3 in a period Ts, and the LED lamps are sequentially lighted in sequence, in the process, the duration of the high level and the low level needs to be reasonably set, namely, the duration of the low level is properly reduced to ensure that the LED lamps present a continuous scrolling pattern on the video signal generating means 3 in the view of human eyes by setting the time for ensuring that each lighted LED lamp has a visual brightness maintaining effect. When N × M square wave signals are sent, i.e. N × M LED lamps are lit once, the processing means stops sending the graphics command signals, at which point the camera 6 stops taking pictures and the camera 6 has transmitted the taken pictures to the computing device 7 during the shooting process. The processing means in the computing device 7 extracts the last captured image and locates the LED lights that are lit in the image, for example, in the usual way such as looking up the counts in the order in which the lights are lit, as long as the number of lights that are lit can be accurately calculated.
If the processing means determines that the nth LED lamp is lit, the delay time between videos is calculated according to the following formula:
Δta=(N*M-n)Ts/(N*M);
where Ts represents a cycle time for lighting the N x M LED lamps. This time may be set in advance by the computing device 7.
The above is the delay time of the video from transmission to reception, i.e. the delay time between video and video.
As one embodiment of the present invention, the following is a mode of the present invention for calculating the delay time of a sound.
The processing device generates a sound command signal while generating a graphic command signal, the sound command signal is sent to the audio generation and acquisition instrument 8 in the form of a second square wave command signal with fb frequency, the processing device receives the acquired sound test signal, compares the two signals, and calculates the difference of the number of square waves between the sent command signal and the received test signal, for example, p, so that the delay time of the sent audio and the received audio is calculated by the following formula:
Δtb=p/fb。
since the voice command signal and the graphic command signal are simultaneously issued, both have a time basis in common. The time difference between audio and video is therefore:
Δtab=Δta-Δtb。
it should be noted that sufficient time is not reserved between the completion of one audio-video synchronization test and the next test is started to ensure that the next test is not interfered by signals. And carrying out multiple tests to continuously optimize parameters, and carrying out model training to control each delay time within an acceptable range. The first square wave signal and the second pulse signal are actually pulse signals.
Specifically, it is preferable that the value of the lighting frequency f between adjacent LED lamps is greater than or equal to the frame rate that human eyes can distinguish, and the frame rate of the collected image of the image collecting unit is greater than the value of the frequency f, otherwise the image cannot be collected.
The further loudspeaker 2 is a dummy mouth to equalize the sound test signal generated by the audio signal generator and the sound collection unit collects the equalized sound test signal.
Referring to fig. 1, if the conversational communication between buildings is taken as an example, two hands-free terminals installed on the barrier of the visual intercom device of the building to be tested are respectively used as the first device to be tested 1 and the second device to be tested 4.
The distance between the simulation mouth and the first device under test 1 is greater than or equal to 10cm and less than 50cm (to avoid interference caused by a short distance, and a long distance makes the first device under test 1 unable to collect enough sound for testing, and the specific value can be set according to the performance parameters and testing requirements of the actual device).
For simplicity, the following description will be made with respect to a single-channel audio delay test, that is, a control unit generates, by using an audio generation acquisition instrument, a test unit signal including multiple sets (e.g., four sets) of CSS (composite source signals), where the first sets (e.g., the first three sets) of CSS signals may be used for training to enable channel transmission to reach a normal state, and then sends a continuous high-level signal portion of one set of CSS signals for delay measurement; preferably, the pulse width of each set of CSS signals is 248.62ms, and the adjacent CSS signals are spaced apart by 101.38ms, so that the length of each test element signal is 1298.62ms (the specific values can be adjusted adaptively according to actual test requirements, and are described herein as examples only).
When the audio and video synchronization performance test is carried out on the building visual intercom device, the test environment noise is not more than 40dB when the audio delay test is carried out, and when the video delay test is carried out, the EUT (electronic equipment) can be in a normal image transmission state, and when the audio communication of the EUT can not be cut off and the audio and video synchronization test is required, the change of the video characteristic parameters and/or the audio characteristic parameters is ensured. For example, the LED image generator can be used as a signal source for performing video frame rate, video delay and lip synchronization tests, and the LED image generator should be placed right in front of the video capturing device, and the distance between the LED matrix and the video receiving device can be adjusted to make the video receiving device on the screen. The LED matrix image above fills the screen in either the horizontal or vertical axis and the image on the screen of the video receiving device should also be able to be focused. Preferably, the LED pattern generator may be composed of a 10 × 10 LED array and a programmed controller, and the LED array may be used to generate a pattern for measurement, i.e., the LED lamps may be lighted at a time by a pulse command from the computing device 7 by sequentially lighting each LED lamp in a longitudinal or transverse direction with a single LED as a basic unit to make the presented image scroll continuously.
In the implementation of the counting scheme of the invention, parameter equipment should be debugged as much as possible so that the graphic test signal and the sound test signal are sent out at the same time as much as possible.
As an embodiment of the invention, the video signal generating means 3 may be, in addition to the time LED array, a display showing a matrix of a, B graphic elements, for example circular, rectangular, etc., which are marked in sequence according to the graphic command signal, in particular by marking the graphic elements with a color that is distinctly different from the unmarked graphic elements. The camera takes the a x B array of graphical elements and sends them to the computing device 7, which locates the marked graphical elements. Namely, the LED array in the scheme is replaced by a display with array graphic elements, and the delay time can be analyzed and calculated.
In addition, the invention also provides an audio and video synchronization test method by adopting the audio and video synchronization test system, which comprises the following specific steps.
Please refer to fig. 1-6, and particularly to fig. 5-7. A method for testing audio and video synchronization comprises the following steps.
In step S11, the computing device 7 generates a graphics command signal, and transmits the graphics command signal to the video signal generation apparatus 3.
In step S21, the video signal generator 3 generates a pattern test signal according to the pattern command signal, and outputs a synchronization trigger signal to the camera 6.
Step S31, the first device under test 1 acquires a pattern test signal and sends the pattern test signal to the second device under test 4 through the network.
In step S41, the second device under test 4 receives and displays the graphic test signal sent by the first device under test 1.
In step S51, the camera 6 receives the synchronization trigger signal, shoots the second device under test 4, and sends the image stream obtained by shooting to the computing device 7.
In step S61, the computing device 7 calculates the video delay time from the capturing of the graphic test signal by the first device under test 1 to the displaying of the graphic test signal by the second device under test 4 according to the image stream.
Further, as another embodiment of the present invention, the following steps are included.
Step S12: the computing device 7 generates a sound command signal synchronized with the graphics command signal to the audio generation and capture instrument 8.
Step S22, the audio frequency generation and acquisition instrument 8 receives the sound command signal to generate a sound test signal, and the sound test signal is emitted through the loudspeaker 2;
step S32, the first device to be tested 1 also collects sound test signals and sends the sound test signals to the second device to be tested 4 through the network;
in step S42, the second device under test 4 receives and plays the sound test signal sent by the first device under test 1;
step S52, the audio generation and collection instrument 8 obtains the audio stream of the sound test signal played by the second device under test 4 through the microphone 5 and forwards the audio stream to the computing device 7;
step S62: the calculating device 7 calculates the audio delay time from the collection of the sound test signal by the first device to be tested 1 to the playing of the sound test signal by the second device to be tested 4 according to the audio stream;
wherein steps S12-S62 are performed simultaneously with steps S11-S61.
Further, another embodiment of the present invention further includes step S7, in which the computing device 7 computes an audio/video time difference between the time when the second device under test 4 plays the sound test signal and the time when the second device under test 4 plays the graphic test signal according to the video delay time and the audio delay time.
Further, step S11 is specifically: the first processing module 71 of the computing device 7 generates and sends graphics command signals to said video signal generating means 3 and to the third processing module 73.
Further, step S12 is specifically: the second processing module 72 of the computing device 7 also generates a sound command signal synchronized with the graphics command signal and sends it to the audio generation and acquisition instrument 8 and to the fourth processing module 74 of the computing device;
further, step S61 is specifically: the third processing module 73 receives the image stream sent by the camera 6, calculates a video delay time according to the image stream and the graphics command signal, and sends the video delay time to a fifth processing module 75 of the computing device;
further, step S62 is specifically: a fourth processing module 74, which receives the audio stream sent by the audio generation and acquisition instrument 8, calculates an audio delay time according to the audio stream and the audio command signal, and sends the audio delay time to the fifth processing module 75;
step S7 specifically includes: the fifth processing module 75 calculates an audio/video time difference between the time when the second device under test 4 plays the sound test signal and the time when the second device under test 4 plays the graphic test signal according to the video delay time and the audio delay time.
Step S21 specifically includes: the video signal generating device 3 is an LED array, the LED array is composed of N M LED lamps, wherein N and M are positive integers, N represents N rows, M represents M columns, the LED lamps are sequentially lightened through a preset time interval according to a graphic command signal, the LED array generates a synchronous triggering signal through a TTL module and outputs the synchronous triggering signal to the camera, when the LED lamps are lightened, the camera 6 is triggered to start shooting, and when the LED lamps are extinguished, the camera 6 is triggered to stop shooting.
Further, step S11 is specifically: the first processing module 71 presets the pulse number W of the graphic command signal, controls the number of the lighted LED lamps to be W, and turns on all the W lamps in sequence to form a one-time test period Ts;
wherein the value range of W is 2-N M;
further, step S61 is specifically: the third processing module 73 extracts the last frame image, and locates the lit LED lamp of the last frame image, and assuming that the lit LED lamp of the last frame image is the nth one, the video delay time Δ ta is calculated as follows:
Δta=(W-n)Ts/(W)。
further, step S62 specifically includes: the fourth processing module 74 calculates the audio delay time Δ tb based on the assumption that the difference between the number of pulses of the sound command signal having a frequency fb and the number of pulses of the received sound test signal, which have been transmitted within Ts, is p, and the following formula:
Δtb=p/fb。
further, step S7 specifically includes: the fifth processing module 75 calculates an audio/video time difference Δ tab according to the video delay time Δ ta and the audio delay time Δ tb, and the calculation formula is as follows:
Δtab=Δta-Δtb。
in the present invention, the computing device 7 may send the synchronized graphics command signal and the voice command signal for multiple times to measure data, but a sufficient time interval is required to be ensured between each time to avoid mutual interference to influence the measurement effect.
While the invention has been described with reference to a preferred embodiment, it will be understood by those skilled in the art that various changes in form and detail may be made therein without departing from the spirit and scope of the invention.

Claims (16)

1. An audio and video synchronization test system is characterized in that:
the computing equipment is respectively electrically connected with a video signal generating device, an audio generating and collecting instrument and a camera;
the video signal generating device is electrically connected with the camera, and the audio generating and collecting instrument is also electrically connected with a microphone and a loudspeaker;
the loudspeaker and the video signal generating device are close to a first device to be tested;
the camera and the microphone are close to a second device to be tested;
the first equipment to be tested and the second equipment to be tested are connected through a network;
the computing equipment is used for generating a graphic command signal and sending the graphic command signal to the video signal generating device;
the video signal generating device is used for generating a graphic test signal according to the graphic command signal and outputting a synchronous trigger signal to the camera;
the first device to be tested acquires the graph test signal and sends the graph test signal to the second device to be tested through a network;
the second device to be tested is used for receiving and displaying the graphic test signal sent by the first device to be tested;
the camera is used for receiving the synchronous trigger signal, shooting the second device to be tested and sending an image stream obtained by shooting to the computing device;
the computing equipment is further used for computing video delay time from the first equipment to be tested to the second equipment to be tested to collect the graphic test signal according to the image stream.
2. An audio-video synchronization test system as claimed in claim 1, characterized in that:
the computing equipment is also used for generating a sound command signal synchronous with the graphic command signal and sending the sound command signal to the audio generation and acquisition instrument;
the audio generation and acquisition instrument is used for receiving the sound command signal and generating a sound test signal, and the sound test signal is externally emitted through the loudspeaker;
the first device to be tested collects the sound test signal and sends the sound test signal to the second device to be tested through a network;
the second device to be tested is used for receiving and playing the sound test signal sent by the first device to be tested;
the audio generation and acquisition instrument is also used for acquiring an audio stream of the sound test signal played by the second device to be tested through a microphone and forwarding the audio stream to the computing device;
the computing equipment is further used for calculating the audio delay time from the collection of the sound test signal by the first device to be tested to the playing of the sound test signal by the second device to be tested according to the audio stream.
3. An audio-video synchronization test system as claimed in claim 2, characterized in that: and the computing equipment is also used for computing the audio and video time difference between the time of playing the sound test signal and the time of playing the graphic test signal by the second equipment to be tested according to the video delay time and the audio delay time.
4. An audio-video synchronization test system as claimed in claim 3, characterized in that: the computing device includes:
the first processing module is respectively and electrically connected with a third processing module and the video signal generating device and is used for generating the synchronous graphic command signal and sending the synchronous graphic command signal to the video signal generating device and the third processing module;
the second processing module is respectively and electrically connected with a fourth processing module, the audio generation and acquisition instrument and the first processing module, and is used for generating the sound command signal synchronous with the graphic command signal and sending the sound command signal to the audio generation and acquisition instrument and the fourth processing module;
the third processing module is electrically connected with the camera and is used for receiving the image stream sent by the camera, calculating the video delay time according to the image stream and the graphics command signal, and sending the video delay time to a fifth processing module;
the fourth processing module is further electrically connected to the audio generation and acquisition instrument, and is configured to receive the audio stream sent by the audio generation and acquisition instrument, calculate the audio delay time according to the audio stream and the sound command signal, and send the audio delay time to a fifth processing module;
and the fifth processing module is respectively connected with the third processing module and the fourth processing module, and calculates the audio and video time difference between the time of playing the sound test signal and the time of playing the graphic test signal by the second device to be tested according to the video delay time and the audio delay time.
5. An audio-video synchronization test system according to claim 4,
the video signal generating device is an LED array, the LED array is composed of N x M LED lamps, wherein N and M are positive integers, N represents N rows, and M represents M columns, and the LED lamps are used for being sequentially lightened at a preset time interval according to the graphic command signal;
the LED array generates a synchronous trigger signal through the TTL module and outputs the synchronous trigger signal to the camera, when the LED lamp is lightened, the camera is triggered to start shooting, and when the LED lamp is extinguished, the camera is triggered to stop shooting.
6. The audio-video synchronization test system according to claim 5, wherein the first processing module is configured to preset the number W of pulses of the graphic command signal, control the number of the lighted LED lamps to be W, and turn on all the W lamps in sequence to form a one-time test period Ts;
wherein the value range of W is 2-N M;
the third processing module is configured to extract a last frame image, locate a lit LED lamp of the last frame image, and if the lit LED lamp of the last frame image is the nth one, the video delay time Δ ta is calculated according to the following formula:
Δta=(W-n)Ts/(W)。
7. an audio-video synchronization test system according to claim 6,
the fourth processing module is configured to calculate the audio delay time Δ tb according to a difference assumed to be p between the number of pulses of the sound command signal with frequency fb that has been transmitted within Ts time and the number of pulses of the received sound test signal, and the formula is as follows:
Δtb=p/fb。
8. an audio-video synchronization test system according to claim 7,
the fifth processing module is configured to calculate the audio/video time difference Δ tab according to the video delay time Δ ta and the audio delay time Δ tb, where a calculation formula is as follows:
Δtab=Δta-Δtb。
9. an audio-video synchronization test method, characterized by using the audio-video synchronization test system as claimed in claim 1, and comprising the following steps:
step S11, the computing equipment generates a graphic command signal and sends the graphic command signal to a video signal generating device;
step S21, the video signal generating device generates a pattern test signal according to the pattern command signal and outputs a synchronous trigger signal to the camera;
step S31, the first device to be tested collects the graphic test signal and sends the graphic test signal to the second device to be tested through a network;
step S41, the second device to be tested receives and displays the graphic test signal sent by the first device to be tested;
step S51, the camera receives the synchronous trigger signal, shoots the second device to be tested, and sends the image stream obtained by shooting to the computing device;
step S61, the computing device calculates, according to the image stream, a video delay time from the acquisition of the graphics test signal by the first device under test to the display of the graphics test signal by the second device under test.
10. An audio-video synchronization test method as claimed in claim 9, characterized in that:
step S12: the computing equipment generates a sound command signal synchronous with the graphic command signal and sends the sound command signal to the audio generation and acquisition instrument;
step S22, the audio frequency generation and acquisition instrument receives the sound command signal to generate a sound test signal, and the sound test signal is emitted through the loudspeaker;
step S32, the first device to be tested also collects the sound test signal and sends the sound test signal to the second device to be tested through the network;
in step S42, the second device under test receives and plays the sound test signal sent by the first device under test;
step S52, the audio generation and acquisition instrument acquires an audio stream of the sound test signal played by the second device under test via a microphone and forwards the audio stream to the computing device;
step S62: the computing equipment calculates the audio delay time from the collection of the sound test signal by the first equipment to be tested to the playing of the sound test signal by the second equipment to be tested according to the audio stream;
wherein the steps S12-S62 are performed simultaneously with the steps S11-S61.
11. An audio-video synchronization test method according to claim 10, characterized in that: includes step S7: and the computing equipment computes the audio-video time difference between the time of playing the sound test signal and the time of playing the graphic test signal by the second equipment to be tested according to the video delay time and the audio delay time.
12. An audio-video synchronization test method according to claim 11, characterized in that:
the step S11 specifically includes: the first processing module of the computing equipment generates the graphics command signal and sends the graphics command signal to the video signal generating device and the third processing module; the step S12 specifically includes: the second processing module of the computing equipment also generates a sound command signal synchronous with the graphic command signal and sends the sound command signal to the audio generation and acquisition instrument and a fourth processing module of the computing equipment;
the step S61 specifically includes: the third processing module receives the image stream sent by the camera, calculates the video delay time according to the image stream and the graphics command signal, and sends the video delay time to a fifth processing module of the computing device;
the step S62 specifically includes: the fourth processing module receives the audio stream sent by the audio generation and acquisition instrument, calculates the audio delay time according to the audio stream and the sound command signal, and sends the audio delay time to a fifth processing module;
the step S7 specifically includes: and the fifth processing module calculates the audio and video time difference between the time of playing the sound test signal and the time of playing the graphic test signal by the second device to be tested according to the video delay time and the audio delay time.
13. An audio-video synchronization test method according to claim 12,
the step S21 specifically includes: the video signal generating device is an LED array, the LED array is composed of N x M LED lamps, N and M are positive integers, N represents N rows, M represents M columns, the LED lamps are sequentially lightened through a preset time interval according to the graphic command signal, the LED array generates a synchronous triggering signal through a TTL module and outputs the synchronous triggering signal to the camera, when the LED lamps are lightened, the camera is triggered to start shooting, and when the LED lamps are extinguished, the camera is triggered to stop shooting.
14. An audio-video synchronization test method according to claim 13,
the step S11 specifically includes: the first processing module presets the pulse number W of the graphic command signals, controls the number of the lighted LED lamps to be W, and lights the W lamps in sequence to be lighted into a one-time test period Ts;
wherein the value range of W is 2-N M;
the step S61 specifically includes: the third processing module extracts a last frame image, positions the lighted LED lamp of the last frame image, and assumes that the lighted LED lamp of the last frame image is the nth one, the video delay time Δ ta is calculated according to the following formula:
Δta=(W-n)Ts/(W)。
15. an audio-video synchronization test method according to claim 14,
the step S62 specifically includes: the fourth processing module calculates the audio delay time Δ tb according to the assumption that the difference between the number of pulses of the sound command signal with frequency fb that has been transmitted within Ts time and the number of pulses of the received sound test signal is p, and the formula is as follows:
Δtb=p/fb。
16. an audio-video synchronization test method according to claim 15,
the step S7 specifically includes: the fifth processing module calculates the audio/video time difference Δ tab according to the video delay time Δ ta and the audio delay time Δ tb, and the calculation formula is as follows:
Δtab=Δta-Δtb。
CN202010148288.2A 2020-03-05 2020-03-05 System and method for audio and video synchronization test Pending CN111277823A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010148288.2A CN111277823A (en) 2020-03-05 2020-03-05 System and method for audio and video synchronization test

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010148288.2A CN111277823A (en) 2020-03-05 2020-03-05 System and method for audio and video synchronization test

Publications (1)

Publication Number Publication Date
CN111277823A true CN111277823A (en) 2020-06-12

Family

ID=71002363

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010148288.2A Pending CN111277823A (en) 2020-03-05 2020-03-05 System and method for audio and video synchronization test

Country Status (1)

Country Link
CN (1) CN111277823A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112929654A (en) * 2021-03-16 2021-06-08 腾讯音乐娱乐科技(深圳)有限公司 Method, device and equipment for detecting sound and picture synchronization and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102905056A (en) * 2012-10-18 2013-01-30 利亚德光电股份有限公司 Method and device for processing video image
CN104023229A (en) * 2014-06-23 2014-09-03 公安部第三研究所 Detection method and system for performances of non-contact type image system
CN106060534A (en) * 2016-06-03 2016-10-26 公安部第三研究所 System and method for testing synchronization of audio and video
US20170163978A1 (en) * 2015-12-08 2017-06-08 Electronics And Telecommunications Research Institute System and method for synchronizing audio signal and video signal

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102905056A (en) * 2012-10-18 2013-01-30 利亚德光电股份有限公司 Method and device for processing video image
CN104023229A (en) * 2014-06-23 2014-09-03 公安部第三研究所 Detection method and system for performances of non-contact type image system
US20170163978A1 (en) * 2015-12-08 2017-06-08 Electronics And Telecommunications Research Institute System and method for synchronizing audio signal and video signal
CN106060534A (en) * 2016-06-03 2016-10-26 公安部第三研究所 System and method for testing synchronization of audio and video

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112929654A (en) * 2021-03-16 2021-06-08 腾讯音乐娱乐科技(深圳)有限公司 Method, device and equipment for detecting sound and picture synchronization and storage medium
CN112929654B (en) * 2021-03-16 2022-03-29 腾讯音乐娱乐科技(深圳)有限公司 Method, device and equipment for detecting sound and picture synchronization and storage medium

Similar Documents

Publication Publication Date Title
EP3468183B1 (en) System and method for audio and video synchronization test
US20080263612A1 (en) Audio Video Synchronization Stimulus and Measurement
JP6958360B2 (en) Information processing equipment, information processing methods, and programs
CN109963184A (en) A kind of method, apparatus and electronic equipment of audio-video network broadcasting
JPWO2017098772A1 (en) Information processing apparatus, information processing method, and program
Lindau The perception of system latency in dynamic binaural synthesis
CN102724604A (en) Sound processing method for video meeting
CN113692091B (en) Equipment control method, device, terminal equipment and storage medium
CN111277823A (en) System and method for audio and video synchronization test
CN111614929A (en) Acoustic positioning video conference method, system and storage medium based on 5G transmission
JP4501037B2 (en) COMMUNICATION CONTROL SYSTEM, COMMUNICATION DEVICE, AND COMMUNICATION METHOD
CN109034267B (en) Piece caudal flexure intelligent selection device
CN102894959B (en) Device for inspection, listening and inquiry in traditional Chinese medical science and control method
JP5967848B1 (en) Conference system
TWI548278B (en) Audio/video synchronization device and audio/video synchronization method
JP2001326950A (en) Line time difference measuring device and signal generator for line time difference measuring device
CN109561239B (en) Piece caudal flexure intelligent selection platform
US20220358932A1 (en) Information processing apparatus, information processing method, and program
US11826648B2 (en) Information processing apparatus, information processing method, and recording medium on which a program is written
US20240107252A1 (en) Insertion of forced gaps for pervasive listening
US20210152813A1 (en) Method of measuring video latency
Yadav et al. Detection of headtracking in room acoustic simulations for one’s own voice
CN117793607A (en) Playing control method and device
JPH10145645A (en) Video signal/audio signal generator for detecting time difference between image and sound
KR20070087680A (en) Interactive video communication system with chat

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20200612