WO2007116205A1 - Method and apparatus for measuring audio/video sync delay - Google Patents
Method and apparatus for measuring audio/video sync delay Download PDFInfo
- Publication number
- WO2007116205A1 WO2007116205A1 PCT/GB2007/001191 GB2007001191W WO2007116205A1 WO 2007116205 A1 WO2007116205 A1 WO 2007116205A1 GB 2007001191 W GB2007001191 W GB 2007001191W WO 2007116205 A1 WO2007116205 A1 WO 2007116205A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- audio
- encoded
- video
- time
- timestamps
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 31
- 230000005236 sound signal Effects 0.000 claims abstract description 13
- 230000000007 visual effect Effects 0.000 claims abstract description 10
- 230000001360 synchronised effect Effects 0.000 claims abstract description 6
- 238000001514 detection method Methods 0.000 description 10
- 230000001934 delay Effects 0.000 description 5
- 230000005540 biological transmission Effects 0.000 description 3
- 238000013500 data storage Methods 0.000 description 3
- 238000004519 manufacturing process Methods 0.000 description 2
- 230000000630 rising effect Effects 0.000 description 2
- 241001482107 Alosa sapidissima Species 0.000 description 1
- 210000004556 brain Anatomy 0.000 description 1
- 238000013144 data compression Methods 0.000 description 1
- 238000000354 decomposition reaction Methods 0.000 description 1
- 230000003111 delayed effect Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000002045 lasting effect Effects 0.000 description 1
- 238000001228 spectrum Methods 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/434—Disassembling of a multiplex stream, e.g. demultiplexing audio and video streams, extraction of additional data from a video stream; Remultiplexing of multiplex streams; Extraction or processing of SI; Disassembling of packetised elementary stream
- H04N21/4341—Demultiplexing of audio and video streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N17/00—Diagnosis, testing or measuring for television systems or their details
- H04N17/004—Diagnosis, testing or measuring for television systems or their details for digital television systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/236—Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
- H04N21/2368—Multiplexing of audio and video streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/8106—Monomedia components thereof involving special audio data, e.g. different tracks for different languages
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/85—Assembly of content; Generation of multimedia applications
- H04N21/854—Content authoring
- H04N21/8547—Content authoring involving timestamps for synchronizing content
Definitions
- a common requirement is to pass the digital data through one or more encoding processes, for example prior to the broadcast transmission of the digital audio-visual data, for example the broadcast of a television programme.
- the coding processes habitually involve data compression and the use of digital audio filters to process the audio signal.
- the encoding process may also typically involve multiplexing a plurality of separate data streams together.
- the audio data will be processed differently from the video data, and each of the different stages in the encoding process can potentially introduce a time delay to the digital data signal
- the overall encoding process can potentially introduce a loss of synchronisation between the audio and video data, which will be most noticeable as a loss of lip-sync in video footage of speaking characters.
- the human brain can perceive even quite small time delays between the video and audio data, with the circumstances in which the audio signal leads the video signal being most noticeable.
- the applicable encoding and transmission standards stipulate maximum time delays between the audio and video data. For example, according to some standards the audio signal must not lead the corresponding video signal by a time delay greater than 40ms.
- a method of determining the delay between an audio and visual signal comprising:
- Encoding the audio and video signals to generate digitally encoded audio and video data streams; Analysing the encoded video and audio data streams to extract each of the audibly and visually encoded timestamps; and
- the audio and video timestamps are encoded as a binary code, for example Gray code.
- Each visually encoded timestamp preferably comprises a plurality of display segments, the colour or shade of each segment being representative of a binary state.
- the display segments comprise a portion of a macro block.
- Each audibly encoded timestamp preferably comprises an audio tone having a plurality of predetermined frequency components, the presence of a frequency component being representative of a binary state.
- each encoded time stamp comprises a frame count.
- apparatus for determining the delay between a digitally encoded audio and video signal, the video signal having a plurality of sequential timestamps visually encoded thereon and the audio signal having a corresponding plurality of timestamps audibly encoded thereon, the audio and video signals being synchronised to one another, the apparatus comprising:
- a video timestamp detector arranged to detect each of the timestamps encoded in the encoded video signal, decode the timestamp and provide a first time signal representative of the actual time of receipt of the video timestamp;
- An audio timestamp detector arranged to detect each of the timestamps encoded in the encoded audio signal, decode the timestamp and provide a second time signal representative of the actual time of receipt of the audio timestamp;
- a timestamp comparator arranged to receive the first and second time signals and measure any delay between their time of receipt.
- Figure 1 schematically illustrates the timing and duration of audio and video events included in a possible test signal
- Figure 2 schematically illustrates a time delay analysis system for determining the time delays between the audio and video signals shown in Figure 1 ;
- Figure 3 schematically illustrates the relative timings of an audio and video signal as shown in Figure 1 in which there is a delay between the audio and video signals;
- Figure 4 schematically illustrates a method of visually encoding a time stamp according to an embodiment of the present information
- Figure 5 schematically illustrates a method of audio encoding a time stamp according to an embodiment of the present information
- Figure 6 schematically illustrates a time delay analysis system according to an embodiment of the present invention for determining the time delays between audio and video signals having time stamps encoded therein of the kind illustrated in Figures 4 & 5.
- any time delay between audio and video data subsequent to an encoding process having been performed on the originally available audio and video data is determined utilising a predetermined video sequence having known timing properties.
- the video/audio data sequence is provided in either an uncompressed data format or in a standard encoded data format, such as for example MPEG-2 video or audio.
- the predetermined audio/video sequence comprises a series of visible "flashes" having a predetermined duration and time interval between each flash.
- the sequence also comprises a corresponding number of audible tones whose duration and time interval between tones exactly corresponds to the occurrences of the visible flashes.
- An example of an appropriate timing diagram for the visible and audible signals is schematically illustrated in
- the upper signal trace 2 represents the binary levels for the visible signal, with the signal either being totally black or totally white in visible appearance.
- the lower signal trace 4 represents the audible signal, with the upper signal level representing a production of an audible tone and the lower signal level representing the absence of a tone.
- a visible flash and audible tone of duration of 1 unit is subsequently produced. This is followed by a further time period during which no visible flash or audible tone is produced, this second time period having a duration of 2 units.
- the total sequence comprises five periods during which a visible flash and audible tone are produced, each period lasting one time unit longer than the preceding period, with correspondingly increasing time periods in between during which no visible flash or audible tone is produced.
- the entire sequence lasts for a total of 30 time units, which will typically be 30 seconds.
- the entire sequence preferably continually repeats.
- the visible flash is produced in at least the macro block, or at least an integer multiple thereof, that is shown at the top left hand corner of the display screen.
- a 4x4 array of blocks i.e. 32x32 pixels, is used to encode the visible flash.
- This location is carefully chosen since, due to the scanning method of generating a displayed image as will be appreciated by those skilled in the art, the digital data representing this part of the display screen will occur very early in the relevant data stream and will consequently practically always be correctly encoded.
- the selection of the visible flash as a 32x32 pixel area will also tend to ensure the correct encoding of this video data.
- the use of only black and white shades for the visible flash will maximise the likelihood of the video data being correctly encoded since these are "basic" digital values unlikely to be corrupted by the encoding process.
- the audio tone is provided as a tone with only a single frequency component, for example at 10KHz, or some other single frequency. Since only a single frequency component is utilised for the audio tone, it should be faithfully encoded by any audio encoder included within the encoding system under test.
- visual data may be provided to the user, for example a larger visual representation of the visible flash, for example as a series of rotating circular segments, each segment being representative of a single time unit such that a complete sequence requires a full "revolution" through the multiple segments.
- a larger visual representation of the visible flash for example as a series of rotating circular segments, each segment being representative of a single time unit such that a complete sequence requires a full "revolution" through the multiple segments.
- the predetermined audio visual sequence is passed through the encoding system under test and the encoded digital data stream subsequently analysed.
- the analysis process comprises detecting one or both of the beginning and end of one of the visible flashes by detecting the point in time within the encoded data stream at which the 32x32 pixel macro block integer changes from "black” to "white” or vice versa.
- the time at which this occurs is accurate to within the duration of 1 frame of visual data, since the display is only refreshed every frame, A typical frame rate is 25 frames per second.
- Concurrently the encoded audio signal is analysed to determine one or both of the beginning and end of the audio tones.
- a preferred method of detecting the beginning or end of the audio tone is to detect the sharply rising or falling amplitude of the tone as each transition from "tone” to "no tone” or vice versa occurs.
- the analysis process can thus determine any time delay between the video and audio "events" (an event being rising or falling audio or video signal edge).
- any determined delay that falls outside a predetermined set of parameters, such as those set by one or more transmission standards, causes an alert to be automatically generated.
- a system in accordance with the applicant's co-pending analysis scheme for determining any loss of video/audio synchronisation in an encoded data stream is schematically represented in Figure 2.
- a predetermined video stream 10 as described above in an un- encoded state is stored on a data storage medium such as a hard disk 12 and is provided as an input to the encoding system 14 to be tested.
- the encoding system will generally output an encoded data stream that can be decomposed to separate video 16 and audio 18 streams.
- Each of the video and audio streams are provided as inputs to an analysis engine 20 and input to separate video and audio event detection units 22, 24.
- each event detection unit is arranged to detect the relevant video or audio 'events' of the encoded test data stream, these being the beginning or ends, or both, of the visible 'flashes' and audio tones as discussed above in relation to Figure 1, and to provide an output signal indicative of when each event occurs.
- the output signals from each of the audio and video event detection units 22, 24 is provided to a time comparison, unit 26 that is arranged to measure any time interval present between the output signals from the event detection units and thus any time interval, be it lag or lead, between the occurrences of the audio and video 'events'.
- This time interval data is provided from the time comparison unit 26 to an output interface unit 28 that is arranged to provide the time interval data to an appropriate user interface.
- the output interface unit 28 is also arranged to compare any time delay between the audio and video signals with defined maximum permitted delays that may be stored in a further data storage area 30 or may be stored internally to the output interface unit. If a detected time interval exceeds a predefined value then the output interface unit may be arranged to provide an alarm signal.
- the sequence of visible flashes and audible tones comprises 'events' with increasing time intervals between each 'event'. This ensures that should the time delay between the video and audio signals be great enough for one of the video events to coincide with an audio event this 'false' synchronisation, which would not cause the analysis engine to generate a report or alarm, will notjbe maintained at the next occurrence of a video and audio event.
- Figure 3 This is schematically illustrated in Figure 3, in which the upper trace 32 represents the video event signal and the lower trace 34 represents the audio event signal.
- the audio event signal has been delayed relative to the video signal by the encoding process by a time period of 3 time units, say seconds, as represented by arrow A.
- the beginning of the second video event 36 occurs at the same time as the beginning of the first audio event 38. If these are the first video and audio events detected by the analysis engine then a false report of synchronisation between the audio and video streams may be provided. However, at the beginning of the next video event 40 it can be seen that the audio stream is out of synchronisation, since the events are not evenly spaced apart and do not have a constant duration. Consequently the analysis engine is able to determine that in fact the video and audio streams are not in synchronisation. If the analysis engine detects both the beginning and end of the video and audio events then the loss of synchronisation will be detected sooner since the end of the first audio event 38 will occur before the end of the second video event 36, even though the beginning of both events coincided. In this instance the loss of synchronisation between the audio and video streams is detected by the analysis engine within one time unit, for example one second.
- any loss in synchronisation can only be determined, and the delay measured, at best when an audio or video event occurs and as described above in relation to Figure 3 the time required to determine a loss in synchronisation can be multiple time units if a 'miss-match' between video and audio events occurs due to a gross loss of synchronisation.
- This is wasteful of system resources since each second of audio-visual data will comprise, typically, 25 frames of data. In other words, the same data is being processed 25 times a second.
- an analysis scheme is provided that allows improved determination of the time delay between audio and video data streams. This is accomplished by providing a predetermined audio-visual test sequence to be encoded in the encoding system under test that includes audio and visual data that allows each frame to be identified.
- the visual encoding is accomplished using a pattern of black and white squares to represent a binary code.
- a preferred binary code is Gray code, since a well known property of Gray code is that when presented in sequence only one bit of the binary word changes at a time.
- An example of a possible sequence of black and white squares is illustrated in Figure 4, in which the sequences of squares for three consecutive frames of audio-visual data are shown.
- the first 5 square sequence represents the Gray code 00101, which is decimal 7, and hence is used to identify the frame as frame number 7.
- the second and third sequences represent 00100 and 01100, decimal 8 and 9, respectively.
- the individual squares are encoded as discrete blocks or integer parts of a macroblock, such as 2x2 block of 32x32 pixels, so as to facilitate the reliable error free encoding of the sequence of squares, thus reliably maintaining the encoded frame identification code.
- the audio signal is also encoded with a timing sequence that serves to identify which frame of the video signal the particular section of audio data should be synchronised to.
- a timing sequence that serves to identify which frame of the video signal the particular section of audio data should be synchronised to.
- this is accomplished by the inclusion of an audio tone that is made up of a number of separate discrete frequencies, each frequency representing one bit in the data word, in an analogous manner to each square of the video code representing a single bit of the Gray code.
- This allows the encoded tone to be analysed using Fourier analysis techniques to determine the presence or otherwise of the individual frequency components and thus the binary code represented by the tone.
- An example of the frequency analysis of such an encoded tone is schematically illustrated in Figure 5.
- the horizontal axis represents the frequency of detected frequency components, in KHz, whilst the vertical axis represents the power of the component.
- two frequency components 40 are shown at 9KHz and 12 KHz. If the selected code is a 5 bit code with the most significant bit represented by the frequency component centred at 3KHZ and the least significant bit at 15KHz, then the frequency spectrum shown in Figure 5 is taken to represent the binary word 00101, or decimal 7 in Gray code. Care must be taken in the selection of the frequency components selected to represent individual bits of the encoded timing word since it is common practice for audio encoders to discard certain frequency components of a signal based on an analysis of what frequencies the human ear will and will not be able to hear. The frequency components selected for the timing word must therefore be such that they will not be discarded by such encoding techniques.
- the audio code may be encoded as a series of short audio tones in a predetermined time interval, each tone in the series representing a bit within the timing word and the presence or not of a tone representing the binary state of the bit. So for example, a series of eight audio tones may be used to represent an 8 bit binary word. The frequency of the audio tones may be pre-selected to facilitate their detection.
- FIG. 6 schematically illustrates an analysis engine for analysing a test audio-visual data stream of the format described above according to embodiments of the present invention after it has been encoded by an encoding system under test.
- the basic components are the same as for the system illustrated in Figure 2.
- the individual video and audio data streams 616, 618 are provided as inputs to respective time code detection units 622, 624.
- Each time code detection unit is arranged to identify and decode the embedded video and audio time codes. Consequently in preferred embodiments the video time code detection unit 622 will be arranged to locate the sequence of coded black and white squares and determine the binary code represented by the particular sequence, thus identifying the individual frame number. The point in time at which each frame is received is also determined.
- the audio time code detection unit 624 is preferably arranged to perform the necessary frequency analysis on the embedded audio time code to determine the present frequency components and thus the represented binary code.
- the relevant output signals from the time code detection units are provided to a time comparison unit 626 that is arranged to determine any time delay between the audio and video data streams on the basis of the time of receipt of corresponding portions of the data streams, as identified by the relevant embedded time codes. Any time delay is provided as an input to a report and/or alert unit 628 that is arranged to determine if the time delay exceeds certain predetermined parameters that may, for example, be stored as a look up table in local data storage 630.
- the analysis engine is capable of determining the relative positions of the separate audio and video data streams within the space of a single frame and can provide audio/video time delay information for each frame, as opposed to delay information only for each video/audio 'event' as is the case with the scheme discussed in the applicant's co-pending application. Consequently, the apparatus and method of the present invention allows improved speed of providing the delay information and improved resolution of the delay information.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- General Health & Medical Sciences (AREA)
- Computer Security & Cryptography (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Television Receiver Circuits (AREA)
- Television Signal Processing For Recording (AREA)
- Synchronisation In Digital Transmission Systems (AREA)
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP07732245A EP2005762A1 (en) | 2006-04-10 | 2007-03-30 | Method and apparatus for measuring audio/video sync delay |
JP2009504802A JP5025722B2 (ja) | 2006-04-10 | 2007-03-30 | オーディオ/ビデオ同期遅延測定方法及び装置 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
GB0607215.1 | 2006-04-10 | ||
GB0607215A GB2437123B (en) | 2006-04-10 | 2006-04-10 | Method and apparatus for measuring audio/video sync delay |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2007116205A1 true WO2007116205A1 (en) | 2007-10-18 |
WO2007116205A9 WO2007116205A9 (en) | 2009-04-09 |
Family
ID=36539692
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/GB2007/001191 WO2007116205A1 (en) | 2006-04-10 | 2007-03-30 | Method and apparatus for measuring audio/video sync delay |
Country Status (4)
Country | Link |
---|---|
EP (1) | EP2005762A1 (ja) |
JP (1) | JP5025722B2 (ja) |
GB (1) | GB2437123B (ja) |
WO (1) | WO2007116205A1 (ja) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2009267766A (ja) * | 2008-04-25 | 2009-11-12 | Taito Corp | タイミング補正プログラム、携帯端末及び処理タイミング同期方法 |
CN112351273A (zh) * | 2020-11-04 | 2021-02-09 | 新华三大数据技术有限公司 | 一种视频播放质量检测方法及装置 |
CN112601078A (zh) * | 2020-12-11 | 2021-04-02 | 杭州当虹科技股份有限公司 | 一种基于视频的编码器延时的自动测量方法 |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2009115121A2 (en) * | 2008-03-19 | 2009-09-24 | Telefonaktiebolaget Lm Ericsson (Publ) | Method and apparatus for measuring audio-video time skew and end-to-end delay |
US20110170537A1 (en) * | 2010-01-08 | 2011-07-14 | Marius Ungureanu | One Way and Round Trip Delays Using Telephony In-Band Tones |
CN103796006A (zh) * | 2012-10-30 | 2014-05-14 | 中兴通讯股份有限公司 | 一种唇音同步测试的系统与方法 |
CN106358039B (zh) * | 2016-09-07 | 2019-02-01 | 深圳Tcl数字技术有限公司 | 音画同步测试方法及装置 |
GB2586986B (en) * | 2019-09-10 | 2023-05-24 | Hitomi Ltd | Signal variation measurement |
CN112601077B (zh) * | 2020-12-11 | 2022-07-26 | 杭州当虹科技股份有限公司 | 一种基于音频的编码器延时的自动测量方法 |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0928110A2 (en) * | 1997-12-30 | 1999-07-07 | Sarnoff Corporation | Image signal processing for electronic watermarking |
JP2001298757A (ja) * | 2000-04-11 | 2001-10-26 | Nippon Hoso Kyokai <Nhk> | 映像・音声遅延時間差測定装置 |
JP2003259314A (ja) * | 2002-02-26 | 2003-09-12 | Nippon Hoso Kyokai <Nhk> | 映像音声同期方法及びそのシステム |
JP2004158913A (ja) * | 2002-11-01 | 2004-06-03 | Canon Inc | 音声画像処理装置 |
WO2005004470A1 (en) * | 2003-07-01 | 2005-01-13 | Lg Electronics Inc. | Method and apparatus for testing lip-sync of digital television receiver |
US20050012860A1 (en) * | 1995-12-07 | 2005-01-20 | Cooper J. Carl | A/V timing measurement for MPEG type television |
US20050219366A1 (en) * | 2004-03-31 | 2005-10-06 | Hollowbush Richard R | Digital audio-video differential delay and channel analyzer |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4326102A (en) * | 1980-02-04 | 1982-04-20 | Msi Data Corporation | Audio data transmission device with coupling device |
JPH05236513A (ja) * | 1992-02-21 | 1993-09-10 | Shibasoku Co Ltd | テレビジョン映像信号と音声信号の伝送遅延時間差の測定方法 |
US5818520A (en) * | 1996-02-12 | 1998-10-06 | Tektronix, Inc. | Programmable instrument for automatic measurement of compressed video quality |
EP0888019A1 (en) * | 1997-06-23 | 1998-12-30 | Hewlett-Packard Company | Method and apparatus for measuring the quality of a video transmission |
GB9804071D0 (en) * | 1998-02-27 | 1998-04-22 | Ridgeway Systems And Software | Audio-video telephony |
US6414960B1 (en) * | 1998-12-29 | 2002-07-02 | International Business Machines Corp. | Apparatus and method of in-service audio/video synchronization testing |
JP2004242130A (ja) * | 2003-02-07 | 2004-08-26 | Nippon Hoso Kyokai <Nhk> | 映像・音声伝送時間差測定用の信号発生装置およびその方法、並びに信号解析装置およびその方法 |
WO2005025224A1 (ja) * | 2003-09-02 | 2005-03-17 | Sony Corporation | コンテンツ受信装置、ビデオオーディオ出力タイミング制御方法及びコンテンツ提供システム |
CN100442858C (zh) * | 2005-10-11 | 2008-12-10 | 华为技术有限公司 | 分组网络中多媒体实时传输的唇同步方法及其装置 |
-
2006
- 2006-04-10 GB GB0607215A patent/GB2437123B/en active Active
-
2007
- 2007-03-30 WO PCT/GB2007/001191 patent/WO2007116205A1/en active Application Filing
- 2007-03-30 EP EP07732245A patent/EP2005762A1/en not_active Withdrawn
- 2007-03-30 JP JP2009504802A patent/JP5025722B2/ja active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050012860A1 (en) * | 1995-12-07 | 2005-01-20 | Cooper J. Carl | A/V timing measurement for MPEG type television |
EP0928110A2 (en) * | 1997-12-30 | 1999-07-07 | Sarnoff Corporation | Image signal processing for electronic watermarking |
JP2001298757A (ja) * | 2000-04-11 | 2001-10-26 | Nippon Hoso Kyokai <Nhk> | 映像・音声遅延時間差測定装置 |
JP2003259314A (ja) * | 2002-02-26 | 2003-09-12 | Nippon Hoso Kyokai <Nhk> | 映像音声同期方法及びそのシステム |
JP2004158913A (ja) * | 2002-11-01 | 2004-06-03 | Canon Inc | 音声画像処理装置 |
WO2005004470A1 (en) * | 2003-07-01 | 2005-01-13 | Lg Electronics Inc. | Method and apparatus for testing lip-sync of digital television receiver |
US20050219366A1 (en) * | 2004-03-31 | 2005-10-06 | Hollowbush Richard R | Digital audio-video differential delay and channel analyzer |
Non-Patent Citations (1)
Title |
---|
See also references of EP2005762A1 * |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2009267766A (ja) * | 2008-04-25 | 2009-11-12 | Taito Corp | タイミング補正プログラム、携帯端末及び処理タイミング同期方法 |
CN112351273A (zh) * | 2020-11-04 | 2021-02-09 | 新华三大数据技术有限公司 | 一种视频播放质量检测方法及装置 |
CN112351273B (zh) * | 2020-11-04 | 2022-03-01 | 新华三大数据技术有限公司 | 一种视频播放质量检测方法及装置 |
CN112601078A (zh) * | 2020-12-11 | 2021-04-02 | 杭州当虹科技股份有限公司 | 一种基于视频的编码器延时的自动测量方法 |
CN112601078B (zh) * | 2020-12-11 | 2022-07-26 | 杭州当虹科技股份有限公司 | 一种基于视频的编码器延时的自动测量方法 |
Also Published As
Publication number | Publication date |
---|---|
EP2005762A1 (en) | 2008-12-24 |
JP2009533920A (ja) | 2009-09-17 |
GB2437123A (en) | 2007-10-17 |
WO2007116205A9 (en) | 2009-04-09 |
JP5025722B2 (ja) | 2012-09-12 |
GB2437123B (en) | 2011-01-26 |
GB0607215D0 (en) | 2006-05-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2007116205A1 (en) | Method and apparatus for measuring audio/video sync delay | |
KR100499037B1 (ko) | 디지털 텔레비젼 수신기의 립 싱크 테스트 방법 및 장치 | |
US9536545B2 (en) | Audio visual signature, method of deriving a signature, and method of comparing audio-visual data background | |
KR102260946B1 (ko) | 미디어 재생 에러들을 검출하기 위한 기술들 | |
EP3171593B1 (en) | Testing system and method | |
US20050219366A1 (en) | Digital audio-video differential delay and channel analyzer | |
CA2428064A1 (en) | Apparatus and method for determining the programme to which a digital broadcast receiver is tuned | |
US20080037954A1 (en) | Automatic Video Glitch Detection and Audio-Video Synchronization Assessment | |
CN102523063A (zh) | 用于监视来自各种源的音频/视觉内容的方法及装置 | |
CN101616331A (zh) | 一种对视频帧率及音视频同步性能进行测试的方法 | |
US20070296870A1 (en) | Pulldown correction for progressive display of audiovisual recordings | |
CN104853244A (zh) | 用于管理音视频、音频或视频内容的方法和装置 | |
CN114666636A (zh) | 音画同步检测方法及计算机可读存储介质 | |
CN112601077B (zh) | 一种基于音频的编码器延时的自动测量方法 | |
CN112601078B (zh) | 一种基于视频的编码器延时的自动测量方法 | |
EP2725578A1 (en) | Loudness log for recovery of gated loudness measurements and associated analyzer | |
GB2437122A (en) | Method and apparatus for measuring audio/video sync delay | |
US10097819B2 (en) | Testing system, testing method, computer program product, and non-transitory computer readable data carrier | |
CN113055711B (zh) | 一种音视频同步检测方法及其检测系统 | |
CN116437068A (zh) | 一种唇音同步的测试方法、装置、电子设备和存储介质 | |
KR101721224B1 (ko) | 실시간 영상 및 음성 왜곡 검출 방법 및 장치 | |
KR100966830B1 (ko) | 오디오 워터마크 삽입 장치 및 오디오 워터마크 검출 장치와 이를 이용한 음향 왜곡 검출 자동화 시스템 | |
JP5205254B2 (ja) | 放送実績取得システム、情報埋め込み装置、情報検出装置及び放送実績取得方法 | |
CN106470339B (zh) | 终端设备、及音频视频同步检测方法 | |
JP2010157914A (ja) | 放送実績取得システム、情報埋め込み装置、情報検出装置及び放送実績取得方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 07732245 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2009504802 Country of ref document: JP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2007732245 Country of ref document: EP |