WO1999052298A1 - Improvements relating to audio-video delay - Google Patents

Improvements relating to audio-video delay Download PDF

Info

Publication number
WO1999052298A1
WO1999052298A1 PCT/GB1999/001041 GB9901041W WO9952298A1 WO 1999052298 A1 WO1999052298 A1 WO 1999052298A1 GB 9901041 W GB9901041 W GB 9901041W WO 9952298 A1 WO9952298 A1 WO 9952298A1
Authority
WO
WIPO (PCT)
Prior art keywords
audio
video
synchronisation
delay
time
Prior art date
Application number
PCT/GB1999/001041
Other languages
French (fr)
Inventor
Barry Flannaghan
Original Assignee
Snell & Wilcox Limited
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Snell & Wilcox Limited filed Critical Snell & Wilcox Limited
Priority to AU33394/99A priority Critical patent/AU3339499A/en
Priority to EP99914665A priority patent/EP1068737A1/en
Publication of WO1999052298A1 publication Critical patent/WO1999052298A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/4302Content synchronisation processes, e.g. decoder synchronisation
    • H04N21/4307Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen
    • H04N21/43072Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen of multiple content streams on the same device

Definitions

  • This invention relates to the processing of audio and video signals.
  • a common problem in the broadcast environment is the difference in delay experienced by the audio and video processes. With many new digital video processes in the signal chain and the consequential need for re- synchronisation, the delay of the video may often differ significantly from that of the audio. This causes the well known lip-sync error problem.
  • a video synchroniser operates by re-timing and, where necessary, either dropping or inserting fields.
  • Each video synchroniser may therefore add up to 40ms of video delay in a 50 field per second system such as PAL or 34ms in NTSC or other 60Hz systems. The precise delay will depend on often arbitrary system clocks. In new installations the audio is often "embedded" in the video channel and so when passing through the synchroniser it will experience the same delay as the video.
  • the present invention consists in one aspect in a digital video synchronisation process in which digital video and audio signals are delayed by the same varying amount to ensure synchronisation with a timing reference and, on dropping or repeating of a video field, the audio signal is time compressed or expanded to recover audio/video synchronisation over a time period governed by the maximum acceptable pitch change or other 2 degradation, characterised in that the trend in delay is monitored to enable prediction of the dropping or repeating of a video field and in that said time compression or expansion is initiated a selected time interval in advance thereof.
  • this time interval is selected to minimise the perceived loss of audio/video synchronisation. That is to say, regard is had to the asymmetry in the thresholds for perceived loss of synchronisation for advanced and retarded audio.
  • the video synchroniser shown in the drawing receives a digital video input signal with embedded audio and provides as an output synchronised with an externally generated timing reference, a digital video signal, again with embedded audio.
  • Digital video first passes into block 10; this extracts the audio data and the timing signals.
  • Video passes to the main memory 12 to be delayed until it is co-timed with the output timing signal extracted by block 14 from the reference input.
  • Block 16 measures the delay between the input and output timing signals and passes this figure to a microprocessor 18. The calculated delay is passed to the audio data memory 20.
  • the audio data and video data memories now give an identical delay.
  • a re-sampling digital filter 22 alters the audio data sampling rate to match the outgoing video in order that the audio data can be synchronously inserted by block 24 into the outgoing video data stream.
  • a synchroniser such as this when its input and output timing references are of a different frequency, as is usual, will occasionally drop or repeat a video frame. In effect, it gains or loses 40ms (at 50Hz). Whereas the resulting picture disturbance is often imperceptible, the same is not true for audio - it is not possible to cut or add 40ms of audio without major disturbance.
  • a typical implementation will compensate by re-sampling the audio to a higher or lower frequency.
  • the microprocessor 18 will initiate a process of reading extra audio samples from the audio memory 20.
  • a pitch change results, but if this is kept to less than 1%, it is unlikely to be noticed.
  • the audio will initially be adrift by a noticeable 40ms.
  • 8 seconds assuming 0.5% pitch change
  • the audio and video will once again be co-timed.
  • the microprocessor 18 will continually monitor the video delay and calculate the rate of change of delay over time.
  • the processor will initiate an increase in the audio delay to give a 0.5% pitch change.
  • the audio will be 30ms (at 50Hz) delayed with respect to the video. If the above discussed thresholds for perception of an audio delay are correct, this loss in synchronisation is imperceptible. Immediately after the frame loss the audio will be 10ms early. If the above- discussed thresholds for perception of an audio advance are correct, this loss in synchronisation is similarly imperceptible.
  • the audio and video will be co-timed.
  • the pitch change and the balance between worst case advance and delay may be controlled by the user.
  • Use of the invention has in this example and on the assumed perception thresholds, replaced a synchronisation error which is perceptible for up to 6 seconds, by a synchronisation error which is not perceptible at all.
  • rate conversion is only one example of a technique for time compression or expansion of the audio signal.
  • Numerous alternative techniques, such as silence compression, will be known to the skilled reader and fall within the scope of the claimed invention.

Abstract

In digital video synchronisation, the trend in delay is monitored to enable a prediction to be made of the dropping or repeating of a video field. The audio time compression or expansion is then initiated a selected time interval in advance of the field the dropping or repeating. The loss of audio synchronisation can in this way be kept below perceptible thresholds.

Description

IMPROVEMENTS RELATING TO AUDIO-VIDEO DELAY
This invention relates to the processing of audio and video signals. A common problem in the broadcast environment is the difference in delay experienced by the audio and video processes. With many new digital video processes in the signal chain and the consequential need for re- synchronisation, the delay of the video may often differ significantly from that of the audio. This causes the well known lip-sync error problem.
A video synchroniser operates by re-timing and, where necessary, either dropping or inserting fields. Each video synchroniser may therefore add up to 40ms of video delay in a 50 field per second system such as PAL or 34ms in NTSC or other 60Hz systems. The precise delay will depend on often arbitrary system clocks. In new installations the audio is often "embedded" in the video channel and so when passing through the synchroniser it will experience the same delay as the video.
Since a discontinuity of 40ms (or 34ms at 60Hz) is unacceptable in audio, a temporary loss of lip sync is inevitable. In order to recover lip sync, the audio signal is time compressed or expanded for a period of time. This period of time must be sufficiently long that the resultant pitch change or other degradation remains imperceptible and will amount to several seconds. Over these several seconds, the loss of lip sync can be seriously objectionable. It can be shown that a loss of synchronisation in which the audio arrives early, is particularly noticeable. It has been suggested that in the case of the audio being earlier than the video a delay much above 10ms is perceptible. In the opposite sense, with the audio being delayed, a delay of up to 30ms may be tolerated. It is an object of the present invention to provide an improved method of managing differential delay of digital audio and video signals.
Accordingly, the present invention consists in one aspect in a digital video synchronisation process in which digital video and audio signals are delayed by the same varying amount to ensure synchronisation with a timing reference and, on dropping or repeating of a video field, the audio signal is time compressed or expanded to recover audio/video synchronisation over a time period governed by the maximum acceptable pitch change or other 2 degradation, characterised in that the trend in delay is monitored to enable prediction of the dropping or repeating of a video field and in that said time compression or expansion is initiated a selected time interval in advance thereof. In one form of the invention, this time interval is selected to minimise the perceived loss of audio/video synchronisation. That is to say, regard is had to the asymmetry in the thresholds for perceived loss of synchronisation for advanced and retarded audio.
The invention will now be described by way of example with reference to the accompanying drawing which is a block diagram of a digital video synchroniser according to the invention.
The video synchroniser shown in the drawing receives a digital video input signal with embedded audio and provides as an output synchronised with an externally generated timing reference, a digital video signal, again with embedded audio.
Digital video first passes into block 10; this extracts the audio data and the timing signals. Video passes to the main memory 12 to be delayed until it is co-timed with the output timing signal extracted by block 14 from the reference input. Block 16 measures the delay between the input and output timing signals and passes this figure to a microprocessor 18. The calculated delay is passed to the audio data memory 20. The audio data and video data memories now give an identical delay. A re-sampling digital filter 22 alters the audio data sampling rate to match the outgoing video in order that the audio data can be synchronously inserted by block 24 into the outgoing video data stream.
A synchroniser such as this, when its input and output timing references are of a different frequency, as is usual, will occasionally drop or repeat a video frame. In effect, it gains or loses 40ms (at 50Hz). Whereas the resulting picture disturbance is often imperceptible, the same is not true for audio - it is not possible to cut or add 40ms of audio without major disturbance.
A typical implementation will compensate by re-sampling the audio to a higher or lower frequency. In the case where the video memory has lost a 3 frame, the microprocessor 18 will initiate a process of reading extra audio samples from the audio memory 20. A pitch change results, but if this is kept to less than 1%, it is unlikely to be noticed. Thus, after a synchroniser drops or repeats a frame the audio will initially be adrift by a noticeable 40ms. After 8 seconds (assuming 0.5% pitch change) the audio and video will once again be co-timed. As a result, there will be a perceptible error for up to 6 seconds. In the improved system according to this invention, the microprocessor 18 will continually monitor the video delay and calculate the rate of change of delay over time. With the highly stable timing references that are generally in use, it will usually be possible to predict accurately the time that the video memory will drop or gain a frame. Where it is predicted that, at the current rate of change, a video frame is due to be lost in 6 seconds time, the processor will initiate an increase in the audio delay to give a 0.5% pitch change. Just before the frame loss, the audio will be 30ms (at 50Hz) delayed with respect to the video. If the above discussed thresholds for perception of an audio delay are correct, this loss in synchronisation is imperceptible. Immediately after the frame loss the audio will be 10ms early. If the above- discussed thresholds for perception of an audio advance are correct, this loss in synchronisation is similarly imperceptible. After a further 2 seconds, the audio and video will be co-timed. Of course, the pitch change and the balance between worst case advance and delay may be controlled by the user. Use of the invention has in this example and on the assumed perception thresholds, replaced a synchronisation error which is perceptible for up to 6 seconds, by a synchronisation error which is not perceptible at all. When it is predicted that, at the current rate of change, a video frame is due to be repeated in 2 seconds time the processor will initiate an decrease in the audio delay to give a 0.5% pitch change. Once again the error will be imperceptible.
It will be understood that the invention has been described by way of examples only. Thus, rate conversion is only one example of a technique for time compression or expansion of the audio signal. Numerous alternative techniques, such as silence compression, will be known to the skilled reader and fall within the scope of the claimed invention.

Claims

4 CLAIMS
1. A digital video synchronisation process in which digital video and audio signals are delayed by the same varying amount to ensure synchronisation with a timing reference and, on dropping or repeating of a video field, the audio signal is time compressed or expanded to recover audio/video synchronisation over a time period governed by the maximum acceptable pitch change or other degradation, characterised in that the trend in delay is monitored to enable prediction of the dropping or repeating of a video field and in that said time compression or expansion is initiated a selected time interval in advance thereof.
2. A process according to Claim 1 , wherein said time interval is selected to minimise the absolute loss of audio/video synchronisation.
3. A process according to Claim 1 , wherein said time interval is selected to minimise the perceived loss of audio/video synchronisation.
4. A process according to Claim 3, wherein said time interval is selected such that the period for which the audio is delayed with respect to the video is greater than the period the period for which the audio is advanced with respect to the video.
PCT/GB1999/001041 1998-04-03 1999-04-06 Improvements relating to audio-video delay WO1999052298A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
AU33394/99A AU3339499A (en) 1998-04-03 1999-04-06 Improvements relating to audio-video delay
EP99914665A EP1068737A1 (en) 1998-04-03 1999-04-06 Improvements relating to audio-video delay

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
GBGB9807295.2A GB9807295D0 (en) 1998-04-03 1998-04-03 Improvements relating to audio-video delay
GB9807295.2 1998-04-03

Publications (1)

Publication Number Publication Date
WO1999052298A1 true WO1999052298A1 (en) 1999-10-14

Family

ID=10829887

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/GB1999/001041 WO1999052298A1 (en) 1998-04-03 1999-04-06 Improvements relating to audio-video delay

Country Status (4)

Country Link
EP (1) EP1068737A1 (en)
AU (1) AU3339499A (en)
GB (1) GB9807295D0 (en)
WO (1) WO1999052298A1 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2000041400A2 (en) * 1999-01-06 2000-07-13 Koninklijke Philips Electronics N.V. System for the presentation of delayed multimedia signals packets
WO2002001878A1 (en) * 2000-06-28 2002-01-03 Grundig Ag Method and device for the time-synchronised relaying of signals
EP1289305A2 (en) * 2001-08-07 2003-03-05 Hitachi Ltd. Digital signal processing device, DV decoder, recording device using DV decoder, and signal processing method
EP1414247A3 (en) * 2002-10-22 2004-05-19 Broadcom Corporation Data rate management system and method for A/V decoder
CN109906613A (en) * 2016-09-14 2019-06-18 Dts有限责任公司 The multi-modal synchronization of audio and video renders

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1994029979A1 (en) * 1993-06-09 1994-12-22 Intelligence At Large, Inc. Method and apparatus for multiple media digital communication system
EP0648056A2 (en) * 1993-09-30 1995-04-12 Thomson Consumer Electronics, Inc. Audio/video synchronization in a digital transmission system
US5603016A (en) * 1994-08-03 1997-02-11 Intel Corporation Method for synchronizing playback of an audio track to a video track

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1994029979A1 (en) * 1993-06-09 1994-12-22 Intelligence At Large, Inc. Method and apparatus for multiple media digital communication system
EP0648056A2 (en) * 1993-09-30 1995-04-12 Thomson Consumer Electronics, Inc. Audio/video synchronization in a digital transmission system
US5603016A (en) * 1994-08-03 1997-02-11 Intel Corporation Method for synchronizing playback of an audio track to a video track

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
"METHOD FOR SYNCHRONIZING AN AUDIO DIGITAL DATA STREAM WITH ANOTHER INFORMATION STREAM", IBM TECHNICAL DISCLOSURE BULLETIN, vol. 38, no. 4, 1 April 1995 (1995-04-01), pages 141, XP000516100, ISSN: 0018-8689 *
KAWAHARA K ET AL: "A SINGLE CHIP MPEG1 DECODER", 1995 SYMPOSIUM ON VLSI CIRCUITS, KYOTO, JUNE 8 - 10, 1995, 8 June 1995 (1995-06-08), INSTITUTE OF ELECTRICAL AND ELECTRONICS ENGINEERS, pages 65/66, XP000557803, ISBN: 0-7803-2600-8 *
KIRBY D G ET AL: "A NEW TECHNIQUE TO MAINTAIN SOUND AND PICTURE SYNCHRONISATION", EBU REVIEW- TECHNICAL, no. 264, 1 January 1995 (1995-01-01), pages 13 - 21, XP000670474, ISSN: 0251-0936 *

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2000041400A2 (en) * 1999-01-06 2000-07-13 Koninklijke Philips Electronics N.V. System for the presentation of delayed multimedia signals packets
WO2000041400A3 (en) * 1999-01-06 2001-02-01 Koninkl Philips Electronics Nv System for the presentation of delayed multimedia signals packets
WO2002001878A1 (en) * 2000-06-28 2002-01-03 Grundig Ag Method and device for the time-synchronised relaying of signals
US7418012B2 (en) 2000-06-28 2008-08-26 Grundig Multimedia B.V. Method and device for time-synchronized relaying of signals
EP1289305A2 (en) * 2001-08-07 2003-03-05 Hitachi Ltd. Digital signal processing device, DV decoder, recording device using DV decoder, and signal processing method
EP1289305A3 (en) * 2001-08-07 2005-10-05 Hitachi Ltd. Digital signal processing device, DV decoder, recording device using DV decoder, and signal processing method
EP1414247A3 (en) * 2002-10-22 2004-05-19 Broadcom Corporation Data rate management system and method for A/V decoder
US7466362B2 (en) 2002-10-22 2008-12-16 Broadcom Corporation Data rate management system and method for A/V decoder
CN109906613A (en) * 2016-09-14 2019-06-18 Dts有限责任公司 The multi-modal synchronization of audio and video renders
CN109906613B (en) * 2016-09-14 2022-09-13 Dts有限责任公司 Multi-mode synchronized rendering of audio and video

Also Published As

Publication number Publication date
EP1068737A1 (en) 2001-01-17
AU3339499A (en) 1999-10-25
GB9807295D0 (en) 1998-06-03

Similar Documents

Publication Publication Date Title
KR100538135B1 (en) Method and apparatus for information stream frame synchronization
JP4492972B2 (en) Video encoder and method for synchronizing timing from encoder to decoder
US6330286B1 (en) Flow control, latency control, and bitrate conversions in a timing correction and frame synchronization apparatus
US5602592A (en) Moving picture compressed signal changeover apparatus
KR100337212B1 (en) Audio signal and video signal synchronization device in digital transmission system
US6081299A (en) Methods and systems for encoding real time multimedia data
US20030053492A1 (en) Multiplexer, receiver, and multiplex transmission method
EP2306736B1 (en) Method and apparatus for processing video and audio data received in decoding system
JP2002503402A (en) Method and apparatus for connecting compressed information streams
AU751335B2 (en) Video and audio signal processing
KR100984818B1 (en) A system and method for determining lip synchronization between audio and video in a digitized envoronment using buffer calculation
US6380981B1 (en) Frame synchronizer
AU766738B2 (en) Data compression unit control for alignment of output signal
EP1068737A1 (en) Improvements relating to audio-video delay
KR20010015380A (en) Synchronizing apparatus for a compressed audio/video signal receiver
JP2002176643A (en) Synchronization recovery system for video/audio in video/ audio decoding processing
JPH11112947A (en) Device and method for data multiplexing, device and method for data processing, and transmission medium
CA2239317C (en) Method and apparatus for synchronizing signals
KR960028170A (en) Digital TV's Audio / Video Synchronizer
JP2008067179A (en) Receiver and coded data reproduction method
JPH09284759A (en) Timing control circuit
Reid An MPEG-2 digital decoder design: A practical approach with emphasis on elementary stream data flows
JPH06165078A (en) Multiplex control circuit

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AU CA JP US

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE

121 Ep: the epo has been informed by wipo that ep was designated in this application
DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
WWE Wipo information: entry into national phase

Ref document number: 1999914665

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 09647582

Country of ref document: US

WWP Wipo information: published in national office

Ref document number: 1999914665

Country of ref document: EP

WWW Wipo information: withdrawn in national office

Ref document number: 1999914665

Country of ref document: EP