WO2003041382A2 - Scalable video transmissions - Google Patents

Scalable video transmissions Download PDF

Info

Publication number
WO2003041382A2
WO2003041382A2 PCT/EP2002/012323 EP0212323W WO03041382A2 WO 2003041382 A2 WO2003041382 A2 WO 2003041382A2 EP 0212323 W EP0212323 W EP 0212323W WO 03041382 A2 WO03041382 A2 WO 03041382A2
Authority
WO
WIPO (PCT)
Prior art keywords
video
data
errors
error
scalable video
Prior art date
Application number
PCT/EP2002/012323
Other languages
French (fr)
Other versions
WO2003041382A3 (en
Inventor
Jonathan Hare
Catherine Dolbear
Tamer Shanableh
Original Assignee
Motorola Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Motorola Inc filed Critical Motorola Inc
Priority to AU2002340500A priority Critical patent/AU2002340500A1/en
Publication of WO2003041382A2 publication Critical patent/WO2003041382A2/en
Publication of WO2003041382A3 publication Critical patent/WO2003041382A3/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/65Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using error resilience
    • H04N19/68Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using error resilience involving the insertion of resynchronisation markers into the bitstream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/36Scalability techniques involving formatting the layers as a function of picture distortion after decoding, e.g. signal-to-noise [SNR] scalability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/89Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving methods or arrangements for detection of transmission errors at the decoder

Definitions

  • This invention relates to video transmission systems and video encoding/decoding techniques.
  • the invention is applicable to a video compression system where the video has been compressed using a scalable compression technique.
  • a scalable video bit-stream refers to the ability to transmit and receive video signals of more than one resolution and/or quality simultaneously.
  • a scalable video bit-stream is one that may be decoded at different rates, according to the bandwidth available at the decoder. This enables the user with access to a higher bandwidth channel to decode high quality video, whilst a lower bandwidth user is still able to view the same video, albeit at a lower quality.
  • the main application for scalable video transmissions is for systems where multiple decoders with access to differing bandwidths are receiving images from a single encoder. Scalable video transmissions can also be used for bit-rate adaptability where the available bit rate is fluctuating in time. Other applications include video multicasting to a number of end-systems with different network and/or device characteristics.
  • scalable video can also be used to provide subscribers of a particular service with different video qualities depending on their tariffs and preferences. Therefore, in these applications it is imperative to protect the enhancement layer from transmission errors. Otherwise, the subscribers may lose confidence in their network operator's ability to provide an acceptable service.
  • enhancements to the video signal may be added to a base layer either by:
  • Such enhancements may be applied to the whole picture or to an arbitrarily shaped object within the picture, which is termed object-based scalability.
  • the H.263+ ITU H.263 ITU-T Recommendation, H.263, "Video Coding for Low Bit Rate Communication"] standard dictates that pictures included in the temporal scalability mode should be bi-directionally predicted (B) pictures. These are as shown in the video stream of FIG. 1.
  • FIG. 1 shows a schematic illustration of a scalable video arrangement 100 illustrating B picture prediction dependencies, as known in the field of video coding techniques.
  • An initial intra-coded frame (Ii) 110 is followed by a bi-directionally predicted frame (B 2 ) 120.
  • FIG. 2 is a schematic illustration of a layered video arrangement, known in the field of video coding techniques.
  • a layered video bit stream includes a base layer 205 and one or more enhancement layers 235.
  • the base layer (layer-1) includes one or more intra-coded pictures (I pictures) 210 sampled, coded and/or compressed from the original video signal pictures. Furthermore, the base layer will include a plurality of subsequent predicted inter-coded pictures (P pictures) 220, 230 predicted from the intra-coded picture (s) 210. In the enhancement layers (layer-2 or layer-3 or higher layer (s) ) 235, three types of picture may be used: (i) Bi-directionally predicted (B) pictures (not shown) ; (ii) Enhanced intra-coded (El) pictures 240 based on the intra-coded picture (s) 210 of the base layer 205; and
  • the vertical arrows from the lower, base layer illustrate that the picture in the enhancement layer is predicted from a reconstructed approximation of that picture in the reference (lower) layer.
  • the enhancement layer picture is referred to as an El picture. It is possible, however, to create a modified bi- directionally predicted picture using both a prior enhancement layer picture and a temporally simultaneous lower layer reference picture. This type of picture is referred to as an EP picture or "Enhancement" P-picture.
  • an El picture in an enhancement layer may have a P picture as its lower layer reference picture
  • an EP picture may have an I picture as its lower-layer enhancement picture.
  • bit-streams For both El and EP pictures, the prediction from the reference layer uses no motion vectors. However, as with normal P pictures, EP pictures use motion vectors when predicting from their temporally, prior-reference picture in the same layer.
  • Current standards incorporating the aforementioned scalability techniques include MPEG-4 and H.263. These standards create highly compressed bit-streams, which represent the coded video. However, due to this high compression, the bit-streams are very prone to corruption by network errors as they are transmitted. For example, in the case of streaming video over an error prone network, even with existing network level error protection tools employed, it is inevitable that some bit-level corruption will occur in the bit-stream and be passed on to the decoder .
  • the coding standards have been designed with various tools incorporated that allow the decoder to cope with the errors. These tools enable the decoder to localise and conceal the errors within the bit-stream.
  • the MPEG-4 standard does define three tools for error resilience of video bit-streams. These are re- synchronisation markers, data partitioning (DP) and reversible variable length codes (RVLCs) . These tools are defined for use in the base layer.
  • DP data partitioning
  • RVLCs reversible variable length codes
  • Data Partitioning separates motion parameters from the texture information, i.e. discrete cosine transform (DCT) coefficients, within a video transmission.
  • Data partitioning is achieved by separating the motion and macroblock header information away from the texture information. This approach requires that a second re- synchronisation marker be inserted between motion and texture information.
  • the use of data partitioning in the same manner as the use of RVLCs, is signalled to the decoder in the Video Object Layer's (VOL's) header. If the texture information is lost, this approach utilises some motion information to conceal these errors. That is, due to the errors the texture information is discarded, while the motion is used to motion compensate the previously decoded Video Object Plane (VOP) .
  • VOP Video Object Plane
  • This separation is performed on a Video Packet basis where an additional re-synchronisation marker called a Motion Marker separates the motion parameters from the DCT coefficients. If an error occurs in the DCT coefficients portion, then the correctly decoded motion vectors are employed to conceal the remaining portion of the Video Packet .
  • the MPEG-4 visual standard associates another useful tool with data partitioning called Reversible Variable Length Codes (RVLCs) .
  • RVLC Reversible Variable Length Codes
  • DCT coefficients are coded with RVLCs rather than the traditional Variable Length Codes (VLCs) .
  • VLCs Variable Length Codes
  • Enhancement layer information contains visual information that enhances the decoding quality of the more important base layer. Hence, as enhancement layer information was not deemed essential, no further resiliency was anticipated.
  • the focus for higher levels of protection in a video bit sequence in current communications systems is the base layer.
  • the decoder wishing to keep the enhancement layer, has to conceal much more data, potentially in error, than it would have to if the error resilience tools could be used.
  • the present invention provides a method for improving a quality of scalable video enhancement layers transmitted over an error-prone network, as claimed in claim 1, a method for improving a quality of scalable video enhancement layers transmitted over an error-prone network, as claimed in claim 2, a method for improving a quality of scalable video enhancement layers transmitted over an error-prone network, as claimed in claim 6, a video communication system, as claimed in claim 10, a fifth aspect of the present invention, there is provided a video communication system, as claimed in claim 13, a video communication unit, as claimed in claim 16, a video encoder or a video decoder, as claimed in claim 17, and a mobile radio device, as claimed in claim 19.
  • This invention provides a method by which re-synchronisation markers are added to enhancement layer bit-streams in scalable MPEG-4 or similar video compression systems, or H.263+ or similar video compression systems. This is in order to promote better bit-level error resilience capabilities, and reduce the amount of concealment of error-free data that takes place.
  • this invention proposes the use of the data partitioning, and preferably a Reversible Variable Length Coding, error resilience tool in the video enhancement layers of the MPEG-4 Simple Scalable Profile.
  • One aim is to reduce or to substantially eliminate the macroblock displacements, caused by error concealment utilising estimated motion vectors.
  • a further aim is to minimise the number of discarded bits by performing reverse decoding.
  • FIG. 1 is a schematic illustration of a video coding arrangement showing picture prediction dependencies, as known in the field of video coding techniques.
  • FIG. 2 is a schematic illustration of a known layered video coding arrangement . Exemplary embodiments of the present invention will now be described, with reference to the accompanying drawi gs, in which:
  • FIG. 3 is a schematic representation of a scalable video communication system adapted to use re-synchronisation markers in an enhancement layer of a video sequence in accordance with the preferred embodiment of the present invention.
  • FIG. 4 illustrates a first enhancement layer bit-stream frame configuration.
  • FIG. 5 illustrates a second enhancement layer bit-stream frame configuration.
  • FIG. 6 illustrates a first enhancement layer bit-stream frame configuration adapted to incorporate the preferred embodiment of the present invention.
  • FIG. 7 illustrates a second enhancement layer bit-stream frame configuration adapted to incorporate the preferred embodiment of the present invention.
  • FIG. 8 illustrates a third enhancement layer bit-stream frame configuration adapted to incorporate the preferred embodiment of the present invention.
  • FIG. 9 illustrates an example of syntax code for use in the MPEG-4 standard, for implementing a preferred embodiment of the present invention.
  • FIG. 10 illustrates a graph highlighting the improvement of decoded SNR scalable video quality, in accordance with a preferred embodiment of the present invention.
  • FIG. 11 illustrates a graph highlighting the improvement of decoded temporal scalable video quality, in accordance with a preferred embodiment of the present invention.
  • FIG. 12 illustrates a data format of an enhancement layer bit stream that can be adapted to incorporate the use of data partitioning and RVLCs in accordance with the preferred further embodiment of the present invention.
  • FIG. 13 illustrates at least one benefit from introducing data partitioning, and preferably reverse variable length coding, in an enhancement layer bit stream, for example the data format of FIG. 12.
  • FIG. 14 illustrates the objective degradations caused by estimating values of lost motion vectors, or using original motion vectors, associated with the error scenarios of FIG. 13.
  • FIG. 15 illustrates proposed syntax amendments to section 6.2.5 "Video Object Plane and Video Plane with Short Header" of the MPEG-4 standard.
  • inventive concepts described herein can be applied to other types of scalable video sequences such as temporal scalability, spatial scalability and Fine Granular scalability (FGS) .
  • FGS Fine Granular scalability
  • inventive concepts herein described find particular application in the current MPEG technology arena, and in future versions of MPEG technology.
  • the inventors have also recognised the applicability of the present invention to H.263+ technology.
  • the preferred embodiment of the present invention illustrates a technique for introducing re-synchronisation markers in enhancement layer bit-streams to improve their error resilience.
  • the video bit- stream syntax is modified to enable the use of the standard base layer error resilience tools within the enhancement layer bit-stream.
  • the decoder can then make use of the re-synchronisation markers in the enhancement layer bit-stream in the same manner as for a base layer bit-stream.
  • FIG. 3 a schematic representation of a video communication system 300, including video encoder 315 and video decoder 325, adapted to incorporate the preferred embodiment of the present invention, is shown.
  • a video picture Fo is compressed 310 in a video encoder 315 to produce the base layer bit stream signal to be transmitted at a rate r__ kilobits per second (kbps) .
  • This signal is decompressed 320 at a video decoder 325 to produce the reconstructed base layer picture Fo' .
  • the compressed base layer bit stream is also decompressed at 330 in the video encoder 315 and compared with the original picture Fo at 340 to produce a difference signal
  • This difference signal is compressed at 360 and transmitted as the enhancement layer bit stream at a rate T 2 kbps .
  • This enhancement layer bit stream is decompressed at 370 in the video decoder 325 to produce the enhancement layer picture Fo'' which is added to the reconstructed base layer picture Fo ' at 380 to produce the final reconstructed picture Fo ' " •
  • the compression function 360 in the video encoder 315 has been adapted to incorporate one or more re- synchronisation markers in the enhancement layer bit- stream.
  • the decompression function 370 in the video decoder 325 has been adapted to determine the location of these one or more re-synchronisation markers in the enhancement layer bit-stream.
  • the video decoder 325 has also been adapted to conceal errors in the received scalable video sequence up to the next available error-free re-synchronisation marker. The use of such re- synchronisation markers is further described with regard to FIG's 6 to 8.
  • an MPEG-4 enhancement layer bit-stream does not have the facility to include any re-synchronisation markers within it. Hence, if the video decoder encounters an error, it has no choice but to conceal all of the data from the point at which the error occurred to the start of the next readable start code (in the case of MPEG-4 this would be a VOP_Start_Code) at the beginning of a subsequent frame of video .
  • the first enhancement layer bit-stream frame 400 includes a start code 410, followed by image data of a frame ⁇ N' 420.
  • the image data of frame 'N' 420 is shown to contain errors 430 within the frame 420.
  • FIG. 4 highlights a problem associated with such an enhancement layer bit-stream frame in that a substantial amount of error-free data 440 subsequent to the errors 430 within the frame 420 has to be concealed. Clearly, it is undesirable, as well as inefficient, to conceal such error-free data.
  • a second enhancement layer bit- stream frame 500 is illustrated that highlights a further problem with errors encountered in enhancement layer bit- streams.
  • the second enhancement layer bit-stream frame 500 illustrates errors being encountered within image data of frame 'N' 520 as well as within the subsequent start code 510 of the next N+l' frame 525.
  • errors 530 occur over the start code in this manner, a whole frame 540 may end up being concealed, as shown.
  • the effect to the viewer of the aforementioned concealed error-free frames depends on the type of scalability being used.
  • the viewer may witness frames that are enhanced at the top of the image, but not enhanced at the bottom.
  • the viewer may experience an annoying flicker, as the enhancement layer disappears entirely over a frame.
  • the concealment can lead to some strange effects, with the viewer seeing part of the current frame and parts of previously decoded frames.
  • re-synchronisation markers 660, 760, 860 are introduced to the enhancement layer bit-stream transmitted by the video encoder 315, to improve the quality of the decoded video image.
  • the improvement is achieved by reducing the amount of error-free data that has to be concealed by the decoder.
  • the concealment instead of the video decoder 325 having to conceal all of the data in a frame, from the point the error occurred to the end of the frame, the concealment only has to continue up to the first occurrence of either: (i) The first valid re-synchronisation marker, bearing in mind that errors could occur in the markers themselves; or (ii) The next valid start code.
  • FIG. 6 a first enhancement layer bit- stream frame configuration 600 is shown, adapted by the introduction of re-synchronisation markers 660 in accordance with the preferred embodiment of the present invention.
  • the configuration is similar to that shown in FIG. 4, where the errors 630 occur within the image data frame 620 and are determined to be located between two start codes 610, 650.
  • the video decoder only conceals error-free data up to the re- synchronisation marker 660 that immediately follows the errors 630.
  • the amount of concealed error-free data 640 is reduced significantly, when compared to the prior art arrangement described with respect to FIG. 4.
  • FIG. 7 a second enhancement layer bit- stream frame configuration 700 is shown, adapted by the introduction of re-synchronisation markers 760 in accordance with the preferred embodiment of the present invention.
  • the configuration is similar to that shown in FIG. 5, where the errors 730 occur across a start code 710, as well as within the image data frame N 720.
  • the video decoder only conceals error-free data up to the first re-synchronisation marker 760, located in the subsequent image data frame N+l' 725.
  • VOP Video Object Plane's
  • a third enhancement layer bit- stream frame configuration 800 is shown, adapted by the introduction of re-synchronisation markers 860 in accordance with the preferred embodiment of the present invention.
  • the errors may occur around, and affecting, the re-synchronisation markers 860.
  • This may be, for example, around the first re-synchronisation marker 860 within the image data frame N 820.
  • the video decoder only conceals error-free data up to the second re-synchronisation marker 860, located within the same image data frame N' 820.
  • the amount of concealed error-free data is still reduced significantly, and only extends into the period between re- synchronisation markers 860.
  • the existing MPEG-4 syntax code 900 can be readily adapted to incorporate the preferred embodiment of the present invention using re-synchronisation markers in one or more enhancement layers .
  • the required change 910 to the MPEG-4 syntax code, to accommodate the preferred embodiment of the present invention, is illustrated in FIG. 9.
  • a processor in a video communication unit may be re-programmed, and thereby adapted, to incorporate the inventive concepts hereinbefore described.
  • the adaptation may be implemented in the respective video communication unit in any suitable manner.
  • new apparatus may be added to a conventional video communication unit, or alternatively existing parts of a conventional video communication unit may be adapted, for example by reprogramming one or more processors therein.
  • the required adaptation may be implemented in the form of processor-implementable instructions stored on a storage medium, such as a floppy disk, hard disk, PROM, RAM or any combination of these or other storage multimedia.
  • any video communication unit operating in a video communication system may be, for example, user equipment such as mobile or portable radios or telephones, or wireless or wired serving communication units such as base transceiver or intermediate (repeater) equipment .
  • user equipment such as mobile or portable radios or telephones
  • wireless or wired serving communication units such as base transceiver or intermediate (repeater) equipment .
  • re-synchronisation markers can be placed in the enhancement layer bit-stream, and used to reduce the amount of error-free data concealment.
  • the MPEG-4 standard states the maximum size of Video Packets, or the number of bits between two consecutive resynchronisation markers, in a given base layer VOP. This aids the decoder, as it roughly knows when to expect the next marker.
  • the period between re-synchronisation markers in the enhancement layer may be user and/or implementation defined.
  • a smaller distance between re-synchronisation markers leads to a reduction in the amount of data that has to be concealed, should an error occur.
  • the distance between the re-synchronisation markers also affects the overall number of bits in the video bit-streams. So, in a fixed bit-rate environment the quality of the enhancement layer has to be reduced, in order for the video to meet the bit budget of the transmission system.
  • the inventors of the present invention have performed extensive experiments that have shown that by placing re- synchronisation markers at approximately every eight hundred bits, the overall increase in bit-stream size is just less than four percent.
  • the current MPEG standard both for scalable MPEG-4 and FGS MPEG-4, explicitly states that error resilience tools are not to be used in the enhancement layer (s) .
  • the standard as well as public domain literature, do not allow the use of re-synchronisation markers in an enhancement layer of scalable MPEG-4.
  • the MPEG 4 standard as well as public domain literature, teaches away from the present invention.
  • the current teaching also states that no foreseeable benefit can be gained in using error resilience tools in scalable (including FGS) MPEG-4 video communication systems. Furthermore, it is generally understood in the field of video communication that MPEG-4 enhancement layers would not be used in a mobile communication environment .
  • a graph 1000 highlights the improvement of decoded SNR scalable video quality, in implementing the preferred embodiment of the present invention.
  • the peak signal to noise ratio (PSNR) is substantially improved when re- synchronisation markers are incorporated into the enhancement layer bit-stream, as compared to the PSNR levels without re-synchronisation markers.
  • a graph 1100 highlights the improvement of decoded temporal scalable video quality, when implementing the preferred embodiment of the present invention.
  • the peak signal to noise ratio (PSNR) is substantially improved when resynchronisation markers are incorporated into the enhancement layer bit-stream, as compared to the PSNR levels without re-synchronisation markers.
  • FIG. 10 and FIG. 11 were made by adding re-synchronisation markers to the enhancement layers of SNR and temporally scalable video bit-streams transmitted over a GPRS network.
  • a noteworthy feature is the relative smoothness of the lines. For example, the lines illustrating the video without re-synchronisation markers tend to oscillate much more because of the larger amounts of data concealment taking place. In contrast, the lines illustrating video with re-synchronisation markers are much flatter due to the reduced amount of data concealment that has to take place.
  • JVT Joint Video Team
  • inventive concepts may be applied to any video communication unit and/or video communication system.
  • inventive concepts find particular use in wireless (radio) devices, such as mobile telephones/mobile radio units and associated wireless communication systems.
  • wireless communication units may include a portable or mobile PMR radio, a personal digital assistant, a laptop computer or a wirelessly networked PC.
  • scalable video system technology may be implemented in the 3 rd generation (3G) of digital cellular telephones, commonly referred to as the Universal Mobile Telecommunications Standard (UMTS) .
  • Scalable video system technology may also find applicability in the packet data variants of both the current 2 nd generation of cellular telephones, commonly referred to as the general packet-data radio system (GPRS) , and the TErrestrial Trunked RAdio (TETRA) standard for digital private and public mobile radio systems.
  • GPRS general packet-data radio system
  • TETRA TErrestrial Trunked RAdio
  • scalable video system technology may also be utilised in the Internet. The aforementioned inventive concepts will therefore find applicability in, and thereby benefit, all these emerging technologies.
  • the video transmission arrangement described above provides at least the advantage of achieving a reduction in the amount of error-free data that has to be discarded. This, in turn, reduces the amount of visual degradation in the video transmission. This is especially important when transmitting video over wireless channels and the Internet, where the errors can be severe.
  • Applications of digital video where the preferred method improves decoder performance are one-way and two-way video communications, surveillance applications, and video streaming applications.
  • the method includes the steps of inserting one or more re-synchronisation markers into one or more enhancement layers of a scalable video sequence in a video encoder.
  • a video decoder detects one or more errors in the one or more enhancement layers of the scalable video sequence; and, in response to detecting one or more errors, conceals data in the received scalable video sequence from a video bit position at which the one or more errors is detected, to a video bit position of a subsequent error- free re-synchronisation marker.
  • H.263+ video is generally used for lower bandwidth transmission systems than MPEG-4 video.
  • the H.263+ standard actually differs from the MPEG-4 standard in several aspects. One of these aspects is that the H.263+ standard does not expressly state that error resilience tools should not be used in the enhancement layers .
  • a video communication system includes a video encoder comprising a processor for encoding a video sequence into an MPEG-4 (or similar) scalable video sequence having a plurality of enhancement layers; and insertion means to insert one or more re- synchronisation markers into one or more enhancement layers.
  • a transmitter transmits the scalable video sequence containing said one or more re-synchronisation markers .
  • a video decoder includes a receiver for receiving the scalable video sequence and a detector that detects one or more errors in the one or more enhancement layers of the scalable video sequence.
  • a processor is operably coupled to the detector for concealing data in the received scalable video sequence, when one or more errors are detected. The video data is concealed from a video bit position at which said one or more errors is detected to a video bit position of a subsequent error-free resynchronisation marker.
  • a video communication unit an adapted video encoder, an adapted video decoder, and a mobile radio device incorporating any one of these units, have also been described.
  • inventive concepts contained herein are equally applicable to any suitable video or image transmission system. Whilst specific, and preferred, implementations of the present invention are described above, it is clear that one skilled in the art could readily apply variations and modifications of such inventive concepts.
  • the inventors have also considered a further enhancement of the above invention.
  • the enhanced solution may further improve the subjective quality of video for a viewer.
  • the inventors have considered the problem that error concealment techniques in the temporally scalable enhancement layer conceal data from the point that the error occurred to the next re-synchronisation marker, or the start code at the beginning of the next frame. If a large amount of motion occurred between the current frame and the previous frame, then the area that is concealed can look discontinuous from the rest of the frame.
  • the inventors have developed a method to selectively conceal an entire frame of video, in order to reduce the discontinuities in the image, thus improving subjective quality for the viewer.
  • This method improves the image quality by adaptively deciding on how much data should be concealed, to give the best image quality. If the motion between the previous frame and the current frame with errors is large, then the method involves a decision to conceal the whole frame to avoid any discontinuity in the decoded image. If the motion is small, then the decision is to conceal as little data as possible.
  • a detailed arrangement starts from the realisation that discontinuities in the decoded image can be very visible to a viewer, using concealment techniques that conceal errors from the point the error occurred to the next re- synchronisation point, i.e. re-synchronisation marker or start code.
  • the visibility of the discontinuity between the unconcealed and concealed areas depends on the amount of motion that is taking place in the scene. For example, visualise a scene with little background motion, but people running in the foreground. The people create lots of foreground motion, which leads to a large discontinuity, which is visually disturbing.
  • the inventors propose a method in which a normalised Sum of Absolute Difference (SAD) is calculated between the current frame and previous frame over the non-concealed area. If the normalised SAD exceeds a pre-defined threshold (indicating a large amount of motion) then the whole frame is concealed.
  • SAD Sum of Absolute Difference
  • the enhancement using the frame suppression method, can in fact even be used in a video transmission system that does not have re-synchronisation markers in the enhancement layers . So it can be seen both as an enhancement to the claimed invention, and as a potential alternative solution if employed alone.
  • the inventors of the present invention have further recognised a potential problem in the concealment of corrupted visual data by means of estimating the values of lost motion vectors and motion compensating previous VOPs.
  • lost motion vectors are estimated from either the surrounding macroblocks, or the co-located macroblocks from previously decoded VOPs, it follows that these motion vectors are not accurate, resulting in non-accurate motion compensation.
  • estimated motion vectors may often cause the displacement of concealed macroblocks.
  • the inventors of the present invention propose an enhancement to the preferred embodiment of the present invention described above.
  • data partitioning currently limited to use in a base layer of a video data stream, is also employed in video enhancement layers.
  • the data partitioning error resilience tool is supplemented with reversible variable length coding.
  • the video bit- stream syntax is modified to enable the use of these standard base layer error resilience tools within the enhancement layer bit-stream.
  • the decoder and decoding operation are further adapted to make use of the data partitioning, into two or more portions of bit streamed data, in the enhancement layer bit-stream in the same manner as for a base layer bit-stream.
  • the focus for the enhanced further embodiment is that by employing data partitioning in the enhancement layer, either in isolation or in addition to the embodiments described with reference to FIG's 4 to 6, a (further) reduction in the amount of data discarded when an error is detected can be achieved.
  • the inventors of the present invention have recognised a potential deficiency of using a specific (resynchronisation) Motion Marker to separate the motion parameters from DCT coefficients within a video packet header in the data partitioning. If an error occurs in the DCT coefficients part, then the original motion vectors are employed, as opposed to the estimated motion vectors, to conceal the remaining part of the video packet .
  • VLC reversible variable length codes
  • the compression function 360 in the video encoder 315 of FIG. 3 has been adapted to incorporate data partitioning in the enhancement layer bit-stream.
  • the decompression function 370 in the video decoder 325 of FIG. 3 has been adapted to determine that data partitioning in the enhancement layer bit-stream has been performed.
  • the video decoder 325 has also been adapted to conceal errors within one or more data portions of the partitioned data that are detected as having errors.
  • the video decoder 325 has further been adapted to perform RVLC on the enhancement layer, when an error in one or more data portions of a data-partitioned sequence is detected, in a similar manner to that currently performed on a base layer of a video sequence .
  • the use of such data partitioning and RVLC is further described with regard to FIG's 12 to 14.
  • FIG. 12 illustrates a data format 1200 of an enhancement layer bit stream that can be adapted to incorporate the use of data partitioning and RVLCs in accordance with the preferred further embodiment of the present invention.
  • the data format 1200 includes a video packet header 1210.
  • the data format 1200 of the enhancement layer has been adapted to incorporate data partitioning, whereby the motion and header information 1220 has been separated from the texture information 1240.
  • a re-synchronisation motion marker 1230 indicates the location where the motion parameters are separated from the DCT coefficients .
  • a video packet header 1250 of a subsequent video packet, VOP start code or GOV start code may follow the data-partitioned data format.
  • a texture header 1235 has been introduced to indicate texture related information to the decoder. Errors 1244 within the texture information can then be localised by decoding in the forward direction 1242 and in the reverse direction 1246, by use of RVLCs.
  • FIG. 13 illustrates the benefits of introducing data partitioning and RVLC in an enhancement layer bit stream, for example the data format of FIG. 12.
  • Three enhancement layer data formats are shown, all of which include initial re-synchronisation markers 1320, as described above, and video packet headers 1210.
  • the first data format 1310 indicates a data structure that does not employ data partitioning (and thereby does not employ RVLC) .
  • the body information contains both motion parameters and DCT coefficients 1330.
  • An asterisk indicates an example of an error location 1244.
  • a remaining part of the Video Packet 1340, subsequent to the error 1244 is concealed.
  • the concealed part may include motion parameters and DCT coefficients. Hence, as motion vectors may be corrupted, estimated motion vector values must be used in the concealing process.
  • the second data format 1350 has added data partitioning only, as described with regard to FIG. 12.
  • motion parameters 1220 have been separated from the DCT coefficients 1240 by a motion marker 1230.
  • an asterisk indicates an example of an error location 1244.
  • a remaining part of the Video Packet 1360, subsequent to the error 1244 is concealed.
  • the concealed part only includes DCT coefficients, as the original motion vectors 1220 have been decoded prior to the motion marker 1230.
  • the original motion vectors 1220 may be employed to produce a higher quality and less disturbing concealment.
  • RVLC can be employed as a further enhancement to the use of data partitioning in data format 1350.
  • the decoder starts the reverse decoding from the end of the video packet until it encounters another invalid VLC code. If only one error occurs in the second data format 1350, the use of RVLC, to decode what would otherwise be concealed DCT coefficients, results in a minimal or substantially eliminated concealment process.
  • the third data format 1370 illustrates an example of this limitation, where two errors 1244 and 1245 are encountered in two different macroblocks.
  • the forward decoding stops at the first error 1244, after the motion marker 1230.
  • the reverse decoding process 1385 initiated by the decoder using RVLC, decodes the DCT coefficients from the end of the video packet back to the second error 1245. This leaves the concealed parts of the DCT coefficients being limited to the intermediate region 1365.
  • FIG. 14 illustrates the objective degradations caused by estimating values of lost motion vectors, or using original motion vectors, associated with the error scenarios of the first and second data formats of FIG. 13. Three plots are illustrated:
  • test sequence Foreman is coded at 18.8 kbit/s and 17.2 Kbit/s for the base and the temporal enhancement layers respectively.
  • the size of the Video Packets was set to 800 bits. Errors in the enhancement layer were generated using a General Packet Radio System
  • the inventive concepts of the further preferred embodiment enhances the concealment quality of motion compensated macroblocks of the MPEG-4 enhancement layers. Additionally, the use of RVLCs reduces the number of concealed macroblocks by reverse decoding from the end of a Video Packet up to the erroneous VLC code .
  • inventive concepts described above enable error resilience tools to be used within scalable video transmissions in a wireless, error prone environment. Consequently, as the inventive concepts described herein enable error resilience tools to be employed, it is now practically feasible to use scalable video in conjunction with network Quality of Service (QoS) information in order to deliver optimal quality to users in situations where network throughput and BER are likely to vary.
  • QoS Quality of Service

Abstract

A method for improving a quality of MPEG-4, or similar, scalable video enchancement layers transmitted over an error-prone network. The method includes the steps of inserting one or more re-synchronisation marker (660, 760, 860) into one or more enchancement layers of a scalable video sequence. When one or more errors (630, 730, 830) is detected in said one or more enhancement layers of said scalable video sequence, data is concealed (640, 740, 840) in the video sequence from a video bit position at which said one or more errors is detected, to a video bit position of a subsequent error free re-synchronisation marker. In this manner, a reduction in the amount of error-free data that has to be discarded can be achieved. This, in turn, reduces the amount of visual degradation in the video transmission.

Description

Scalable Video Transmissions
Field of the Invention
This invention relates to video transmission systems and video encoding/decoding techniques. The invention is applicable to a video compression system where the video has been compressed using a scalable compression technique.
Background of the Invention In the field of video technology, it is known that video is transmitted as a series of still images/pictures. Since the quality of a video signal can be affected during coding or compression of the video signal, it is known to include additional information or 'layers' based on the difference between the video signal and the encoded video bit stream. The inclusion of additional layers enables the quality of the received signal, following decoding and/or decompression, to be enhanced. Hence, a hierarchy of base pictures and enhancement pictures, partitioned into one or more layers, is used to produce a layered video bit stream.
A scalable video bit-stream refers to the ability to transmit and receive video signals of more than one resolution and/or quality simultaneously. A scalable video bit-stream is one that may be decoded at different rates, according to the bandwidth available at the decoder. This enables the user with access to a higher bandwidth channel to decode high quality video, whilst a lower bandwidth user is still able to view the same video, albeit at a lower quality. The main application for scalable video transmissions is for systems where multiple decoders with access to differing bandwidths are receiving images from a single encoder. Scalable video transmissions can also be used for bit-rate adaptability where the available bit rate is fluctuating in time. Other applications include video multicasting to a number of end-systems with different network and/or device characteristics. More importantly, scalable video can also be used to provide subscribers of a particular service with different video qualities depending on their tariffs and preferences. Therefore, in these applications it is imperative to protect the enhancement layer from transmission errors. Otherwise, the subscribers may lose confidence in their network operator's ability to provide an acceptable service.
In a layered (scalable) video bit stream, enhancements to the video signal may be added to a base layer either by:
(i) Increasing the resolution of the picture (spatial scalability) ;
(ii) Including error information to improve the Signal to Noise Ratio of the picture (SNR scalability) ;
(iii) Including extra pictures to increase the frame rate (temporal scalability) ; or
(iv) Providing a continuous enhancement that may be truncated at any chosen bit rate (Fine Granular Scalability) .
Such enhancements may be applied to the whole picture or to an arbitrarily shaped object within the picture, which is termed object-based scalability.
In order to preserve the disposable nature of the temporal enhancement layer, the H.263+ ITU H.263 [ITU-T Recommendation, H.263, "Video Coding for Low Bit Rate Communication"] standard dictates that pictures included in the temporal scalability mode should be bi-directionally predicted (B) pictures. These are as shown in the video stream of FIG. 1.
FIG. 1 shows a schematic illustration of a scalable video arrangement 100 illustrating B picture prediction dependencies, as known in the field of video coding techniques. An initial intra-coded frame (Ii) 110 is followed by a bi-directionally predicted frame (B2) 120.
This, in turn, is followed by a (uni-directional) predicted frame (P3) 130, and again followed by a second bi- directionally predicted frame (B4) 140. This again, in turn, is followed by a (uni-directional) predicted frame (P5) 150, and so on.
As an enhancement to the arrangement of FIG. 1, a layered video bit stream may be used. FIG. 2 is a schematic illustration of a layered video arrangement, known in the field of video coding techniques. A layered video bit stream includes a base layer 205 and one or more enhancement layers 235.
The base layer (layer-1) includes one or more intra-coded pictures (I pictures) 210 sampled, coded and/or compressed from the original video signal pictures. Furthermore, the base layer will include a plurality of subsequent predicted inter-coded pictures (P pictures) 220, 230 predicted from the intra-coded picture (s) 210. In the enhancement layers (layer-2 or layer-3 or higher layer (s) ) 235, three types of picture may be used: (i) Bi-directionally predicted (B) pictures (not shown) ; (ii) Enhanced intra-coded (El) pictures 240 based on the intra-coded picture (s) 210 of the base layer 205; and
(iii) Enhanced predicted (EP) pictures 250, 260, based on the inter-coded predicted pictures 220, 230 of the base layer 205.
The vertical arrows from the lower, base layer illustrate that the picture in the enhancement layer is predicted from a reconstructed approximation of that picture in the reference (lower) layer.
If prediction is only formed from the lower layer, then the enhancement layer picture is referred to as an El picture. It is possible, however, to create a modified bi- directionally predicted picture using both a prior enhancement layer picture and a temporally simultaneous lower layer reference picture. This type of picture is referred to as an EP picture or "Enhancement" P-picture.
The prediction flow for El and EP pictures is shown in FIG. 2. Although not specifically shown in FIG. 2, an El picture in an enhancement layer may have a P picture as its lower layer reference picture, and an EP picture may have an I picture as its lower-layer enhancement picture.
For both El and EP pictures, the prediction from the reference layer uses no motion vectors. However, as with normal P pictures, EP pictures use motion vectors when predicting from their temporally, prior-reference picture in the same layer. Current standards incorporating the aforementioned scalability techniques include MPEG-4 and H.263. These standards create highly compressed bit-streams, which represent the coded video. However, due to this high compression, the bit-streams are very prone to corruption by network errors as they are transmitted. For example, in the case of streaming video over an error prone network, even with existing network level error protection tools employed, it is inevitable that some bit-level corruption will occur in the bit-stream and be passed on to the decoder .
To counter these bit-level errors, the coding standards have been designed with various tools incorporated that allow the decoder to cope with the errors. These tools enable the decoder to localise and conceal the errors within the bit-stream.
However, the current MPEG-4 standard does not allow the use of these error resilience tools within the scalable enhancement layers .
The MPEG-4 standard does define three tools for error resilience of video bit-streams. These are re- synchronisation markers, data partitioning (DP) and reversible variable length codes (RVLCs) . These tools are defined for use in the base layer.
Data Partitioning (DP) separates motion parameters from the texture information, i.e. discrete cosine transform (DCT) coefficients, within a video transmission. Data partitioning is achieved by separating the motion and macroblock header information away from the texture information. This approach requires that a second re- synchronisation marker be inserted between motion and texture information. The use of data partitioning, in the same manner as the use of RVLCs, is signalled to the decoder in the Video Object Layer's (VOL's) header. If the texture information is lost, this approach utilises some motion information to conceal these errors. That is, due to the errors the texture information is discarded, while the motion is used to motion compensate the previously decoded Video Object Plane (VOP) . This separation is performed on a Video Packet basis where an additional re-synchronisation marker called a Motion Marker separates the motion parameters from the DCT coefficients. If an error occurs in the DCT coefficients portion, then the correctly decoded motion vectors are employed to conceal the remaining portion of the Video Packet .
The MPEG-4 visual standard associates another useful tool with data partitioning called Reversible Variable Length Codes (RVLCs) . It is noteworthy that, according to the MPEG-4 video standard, RVLC can only be used in conjunction with data partitioning within the base layer. Here the DCT coefficients are coded with RVLCs rather than the traditional Variable Length Codes (VLCs) . As such, parts of a bit stream, which cannot be decoded in the forward direction due to the presence of errors, can often be decoded in the reverse or backward direction. Hence, the number of discarded bits can be reduced.
It is worth re-iterating that the MPEG-4 bit stream syntax and the Simple Scalable Profile do MOT currently allow the use of error resiliency tools in the Enhancement layer. This was because enhancement layers were considered as an error resilience tool . Enhancement layer information contains visual information that enhances the decoding quality of the more important base layer. Hence, as enhancement layer information was not deemed essential, no further resiliency was anticipated.
Thus, the focus for higher levels of protection in a video bit sequence in current communications systems is the base layer. This means that when an error occurs in an enhancement layer bit-stream, the decoder, wishing to keep the enhancement layer, has to conceal much more data, potentially in error, than it would have to if the error resilience tools could be used.
In summary, there exists a need in the field of video communications, and in particular in scalable video communications, for an apparatus and a method for improving the quality of scalable video enhancement layers transmitted over an error-prone network, where the abovementioned disadvantages with prior art arrangements may be alleviated.
Known prior art documents include: (i) US-A-200221761 (Zhang et al . ) , published 21 February
2002;
(ii) 'Error Resilience Methods for FGS Video Enhancement
Bitstream' , R. Yan et al , The First Pacific Rim Conference on Multimedia (IEEE-PCM 2000), Dec 13-15, 2000; (iϋ) 'Robust Video Coding Algorithms and Systems' , J.D.
Villasenor, et al . , Proceedings of the IEEE, Vol. 87, No.
10, October 1999. Statement of Invention
The present invention provides a method for improving a quality of scalable video enhancement layers transmitted over an error-prone network, as claimed in claim 1, a method for improving a quality of scalable video enhancement layers transmitted over an error-prone network, as claimed in claim 2, a method for improving a quality of scalable video enhancement layers transmitted over an error-prone network, as claimed in claim 6, a video communication system, as claimed in claim 10, a fifth aspect of the present invention, there is provided a video communication system, as claimed in claim 13, a video communication unit, as claimed in claim 16, a video encoder or a video decoder, as claimed in claim 17, and a mobile radio device, as claimed in claim 19.
Further aspects of the present invention are as claimed in the dependent claims.
In summary, an apparatus and methods for improving the quality of scalable video enhancement layers transmitted over an error-prone network by the use of re- synchronisation markers are described. This invention provides a method by which re-synchronisation markers are added to enhancement layer bit-streams in scalable MPEG-4 or similar video compression systems, or H.263+ or similar video compression systems. This is in order to promote better bit-level error resilience capabilities, and reduce the amount of concealment of error-free data that takes place.
In addition, or in the alternative, this invention proposes the use of the data partitioning, and preferably a Reversible Variable Length Coding, error resilience tool in the video enhancement layers of the MPEG-4 Simple Scalable Profile. One aim is to reduce or to substantially eliminate the macroblock displacements, caused by error concealment utilising estimated motion vectors. A further aim is to minimise the number of discarded bits by performing reverse decoding.
Brief Description of the Drawings FIG. 1 is a schematic illustration of a video coding arrangement showing picture prediction dependencies, as known in the field of video coding techniques.
FIG. 2 is a schematic illustration of a known layered video coding arrangement . Exemplary embodiments of the present invention will now be described, with reference to the accompanying drawi gs, in which:
FIG. 3 is a schematic representation of a scalable video communication system adapted to use re-synchronisation markers in an enhancement layer of a video sequence in accordance with the preferred embodiment of the present invention.
FIG. 4 illustrates a first enhancement layer bit-stream frame configuration. FIG. 5 illustrates a second enhancement layer bit-stream frame configuration.
FIG. 6 illustrates a first enhancement layer bit-stream frame configuration adapted to incorporate the preferred embodiment of the present invention. FIG. 7 illustrates a second enhancement layer bit-stream frame configuration adapted to incorporate the preferred embodiment of the present invention. FIG. 8 illustrates a third enhancement layer bit-stream frame configuration adapted to incorporate the preferred embodiment of the present invention.
FIG. 9 illustrates an example of syntax code for use in the MPEG-4 standard, for implementing a preferred embodiment of the present invention.
FIG. 10 illustrates a graph highlighting the improvement of decoded SNR scalable video quality, in accordance with a preferred embodiment of the present invention. FIG. 11 illustrates a graph highlighting the improvement of decoded temporal scalable video quality, in accordance with a preferred embodiment of the present invention. FIG. 12 illustrates a data format of an enhancement layer bit stream that can be adapted to incorporate the use of data partitioning and RVLCs in accordance with the preferred further embodiment of the present invention. FIG. 13 illustrates at least one benefit from introducing data partitioning, and preferably reverse variable length coding, in an enhancement layer bit stream, for example the data format of FIG. 12.
FIG. 14 illustrates the objective degradations caused by estimating values of lost motion vectors, or using original motion vectors, associated with the error scenarios of FIG. 13. FIG. 15 illustrates proposed syntax amendments to section 6.2.5 "Video Object Plane and Video Plane with Short Header" of the MPEG-4 standard.
Description of Preferred Embodiments This invention applies to SNR scalable encoded video.
However, it is within the contemplation of the invention that the inventive concepts described herein can be applied to other types of scalable video sequences such as temporal scalability, spatial scalability and Fine Granular scalability (FGS) .
The inventive concepts herein described find particular application in the current MPEG technology arena, and in future versions of MPEG technology. However, the inventors have also recognised the applicability of the present invention to H.263+ technology.
The preferred embodiment of the present invention illustrates a technique for introducing re-synchronisation markers in enhancement layer bit-streams to improve their error resilience.
In the example context of MPEG-4 video, the video bit- stream syntax is modified to enable the use of the standard base layer error resilience tools within the enhancement layer bit-stream. Advantageously, the decoder can then make use of the re-synchronisation markers in the enhancement layer bit-stream in the same manner as for a base layer bit-stream.
An analogous technique can be used with H263+ video.
Referring first to FIG. 3, a schematic representation of a video communication system 300, including video encoder 315 and video decoder 325, adapted to incorporate the preferred embodiment of the present invention, is shown.
In FIG. 3, a video picture Fo is compressed 310 in a video encoder 315 to produce the base layer bit stream signal to be transmitted at a rate r__ kilobits per second (kbps) . This signal is decompressed 320 at a video decoder 325 to produce the reconstructed base layer picture Fo' .
The compressed base layer bit stream is also decompressed at 330 in the video encoder 315 and compared with the original picture Fo at 340 to produce a difference signal
350. This difference signal is compressed at 360 and transmitted as the enhancement layer bit stream at a rate T2 kbps . This enhancement layer bit stream is decompressed at 370 in the video decoder 325 to produce the enhancement layer picture Fo'' which is added to the reconstructed base layer picture Fo ' at 380 to produce the final reconstructed picture Fo' "
In accordance with the preferred embodiment of the present invention, the compression function 360 in the video encoder 315 has been adapted to incorporate one or more re- synchronisation markers in the enhancement layer bit- stream. Furthermore, the decompression function 370 in the video decoder 325 has been adapted to determine the location of these one or more re-synchronisation markers in the enhancement layer bit-stream. The video decoder 325 has also been adapted to conceal errors in the received scalable video sequence up to the next available error-free re-synchronisation marker. The use of such re- synchronisation markers is further described with regard to FIG's 6 to 8.
It is within the contemplation of the invention that alternative encoding and decoding configurations could be adapted to use re-synchronisation markers within the enhancement layer bit-stream. As a result, the inventive concepts hereinafter described should not be viewed as being limited to the example configuration provided in FIG. 3.
As discussed above, in prior art arrangements, an MPEG-4 enhancement layer bit-stream does not have the facility to include any re-synchronisation markers within it. Hence, if the video decoder encounters an error, it has no choice but to conceal all of the data from the point at which the error occurred to the start of the next readable start code (in the case of MPEG-4 this would be a VOP_Start_Code) at the beginning of a subsequent frame of video .
The problem with this prior art arrangement is that the actual amount of error-free data concealed could be very large, depending on the position of the error. This is illustrated in FIG. 4 and FIG. 5.
Referring now to FIG. 4, a first enhancement layer bit- stream frame 400 is illustrated. The first enhancement layer bit-stream frame 400 includes a start code 410, followed by image data of a frame λN' 420. The image data of frame 'N' 420 is shown to contain errors 430 within the frame 420. FIG. 4 highlights a problem associated with such an enhancement layer bit-stream frame in that a substantial amount of error-free data 440 subsequent to the errors 430 within the frame 420 has to be concealed. Clearly, it is undesirable, as well as inefficient, to conceal such error-free data.
Referring next to FIG. 5, a second enhancement layer bit- stream frame 500 is illustrated that highlights a further problem with errors encountered in enhancement layer bit- streams. The second enhancement layer bit-stream frame 500 illustrates errors being encountered within image data of frame 'N' 520 as well as within the subsequent start code 510 of the next N+l' frame 525. When errors 530 occur over the start code in this manner, a whole frame 540 may end up being concealed, as shown.
Typically, the effect to the viewer of the aforementioned concealed error-free frames depends on the type of scalability being used. In the case of spatial or SNR scalability, where the enhancement layer increases the spatial resolution or quality of the base image, the viewer may witness frames that are enhanced at the top of the image, but not enhanced at the bottom. Alternatively, if the errors affect a start code, the viewer may experience an annoying flicker, as the enhancement layer disappears entirely over a frame. In the case of temporal scalability, the concealment can lead to some strange effects, with the viewer seeing part of the current frame and parts of previously decoded frames.
In accordance with the preferred embodiment of the present invention, re-synchronisation markers 660, 760, 860 are introduced to the enhancement layer bit-stream transmitted by the video encoder 315, to improve the quality of the decoded video image. The improvement is achieved by reducing the amount of error-free data that has to be concealed by the decoder. Instead of the video decoder 325 having to conceal all of the data in a frame, from the point the error occurred to the end of the frame, the concealment only has to continue up to the first occurrence of either: (i) The first valid re-synchronisation marker, bearing in mind that errors could occur in the markers themselves; or (ii) The next valid start code.
In this manner, the amount of error-free data concealment is reduced substantially.
Referring next to FIG. 6, a first enhancement layer bit- stream frame configuration 600 is shown, adapted by the introduction of re-synchronisation markers 660 in accordance with the preferred embodiment of the present invention. The configuration is similar to that shown in FIG. 4, where the errors 630 occur within the image data frame 620 and are determined to be located between two start codes 610, 650. As shown, by introducing re- synchronisation markers in the manner described above, the video decoder only conceals error-free data up to the re- synchronisation marker 660 that immediately follows the errors 630. Hence, the amount of concealed error-free data 640 is reduced significantly, when compared to the prior art arrangement described with respect to FIG. 4.
Referring next to FIG. 7, a second enhancement layer bit- stream frame configuration 700 is shown, adapted by the introduction of re-synchronisation markers 760 in accordance with the preferred embodiment of the present invention. The configuration is similar to that shown in FIG. 5, where the errors 730 occur across a start code 710, as well as within the image data frame N 720. As shown, by introducing re-synchronisation markers in the manner described above, the video decoder only conceals error-free data up to the first re-synchronisation marker 760, located in the subsequent image data frame N+l' 725. This is because the corrupted VOP header is replicated in the Header Extension Code (HEC) of the Video Packet Header and therefore the Video Object Plane's (VOP's) header can be restored. Hence, the amount of concealed error-free data is reduced significantly, when compared to the prior art arrangement described with respect to FIG. 5.
Referring next to FIG. 8, a third enhancement layer bit- stream frame configuration 800 is shown, adapted by the introduction of re-synchronisation markers 860 in accordance with the preferred embodiment of the present invention.
In FIG. 8, it is shown that the errors may occur around, and affecting, the re-synchronisation markers 860. This may be, for example, around the first re-synchronisation marker 860 within the image data frame N 820. As shown for this situation, the video decoder only conceals error-free data up to the second re-synchronisation marker 860, located within the same image data frame N' 820. Hence, the amount of concealed error-free data is still reduced significantly, and only extends into the period between re- synchronisation markers 860.
For completeness, it should also be noted that state of the art techniques for error resilience have been developed that allow decoding of video frames even if the frame header is corrupted by errors, provided re-synchronisation markers exist in the bit-stream. It is envisaged that such techniques can also be used to complement the performance of the present invention. Referring now to FIG. 9, the existing MPEG-4 syntax code 900 can be readily adapted to incorporate the preferred embodiment of the present invention using re-synchronisation markers in one or more enhancement layers . The required change 910 to the MPEG-4 syntax code, to accommodate the preferred embodiment of the present invention, is illustrated in FIG. 9. Thus, in accordance with the preferred embodiment of the invention, a processor in a video communication unit may be re-programmed, and thereby adapted, to incorporate the inventive concepts hereinbefore described.
More generally, the adaptation may be implemented in the respective video communication unit in any suitable manner. For example, new apparatus may be added to a conventional video communication unit, or alternatively existing parts of a conventional video communication unit may be adapted, for example by reprogramming one or more processors therein. As such, the required adaptation may be implemented in the form of processor-implementable instructions stored on a storage medium, such as a floppy disk, hard disk, PROM, RAM or any combination of these or other storage multimedia.
It is also within the contemplation of the invention that such adaptation of a video encoding or video operations decoding may be facilitated by any video communication unit operating in a video communication system. This may be, for example, user equipment such as mobile or portable radios or telephones, or wireless or wired serving communication units such as base transceiver or intermediate (repeater) equipment . With the revised syntax code in place, re-synchronisation markers can be placed in the enhancement layer bit-stream, and used to reduce the amount of error-free data concealment. Based on Profiles and levels, the MPEG-4 standard states the maximum size of Video Packets, or the number of bits between two consecutive resynchronisation markers, in a given base layer VOP. This aids the decoder, as it roughly knows when to expect the next marker.
It is envisaged that the period between re-synchronisation markers in the enhancement layer may be user and/or implementation defined. A smaller distance between re- synchronisation markers leads to a reduction in the amount of data that has to be concealed, should an error occur.
However, the distance between the re-synchronisation markers also affects the overall number of bits in the video bit-streams. So, in a fixed bit-rate environment the quality of the enhancement layer has to be reduced, in order for the video to meet the bit budget of the transmission system.
The inventors of the present invention have performed extensive experiments that have shown that by placing re- synchronisation markers at approximately every eight hundred bits, the overall increase in bit-stream size is just less than four percent.
The current MPEG standard, both for scalable MPEG-4 and FGS MPEG-4, explicitly states that error resilience tools are not to be used in the enhancement layer (s) . Hence, the standard, as well as public domain literature, do not allow the use of re-synchronisation markers in an enhancement layer of scalable MPEG-4. In fact, the MPEG 4 standard, as well as public domain literature, teaches away from the present invention.
The current teaching also states that no foreseeable benefit can be gained in using error resilience tools in scalable (including FGS) MPEG-4 video communication systems. Furthermore, it is generally understood in the field of video communication that MPEG-4 enhancement layers would not be used in a mobile communication environment .
Despite the prior art teaching away from the present invention, the results obtained by the inventors of the present invention, shown below in FIG. 10 and FIG. 11, have shown that resynchronisation markers in the enhancement layer (s) are of great benefit to the use of MPEG-4 scalable video in a mobile environment. As such, it is envisaged that emerging 3rd generation 3G mobile communication systems will require scalable video to accommodate different bit ' rates at different parts of the coverage area. It is further envisaged that future video technology, similar to today's current MPEG-4 standard, will similarly benefit from the inventive concepts described herein.
Referring now to FIG. 10, a graph 1000 highlights the improvement of decoded SNR scalable video quality, in implementing the preferred embodiment of the present invention. As clearly shown, the peak signal to noise ratio (PSNR) is substantially improved when re- synchronisation markers are incorporated into the enhancement layer bit-stream, as compared to the PSNR levels without re-synchronisation markers. Referring now to FIG. 11, a graph 1100 highlights the improvement of decoded temporal scalable video quality, when implementing the preferred embodiment of the present invention. As clearly shown, the peak signal to noise ratio (PSNR) is substantially improved when resynchronisation markers are incorporated into the enhancement layer bit-stream, as compared to the PSNR levels without re-synchronisation markers.
The measurements of FIG. 10 and FIG. 11 were made by adding re-synchronisation markers to the enhancement layers of SNR and temporally scalable video bit-streams transmitted over a GPRS network. In particular, a noteworthy feature is the relative smoothness of the lines. For example, the lines illustrating the video without re-synchronisation markers tend to oscillate much more because of the larger amounts of data concealment taking place. In contrast, the lines illustrating video with re-synchronisation markers are much flatter due to the reduced amount of data concealment that has to take place.
Finally, the applicant notes that future versions of the MPEG communication standard, such as the Joint Video Team (JVT) (from MEPG-4 and H26L) configuration are currently under development. The present invention is not limited to the MPEG-4 standard, and is envisaged by the inventors as applying to future versions of scalable video compression.
It is within the contemplation of the present invention that the aforementioned inventive concepts may be applied to any video communication unit and/or video communication system. In particular, the inventive concepts find particular use in wireless (radio) devices, such as mobile telephones/mobile radio units and associated wireless communication systems. Such wireless communication units may include a portable or mobile PMR radio, a personal digital assistant, a laptop computer or a wirelessly networked PC.
Although the preferred embodiment of the present invention has been described with reference to the MPEG-4 standard, scalable video system technology may be implemented in the 3rd generation (3G) of digital cellular telephones, commonly referred to as the Universal Mobile Telecommunications Standard (UMTS) . Scalable video system technology may also find applicability in the packet data variants of both the current 2nd generation of cellular telephones, commonly referred to as the general packet-data radio system (GPRS) , and the TErrestrial Trunked RAdio (TETRA) standard for digital private and public mobile radio systems. Furthermore, scalable video system technology may also be utilised in the Internet. The aforementioned inventive concepts will therefore find applicability in, and thereby benefit, all these emerging technologies.
It will be understood that the video transmission arrangement described above provides at least the advantage of achieving a reduction in the amount of error-free data that has to be discarded. This, in turn, reduces the amount of visual degradation in the video transmission. This is especially important when transmitting video over wireless channels and the Internet, where the errors can be severe. Applications of digital video where the preferred method improves decoder performance are one-way and two-way video communications, surveillance applications, and video streaming applications. Method of the invention
Summarising the discussion above, a method for improving a quality of MPEG-4 (or similar) scalable video enhancement layers transmitted over an error-prone network has been described. The method includes the steps of inserting one or more re-synchronisation markers into one or more enhancement layers of a scalable video sequence in a video encoder. A video decoder detects one or more errors in the one or more enhancement layers of the scalable video sequence; and, in response to detecting one or more errors, conceals data in the received scalable video sequence from a video bit position at which the one or more errors is detected, to a video bit position of a subsequent error- free re-synchronisation marker.
The inventors have furthermore realised that the present invention may also be applied to video encoding in accordance with the H.263+ standard. This is the method claimed in appended claim 2. H.263+ video is generally used for lower bandwidth transmission systems than MPEG-4 video. The H.263+ standard actually differs from the MPEG-4 standard in several aspects. One of these aspects is that the H.263+ standard does not expressly state that error resilience tools should not be used in the enhancement layers .
Apparatus of the invention
A video communication system has been described that includes a video encoder comprising a processor for encoding a video sequence into an MPEG-4 (or similar) scalable video sequence having a plurality of enhancement layers; and insertion means to insert one or more re- synchronisation markers into one or more enhancement layers. A transmitter transmits the scalable video sequence containing said one or more re-synchronisation markers . A video decoder includes a receiver for receiving the scalable video sequence and a detector that detects one or more errors in the one or more enhancement layers of the scalable video sequence. A processor is operably coupled to the detector for concealing data in the received scalable video sequence, when one or more errors are detected. The video data is concealed from a video bit position at which said one or more errors is detected to a video bit position of a subsequent error-free resynchronisation marker.
A video communication unit, an adapted video encoder, an adapted video decoder, and a mobile radio device incorporating any one of these units, have also been described.
Generally, the inventive concepts contained herein are equally applicable to any suitable video or image transmission system. Whilst specific, and preferred, implementations of the present invention are described above, it is clear that one skilled in the art could readily apply variations and modifications of such inventive concepts.
Thus, an improved apparatus and methods for improving the quality of scalable video enhancement layers transmitted over an error-prone network have been provided, whereby the aforementioned disadvantages with prior art arrangements have been substantially alleviated. Enhanced approach
The inventors have also considered a further enhancement of the above invention. The enhanced solution may further improve the subjective quality of video for a viewer.
The inventors have considered the problem that error concealment techniques in the temporally scalable enhancement layer conceal data from the point that the error occurred to the next re-synchronisation marker, or the start code at the beginning of the next frame. If a large amount of motion occurred between the current frame and the previous frame, then the area that is concealed can look discontinuous from the rest of the frame.
Therefore, the inventors have developed a method to selectively conceal an entire frame of video, in order to reduce the discontinuities in the image, thus improving subjective quality for the viewer.
This method improves the image quality by adaptively deciding on how much data should be concealed, to give the best image quality. If the motion between the previous frame and the current frame with errors is large, then the method involves a decision to conceal the whole frame to avoid any discontinuity in the decoded image. If the motion is small, then the decision is to conceal as little data as possible.
A detailed arrangement starts from the realisation that discontinuities in the decoded image can be very visible to a viewer, using concealment techniques that conceal errors from the point the error occurred to the next re- synchronisation point, i.e. re-synchronisation marker or start code.
The visibility of the discontinuity between the unconcealed and concealed areas depends on the amount of motion that is taking place in the scene. For example, visualise a scene with little background motion, but people running in the foreground. The people create lots of foreground motion, which leads to a large discontinuity, which is visually disturbing. In order to improve the quality of the video for the viewer (subjective quality will increase, but objective quality measures such as PSNR will decrease) , the inventors propose a method in which a normalised Sum of Absolute Difference (SAD) is calculated between the current frame and previous frame over the non-concealed area. If the normalised SAD exceeds a pre-defined threshold (indicating a large amount of motion) then the whole frame is concealed.
The enhancement, using the frame suppression method, can in fact even be used in a video transmission system that does not have re-synchronisation markers in the enhancement layers . So it can be seen both as an enhancement to the claimed invention, and as a potential alternative solution if employed alone.
The inventors of the present invention have further recognised a potential problem in the concealment of corrupted visual data by means of estimating the values of lost motion vectors and motion compensating previous VOPs. As lost motion vectors are estimated from either the surrounding macroblocks, or the co-located macroblocks from previously decoded VOPs, it follows that these motion vectors are not accurate, resulting in non-accurate motion compensation. Hence, estimated motion vectors may often cause the displacement of concealed macroblocks.
Thus, the inventors of the present invention propose an enhancement to the preferred embodiment of the present invention described above. In this enhanced further embodiment of the present invention, it is proposed that data partitioning, currently limited to use in a base layer of a video data stream, is also employed in video enhancement layers. In a preferred further embodiment, the data partitioning error resilience tool is supplemented with reversible variable length coding.
In the example context of MPEG-4 video, the video bit- stream syntax is modified to enable the use of these standard base layer error resilience tools within the enhancement layer bit-stream. Advantageously, the decoder and decoding operation are further adapted to make use of the data partitioning, into two or more portions of bit streamed data, in the enhancement layer bit-stream in the same manner as for a base layer bit-stream.
An analogous technique can be used with H263+ video.
The focus for the enhanced further embodiment is that by employing data partitioning in the enhancement layer, either in isolation or in addition to the embodiments described with reference to FIG's 4 to 6, a (further) reduction in the amount of data discarded when an error is detected can be achieved. Further, the inventors of the present invention have recognised a potential deficiency of using a specific (resynchronisation) Motion Marker to separate the motion parameters from DCT coefficients within a video packet header in the data partitioning. If an error occurs in the DCT coefficients part, then the original motion vectors are employed, as opposed to the estimated motion vectors, to conceal the remaining part of the video packet .
Thus, a corruption of one or more variable length coded
(VLC) DCT coefficients impairs the rest of a video packet regardless of its error status. In this regard, the further embodiment of the present invention proposes to additionally use reversible variable length codes (RVLCs) to allow parts of a video sequence bit stream that could not be decoded in the forward direction due to the presence of errors, to be decoded in the reverse (backward) diction. A preferred example would be within enhancement layers of the MPEG-4 Simple Scalable Profile.
In accordance with the preferred further embodiment of the present invention, the compression function 360 in the video encoder 315 of FIG. 3 has been adapted to incorporate data partitioning in the enhancement layer bit-stream. Furthermore, the decompression function 370 in the video decoder 325 of FIG. 3 has been adapted to determine that data partitioning in the enhancement layer bit-stream has been performed. The video decoder 325 has also been adapted to conceal errors within one or more data portions of the partitioned data that are detected as having errors. The video decoder 325 has further been adapted to perform RVLC on the enhancement layer, when an error in one or more data portions of a data-partitioned sequence is detected, in a similar manner to that currently performed on a base layer of a video sequence . The use of such data partitioning and RVLC is further described with regard to FIG's 12 to 14.
It is within the contemplation of the invention that alternative encoding and decoding configurations could be adapted to use DP, and preferably RVLC within the enhancement layer bit-stream. As a result, the inventive concepts hereinafter described should not be viewed as being limited to the example configuration provided in FIG. 3.
The introduction of such further error resiliency tools within the enhancement layer reduces, or substantially eliminates, the macroblock displacements caused by error concealment utilising estimated motion vectors. Additionally, the use of DP and RVLCs minimises the number of discarded bits by facilitating the performance of reverse decoding on a video bit stream.
FIG. 12 illustrates a data format 1200 of an enhancement layer bit stream that can be adapted to incorporate the use of data partitioning and RVLCs in accordance with the preferred further embodiment of the present invention. The data format 1200 includes a video packet header 1210. The data format 1200 of the enhancement layer has been adapted to incorporate data partitioning, whereby the motion and header information 1220 has been separated from the texture information 1240. A re-synchronisation motion marker 1230 indicates the location where the motion parameters are separated from the DCT coefficients . A video packet header 1250 of a subsequent video packet, VOP start code or GOV start code, may follow the data-partitioned data format.
In addition, a texture header 1235 has been introduced to indicate texture related information to the decoder. Errors 1244 within the texture information can then be localised by decoding in the forward direction 1242 and in the reverse direction 1246, by use of RVLCs.
FIG. 13 illustrates the benefits of introducing data partitioning and RVLC in an enhancement layer bit stream, for example the data format of FIG. 12. Three enhancement layer data formats are shown, all of which include initial re-synchronisation markers 1320, as described above, and video packet headers 1210. The first data format 1310 indicates a data structure that does not employ data partitioning (and thereby does not employ RVLC) . In this regard, the body information contains both motion parameters and DCT coefficients 1330.
An asterisk indicates an example of an error location 1244. In the first data format 1310, a remaining part of the Video Packet 1340, subsequent to the error 1244, is concealed. The concealed part may include motion parameters and DCT coefficients. Hence, as motion vectors may be corrupted, estimated motion vector values must be used in the concealing process.
However, the second data format 1350 has added data partitioning only, as described with regard to FIG. 12. In this regard, motion parameters 1220 have been separated from the DCT coefficients 1240 by a motion marker 1230. Again, an asterisk indicates an example of an error location 1244. In the second data format 1350, a remaining part of the Video Packet 1360, subsequent to the error 1244, is concealed. However, in this regard, the concealed part only includes DCT coefficients, as the original motion vectors 1220 have been decoded prior to the motion marker 1230.
Hence, the original motion vectors 1220 may be employed to produce a higher quality and less disturbing concealment.
As a further enhancement to the use of data partitioning in data format 1350, RVLC can be employed. In this case, if an illegal or invalid VLC code is encountered, the decoder starts the reverse decoding from the end of the video packet until it encounters another invalid VLC code. If only one error occurs in the second data format 1350, the use of RVLC, to decode what would otherwise be concealed DCT coefficients, results in a minimal or substantially eliminated concealment process.
Note that the benefits gained from using RVLC in addition to data partitioning are dependent upon the number of errors occurring within a video packet. The third data format 1370 illustrates an example of this limitation, where two errors 1244 and 1245 are encountered in two different macroblocks. In this regard, the forward decoding stops at the first error 1244, after the motion marker 1230. The reverse decoding process 1385 initiated by the decoder using RVLC, decodes the DCT coefficients from the end of the video packet back to the second error 1245. This leaves the concealed parts of the DCT coefficients being limited to the intermediate region 1365. Clearly, this is a substantial improvement over the second data format 1350, data- partitioning only, case, where no reverse decoding is performed. FIG. 14 illustrates the objective degradations caused by estimating values of lost motion vectors, or using original motion vectors, associated with the error scenarios of the first and second data formats of FIG. 13. Three plots are illustrated:
(i) An error free transmission of the enhancement layer;
(ii) An enhancement layer with video packets where data partitioning is employed with an error 1244, and hence the original motion vectors are used for the concealment; and
(iii) An enhancement layer with video packets where motion parameters and DCT coefficients are not separated with an error 1244 introduced. Hence, estimated motion vectors are used for the concealment process .
In FIG. 14, the test sequence Foreman is coded at 18.8 kbit/s and 17.2 Kbit/s for the base and the temporal enhancement layers respectively. The size of the Video Packets was set to 800 bits. Errors in the enhancement layer were generated using a General Packet Radio System
(GPRS) physical link layer simulator. The resultant Frame Erasure Rate (FER) is 6.8% and the Residual Bit Error Rate
(RBER) is 0.1%.
Although it is evident from the results illustrated in FIG. 14, that data partitioning enhances the quality of concealed macroblocks, the absence of the DCT coefficients (or the prediction error in this case) is still noticeable. Any such degradation is preferably overcome by reducing the number of concealed macroblocks by the use of RVLCs . RVLCs can only be used in conjunction with the data-partitioning tool. Hence, by superimposing the reverse decoding technique of RVLC on the arrangements in FIG. 13, the DCT coefficients concealed due to error 1244 can be minimised or substantially eliminated.
For the use of data partitioning and RVLCs in the enhancement layers, a simple amendment to the current MPEG-4 video bit stream syntax (section 6.2.5 "Video object plane and video plane with short header", VideoObjectPlane () ) is needed. The syntax code change, which is a further enhancement to that illustrated in FIG. 9, is illustrated in FIG. 15.
It is noteworthy that the benefits to be gained from employing data partitioning (and subsequently RVLC) also varies dependent upon the scalability mode used. For instance in Spatial and SNR scalabilities, predicted VOPs are always reconstructed without motion compensation, i.e. with nil motion vectors. Separating motion parameters from texture data in this case does not make much sense unless RVLC is employed. This is because corrupted macroblocks are often concealed by copying them from the corresponding co- located ones of the base layer VOP without motion compensation. On the other hand, for all other cases where the macroblocks are motion compensated for prediction, data partitioning is a useful tool regardless of the use of RVLCs. Hence, it is within the contemplation of the invention that data partitioning can be used in isolation, or in tandem with RVLC, within enhancement layer video bit streams .
As described above, the inventive concepts of the further preferred embodiment enhances the concealment quality of motion compensated macroblocks of the MPEG-4 enhancement layers. Additionally, the use of RVLCs reduces the number of concealed macroblocks by reverse decoding from the end of a Video Packet up to the erroneous VLC code .
Advantageously, the inventive concepts described above enable error resilience tools to be used within scalable video transmissions in a wireless, error prone environment. Consequently, as the inventive concepts described herein enable error resilience tools to be employed, it is now practically feasible to use scalable video in conjunction with network Quality of Service (QoS) information in order to deliver optimal quality to users in situations where network throughput and BER are likely to vary.
It is within the contemplation of the enhanced further embodiment of the present invention, with the use of DP within an enhancement layer of a video sequence, for DP to be used in isolation, or together with RVLC, or in addition to the re-synchronisation marker technique described above. Clearly, wherever practically possible, it is preferred to use all three error resilience tools within an enhancement layer video sequence. However, improved video quality benefits can still be gained by using re-synchronisation markers with or without data partitioning.

Claims

Claims
1. A method for improving a quality of MPEG-4, or similar, scalable video enhancement layers transmitted over an error-prone network, the method comprising the steps of: inserting one or more re-synchronisation markers (660, 760, 860) into one or more enhancement layers of a scalable video sequence; detecting one or more errors (630, 730, 830) in said one or more enhancement layers of said scalable video sequence; and concealing data (640, 740, 840) in said received scalable video sequence when said error is detected, from a video bit position at which said one or more errors is detected, to a video bit position of a subsequent error-free resynchronisation marker.
2. A method for improving a quality of H.263+, or similar, scalable video enhancement layers transmitted over an error-prone network, the method comprising the steps of: inserting one or more re-synchronisation markers (660, 760, 860) into one or more enhancement layers of a scalable video sequence; detecting one or more errors (630, 730, 830) in said one or more enhancement layers of said scalable video sequence; and concealing data (640, 740, 840) in said received scalable video sequence when said error is detected, from a video bit position at which said one or more errors is detected, to a video bit position of a subsequent error-free resynchronisation marker.
3. A method in accordance with claim 1 or claim 2, wherein said one or more errors (630, 730, 830) are detected over a frame start code (710) or across a resynchronisation marker (660, 760, 860) .
4. A method in accordance with any previous claim, wherein a number of said one or more re-synchronisation markers to be inserted into said one or more enhancement layers is user definable or pre-determined, for example to provide a trade off between error resiliency and bit efficiency.
5. A method in accordance with any previous claim, wherein a location of said one or more re-synchronisation markers is user definable or pre-determined, for example to provide a trade off between error resiliency and bit efficiency.
6. A method for improving a quality of MPEG-4 or H.263+ or similar, scalable video sequence transmitted over an error-prone network, in particular a method for improving a quality of MPEG-4, or similar, scalable video enhancement layer transmitted over an error-prone network according to any preceding Claim, the method comprising the steps of :
data partitioning (1220, 1240) of a video bit stream into two or more data portions in one or more enhancement layers of said scalable video sequence; detecting one or more errors (1244, 1245) in one or more of said data portions; and concealing data (1340, 1360) in one or more data portions where said one or more errors is detected.
7. A method according to Claim 6 , the method further comprising the step of: inserting one or more re-synchronisation markers (1230) into one or more enhancement layers of a scalable video sequence to indicate data partitioning between said two or more data portions.
8. A method according to Claim 6 or Claim 7, wherein said two or more data portions in said one or more enhancement layers include a data portion of motion parameters and a data portion of DCT coefficients.
9. A method according to any of preceding Claims 6 to 8, the method further comprising the step of: performing reversible variable length coding on said one or more of said data portions of said scalable video sequence when said one or more errors (1244, 1245) is detected in said one or more of said data portions.
10. A video communication system (300) comprising:
a video encoder (315) comprising: a processor for encoding a video sequence into an MPEG-4, or similar, scalable video sequence having a plurality of enhancement layers; insertion means to insert one or more re-synchronisation markers into one or more enhancement layers; and a transmitter for transmitting said scalable video sequence containing said one or more re-synchronisation markers; and
a video decoder (325) comprising: a receiver for receiving said scalable video sequence containing said one or more re-synchronisation markers from said video encoder; a detector detecting one or more errors in said one or more enhancement layers of said received scalable video sequence ; and a processor operably coupled to said detector for concealing data in said received scalable video sequence when said one or more errors is detected, wherein said video data is concealed from a video bit position at which said one or more errors is detected, to a video bit position of a subsequent error free re-synchronisation marker .
11. The video communication system (300) according to claim 10, wherein said detector detects one or more errors over a frame start code or across a re-synchronisation marker contained within said received scalable video sequence .
12. The video communication system (300) according to claim 10 or claim 11, wherein said video encoder further comprises user input means to enable a user to select : a number of said one or more re-synchronisation markers to be inserted into said one or more enhancement layers; and/or a location of said one or more re-synchronisation markers.
13. A video communication system (300), in particular a video communication system according to any of preceding
Claims 10 to 12, the video communication system comprising:
a video encoder (315) comprising: a processor for encoding a video sequence into an MPEG-4, or similar, scalable video sequence having a plurality of enhancement layers; data partitioning means to partition data of a video bit stream into two or more data portions in one or more enhancement layers of said scalable video sequence; and a transmitter for transmitting said scalable video sequence containing said one or more data portions; and
a video decoder (325) comprising: a receiver for receiving said scalable video sequence containing said one or more data portions from said video encoder; a detector detecting one or more errors in said one or more enhancement layers of said received scalable video sequence; and a processor operably coupled to said detector for concealing data in said received scalable video sequence when said one or more errors is detected, wherein said video data is concealed in one or more data portions where said one or more errors is detected.
14. The video communication system (300) according to claim 13, wherein said two or more data portions in said one or more enhancement layers include a data portion of motion parameters and a data portion of DCT coefficients .
15. The video communication system (300) according to Claim 13 or Claim 14, wherein said video encoder further comprises means to perform reversible variable length coding on said one or more of said data portions of said scalable video sequence when said one or more errors (1244, 1245) is detected in said one or more of said data portions .
16. A video communication unit (315, 325) adapted for use in the method of any of claims 1 to 9 or adapted for use in the communication system of any of claims 10 to 15.
17. A video encoder (315) or a video decoder (325) adapted for use in the method of any of claims 1 to 9 or adapted for use in the communication system of any of claims 10 to 15.
18. A mobile radio device comprising a video communication unit in accordance with claim 16 or a video encoder or a video decoder in accordance with claim 17.
19. A mobile radio device according to claim 13, wherein the mobile radio device is a mobile phone, a portable or mobile PMR radio, a personal digital assistant, a lap-top computer or a wirelessly networked PC.
PCT/EP2002/012323 2001-11-07 2002-11-04 Scalable video transmissions WO2003041382A2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
AU2002340500A AU2002340500A1 (en) 2001-11-07 2002-11-04 Scalable video transmissions

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
GB0126747A GB2381980A (en) 2001-11-07 2001-11-07 Error concealment in scalable video transmissions
GB0126747.5 2001-11-07

Publications (2)

Publication Number Publication Date
WO2003041382A2 true WO2003041382A2 (en) 2003-05-15
WO2003041382A3 WO2003041382A3 (en) 2003-09-25

Family

ID=9925341

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2002/012323 WO2003041382A2 (en) 2001-11-07 2002-11-04 Scalable video transmissions

Country Status (3)

Country Link
AU (1) AU2002340500A1 (en)
GB (2) GB2381980A (en)
WO (1) WO2003041382A2 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2005027520A1 (en) 2003-09-12 2005-03-24 Institute Of Computing Technology Chinese Academy Of Sciences Bi-directional predicting method for video coding/decoding
CN100396104C (en) * 2004-09-15 2008-06-18 乐金电子(惠州)有限公司 Method and device for processing image error data in block network
CN100433838C (en) * 2004-02-05 2008-11-12 三星电子株式会社 Decoding method and decoding apparatus
US7653136B2 (en) 2004-01-14 2010-01-26 Samsung Electronics Co., Ltd. Decoding method and decoding apparatus
WO2014117182A1 (en) * 2013-01-28 2014-07-31 Microsoft Corporation Conditional concealment of lost video data
US9237289B2 (en) 2011-10-25 2016-01-12 Microsoft Technology Licensing, Llc Estimating quality of a video signal

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7929776B2 (en) 2005-03-10 2011-04-19 Qualcomm, Incorporated Method and apparatus for error recovery using intra-slice resynchronization points

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020021761A1 (en) * 2000-07-11 2002-02-21 Ya-Qin Zhang Systems and methods with error resilience in enhancement layer bitstream of scalable video coding

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020021761A1 (en) * 2000-07-11 2002-02-21 Ya-Qin Zhang Systems and methods with error resilience in enhancement layer bitstream of scalable video coding

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
"OVERVIEW OF THE MPEG-4 STANDARD" ISO/IEC JT1/SC29/WG11 N3747, XX, XX, October 2000 (2000-10), pages 1-60, XP000925987 *
BUDAGAVI M ET AL: "Wireless MPEG-4 video communication on DSP chips" IEEE SIGNAL PROCESSING MAGAZINE, IEEE INC. NEW YORK, US, vol. 17, no. 1, January 2000 (2000-01), pages 36-53, XP002144403 ISSN: 1053-5888 *
VILLASENOR J D ET AL: "ROBUST VIDEO CODING ALGORITHMS AND SYSTEMS" PROCEEDINGS OF THE IEEE, IEEE. NEW YORK, US, vol. 87, no. 10, October 1999 (1999-10), pages 1724-1733, XP000927244 ISSN: 0018-9219 cited in the application *
WANG Y ET AL: "ERROR RESILIENT VIDEO CODING TECHNIQUES REAL-TIME VIDEO COMMUNICATIONS OVER UNRELIABLE NETWORKS" IEEE SIGNAL PROCESSING MAGAZINE, IEEE INC. NEW YORK, US, July 2000 (2000-07), pages 61-82, XP002951457 ISSN: 1053-5888 *
YAN R ET AL: "ERROR RESILIENCE METHODS IN THE FGS ENHANCEMENT BITSTREAM" ISO/IEC JTC1/SC29/WG11 MPEG00/M6207, XX, XX, July 2000 (2000-07), page COMPLETE XP001112952 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2005027520A1 (en) 2003-09-12 2005-03-24 Institute Of Computing Technology Chinese Academy Of Sciences Bi-directional predicting method for video coding/decoding
US7653136B2 (en) 2004-01-14 2010-01-26 Samsung Electronics Co., Ltd. Decoding method and decoding apparatus
CN100433838C (en) * 2004-02-05 2008-11-12 三星电子株式会社 Decoding method and decoding apparatus
CN100396104C (en) * 2004-09-15 2008-06-18 乐金电子(惠州)有限公司 Method and device for processing image error data in block network
US9237289B2 (en) 2011-10-25 2016-01-12 Microsoft Technology Licensing, Llc Estimating quality of a video signal
US10038898B2 (en) 2011-10-25 2018-07-31 Microsoft Technology Licensing, Llc Estimating quality of a video signal
WO2014117182A1 (en) * 2013-01-28 2014-07-31 Microsoft Corporation Conditional concealment of lost video data
US9609321B2 (en) 2013-01-28 2017-03-28 Microsoft Technology Licensing, Llc Conditional concealment of lost video data

Also Published As

Publication number Publication date
GB0126747D0 (en) 2002-01-02
GB0205102D0 (en) 2002-04-17
GB2381981A (en) 2003-05-14
AU2002340500A1 (en) 2003-05-19
GB2381980A (en) 2003-05-14
WO2003041382A3 (en) 2003-09-25

Similar Documents

Publication Publication Date Title
JP5007012B2 (en) Video encoding method
JP4362259B2 (en) Video encoding method
US20050163211A1 (en) Scalable video transmission
KR100878057B1 (en) Video coding
JP2004531925A (en) System and method for encoding and decoding redundant motion vectors in a compressed video bitstream
Kim et al. Multiple description motion coding algorithm for robust video transmission
WO2003041382A2 (en) Scalable video transmissions
Katsaggelos et al. Error resilience and concealment in video coding
WO2002019708A1 (en) Dual priority video transmission for mobile applications
Aladrovic et al. An error resilience scheme for layered video coding
Kaup Error concealment for SNR scalable video coding in wireless communication
WO2001015458A2 (en) Dual priority video transmission for mobile applications
Navarro et al. Significance analysis of MPEG-4 video syntactical elements

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ OM PH PL PT RO RU SD SE SG SI SK SL TJ TM TN TR TT TZ UA UG US UZ VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR IE IT LU MC NL PT SE SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
122 Ep: pct application non-entry in european phase
NENP Non-entry into the national phase in:

Ref country code: JP

WWW Wipo information: withdrawn in national office

Country of ref document: JP