WO1997010675A1 - Stereoscopic image encoding - Google Patents

Stereoscopic image encoding Download PDF

Info

Publication number
WO1997010675A1
WO1997010675A1 PCT/GB1996/002261 GB9602261W WO9710675A1 WO 1997010675 A1 WO1997010675 A1 WO 1997010675A1 GB 9602261 W GB9602261 W GB 9602261W WO 9710675 A1 WO9710675 A1 WO 9710675A1
Authority
WO
WIPO (PCT)
Prior art keywords
sub
image
images
redundancies
data
Prior art date
Application number
PCT/GB1996/002261
Other languages
French (fr)
Inventor
Malcolm Mccormick
Neil Davies
Amar Aggoun
Matthew Charles Forman
Original Assignee
De Montfort University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by De Montfort University filed Critical De Montfort University
Priority to DE69603740T priority Critical patent/DE69603740T2/en
Priority to EP96930270A priority patent/EP0850539B1/en
Priority to US09/043,393 priority patent/US6535629B2/en
Priority to AU69384/96A priority patent/AU6938496A/en
Priority to JP9511762A priority patent/JPH11512580A/en
Publication of WO1997010675A1 publication Critical patent/WO1997010675A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/207Image signal generators using stereoscopic image cameras using a single 2D image sensor
    • H04N13/225Image signal generators using stereoscopic image cameras using a single 2D image sensor using parallax barriers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/207Image signal generators using stereoscopic image cameras using a single 2D image sensor
    • H04N13/229Image signal generators using stereoscopic image cameras using a single 2D image sensor using lenticular lenses, e.g. arrangements of cylindrical lenses
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/207Image signal generators using stereoscopic image cameras using a single 2D image sensor
    • H04N13/232Image signal generators using stereoscopic image cameras using a single 2D image sensor using fly-eye lenses, e.g. arrangements of circular lenses
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/302Image reproducers for viewing without the aid of special glasses, i.e. using autostereoscopic displays
    • H04N13/305Image reproducers for viewing without the aid of special glasses, i.e. using autostereoscopic displays using lenticular lenses, e.g. arrangements of cylindrical lenses
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/302Image reproducers for viewing without the aid of special glasses, i.e. using autostereoscopic displays
    • H04N13/307Image reproducers for viewing without the aid of special glasses, i.e. using autostereoscopic displays using fly-eye lenses, e.g. arrangements of circular lenses
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/302Image reproducers for viewing without the aid of special glasses, i.e. using autostereoscopic displays
    • H04N13/31Image reproducers for viewing without the aid of special glasses, i.e. using autostereoscopic displays using parallax barriers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding

Definitions

  • This invention relates to storing and/or transmitting 3D images.
  • 3D images can be formed in a variety of ways. No matter how they are formed, there is substantially more information content in a 3D image than in a corresponding 2D image - the depth information is additional. Storing and/or transmitting 3D images therefore is more demanding of storage space or bandwidth than for 2D images, much as colour images are more demanding than monochrome images. Coloured 3D images would appear on the face of it to be very demanding, but the problems can be eased by data compression techniques, su ⁇ risingly to such a degree as brings 3D television into immediate prospect.
  • Imaging techniques produce images which are particularly well adapted, as it turns out, to compression, and the present invention is particularly concerned with such imaging techniques.
  • the invention comprises a method for storing and/or transmitting 3D image information comprising the steps of :
  • the image may be of a scene and produced using an optical imaging system comprising a microlens or lenticular array of small spherical or cylindrical lenses each of which images the scene from a slightly different viewpoint.
  • the image may however be electronically generated or partially electronically generated. Photographic images may be electronically scanned and captured as electronic data.
  • Small sub-image data sectors generated by the optical system are fed successively into an encoder where a previously fed sub-image is substracted from the most recently fed-in sub-image by a differential pulse code modulation (D P C M) coding technique to remove redundancies between the sub-images.
  • D P C M differential pulse code modulation
  • Redundancies may be eliminated within the sub-images themselves by techniques for example normally used in compression of two dimensional image data such for example as a discrete cosine transform (DCT) coding scheme.
  • DCT discrete cosine transform
  • a 3D - DCT coding scheme may be applied directly to groups of sub- images, the use of the third transform dimension eliminating inter-sub-group redundancies with the first two transform dimensions used to remove intra-sub-image redundancies.
  • a quantisation function may be applied to the coded data that sets small values to zero and transforms all other non-zero values to nearest values in a set of preferred values.
  • the coded data may then be entropy encoded.
  • the above coding schemes are suitable for compressing still image data.
  • a DPCM/3D-DCT coding scheme may be used, the DPCM coding decorrelating image data in the temporal domain and the 3D-DCT scheme eliminating spatial redundancies.
  • a hybrid DPCM2/DCT scheme may be used for compression of moving 3D image information, in which a 2D-DCT scheme decorrelates and hence removes redundancies within each sub-image and two DPCM loops are used, one to remove redundancies between sub-images in a spatial sense while the second is used to remove temporal (inter-frame) redundancies.
  • Both of these moving 3D-image compression schemes may make use of motion compensation to achieve greater overall image reduction.
  • Figure 1 is a diagrammatic illustration of an optical system casting an image on an electronic imaging device and display arrangements for images therefrom;
  • Figure 2 is a diagrammatic illustration of the imaging process in the optical system of Figure 1 ;
  • Figure 3 is a section of a lenticular - integral image
  • Figure 4 is a section of a full-integral image
  • Figure 5 is a diagrammatic illustration of a first coding scheme for still 3D image data compression
  • Figure 6 is a diagrammatic illustration of a second coding scheme for still 3D image data compression
  • Figure 7 is a diagrammatic illustration of a coding scheme for moving 3D image data compression
  • Figure 8 is a diagrammatic illustration of a scheme for an input structure for the coding schemes of Figures 5 to 7;
  • Figure 9 is a diagrammatic illustration of a scanning strategy used in the coding schemes of Figures 5 to 7.
  • the drawings illustrate methods for storing and/or transmitting (and, of course, displaying or replaying) 3D image information.
  • Figures 1 and 2 illustrate imaging a scene S to be stored and/or transmitted using an optical imaging system 1 1 comprising a microlens or lenticular array 12 of small spherical or lenticular (i.e. cylindrical) lenses each of which images the scene S from a slightly different viewpoint to produce an array of strongly correlated sub-images.
  • the optical imaging system 11 comprises a front-end optical arrangement 13 comprising a segmented wide aperture input lens 14, a microtelescopic array 15 (a double integral, autocollimating microlens array, see Figure 2), and a segmented output macrolens array 16.
  • each segment ofthe input lens array 14 individually transposes its image at the focusing screen 17 ofthe array 15 (see Figure 2).
  • the screen 17 comprises a double microlens screen.
  • Each segment of the output lens array 16 projects the transposed images to be supe ⁇ osed at a particular plane.
  • the initial transposition and projection will also produce a reversal of parallax between the individual supe ⁇ osed image fields.
  • As the initial transposed image is formed on the double integral microlens screen 17 it is presented to the output lenses as a spatially reversed 3D optical model.
  • the resulting supe ⁇ osed 3D image is constructed from the integration of all the spatially reversed optical models projected by each lens segment, and consequently continuity of parallax throughout the viewing angle is achieved.
  • the recorded image is a planar 2D image, which contains all the 3D information relating to the scene S.
  • This image cast on the microlens encoding screen 12, is reduced in size by a copy lens arrangement 19 to form a reduced image on an electronic image capture device such as a high resolution CCD array 21.
  • an electronic image capture device such as a high resolution CCD array 21.
  • the same can be achieved by imaging directly on to a high resolution CCD array which is overlayed with a microlens encoding system.
  • Figure 1 also illustrates image viewing arrangements, namely a flat panel display 22 and a projection display 23 for viewing the image captured by the device 21 as a 3D image.
  • the flat panel display 22 comprises a high density pixel screen 24 (which may be a liquid crystal display panel or a cathode ray tube or a gas plasma screen) with an adjacent lens array 25 which acts as a decoding screen to decode the coded information produced by the encoding screen 12.
  • the projection arrangement 23 comprises a high resolution video screen 26 and a projection lens arrangement 27 projecting the coded image on the screen 26 on to an integral back projection decoding screen 28 by which the observer sees a 3D image.
  • a light valve/LCD arrangement may be used in place of a video screen.
  • Figures 3 and 4 show sections (magnified) of coded images produced by arrangements such as that illustrated in Figures 1 and 2.
  • a lenticular arrangement can be used, using cylindrical lenses, the lenses being arranged with their axes vertical to correspond to the horizontal spacing ofthe eyes in binocular vision.
  • Figure 3 shows a 64 x 64 pixel section of a lenticular - integral image
  • Figure 4 shows six sub-images of a full-integral image each of 8 x 8 pixels.
  • the minimum bandwidth initially believed to be required for the transmission of moving integral 3D images was 42 GHz.
  • full colour 3D display is possible using a receiver with a bandwidth no greater than is required for HDTV.
  • a compression ratio of approximately 4: 1 is required for the transmission of integral 3D TV pictures.
  • Higher compression rates than 4:1 are possible, enabling higher quality display (more pixels per sub-image) and more efficient use of transmission bandwidth or storage space, i.e. allowing several TN picture channels to have the same broadcast channel.
  • the invention comprises compressing the electronic data produced by the CCD array 21 (or other electronic imaging device) by eliminating redundancies between these strongly correlated sub-images before storing or transmitting the compressed data.
  • the compression is reversible so as to expand the data to recreate the sub- images for viewing as a 3D image through an optical viewing system such as either ofthe systems A, B of Figure 1.
  • Figures 5 and 6 illustrate two still 3D image coding schemes, utilising two different decorrelation techniques namely :
  • small sub-image data sectors (such as the six sectors seen in Figure 4) generated by the optical system are fed successively into an encoder where a previously coded sub-image is subtracted from the most recently fed-in sub-image by a differential pulse code modulation (DPCM) technique, and
  • DPCM differential pulse code modulation
  • FIG. 5 which shows both encoder and decoder (as do Figures 6 and 7) illustrates a hybrid DPCM/DCT coding scheme in which sub-images are input (as electronic data representing colour/brightness values) into a loop 51 in which a previous sub-image is held in a store 52 to be fed to a subtraction unit 53 - the DPCM step. The resulting partially decorrelated sub-image is fed to a DCT stage 54 where intra-sub- image redundancies are eliminated.
  • DCT discrete cosine transform
  • the thus further decorrelated sub-image data then go to a quantiser 55 where all low value pixel values are set to zero and all others are reduced to the nearest one of a small set of discrete values.
  • the quantised data are then fed to an entropy encoder 56 that achieves further gain in compression by evaluating the statistics of the occurrence of non-zero coefficient values and representing the quantised data with regard to their statistical significance is such a way as to minimise output data quantity.
  • the quantised data are fed, in the loop 51, to a de-quantiser 57 that restores the pre-quantised values, then to an inverse DCT unit 58 that effectively restores the sub- image to be passed to the store 52.
  • the final coded sub-image data is output from the entropy coder 56.
  • the DCT coder 54 for intra - sub-image correlates applies the formula
  • f(l,m) is the input data array formed from an 8x8 sub-image difference
  • the decoder illustrated in Figure 5 comprises an entropy decoder 61 receiving the input coded sub-images and passing them to a de-quantiser 62 thence to an inverse DCT stage 63 and finally into a loop 64 with a sub-image store 65 for the inverse DPCM stage, the restored sub-image being output from the loop 64.
  • coder and decoder In between the coder and decoder, of course, will be a UHF transmission of the compressed data and/or a storage on e.g. magnetic or video disc recording medium.
  • Figure 6 shows a purely DCT encoding scheme comprising a 3D - DCT stage 66, quantiser 67 and entropy encoder 68 outputting the compressed data.
  • the decoder comprises the inverse elements, namely entropy decoder 71, de-quantiser 72 and 3D inverse DCT stage 73.
  • the 3D - DCT stage for four 8 x 8 pixel sub-images is
  • the third transform dimension takes account of inter-sub-image redundancies such that a small group of sub-images is completely decorrelated in a single transform calculation.
  • FIG. 5 illustrates a further coding scheme for moving integral 3D images.
  • a DPCM stage 74 has a substractor 75 and sub-image store 76, as before. This passes DPCM decorrelated sub-images to a hybrid DPCM/DCT coding arrangement 77 with a DCT coder 78, quantiser 79 and entropy coder 81 together with a DPCM loop 82 with dequantiser 83 and IDCT stage 84, sub-image store 85 and, additionally, a motion compensation stage 87.
  • DPCM is used in this scheme to decorrelate in the temporal and inter-sub-image domains, DCT to decorrelate intra-sub-image redundancies.
  • the decoder has an entropy decoder 87, a dequantiser 88 and IDCT stage 89 with two inverse DPCM loops 91, 92.
  • Figure 8 illustrates, for 3D-DCT-based schemes, a strategy for the extraction of a group of sub-images from an integral image, the sub-images I, II, III, IN being extracted one after the other and assembled (notionally, as frame data) in the order to be fed to the coding arrangement. Selection of groups of neighbouring sub-images maximises the inter-sub-image correlation, allowing decorrelation to reduce the information required to be transmitted to a minimum.
  • Figure 9 illustrates an entropy-coder scanning strategy for four 8 x 8 pixel transformed and quantised sub-images. Each plane ofthe processed sub-image group is scanned in turn over the u and v axis directions according to the scanning scheme illustrated in the right hand part of this figure. This zig-zag diagonal scanning statistically increases the zero run lengths leading to enhancement of compression by entropy coding.
  • moving 3D colour pictures can be transmitted within a standard UHF terrestrial TN bandwidth, the images being true colour, unlike holography, and the 3D imaging being integral, with no flipping over a wide viewing angle.
  • the system described is compatible with 2D television receivers inasmuch as without the optical decoding arrangements the picture is identical to what the comparable 2D picture would be but would appear slightly out of focus. This can be compensated for by an image enhancement technique to sha ⁇ en the focus.

Abstract

There is disclosed a method for storing and/or transmitting 3D image information comprising the steps of: producing an image to be stored and/or transmitted comprising an array of strongly correlated neighbouring sub-images; casting the sub-images on to a pixel screen capturing the sub-images as electronic data; compressing the electronic data by eliminating redundancies associated with the sub-images; storing and/or transmitting the compressed data; the compression being reversible so as to expand the data to re-create the sub-images for viewing as a 3D image through an optical viewing system comprising a microlens or lenticular array.

Description

STEREOSCOPIC IMAGE ENCODING
This invention relates to storing and/or transmitting 3D images.
3D images can be formed in a variety of ways. No matter how they are formed, there is substantially more information content in a 3D image than in a corresponding 2D image - the depth information is additional. Storing and/or transmitting 3D images therefore is more demanding of storage space or bandwidth than for 2D images, much as colour images are more demanding than monochrome images. Coloured 3D images would appear on the face of it to be very demanding, but the problems can be eased by data compression techniques, suφrisingly to such a degree as brings 3D television into immediate prospect.
Methods for making (and viewing) 3D images - autostereoscopic images, i.e. not requiring aids such as spectacles to view - are known and involve the use of an optical imaging system comprising a microlens array of small spherical or lenticular (i.e. cylindrical) lenses. Such imaging techniques produce images which are particularly well adapted, as it turns out, to compression, and the present invention is particularly concerned with such imaging techniques.
The invention comprises a method for storing and/or transmitting 3D image information comprising the steps of :
• producing an image to be stored and/or transmitted comprising an array of strongly correlated neighbouring sub-images;
• casting the sub-images on to a pixel screen capturing the sub-images as electronic data;
• compressing the electronic data by eliminating redundancies associated with the sub-images; storing and/or transmitting the compressed data; • the compression being reversible so as to expand the data to re-create the sub-images for viewing as a 3-D image through an optical viewing system comprising a microlens or lenticular array.
The image may be of a scene and produced using an optical imaging system comprising a microlens or lenticular array of small spherical or cylindrical lenses each of which images the scene from a slightly different viewpoint.
The image may however be electronically generated or partially electronically generated. Photographic images may be electronically scanned and captured as electronic data.
Small sub-image data sectors generated by the optical system are fed successively into an encoder where a previously fed sub-image is substracted from the most recently fed-in sub-image by a differential pulse code modulation (D P C M) coding technique to remove redundancies between the sub-images.
Redundancies may be eliminated within the sub-images themselves by techniques for example normally used in compression of two dimensional image data such for example as a discrete cosine transform (DCT) coding scheme.
A 3D - DCT coding scheme may be applied directly to groups of sub- images, the use of the third transform dimension eliminating inter-sub-group redundancies with the first two transform dimensions used to remove intra-sub-image redundancies. A quantisation function may be applied to the coded data that sets small values to zero and transforms all other non-zero values to nearest values in a set of preferred values.
The coded data may then be entropy encoded.
The above coding schemes are suitable for compressing still image data. For storing and/or transmitting moving 3D image information a DPCM/3D-DCT coding scheme may be used, the DPCM coding decorrelating image data in the temporal domain and the 3D-DCT scheme eliminating spatial redundancies.
A hybrid DPCM2/DCT scheme may be used for compression of moving 3D image information, in which a 2D-DCT scheme decorrelates and hence removes redundancies within each sub-image and two DPCM loops are used, one to remove redundancies between sub-images in a spatial sense while the second is used to remove temporal (inter-frame) redundancies.
Both of these moving 3D-image compression schemes may make use of motion compensation to achieve greater overall image reduction.
Methods for storing and/or transmitting 3D image information according to the invention will now be described with reference to the accompanying drawings, in which :
Figure 1 is a diagrammatic illustration of an optical system casting an image on an electronic imaging device and display arrangements for images therefrom; Figure 2 is a diagrammatic illustration of the imaging process in the optical system of Figure 1 ;
Figure 3 is a section of a lenticular - integral image;
Figure 4 is a section of a full-integral image;
Figure 5 is a diagrammatic illustration of a first coding scheme for still 3D image data compression;
Figure 6 is a diagrammatic illustration of a second coding scheme for still 3D image data compression;
Figure 7 is a diagrammatic illustration of a coding scheme for moving 3D image data compression;
Figure 8 is a diagrammatic illustration of a scheme for an input structure for the coding schemes of Figures 5 to 7; and
Figure 9 is a diagrammatic illustration of a scanning strategy used in the coding schemes of Figures 5 to 7.
The drawings illustrate methods for storing and/or transmitting (and, of course, displaying or replaying) 3D image information.
Figures 1 and 2 illustrate imaging a scene S to be stored and/or transmitted using an optical imaging system 1 1 comprising a microlens or lenticular array 12 of small spherical or lenticular (i.e. cylindrical) lenses each of which images the scene S from a slightly different viewpoint to produce an array of strongly correlated sub-images. The optical imaging system 11 comprises a front-end optical arrangement 13 comprising a segmented wide aperture input lens 14, a microtelescopic array 15 (a double integral, autocollimating microlens array, see Figure 2), and a segmented output macrolens array 16.
As seen in Figure 2, each segment ofthe input lens array 14 individually transposes its image at the focusing screen 17 ofthe array 15 (see Figure 2). The screen 17 comprises a double microlens screen. Each segment of the output lens array 16 projects the transposed images to be supeφosed at a particular plane. The initial transposition and projection will also produce a reversal of parallax between the individual supeφosed image fields. As the initial transposed image is formed on the double integral microlens screen 17 it is presented to the output lenses as a spatially reversed 3D optical model. The resulting supeφosed 3D image is constructed from the integration of all the spatially reversed optical models projected by each lens segment, and consequently continuity of parallax throughout the viewing angle is achieved.
The recorded image is a planar 2D image, which contains all the 3D information relating to the scene S. This image, cast on the microlens encoding screen 12, is reduced in size by a copy lens arrangement 19 to form a reduced image on an electronic image capture device such as a high resolution CCD array 21. The same can be achieved by imaging directly on to a high resolution CCD array which is overlayed with a microlens encoding system.
Figure 1 also illustrates image viewing arrangements, namely a flat panel display 22 and a projection display 23 for viewing the image captured by the device 21 as a 3D image. The flat panel display 22 comprises a high density pixel screen 24 (which may be a liquid crystal display panel or a cathode ray tube or a gas plasma screen) with an adjacent lens array 25 which acts as a decoding screen to decode the coded information produced by the encoding screen 12. The projection arrangement 23 comprises a high resolution video screen 26 and a projection lens arrangement 27 projecting the coded image on the screen 26 on to an integral back projection decoding screen 28 by which the observer sees a 3D image. A light valve/LCD arrangement may be used in place of a video screen.
Figures 3 and 4 show sections (magnified) of coded images produced by arrangements such as that illustrated in Figures 1 and 2. Instead of an integral coding arrangement, using spherical microlenses, a lenticular arrangement can be used, using cylindrical lenses, the lenses being arranged with their axes vertical to correspond to the horizontal spacing ofthe eyes in binocular vision. Figure 3 shows a 64 x 64 pixel section of a lenticular - integral image, while Figure 4 shows six sub-images of a full-integral image each of 8 x 8 pixels.
It is with the recording, transmission, storage, reception, retrieval and display of such coded images as are seen in Figures 3 and 4 that this invention is concerned.
The minimum bandwidth initially believed to be required for the transmission of moving integral 3D images was 42 GHz. In fact, full colour 3D display is possible using a receiver with a bandwidth no greater than is required for HDTV. A compression ratio of approximately 4: 1 is required for the transmission of integral 3D TV pictures. Higher compression rates than 4:1 are possible, enabling higher quality display (more pixels per sub-image) and more efficient use of transmission bandwidth or storage space, i.e. allowing several TN picture channels to have the same broadcast channel.
Conventional compression algorithms for HDTN, such as transform and sub-band coding techniques, achieve compression by decorrelating in the spatial and/or temporal correlation domains. A totally white screen, of course, is totally correlated. while a picture with areas of more or less solid colour is less, but still fairly well correlated spatially. Successive frames of a television transmission are usually very well correlated temporally, even with high-action scenarios, and there is substantial scope for compression of ordinary 2D colour picture information on all those accounts.
The addition of the third spatial dimension would appear to pose substantial problems, but the invention overcomes the perceived difficulties by working with the strongly correlated - as will be evident from Figures 3 and 4 - sub-images, i.e. portions of the full image produced by the optical system described with reference to Figures 1 and 2.
The invention comprises compressing the electronic data produced by the CCD array 21 (or other electronic imaging device) by eliminating redundancies between these strongly correlated sub-images before storing or transmitting the compressed data.
The compression, further according to the invention, is reversible so as to expand the data to recreate the sub- images for viewing as a 3D image through an optical viewing system such as either ofthe systems A, B of Figure 1.
Figures 5 and 6 illustrate two still 3D image coding schemes, utilising two different decorrelation techniques namely :
• small sub-image data sectors (such as the six sectors seen in Figure 4) generated by the optical system are fed successively into an encoder where a previously coded sub-image is subtracted from the most recently fed-in sub-image by a differential pulse code modulation (DPCM) technique, and
• intra - sub-image redundancies are eliminated using a discrete cosine transform (DCT) technique. Figure 5, which shows both encoder and decoder (as do Figures 6 and 7) illustrates a hybrid DPCM/DCT coding scheme in which sub-images are input (as electronic data representing colour/brightness values) into a loop 51 in which a previous sub-image is held in a store 52 to be fed to a subtraction unit 53 - the DPCM step. The resulting partially decorrelated sub-image is fed to a DCT stage 54 where intra-sub- image redundancies are eliminated.
The thus further decorrelated sub-image data then go to a quantiser 55 where all low value pixel values are set to zero and all others are reduced to the nearest one of a small set of discrete values. The quantised data are then fed to an entropy encoder 56 that achieves further gain in compression by evaluating the statistics of the occurrence of non-zero coefficient values and representing the quantised data with regard to their statistical significance is such a way as to minimise output data quantity.
The quantised data are fed, in the loop 51, to a de-quantiser 57 that restores the pre-quantised values, then to an inverse DCT unit 58 that effectively restores the sub- image to be passed to the store 52.
The final coded sub-image data is output from the entropy coder 56.
The DCT coder 54 for intra - sub-image correlates applies the formula
Figure imgf000010_0001
where f(l,m) is the input data array formed from an 8x8 sub-image difference;
F (u,v) is the resulting transform coefficient array; and
=l if s = o, γ 2 if S > 0 The decoder illustrated in Figure 5 comprises an entropy decoder 61 receiving the input coded sub-images and passing them to a de-quantiser 62 thence to an inverse DCT stage 63 and finally into a loop 64 with a sub-image store 65 for the inverse DPCM stage, the restored sub-image being output from the loop 64.
In between the coder and decoder, of course, will be a UHF transmission of the compressed data and/or a storage on e.g. magnetic or video disc recording medium.
Figure 6 shows a purely DCT encoding scheme comprising a 3D - DCT stage 66, quantiser 67 and entropy encoder 68 outputting the compressed data. The decoder comprises the inverse elements, namely entropy decoder 71, de-quantiser 72 and 3D inverse DCT stage 73.
The 3D - DCT stage for four 8 x 8 pixel sub-images is
F ττ <(u,v,w -)s = D»D —. —D~ ^ γ Σ - ∑A / ,(,-, ,>* λ) cos π u ( 5-—2m + l ) ^ cos π v ( --—2n + l ) -- cos π w ( *--2/ + l ) '- l /„o m-0 7 -0 l θ l O δ
The third transform dimension takes account of inter-sub-image redundancies such that a small group of sub-images is completely decorrelated in a single transform calculation.
The arrangements discussed with reference to Figures 5 and 6 are suitable for data compression for still pictures. An arrangement for compression of moving integral 3D-TN pictures is essentially the same as the arrangement of Figure 5 except that the 2D DCT stage 54 is replaced by a 3D DCT stage and the 2D IDCT stages 58 and 63 are replaced by 3D IDCT stages. Figure 7 illustrates a further coding scheme for moving integral 3D images.
A DPCM stage 74 has a substractor 75 and sub-image store 76, as before. This passes DPCM decorrelated sub-images to a hybrid DPCM/DCT coding arrangement 77 with a DCT coder 78, quantiser 79 and entropy coder 81 together with a DPCM loop 82 with dequantiser 83 and IDCT stage 84, sub-image store 85 and, additionally, a motion compensation stage 87. DPCM is used in this scheme to decorrelate in the temporal and inter-sub-image domains, DCT to decorrelate intra-sub-image redundancies.
The decoder has an entropy decoder 87, a dequantiser 88 and IDCT stage 89 with two inverse DPCM loops 91, 92.
Figure 8 illustrates, for 3D-DCT-based schemes, a strategy for the extraction of a group of sub-images from an integral image, the sub-images I, II, III, IN being extracted one after the other and assembled (notionally, as frame data) in the order to be fed to the coding arrangement. Selection of groups of neighbouring sub-images maximises the inter-sub-image correlation, allowing decorrelation to reduce the information required to be transmitted to a minimum.
Figure 9 illustrates an entropy-coder scanning strategy for four 8 x 8 pixel transformed and quantised sub-images. Each plane ofthe processed sub-image group is scanned in turn over the u and v axis directions according to the scanning scheme illustrated in the right hand part of this figure. This zig-zag diagonal scanning statistically increases the zero run lengths leading to enhancement of compression by entropy coding.
Using the techniques described and illustrated, moving 3D colour pictures can be transmitted within a standard UHF terrestrial TN bandwidth, the images being true colour, unlike holography, and the 3D imaging being integral, with no flipping over a wide viewing angle.
The system described is compatible with 2D television receivers inasmuch as without the optical decoding arrangements the picture is identical to what the comparable 2D picture would be but would appear slightly out of focus. This can be compensated for by an image enhancement technique to shaφen the focus.

Claims

1. A method for storing and/or transmitting 3D image information comprising the steps of :
• producing an image to be stored and/or transmitted comprising an array of strongly correlated neighbouring sub-images;
• casting the sub-images on to a pixel screen capturing the sub-images as electronic data;
• compressing the electronic data by eliminating redundancies associated with the sub-images;
• storing and/or transmitting the compressed data;
• the compression being reversible so as to expand the data to re-create the sub-images for viewing as a 3-D image through an optical viewing system comprising a microlens or lenticular array.
2. A method according to claim 1 in which the image is of a scene produced using an optical imaging system comprising a microlens or lenticular array of small spherical or cylindrical lenses each of which images the scene from a slightly different viewpoint.
3. A method according to claim 1 or claim 2, in which the image is electronically generated or partially electronically generated.
4. A method according to any one of claims 1 to 3, in which photographic images are electronically scanned and captured as electronic data.
5. A method according to any one of claims 1 to 4, in which redundancies between sub-images are eliminated.
6. A method according to any one of claims 1 to 5, in which redundancies within sub-images are eliminated.
7. A method according to any one of claims 1 to 5, in which small sub-image data sectors generated by the optical system are fed successively into an encoder where a previously coded sub-image is subtracted from the most-recently fed-in sub-image by a differential pulse code modulation (DPCM) coding technique.
8. A method according to claim 7, in which redundancies are eliminated within the sub- images themselves.
9. A method according to claim 8, in which the intra-sub-image redundancies are eliminated using a discrete cosine transform (DCT) coding scheme.
10. A method according to any one of claims 1 to 9, in which a 3D-DCT coding scheme is applied directly to groups of sub-images, the third dimension eliminating inter-sub-image redundancies.
11. A method according to any one of claims 1 to 10, in which a quantisation function is applied to the coded data that sets small values to zero and transforms all other non-zero values to nearest values in a set of preferred values.
12. A method according to any one of claims 1 to 1 l,in which the coded data are further entropy encoded.
13. A method according to any one of claims 1 to 12, for storing and/or transmitting moving 3D image information using a DPCM coding technique decorrelating image data in the temporal domain and 3D-DCT eliminating spatial redundancies.
14. A method according to any one of claims 1 to 13, in which a 2D-DCT scheme correlates intra-sub-image spatial redundancies and a DPCM technique decorrelates inter-sub-image data both in the spatial and temporal domains.
PCT/GB1996/002261 1995-09-16 1996-09-13 Stereoscopic image encoding WO1997010675A1 (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
DE69603740T DE69603740T2 (en) 1995-09-16 1996-09-13 STEREOSCOPIC IMAGE CODING
EP96930270A EP0850539B1 (en) 1995-09-16 1996-09-13 Stereoscopic image encoding
US09/043,393 US6535629B2 (en) 1995-09-16 1996-09-13 Stereoscopic image encoding
AU69384/96A AU6938496A (en) 1995-09-16 1996-09-13 Stereoscopic image encoding
JP9511762A JPH11512580A (en) 1995-09-16 1996-09-13 3D image coding

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
GB9518984.1 1995-09-16
GBGB9518984.1A GB9518984D0 (en) 1995-09-16 1995-09-16 Storing and/or transmitting 3d images

Publications (1)

Publication Number Publication Date
WO1997010675A1 true WO1997010675A1 (en) 1997-03-20

Family

ID=10780836

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/GB1996/002261 WO1997010675A1 (en) 1995-09-16 1996-09-13 Stereoscopic image encoding

Country Status (8)

Country Link
US (1) US6535629B2 (en)
EP (1) EP0850539B1 (en)
JP (1) JPH11512580A (en)
AT (1) ATE183353T1 (en)
AU (1) AU6938496A (en)
DE (1) DE69603740T2 (en)
GB (1) GB9518984D0 (en)
WO (1) WO1997010675A1 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE19949011A1 (en) * 1999-10-11 2001-04-12 Werner Breit Camera and display screen unit has detection units in form of individual hollow acquisition bodies and screen unit with correspondingly arranged individual hollow reproduction bodies
EP1154655A2 (en) 2000-05-12 2001-11-14 Lg Electronics Inc. Apparatus and method for displaying three-dimensional image
WO2002017647A1 (en) * 2000-08-23 2002-02-28 Demontfort University Autostereoscopic display
WO2002017646A1 (en) * 2000-08-23 2002-02-28 Demontfort University Three-dimensional imaging system
WO2004077838A1 (en) * 2003-02-27 2004-09-10 T-Mobile Deutschland Gmbh Method for the compressed transmission of image data for three-dimensional representation of scenes and objects
US10432944B2 (en) 2017-08-23 2019-10-01 Avalon Holographics Inc. Layered scene decomposition CODEC system and methods

Families Citing this family (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU5529299A (en) * 1999-05-19 2000-12-12 Lenslet, Ltd. Image compression
US7515753B2 (en) 1999-05-19 2009-04-07 Lenslet Labs Ltd. Phase extraction in optical processing
US6643386B1 (en) * 2000-08-10 2003-11-04 Omnivision Technologies, Inc. Method and apparatus for adding watermarks to images and/or video data streams
US8103877B2 (en) 2000-12-21 2012-01-24 Digimarc Corporation Content identification and electronic tickets, coupons and credits
US20060256436A1 (en) * 2002-01-23 2006-11-16 The University Of Connecticut Integral three-dimensional imaging with digital reconstruction
JP2006510953A (en) * 2002-12-13 2006-03-30 ヘンツェ−リソチェンコ パテントフェルヴァルトゥングス ゲーエムベーハー ウント コー.カーゲー Device for acquiring and reproducing 3D image information of an object
US7426597B1 (en) * 2003-05-07 2008-09-16 Nvidia Corporation Apparatus, system, and method for bus link width optimization of a graphics system
US7469311B1 (en) 2003-05-07 2008-12-23 Nvidia Corporation Asymmetrical bus
WO2005076985A2 (en) 2004-02-04 2005-08-25 Digimarc Corporation Digital watermarking image signals on-chip and photographic travel logs through digital watermarking
US20060238545A1 (en) * 2005-02-17 2006-10-26 Bakin Dmitry V High-resolution autostereoscopic display and method for displaying three-dimensional images
ES2602091T3 (en) * 2005-06-23 2017-02-17 Koninklijke Philips N.V. Combined exchange of image and related data
TW200823595A (en) * 2006-11-28 2008-06-01 Univ Nat Taiwan Image capture device using programmable aperture
CN102343526B (en) * 2011-06-28 2013-09-04 天津汽车模具股份有限公司 Method for quickly determining machining center of automobile die cast
US9106894B1 (en) * 2012-02-07 2015-08-11 Google Inc. Detection of 3-D videos
FR2989805A1 (en) * 2012-04-19 2013-10-25 France Telecom METHOD FOR ENCODING AND DECODING INTEGRAL IMAGES, DEVICE FOR ENCODING AND DECODING INTEGRAL IMAGES, AND CORRESPONDING COMPUTER PROGRAMS

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0305274A1 (en) * 1987-08-20 1989-03-01 Le Particulier Editions Sa Method and arrangement for generating stereoscopic images
US4999705A (en) * 1990-05-03 1991-03-12 At&T Bell Laboratories Three dimensional motion compensated video coding
EP0588410A1 (en) * 1992-09-14 1994-03-23 Koninklijke KPN N.V. System comprising a first encoder for coding a first digital signal and a second encoder for coding a second digital signal (e.g. stereoscopic video signal)
GB2271240A (en) * 1992-09-30 1994-04-06 Fujitsu Ltd Stereoscopic image information transmission system

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2774738B2 (en) * 1992-05-27 1998-07-09 シャープ株式会社 Image coding restoration system
KR100342410B1 (en) * 1993-05-05 2002-11-18 삐에르 알리오 Automatic Stereoscopic Projection Device
JP3361643B2 (en) * 1995-02-14 2003-01-07 富士通株式会社 Image processing system
US5675377A (en) * 1995-06-30 1997-10-07 Telefonaktiebolaget Lm Ericsson True three-dimensional imaging and display system

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0305274A1 (en) * 1987-08-20 1989-03-01 Le Particulier Editions Sa Method and arrangement for generating stereoscopic images
US4999705A (en) * 1990-05-03 1991-03-12 At&T Bell Laboratories Three dimensional motion compensated video coding
EP0588410A1 (en) * 1992-09-14 1994-03-23 Koninklijke KPN N.V. System comprising a first encoder for coding a first digital signal and a second encoder for coding a second digital signal (e.g. stereoscopic video signal)
GB2271240A (en) * 1992-09-30 1994-04-06 Fujitsu Ltd Stereoscopic image information transmission system

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
DINSTEIN ET AL.: "On Stereo Image Coding", 9TH INT. CONF. ON PATTEN RECOGNITION, 14 November 1988 (1988-11-14), ROMA, ITALY, pages 357 - 359, XP002021653 *
FORMAN ET AL.: "Compression of Integral 3D TV Pictures", 5TH INT. CONF. ON IMAGE PROCESSING AND ITS APPLICATIONS, 4 July 1995 (1995-07-04), HERIOT-WATT UNIVERSITY, UK, pages 584 - 588, XP000613545 *
YAMAGUCHI ET AL.: "Data Compression and Depth Shape Reproduction of Stereoscopic Images", SYSTEMS & COMPUTERS IN JAPAN, vol. 22, no. 12, 1991, NEW YORK US, pages 53 - 64, XP000262424 *
ZIEGLER: "Digital Stereoscopic Imaging & Applications. A Way towards New Dimensions. The RACE II Project DISTIMA", IEE COLLOQUIUM ON STEREOSCOPIC TELEVISION, 15 October 1992 (1992-10-15), LONDON, UK, pages 6/1 - 6/4, XP002021654 *

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE19949011A1 (en) * 1999-10-11 2001-04-12 Werner Breit Camera and display screen unit has detection units in form of individual hollow acquisition bodies and screen unit with correspondingly arranged individual hollow reproduction bodies
DE19949011C2 (en) * 1999-10-11 2001-10-25 Werner Breit Passage of light waves
EP1154655A2 (en) 2000-05-12 2001-11-14 Lg Electronics Inc. Apparatus and method for displaying three-dimensional image
WO2002017647A1 (en) * 2000-08-23 2002-02-28 Demontfort University Autostereoscopic display
WO2002017646A1 (en) * 2000-08-23 2002-02-28 Demontfort University Three-dimensional imaging system
WO2004077838A1 (en) * 2003-02-27 2004-09-10 T-Mobile Deutschland Gmbh Method for the compressed transmission of image data for three-dimensional representation of scenes and objects
US7212662B2 (en) 2003-02-27 2007-05-01 T-Mobile Deutschland Gmbh Method for the compressed transmission of image data for 3-dimensional representation of scenes and objects
US10432944B2 (en) 2017-08-23 2019-10-01 Avalon Holographics Inc. Layered scene decomposition CODEC system and methods
US10972737B2 (en) 2017-08-23 2021-04-06 Avalon Holographics Inc. Layered scene decomposition CODEC system and methods

Also Published As

Publication number Publication date
GB9518984D0 (en) 1995-11-15
JPH11512580A (en) 1999-10-26
DE69603740T2 (en) 2000-02-10
ATE183353T1 (en) 1999-08-15
EP0850539A1 (en) 1998-07-01
US6535629B2 (en) 2003-03-18
US20020054705A1 (en) 2002-05-09
AU6938496A (en) 1997-04-01
EP0850539B1 (en) 1999-08-11
DE69603740D1 (en) 1999-09-16

Similar Documents

Publication Publication Date Title
US6535629B2 (en) Stereoscopic image encoding
US10528004B2 (en) Methods and apparatus for full parallax light field display systems
US7916934B2 (en) Method and system for acquiring, encoding, decoding and displaying 3D light fields
US8044994B2 (en) Method and system for decoding and displaying 3D light fields
US10244223B2 (en) Methods for full parallax compressed light field 3D imaging systems
KR101177663B1 (en) Method and system for digital decoding 3d stereoscopic video images
US6055012A (en) Digital multi-view video compression with complexity and compatibility constraints
US4884131A (en) Stereoscopic television system with predictive coding to minimize data bandwidth
US20080205791A1 (en) Methods and systems for use in 3d video generation, storage and compression
KR19990045067A (en) Visual Offset Estimation for Stereoscopic Video Coding
WO2009011492A1 (en) Method and apparatus for encoding and decoding stereoscopic image format including both information of base view image and information of additional view image
Conti et al. Spatial prediction based on self-similarity compensation for 3D holoscopic image and video coding
US20110268193A1 (en) Encoding and decoding method for single-view video or multi-view video and apparatus thereof
JP4825983B2 (en) Image information compression method and free viewpoint television system
WO2020006035A1 (en) Random access in encoded full parallax light field images
Fecker et al. H. 264/AVC-compatible coding of dynamic light fields using transposed picture ordering
Dick et al. 3D holoscopic video coding using MVC
Forman et al. Quantisation strategies for 3D-DCT-based compression of full parallax 3D images
Aggoun et al. Live immerse video-audio interactive multimedia
Monteiro et al. Optimized reference picture selection for light field image coding
Forman et al. A novel coding scheme for full parallax 3D-TV pictures
Brewin et al. Electronic capture and display of full-parallax 3D images
Sethuraman Stereoscopic image sequence compression using multiresolution and quadtree decomposition-based disparity-and motion-adaptive segmentation
Forman et al. Compression of integral 3D TV pictures
Sgouros et al. Compression of IP images for autostereoscopic 3D imaging applications

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AL AM AT AU AZ BB BG BR BY CA CH CN CZ DE DK EE ES FI GB GE HU IL IS JP KE KG KP KR KZ LK LR LS LT LU LV MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK TJ TM TR TT UA UG US UZ VN AM AZ BY KG KZ MD RU TJ TM

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): KE LS MW SD SZ UG AT BE CH DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA

DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
121 Ep: the epo has been informed by wipo that ep was designated in this application
WWE Wipo information: entry into national phase

Ref document number: 1996930270

Country of ref document: EP

ENP Entry into the national phase

Ref country code: JP

Ref document number: 1997 511762

Kind code of ref document: A

Format of ref document f/p: F

WWE Wipo information: entry into national phase

Ref document number: 09043393

Country of ref document: US

WWP Wipo information: published in national office

Ref document number: 1996930270

Country of ref document: EP

REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

NENP Non-entry into the national phase

Ref country code: CA

WWG Wipo information: grant in national office

Ref document number: 1996930270

Country of ref document: EP