USRE43203E1 - Computation time reduction for the three-dimensional displays - Google Patents

Computation time reduction for the three-dimensional displays Download PDF

Info

Publication number
USRE43203E1
USRE43203E1 US12/129,941 US12994101A USRE43203E US RE43203 E1 USRE43203 E1 US RE43203E1 US 12994101 A US12994101 A US 12994101A US RE43203 E USRE43203 E US RE43203E
Authority
US
United States
Prior art keywords
regions
sub
observer
display means
display
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US12/129,941
Inventor
Douglas A. Levenets
Paul M. Blanchard
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Mind Fusion LLC
Original Assignee
F Poszat HU LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to US12/129,941 priority Critical patent/USRE43203E1/en
Application filed by F Poszat HU LLC filed Critical F Poszat HU LLC
Assigned to HOLOGRAPHIC IMAGING LLC reassignment HOLOGRAPHIC IMAGING LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BLANCHARD, PAUL MICHAEL, PAYNE, DOUGLAS ALAN
Assigned to QINETIQ LIMITED reassignment QINETIQ LIMITED ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HOLOGRAPHIC IMAGING LLC
Assigned to F. POSZAT HU, LLC reassignment F. POSZAT HU, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: QINETIQ LIMITED COMPANY
Assigned to HOLOGRAPHIC IMAGING LLC reassignment HOLOGRAPHIC IMAGING LLC DECLARATION TO CORRECT THE INVENTOR NAME CHANGE PREVIOUSLY RECORDED ON REEL 026815 FRAME 0277. ASSIGNORS HEREBY CONFIRMS THE ASSIGNMENT OF ASSIGNORS INTEREST. Assignors: LEVENETS, DOUGLAS ALAN
Publication of USRE43203E1 publication Critical patent/USRE43203E1/en
Application granted granted Critical
Adjusted expiration legal-status Critical
Assigned to INTELLECTUAL VENTURES ASSETS 191 LLC reassignment INTELLECTUAL VENTURES ASSETS 191 LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: F. POSZAT HU, L.L.C.
Assigned to INTELLECTUAL VENTURES ASSETS 186 LLC, INTELLECTUAL VENTURES ASSETS 191 LLC reassignment INTELLECTUAL VENTURES ASSETS 186 LLC SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MIND FUSION, LLC
Assigned to MIND FUSION, LLC reassignment MIND FUSION, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: INTELLECTUAL VENTURES ASSETS 191 LLC
Expired - Fee Related legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03HHOLOGRAPHIC PROCESSES OR APPARATUS
    • G03H1/00Holographic processes or apparatus using light, infrared or ultraviolet waves for obtaining holograms or for obtaining an image from them; Details peculiar thereto
    • G03H1/04Processes or apparatus for producing holograms
    • G03H1/08Synthesising holograms, i.e. holograms synthesized from objects or objects from holograms
    • G03H1/0808Methods of numerical synthesis, e.g. coherent ray tracing [CRT], diffraction specific
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B30/00Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images
    • G02B30/20Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes
    • G02B30/26Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes of the autostereoscopic type
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03HHOLOGRAPHIC PROCESSES OR APPARATUS
    • G03H1/00Holographic processes or apparatus using light, infrared or ultraviolet waves for obtaining holograms or for obtaining an image from them; Details peculiar thereto
    • G03H1/22Processes or apparatus for obtaining an optical image from holograms
    • G03H1/2294Addressing the hologram to an active spatial light modulator
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03HHOLOGRAPHIC PROCESSES OR APPARATUS
    • G03H1/00Holographic processes or apparatus using light, infrared or ultraviolet waves for obtaining holograms or for obtaining an image from them; Details peculiar thereto
    • G03H1/22Processes or apparatus for obtaining an optical image from holograms
    • G03H1/2202Reconstruction geometries or arrangements
    • G03H1/2205Reconstruction geometries or arrangements using downstream optical component
    • G03H2001/221Element having optical power, e.g. field lens
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03HHOLOGRAPHIC PROCESSES OR APPARATUS
    • G03H1/00Holographic processes or apparatus using light, infrared or ultraviolet waves for obtaining holograms or for obtaining an image from them; Details peculiar thereto
    • G03H1/22Processes or apparatus for obtaining an optical image from holograms
    • G03H1/2202Reconstruction geometries or arrangements
    • G03H2001/2236Details of the viewing window
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03HHOLOGRAPHIC PROCESSES OR APPARATUS
    • G03H1/00Holographic processes or apparatus using light, infrared or ultraviolet waves for obtaining holograms or for obtaining an image from them; Details peculiar thereto
    • G03H1/22Processes or apparatus for obtaining an optical image from holograms
    • G03H1/2202Reconstruction geometries or arrangements
    • G03H2001/2236Details of the viewing window
    • G03H2001/2242Multiple viewing windows
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03HHOLOGRAPHIC PROCESSES OR APPARATUS
    • G03H2210/00Object characteristics
    • G03H2210/303D object
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03HHOLOGRAPHIC PROCESSES OR APPARATUS
    • G03H2210/00Object characteristics
    • G03H2210/40Synthetic representation, i.e. digital or optical object decomposition
    • G03H2210/45Representation of the decomposed object
    • G03H2210/452Representation of the decomposed object into points
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03HHOLOGRAPHIC PROCESSES OR APPARATUS
    • G03H2226/00Electro-optic or electronic components relating to digital holography
    • G03H2226/02Computing or processing means, e.g. digital signal processor [DSP]
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03HHOLOGRAPHIC PROCESSES OR APPARATUS
    • G03H2226/00Electro-optic or electronic components relating to digital holography
    • G03H2226/05Means for tracking the observer

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Optics & Photonics (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
  • Length Measuring Devices By Optical Means (AREA)
  • Liquid Crystal (AREA)
  • Solid-Sorbent Or Filter-Aiding Compositions (AREA)
  • Magnetic Resonance Imaging Apparatus (AREA)

Abstract

A reconfigurable, three-dimensional display (1) wherein knowledge of the viewer's (4) eyes is used to enable the effective exit pupil(s) of the display system to be optimised. The system utilises this knowledge to identify contributing regions (5) within the display (1) that contribute light to the viewer (4). Priority is given to calculating and displaying the part of the display corresponding to the contributing region (5), thereby allowing the system computation requirements to be minimised. Further computation savings are achievable by recognising that only light travelling in a limited range of angles need to be considered.

Description

This application is the US national phase of international application PCT/GB01/02302, filed in English on 24 May 2001, which designated the US. PCT/GB01/02302 claims priority to GB Application No. 0013942.8 filed 9 Jun. 2000. The entire contents of these applications are incorporated herein by reference.
The present invention relates to reconfigurable three-dimensional displays and particularly to displays utilising Computer Generated Holograms (CGH). The invention provides a means of minimising the computation time required to generate a Computer Generated Hologram (CGH).
There is significant interest in producing reconfigurable three-dimensional displays for use in applications such as interactive design, medical imaging and scientific visualisation. For some applications it is desirable that the display can be viewed without any special glasses and by multiple viewers and that the image contains as many of the three-dimensional depth cues as possible. Display methods with these attributes include Computer Generated Holograms (CGH) and auto-stereo systems incorporating lenticular lens sheets or microlens arrays. These methods typically make use of a two-dimensional pixelated display device (transmissive or reflective) such as a liquid crystal display in which the amplitude and/or phase of each pixel (picture element) is used to encode the required pixel value. The pixel values are calculated such that when the device is appropriately illuminated and suitable replay optics are used, a three-dimensional image of a simulated object is obtained. Alternatively, the liquid crystal display may be replaced by a non-pixelated spatial light modulator such as an acousto-optic modulator. In this configuration small regions within the spatial light modulator (SLM) perform the same function as the individual pixels in the liquid crystal display.
Despite advances in computer technology and optical modulation techniques, the generation of near real-time, reconfigurable, high quality, three-dimensional images continues to present major technical challenges. For example, reconfigurable display devices for three-dimensional images remain complex and in the case of CGHs computationally intensive, placing heavy processing demands on underlying computer systems.
For a display system based on a CGH, the complexity of the display device, in terms of the number of constituent picture elements or pixels required, is determined both by the required image size and angle of view (the angle over which the image can be seen by a viewer). Particularly for multi-viewer systems, which tend to need large viewing angles, this results in a need for large numbers of pixels. (e.g. 1012 pixels for a workstation application) and hence very complex display devices. In terms of displaying Computer Generated Holograms, the computation of the pixel values for such a large number of pixels in real-time or even at an interactive rate is also a major challenge, as is the method of transferring data to the display.
In parallel with the aforementioned advances in computer technology, several techniques have been developed for reducing the full system requirements of the display system in an attempt to produce a viable reconfigurable three-dimensional display.
One conventional technique for reducing the requirements of the display system is to incorporate moveable exit pupil(s) within the system. The exit pupil is a region through which all light rays that can be traced all the way through an optical system must pass and is a familiar concept to the skilled person working in the field of optics. In a conventional optical system the exit pupil would typically be a physical aperture or an image of a physical aperture.
A. Schwartze, “Head tracking stereoscopic display”, IEEE International Display Research Conference, 1985 describes a stereoscopic display system which presents two two-dimensional perspectives to the viewer. The system uses two CRT projection modules in conjunction with a simple head position sensor and electromechanically controlled moveable exit pupils. The viewer's head is tracked to ensure the exit pupil for the appropriate view is located over the correct eye. The system accommodates lateral movements made by the observer and allows the user to view a stereoscopic presentation without using special glasses.
Since the system senses the lateral position of the observer relative to the display, it is possible to generate and display different stereoscopic views according to observer position, thus producing the look-around effect of a real object. The full system requirements are therefore reduced since the system is not required to simultaneously produce images for all possible viewing angles, only the images which correspond to the observer's particular lateral position. Thus the computational requirement for the system is also reduced.
Fukaya et. al. [“Eye-position tracking type electro-holographic display using liquid crystal devices”, N. Fukaya, T. Honda, K. Maeno and K. Sato, Proceedings of EOS Topical meeting on Diffractive Optics, 1997] describe an electroholographic display system incorporating eye-position tracking.
In the system proposed by Fukaya et. al., two small LCDs display computer generated holograms. Being small, the three-dimensional image from each liquid crystal display has only a small exit pupil (small field-of-view). The LCDs are positioned relative to each other such that their exit pupils coincide with each of the observer's eyes. A scanning mirror is used to ensure the exit pupils follow the observer's eyes during lateral movements. The computed CGH contains far more pixels than the display modulator can support, and essentially only the part of this data required by the observer is put on the modulator, dictated by the observer's head position. By displaying only part of CGH, the full system requirements are again reduced.
The computational load on the underlying processing system for both of these displays is relatively low, however, both techniques suffer from several inherent disadvantages.
Firstly, in both of the aforementioned systems the exit pupil(s) is scanned by mechanical means to follow the viewer's position. The complexity of the display system is thereby increased over a conventional CRT or LCD display. Furthermore, a disquieting time lag may be introduced due to the tracking control system.
Secondly, the shape and size of the exit pupil(s) for the above systems are limited and only the lateral positions of the exit pupil(s) can be controlled. Furthermore, the displays, as described, are restricted to single observers and are incapable of producing an enlarged exit pupil (field-of-view) to cater for multiple viewers or to reduce any lag in image appearance when the viewer makes rapid head movements.
These two techniques essentially use a low pixel count display devices (with relatively low computational requirements) and increase the field of view through mechanical scanning. The present invention relates to using a high pixel count display with a large field of view that doesn't suffer from the above disadvantages. Such displays would normally have high computational requirements. It is an object of the present invention to enable these requirements to be significantly reduced.
In order to clarify the significance of the exit pupil in such 3D displays the following analogy is provided. Consider viewing an object a short distance behind a window in an opaque screen. The range of positions from which the object can be seen (e.g. to look around its sides) is limited by the size of the window. In the display system under consideration the exit pupil defines the ‘window’ in this analogy and the image provided by the display is the ‘object’. An interpretation of the two previously described display systems is that the ‘window’, or exit pupil, is moved to always be in line with the observer and the ‘object’, or image.
For a given observer position, the ‘window’ can be smaller (without clipping the observer's view of the object) the closer it is to the observer. The present invention can be considered equivalent to enabling the display to produce an image that can only be seen through such a small, adaptable ‘window’. Just as a ‘large window’ (determined by the optics in the system) through which an image can be seen from a wide range of viewing positions is properly termed the ‘exit pupil’, we introduce the term ‘effective exit pupil’ to represent the additional small, adaptable ‘window’. The present invention enables a small effective exit pupil to be positioned near to the viewer(s). The effective exit pupil is ‘moved’ to follow viewer(s) such that the image can be seen from the same range of viewing positions that the exit pupil would allow. This is achieved by appropriate calculation of the display pixel values. The viewers eye positions need to be obtained using some form of head tracking device.
It will be shown that the computation time required to calculate the pixel values in the display is reduced as the size of the above effective exit pupil is reduced.
Unlike the movable exit pupils of the previous examples, the positioning of the effective exit pupil(s) requires no moving parts in the system of the present invention, potentially reducing costs and increasing reliability. Further, the system of the present invention allows the effective exit pupil(s) to be of arbitrary shape and arbitrarily positioned in all three dimensions, improving ease of use. Multiple viewers are also catered for with the present system.
Additionaly, the present invention allows the effective exit pupil(s) or field-of-view of the system to be enlarged when no changes occur in the image. This has the potential to allow the elimination of lag in image appearance that may occur when the user makes rapid head movements.
According to the present invention apparatus for producing a three-dimensional image comprises;
a display means capable of producing a three dimensional image which is capable of being viewed from a range of viewing positions,
a monitoring means for obtaining information about the at least one observer,
and control means responsive to the monitoring means for controlling the display means,
characterised in that the display means comprises a plurality of addressable sub-regions controlled by the control means such that priority is given to producing a complete image for the at least one determined viewing position.
The apparatus provides a viable display by giving priority to only those parts of the image that may be seen by the at least one observer at the at least one determined viewing position. This enables a reduction in the time required to compute the pixel values in the display means thus allowing the image to be updated at a higher frequency. Since no moving parts are introduced, the apparatus remains cost-effective and reliability is enhanced.
The eye positions and pupil sizes of the viewer's eyes are required by the control means for the purpose of identifying the contributing regions in the display means. The eye positions should be obtained by the monitoring means. The pupil sizes may be obtained by the monitoring means or may be assumed to be a typical value such as 4 mm diameter.
Preferably the viewing positions of two or more observers may be determined and priority given to producing a complete image for each of the determined viewing positions. Multiple observers are therefore catered for by the present apparatus.
In a preferred embodiment the control means identifies at least one contributing region, comprising addressable sub-regions of the display means, that contributes to the image formed for the at least one viewing position and gives priority to calculating pixel values of the display means for said at least one contributing region. Parts of the image seen by the observer may be generated by separate areas within the contributing region, whilst separate contributing regions within the display-means may be envisaged for multiple viewers.
In a further preferred embodiment the control means gives priority to calculating pixel values for each of the sub-regions of the display means that comprise substantially the centre of the at least one said contributing region. This is advantageous in that the image appears to be built up from its centre.
In another embodiment the apparatus further comprises means for determining the position of ocular fixation within the display means of the at least one observer. In this embodiment the control means gives priority to calculating pixel values for each of the sub-regions of the display means that contribute substantially to the image at the position of ocular fixation of the at least one observer.
For the purposes of this specification, the position of ocular fixation within the display means shall be defined as the observer's point of fixation on the image. The technique of monitoring an observers point of fixation on an image is commonly known as gaze-tracking and is equivalent to detecting the position of the fovea of the observer's eyes. The information relating to the observer's point of fixation on the image is used advantageously to cause the image to be built up from the point of interest.
In a further preferred embodiment the control means gives priority to calculating pixel values for each of the sub-regions of the display means that correspond substantially to parts of the image that are changing. This approach increases the rate of interaction between an observer and the image being displayed. Usability of the apparatus is generally enhanced by according priority to parts of the image that are changing.
In another embodiment the control means determines the range of angles that sub-regions of the display means must direct light into to contribute to the image formed for the at least one viewing position and the pixel values of the display means are calculated such that priority is given to directing light into said range of angles.
In a further preferred embodiment the control means controls the display means such that the effective exit pupil of the apparatus is optimised for the at least one viewer.
The effective exit pupil of the apparatus is a region through which all the light rays considered in the calculation of the pixel values will pass. The apparatus capitalises on the fact that no image degradation will be perceived provided the observer's eye pupils are located within the effective exit pupil of the display apparatus.
Advantageously the time required to calculate and display the image may be minimised by optimising the effective exit pupil of the apparatus to coincide with the pupils of the observer's eyes. The apparatus allows the effective exit pupil to be of arbitrary shape and size and to be arbitrarily positioned in all three-dimensions. Hence, the apparatus exhibits a high degree of flexibility, allowing the shape, size and position of the effective exit pupil be adjusted.
The effective exit pupil over which the complete image is viewable may be enlarged during periods when the image is substantially unchanging. This allows the observer to make rapid head movements without perceiving any lag in image replay.
The display means may comprise a spatial light modulator (SLM) means and the control means may be used to calculate the modulation required in the sub-regions of the spatial light modulator means.
In a further preferred embodiment the monitoring means collects information relating to the eyes of the at least one observer and the control means uses said information to adapt the effective exit pupil of the apparatus to match the eyes of the at least one observer.
In a particular embodiment the addressable sub-regions of the display means display a computer generated hologram.
According to a second aspect of the present invention, a method of minimising the time required to display a reconfigurable three-dimensional image using a display means comprising a plurality of addressable sub-regions, capable of producing a three dimensional image which is capable of being viewed from a range of viewing positions,
a monitoring means for determining the viewing position of at least one observer,
and control means responsive to the monitoring means for controlling the display means,
characterised by the steps of determining the viewing position of the at least one observer and prioritising the control of the addressable sub-regions within the display means to produce a complete image for the at least one determined viewing position.
Preferably the viewing positions of two or more observers may be determined and the control of the addressable sub-regions within the display means prioritised to produce a complete image for the two or more determined viewing positions.
In a preferred embodiment at least one contributing region is identified, comprising addressable sub-regions of the display means, that contributes to the image formed for the at least one viewing position and priority is given to calculating pixel values for the sub-regions within the display means that comprise substantially the at least one said contributing region.
In a further preferred embodiment priority is given to calculating pixel values for each of the sub-regions of the display means that comprise substantially the centre of the at least one said contributing region.
In another embodiment the method further comprises the steps of determining the position of ocular fixation within the display means of the at least one observer and giving priority to calculating pixel values for each of the sub-regions of the display means that contribute substantially to the image at the position of ocular fixation of the at least one observer.
In a further embodiment priority is given to calculating pixel values for each of the sub-regions of the display means that correspond substantially to parts of the image that are changing.
In a another embodiment the method further comprises the steps of determining the range of angles that sub-regions of the display means must direct light into to contribute substantially to the image formed for the at least one viewing position and calculating the pixel values of the display means such that priority is given to directing light into substantially said range of angles.
In a another embodiment the method further comprises the step of controlling the display means such that the effective exit pupil of the apparatus is optimised for the at least one viewer.
Preferably the step of enlarging the effective exit pupil over which the complete image is viewable may be incorporated during periods when the image is substantially unchanging.
In one embodiment the display means comprises a spatial light modulator (SLM) means and the control means calculates the modulation required in the sub-regions of the spatial light modulator means.
In a further preferred embodiment the monitoring means collects information relating to the eyes of the at least one observer and the control means uses said information to adapt the effective exit pupil of the apparatus to match the eyes of the at least one observer.
In a particular embodiment the addressable sub-regions of the display means display a computer generated hologram.
The invention will now be described, by example only, with reference to the accompanying drawings in which;
FIG. 1 shows a schematic representation of a reconfigurable three-dimensional display system illustrating how only a small region of the display contributes light to the three-dimensional image that reaches the viewer's eyes.
FIG. 2 illustrates a schematic cross-section of a CGH display showing how ray tracing may be used to determine the limited range of angles that each pixel in the CGH contributes light into that reaches the viewer's eyes.
FIG. 3 illustrates a schematic cross-section of a CGH display showing how collimated beams arising from the CGH give rise to points in the focal plane of the replay optics.
Conventional autostereoscopic and holographic three-dimensional displays contain far more information than a single viewer requires because the display simultaneously provides views for every potential viewing position, whether a viewer is there or not. Such displays have very large numbers of pixels, placing significant demands on the computation system used to calculate their values.
The full system requirements for a such a reconfigurable three-dimensional display may therefore be reduced by calculating only the pixel values in the display device required to provide a three-dimensional image to a particular viewer, by making use of knowledge of the viewer's eye positions. This knowledge can be obtained using head/eye tracking systems that have previously been demonstrated.
The present invention uses the knowledge of the viewer's eye positions to enable an effective exit pupil(s) of the display system to be optimised. The system simultaneously controls the shape, size and position of the effective exit pupil(s) of the display so as to coincide with the viewer's eye pupils. However, unlike some display systems which use electromechanical means to move an exit pupil to follow the viewer's eyes, the present invention relies purely on calculating the appropriate pixel transmission values in the display device to control the effective exit pupil of the system.
No moving parts are introduced into the system, providing a cost effective solution and maintaining reliability. Moreover, since the effective exit pupil of the system is ‘moveable’ by recomputing the display pixel values, there is no mechanical inertia in the display system to overcome. Accordingly, the system enables the effective exit pupil of the display to be rapidly repositioned both laterally and vertically in sympathy with the movements of the viewer.
The display system exhibits a high degree of flexibility and may be adapted in real time to cater for multiple concurrent viewers by altering the shape, size and position of the effective exit pupil(s) of the system.
Although the concept of controlling the effective exit pupil(s) of the system provides a concise way of explaining the present invention, the means by which the invention could be practically implemented would involve identifying contributing regions in the display and the angles into which those regions need to contribute light. Hence the following discussion describes the invention from this perspective.
The present invention utilises the fact that for a given viewer location, only a small region of the display contributes light to the image seen, known as ‘the contributing region’. Calculation and population of the pixels within this region alone will provide the viewer with a three-dimensional image of the same quality as if all of the pixels were populated. To take account of the viewer's two eyes, or the additional eyes of multiple viewers, multiple ‘contributing regions’ could be utilised.
For the purpose of the of the following discussion the display shall be described in terms of Computer Generated Hologram (CGH) based systems although the principle is also applicable to other display systems. Furthermore, descriptions of display devices in terms of two-dimensional pixelated liquid crystal displays (transmissive or reflective) are illustrative and are not meant to be limiting. For example, a liquid crystal display may be replaced by a non-pixelated spatial light modulator such as an acousto-optic modulator. In this configuration small regions within the spatial light modulator (SLM) perform the same function as the individual pixels in the liquid crystal display.
For clarity, the following embodiments have been described in terms of a single viewer, although in practice the technique is applicable to three-dimensional displays with multiple viewers.
Referring to FIG. 1, light passes through the CGH (1) and the replay optics (2) to form the three-dimensional image (3). Only a small proportion of this light will pass through the viewer's eye pupils (4), determined by the eye monitoring means (9). This light comes from the contributing region (5) and priority would be given within the controlling means (8) to calculating these pixels within the contributing region first.
The monitoring means (9) monitors the position of the viewer with respect to the CGH (1) and provides position data (for example the co-ordinates of the viewer in x, y and z axes) to the control means (8). The monitoring means (9) may also be used to obtain supplementary information regarding the eye positions and pupil sizes of the viewer's eyes. Furthermore, the viewer's point of fixation on the image produced by the CGH (1) may also be determined by the monitoring means (9). This is equivalent to detecting the position of the fovea of the viewer's eyes. For the purposes of this specification the viewer's point of fixation shall also be referred to as the point of ocular fixation.
The above mentioned supplementary information is communicated to the control means (8) and is used by the control means (8) to determine the optimum contributing region.
The control means (8) utilises the position data from the monitoring means (9) in conjunction with parameters relating to the display system (for example the size and configuration of the display element and optical components within the system) to calculate pixel values for the CGH (1). The control means (8) may include computing means to calculate the pixel values for the CGH (1).
The pixel values calculated by the control means (8) govern the amplitude and/or phase of each pixel within the CGH (1) such that when the CGH (1) is appropriately illuminated, a three-dimensional image (3) is produced by the relay optics (2).
The computational requirement for the above system is reduced since only the pixels within the contributing region need to be calculated in order to present an image to the viewer. However, the image quality remains the same as if all of the pixels in the three-dimensional display were populated. Since the number of pixels within the contributing region is less than the number of pixels in the full display, the time required to calculate and display the image is correspondingly reduced.
The above computational savings may be offset against other system parameters to allow other system functions to be improved; for example enabling the displayed image to be updated at a faster rate, thereby enhancing interaction between the viewer and the display system.
Calculation of the pixel values within the contributing region (5) can be further prioritised such that the most important pixels within the contributing region (5) are calculated first. This will reduce the time taken to present the viewer with an acceptable image. Two approaches are proposed to identify the pixels of most importance.
Firstly, the pixels in the centre (6) of the contributing region (5) may be calculated first. This will cause the image to be built up from the centre (6) of the contributing region.
Alternatively, in addition to knowledge of the viewers eye positions, the viewer's point of ocular fixation on the image may also be measured. This is commonly known as gaze tracking and is equivalent to detecting the position of the fovea of the viewer's eyes. Priority may therefore be given to calculating the pixels that correspond to the viewers point of fixation on the image (7). This will cause the image to be built up from the point of interest (7) within the contributing region (5).
It is envisaged that calculation of pixels outside of the contributing region will be carried out when the image is unchanging, effectively ‘filling out’ the range of potential viewing positions and thus allowing rapid head or viewer motion with no lag in image appearance.
A second level of computational saving is also proposed, which will still provide the viewer with a three-dimensional image of the same quality as if all of the pixels were populated. This is achieved by considering the limited range of angles that pixels within the contributing regions need to direct light into. FIG. 2 shows how these restricted angular ranges may be identified by tracing rays from the CGH, through the replay optics and three-dimensional image and finally through the pupils of the viewer's eyes.
FIG. 2 shows a cross-section of a schematic of a CGH based display system. Light passes through the CGH (20), the replay optics (21) and forms the three-dimensional image around the focal plane (22). The three-dimensional image will typically be contained within a diamond shaped (polyhedral in three-dimensions) volume (23). An eye with a pupil is shown (24). Only light arising from the contributing region (25) reaches the eye pupil. The extremities of the contributing region are defined by the intersections of the light rays (26) that pass through the extremities of the image and the eye pupil. It can be seen that only a limited angular cone of light (28) arising from a single pixel (27) in the CGH reaches the eye pupil. Another feature of CGH based displays is the viewing region, the bounds of which are shown by dashed lines (29). This is the region in which the eye needs to be located in order to see the whole image.
To illustrate how the method may be used to achieve computational savings, consider the design of a CGH using the ‘Coherent Ray Trace’ method.
The desired three-dimensional object would be defined mathematically, perhaps with computer aided design software. Its surface would be uniformly sampled at a number of points. Light rays would be traced from every one of these points to every pixel in the CGH. The required transmission of each CGH pixel is calculated from the coherent addition of all these light rays.
If the viewer's eye positions and pupil sizes are known then only those rays that could be extrapolated back to the eye pupils would need to be traced. This enables the number of rays that need to be traced to be reduced, thus saving computation time.
If only the viewers eye positions are known, then some sensible value for the pupil size may be assumed e.g. a diameter of 4 mm. Using a larger value for the pupil size in the CGH calculation will make the system more tolerant of eye position measurement errors although at the cost of increased computation time.
A rule-of-thumb formula can be derived to estimate the computation savings expected:
If the eye is within the viewing region (29) it can be shown that all light rays from the three-dimensional image that reach the pupil either pass through the focal plane within the image volume or could be extrapolated back to it. Thus, to a good approximation, the maximum number of ray traces will be required for an image that consists of points filling the entire focal plane.
It is very convenient to use this image as points in the focal plane of the replay optics arise from collimated beams from the CGH. FIG. 3 illustrates this phenomenon. As before, light rays can be traced from the CGH (30), through the replay optics (31) to the viewer's eye pupil (33). Rays passing through the same point (34) In in the focal plane (32) can be shown to intersect the CGH at the same angle over a small area (36) of the CGH. Thus each point in the focal plane corresponds to a collimated beam (35) at a different angle in the region between the CGH and replay optics. The area on the CGH (and thus the number of CGH pixels within it) that contributes light at a particular angle is proportional to the area of the eye pupil and is independent of the distance of the CGH from the optics, although its position will vary. This area is approximately the same for all ‘angles’ of the light rays considered. Since the focal plane is completely filled it can be said that an area of the CGH equivalent to this area directs light into all these angles.
Thus, using a CGH that is designed to direct light into all angles over its whole area as a baseline, simple geometrical arguments show that the reduction, R, in the number of ‘traces’ is
R area  of  collimated  beam area  of  CGH = area  of  pupil area  of  CGH × ( f e ) 2
and it is assumed that the pupil is approximately parallel to the lens.
If a CGH based three-dimensional display of a 50×50 cm image with a field of view of ±15°, pupil diameter 4 mm and d=e=f=1 m, then of order 104 times less traces will be needed. If the eye positions are only known approximately then the savings will be less although still useful. For example, where the eye positions are only known to be within a circular region 10 cm in diameter at the above distance, then the reduction in the number of traces would be around 30 times.
Computation savings may also be achieved with other CGH design methods. For example, the diffraction specific method (M. Lucente, “Holographic bandwidth compression using spatial subsampling”, Optical Engineering, Vol. 35, No. 6, June 1996). In this case the bandwidth (range of spatial frequencies) of the fringes encoded into each hogel (holographic element) would be determined by the limited range of angles that each hogel is required to direct light into for a given viewer position.

Claims (41)

The invention claimed is:
1. Apparatus for producing a three-dimensional image, said apparatus An apparatus comprising;:
a display means for producing a three dimensional image, said image capable of being configured to be viewed from a range of viewing positions, wherein said the display means having has an exit pupil of a size determined by values assigned to pixels in a plurality of sub-regions of the display means;
a monitoring means for determining a viewing position of at least one an observer,; and
control means, responsive to the monitoring means, for controlling the display means, wherein the control means provides computational priority to the determined viewing position by computing pixel values in the sub-regions such that a complete image is first provided for said the determined viewing position and, as a consequence, reducing the size of the exit pupil.
2. The apparatus according to claim 1, wherein the viewing positions of two or more observers are determined and wherein priority is given to producing a complete image for each of the determined viewing positions.
3. The apparatus according to claim 1 or 2, wherein the control means identifies at least one a contributing region, comprising addressable sub-regions of the display means, that contributes to the image formed for the at least one viewing position and wherein priority is given to calculating pixel values of the display means for said at least one the contributing region.
4. The apparatus according to claim 3, wherein the control means gives priority to calculating pixel values for each of the sub-regions of the display means that comprise substantially the centre center of the at least one said contributing region.
5. The apparatus according to claim 3 and, further comprising means for determining the position of ocular fixation within the display means of the at least one observer and, wherein the control means gives priority to calculating pixel values for each of the sub-regions of the display means that contribute substantially to the image at the position of ocular fixation of the at least one observer.
6. The apparatus according to any of the preceding claims claim 1, wherein the control means gives priority to calculating pixel values for each of the sub-regions of the display means that correspond substantially to parts of the image that are changing.
7. The apparatus according to claim 1, wherein the control means determines the a range of angles that sub-regions of the display means must direct light into to contribute to the image formed for the at least one viewing position, and wherein priority is given to calculating pixel values of the display means such that light is substantially directed into said the range of angles.
8. The apparatus according to claim 1, wherein the control means controls the display means such that the effective exit pupil of the apparatus is optimised optimized for the at least one viewer observer.
9. The apparatus according to claim 1, wherein the effective exit pupil over which the complete image is viewable is enlarged during periods when the image is substantially unchanging.
10. The apparatus according to claim 1, wherein the display means comprises a spatial light modulator (SLM) means.
11. The apparatus according to claim 10, wherein the control means calculates the modulation required in the sub-regions of the spatial light modulator means.
12. The apparatus according to claim 1, wherein the monitoring means collects information relating to the eyes of the at least one observer, and wherein said the information is used by the control means to adapt the effective exit pupil of the apparatus to match the eyes of the at least one observer.
13. The apparatus according to claim 1 3, wherein the addressable sub-regions of the display means display a computer generated hologram.
14. A method of, comprising:
reducing the a time required to display a reconfigurable three-dimensional image using a display means comprising a plurality of addressable sub-regions, capable of producing a configured to produce the three dimensional image which is capable of being viewed so as to be viewable from a range of viewing positions, a monitoring means for determining the viewing position of at least one observer, and control means responsive to the monitoring means for controlling the display means (1), wherein said method is comprised by the steps of:;
determining the a viewing position of the at least one an observer using a monitoring means; and
prioritising theresponsive to determining the viewing position of the observer, prioritizing control of the addressable sub-regions within the display means to first produce a complete image for the determined viewing position, thereby reducing the an initial computation of pixel values.
15. The method according to claim 14, wherein the viewing positions of two or more observers are determined, and wherein the control of the addressable sub-regions within the display means is prioritised prioritized to produce a the complete image for the two or more determined viewing positions of the two or more observers.
16. The method according to claims claim 14 or 15 and, further comprising the steps of:
identifying at least one a contributing region, comprising the addressable sub-regions of the display means, that contributes to the image formed for the at least one viewing position; and
giving priority to calculating pixel values for the sub-regions within the display means that comprise substantially the at least one said contributing region.
17. The method according to claim 16, wherein priority is given to calculating pixel values for each of the sub-regions of the display means that comprise substantially the centre center of the at least one said contributing region.
18. The method according to claim 16, further comprising the steps of:
determining the a position of ocular fixation of the observer within the display means of the at least one observer; and
giving priority to calculating pixel values for each of the sub-regions of the display means that contribute substantially to the image at the position of ocular fixation of the at least one observer.
19. The method according to claim 14, wherein priority is given to calculating pixel values for each of the sub-regions of the display means that correspond substantially to parts of the image that are changing.
20. The method according to claim 14, further comprising the steps of:
determining the a range of angles that sub-regions of the display means must direct light into to contribute substantially to the image formed for the at least one viewing position; and
calculating the pixel values of the display means such that priority is given to directing light into substantially said the range of angles.
21. The method according to claim 14, further comprising the step of controlling the display means such that the an effective exit pupil of the apparatus is optimised optimized for the at least one viewer observer.
22. The method according to claim 14, further comprising the step of enlarging the an effective exit pupil over which the complete image is viewable during periods when the image is substantially unchanging.
23. The method according to claim 14, wherein the display means comprises a spatial light modulator (SLM) means.
24. The method according to claim 23, wherein the control means calculates the an amount of modulation required in the sub-regions of the spatial light modulator means.
25. The method according to claim 14, wherein the monitoring means collects information relating to the eyes of the at least one observer, and wherein said the information is used by the control means to adapt the an effective exit pupil of the apparatus to match the one or more eyes of the at least one observer.
26. The method according to claim 14, wherein the addressable sub-regions of the display means display a computer generated hologram.
27. A system, comprising:
a display configured to generate a three-dimensional image, wherein the display is comprised of a plurality of sub-regions;
a gaze tracking device configured to monitor an ocular fixation of an observer; and
a controller configured to calculate pixel values for the plurality of sub-regions to determine a viewing angle corresponding to the ocular fixation of the observer, wherein the calculation of the pixel values is prioritized according to which of the plurality of sub-regions include pixels that are visible from within the viewing angle.
28. The system according to claim 27, wherein the calculation of the pixel values is further prioritized according to which of the plurality of sub-regions are located at an approximate center point of the ocular fixation.
29. The system according to claim 27, wherein the gaze tracking device is further configured to monitor a second ocular fixation of a second observer.
30. The system according to claim 29, wherein the viewing angle corresponds to both the first and second ocular fixations.
31. The system according to claim 27, wherein the pixels contribute to the generation of the three-dimensional image.
32. The system according to claim 27, wherein pixel values associated with pixels that are not visible from within the viewing angle are calculated after pixels values that are associated with pixels that are visible from within the viewing angle.
33. A method, comprising:
monitoring an ocular point of fixation of an observer viewing a three-dimensional image;
determining a viewing angle corresponding to the ocular point of fixation;
identifying sub-regions of a display configured to produce the three-dimensional image in the viewing angle; and
calculating pixel values for the sub-regions, wherein the calculation of the pixel values is prioritized according to the identification of the sub-regions which produce the three-dimensional image.
34. The method according to claim 33, further comprising:
monitoring a change in the ocular point of fixation of the observer to determine a new viewing angle; and
recalculating the pixel values to update the three-dimensional image in the new viewing angle.
35. The method according to claim 34, further comprising:
identifying a plurality of sub-regions of the display configured to produce the updated three-dimensional image in the new viewing angle, wherein the plurality of sub-regions includes at least one sub-region that is different than the sub-regions of the display which are configured to produce the three-dimensional image.
36. The method according to claim 33, further comprising:
identifying a change in the pixel values; and
recalculating the pixel values to update the three-dimensional image, wherein the pixel values that change receive a higher priority for calculation.
37. The method according to claim 33, further comprising:
identifying an ocular point of fixation of a second observer viewing the three-dimensional image; and
determining a new viewing angle corresponding to the ocular points of fixation for both the observer and the second observer.
38. The method according to claim 37, wherein the new viewing angle is larger than the viewing angle.
39. The method according to claim 33, wherein the ocular point of fixation is monitored by detecting a position of a fovea of the observer.
40. The method according to claim 33, wherein the ocular point of fixation is positioned at an approximate ocular center point of the observer.
41. The method according to claim 33, wherein the calculation of the pixel values is prioritized according to which sub-regions include pixels that are visible from within the viewing angle.
US12/129,941 2000-06-09 2001-05-24 Computation time reduction for the three-dimensional displays Expired - Fee Related USRE43203E1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/129,941 USRE43203E1 (en) 2000-06-09 2001-05-24 Computation time reduction for the three-dimensional displays

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
GB0013942A GB2363273A (en) 2000-06-09 2000-06-09 Computation time reduction for three dimensional displays
PCT/GB2001/002302 WO2001095016A1 (en) 2000-06-09 2001-05-24 Computation time reduction for three-dimensional displays
US12/129,941 USRE43203E1 (en) 2000-06-09 2001-05-24 Computation time reduction for the three-dimensional displays
US10/297,487 US7053925B2 (en) 2000-06-09 2001-05-24 Computation time reduction for the three-dimensional displays

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US10/297,487 Reissue US7053925B2 (en) 2000-06-09 2001-05-24 Computation time reduction for the three-dimensional displays

Publications (1)

Publication Number Publication Date
USRE43203E1 true USRE43203E1 (en) 2012-02-21

Family

ID=9893215

Family Applications (2)

Application Number Title Priority Date Filing Date
US12/129,941 Expired - Fee Related USRE43203E1 (en) 2000-06-09 2001-05-24 Computation time reduction for the three-dimensional displays
US10/297,487 Ceased US7053925B2 (en) 2000-06-09 2001-05-24 Computation time reduction for the three-dimensional displays

Family Applications After (1)

Application Number Title Priority Date Filing Date
US10/297,487 Ceased US7053925B2 (en) 2000-06-09 2001-05-24 Computation time reduction for the three-dimensional displays

Country Status (9)

Country Link
US (2) USRE43203E1 (en)
EP (1) EP1287400B8 (en)
JP (1) JP4717320B2 (en)
AT (1) ATE377773T1 (en)
AU (1) AU2001262486A1 (en)
CA (1) CA2411837A1 (en)
DE (1) DE60131274T2 (en)
GB (1) GB2363273A (en)
WO (1) WO2001095016A1 (en)

Families Citing this family (37)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS6357663A (en) * 1986-08-29 1988-03-12 Asahi Chem Ind Co Ltd Gasket for syringe, having excellent heat resistance
US6953249B1 (en) 2001-01-29 2005-10-11 Maguire Jr Francis J Method and devices for displaying images for viewing with varying accommodation
EP1563346B1 (en) * 2002-11-13 2009-09-02 SeeReal Technologies GmbH Device for reconstructing video holograms
JP4406824B2 (en) * 2003-08-21 2010-02-03 ソニー株式会社 Image display device, pixel data acquisition method, and program for executing the method
JP4845386B2 (en) * 2004-02-09 2011-12-28 株式会社エヌ・ティ・ティ・ドコモ Stereoscopic image display apparatus and stereoscopic image display method
DE102004063838A1 (en) * 2004-12-23 2006-07-06 Seereal Technologies Gmbh Method and apparatus for calculating computer generated video holograms
DE102005010250B4 (en) * 2005-02-25 2007-02-01 Seereal Technologies Gmbh Method and device for tracking sweet spots
EP1915874A2 (en) 2005-08-17 2008-04-30 SeeReal Technologies GmbH Method and circuit arrangement for recognising and tracking eyes of several observers in real time
US8049941B2 (en) * 2005-12-22 2011-11-01 Seereal Technologies S.A. Method for the compensation of an inhomogeneous brightness perception in holographically reconstructed scenes
DE102006004300A1 (en) 2006-01-20 2007-08-02 Seereal Technologies S.A. Projection device for holographic reconstruction of scenes, comprises reproduction medium, which is used to reproduce Fourier transformation of light from light source modulated by light modulation device onto screen
US8120828B2 (en) 2006-05-12 2012-02-21 Seereal Technologies S.A. Reflective optical system, tracking system and holographic projection system and method
DE102006043297B4 (en) * 2006-09-14 2010-12-09 Seereal Technologies S.A. A playback apparatus and method comprising means for tracking a viewer window
DE102007005822A1 (en) * 2007-01-31 2008-08-07 Seereal Technologies S.A. Holographic reconstruction system with optical wave tracking
DE102007005823A1 (en) * 2007-01-31 2008-08-07 Seereal Technologies S.A. Optical wavefront correction for a holographic projection system
GB0718607D0 (en) * 2007-05-16 2007-10-31 Seereal Technologies Sa Holograms
DE102007025069B4 (en) 2007-05-21 2018-05-24 Seereal Technologies S.A. Holographic reconstruction system
DE102007036127A1 (en) * 2007-07-27 2009-01-29 Seereal Technologies S.A. Holographic reconstruction device
DE102008000116A1 (en) * 2008-01-21 2009-07-30 Seereal Technologies S.A. Illumination unit for a holographic reconstruction system
DE102008000438A1 (en) * 2008-02-28 2009-09-10 Seereal Technologies S.A. Controllable deflection device
DE102008001644B4 (en) * 2008-05-08 2010-03-04 Seereal Technologies S.A. Device for displaying three-dimensional images
KR101595104B1 (en) 2008-07-10 2016-02-17 리얼 뷰 이미징 리미티드 Broad viewing angle displays and user interfaces
US8284234B2 (en) 2009-03-20 2012-10-09 Absolute Imaging LLC Endoscopic imaging using reflection holographic optical element for autostereoscopic 3-D viewing
US8619078B2 (en) * 2010-05-21 2013-12-31 International Business Machines Corporation Parallelized ray tracing
US8629867B2 (en) 2010-06-04 2014-01-14 International Business Machines Corporation Performing vector multiplication
US8692825B2 (en) 2010-06-24 2014-04-08 International Business Machines Corporation Parallelized streaming accelerated data structure generation
US20120200676A1 (en) * 2011-02-08 2012-08-09 Microsoft Corporation Three-Dimensional Display with Motion Parallax
US9134700B2 (en) 2011-04-27 2015-09-15 Panasonic Intellectual Property Management Co., Ltd. Display device
US9191661B2 (en) * 2011-08-29 2015-11-17 Microsoft Technology Licensing, Llc Virtual image display device
FR2986874A1 (en) * 2012-02-15 2013-08-16 France Telecom HOLOGRAPHIC PATTERN ENCODING METHOD, CORRESPONDING ENCODING DEVICE AND COMPUTER PROGRAM
EP2834698B1 (en) * 2012-05-03 2021-06-23 Nokia Technologies Oy Image providing apparatus, method and computer program
US9265458B2 (en) 2012-12-04 2016-02-23 Sync-Think, Inc. Application of smooth pursuit cognitive testing paradigms to clinical drug development
US9380976B2 (en) 2013-03-11 2016-07-05 Sync-Think, Inc. Optical neuroinformatics
FR3015743A1 (en) * 2013-12-23 2015-06-26 Orange METHOD OF PROCESSING A SEQUENCE OF HOLOGRAPHIC IMAGES, DEVICES, SIGNALS, DEVICES AND COMPUTER PROGRAM THEREFOR
CN105223796B (en) * 2015-09-08 2018-09-11 北京邮电大学 Hologram computational methods based on near-eye display device and device
KR102571080B1 (en) * 2016-02-12 2023-08-25 삼성전자주식회사 Method and apparatus for processing holographic image
CN109581850B (en) * 2017-09-29 2021-03-05 京东方科技集团股份有限公司 Holographic display method and holographic display device
CN109727317B (en) * 2019-01-07 2021-02-09 京东方科技集团股份有限公司 Augmented reality system and control method

Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5052765A (en) 1989-08-31 1991-10-01 Hughes Aircraft Company Scanning fiber optic holographic exposure and feedback system
US5495576A (en) 1993-01-11 1996-02-27 Ritchey; Kurtis J. Panoramic image based virtual reality/telepresence audio-visual system and method
JPH08111877A (en) 1994-10-12 1996-04-30 Canon Inc Method and device for displaying stereoscopic image
JPH08167048A (en) 1994-12-13 1996-06-25 Yokogawa Electric Corp Three-dimensional image display device
EP0721131A2 (en) 1994-12-29 1996-07-10 Sharp Kabushiki Kaisha Observer tracking autostereoscopic display and method of tracking an observer
US5583795A (en) * 1995-03-17 1996-12-10 The United States Of America As Represented By The Secretary Of The Army Apparatus for measuring eye gaze and fixation duration, and method therefor
US5673151A (en) 1992-02-07 1997-09-30 Virtual I/O Image correction in virtual reality and heads up displays
US5701132A (en) 1996-03-29 1997-12-23 University Of Washington Virtual retinal display with expanded exit pupil
US5726916A (en) * 1996-06-27 1998-03-10 The United States Of America As Represented By The Secretary Of The Army Method and apparatus for determining ocular gaze point of regard and fixation duration
US5781229A (en) 1997-02-18 1998-07-14 Mcdonnell Douglas Corporation Multi-viewer three dimensional (3-D) virtual display system and operating method therefor
US6157402A (en) 1997-02-13 2000-12-05 Torgeson; W. Lee Autostereoscopic image presentation system using a screen assembly
US6525478B2 (en) 2000-03-15 2003-02-25 Robert Bosch Gmbh Device for checking the evaluation circuit for automatic switching operations of illumination device in vehicles
JP2009068674A (en) 2007-09-18 2009-04-02 Masayoshi Nishida Continuously variable transmission
JP2010067257A (en) 2008-09-12 2010-03-25 Square Enix Co Ltd Apparatus of providing casino game

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0968674A (en) * 1995-09-01 1997-03-11 Tsushin Hoso Kiko Three-dimensional image display device
JPH1067257A (en) * 1996-08-28 1998-03-10 Asahi Glass Co Ltd Head up display

Patent Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5052765A (en) 1989-08-31 1991-10-01 Hughes Aircraft Company Scanning fiber optic holographic exposure and feedback system
US5673151A (en) 1992-02-07 1997-09-30 Virtual I/O Image correction in virtual reality and heads up displays
US5495576A (en) 1993-01-11 1996-02-27 Ritchey; Kurtis J. Panoramic image based virtual reality/telepresence audio-visual system and method
JPH08111877A (en) 1994-10-12 1996-04-30 Canon Inc Method and device for displaying stereoscopic image
JPH08167048A (en) 1994-12-13 1996-06-25 Yokogawa Electric Corp Three-dimensional image display device
EP0721131A2 (en) 1994-12-29 1996-07-10 Sharp Kabushiki Kaisha Observer tracking autostereoscopic display and method of tracking an observer
US5583795A (en) * 1995-03-17 1996-12-10 The United States Of America As Represented By The Secretary Of The Army Apparatus for measuring eye gaze and fixation duration, and method therefor
US5701132A (en) 1996-03-29 1997-12-23 University Of Washington Virtual retinal display with expanded exit pupil
US5726916A (en) * 1996-06-27 1998-03-10 The United States Of America As Represented By The Secretary Of The Army Method and apparatus for determining ocular gaze point of regard and fixation duration
US6157402A (en) 1997-02-13 2000-12-05 Torgeson; W. Lee Autostereoscopic image presentation system using a screen assembly
US5781229A (en) 1997-02-18 1998-07-14 Mcdonnell Douglas Corporation Multi-viewer three dimensional (3-D) virtual display system and operating method therefor
US6525478B2 (en) 2000-03-15 2003-02-25 Robert Bosch Gmbh Device for checking the evaluation circuit for automatic switching operations of illumination device in vehicles
JP2009068674A (en) 2007-09-18 2009-04-02 Masayoshi Nishida Continuously variable transmission
JP2010067257A (en) 2008-09-12 2010-03-25 Square Enix Co Ltd Apparatus of providing casino game

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Alfred Schwartz, Head tracking Stereoscopic Display, Aug. 1986, IEEE, vol. ED-33, No. 8, pp. 1123-1127. *

Also Published As

Publication number Publication date
AU2001262486A1 (en) 2001-12-17
US20040021768A1 (en) 2004-02-05
GB0013942D0 (en) 2000-08-02
ATE377773T1 (en) 2007-11-15
JP2003536297A (en) 2003-12-02
EP1287400B1 (en) 2007-11-07
EP1287400B8 (en) 2008-03-26
GB2363273A (en) 2001-12-12
JP4717320B2 (en) 2011-07-06
WO2001095016A1 (en) 2001-12-13
US7053925B2 (en) 2006-05-30
DE60131274T2 (en) 2008-06-05
DE60131274D1 (en) 2007-12-20
EP1287400A1 (en) 2003-03-05
CA2411837A1 (en) 2001-12-13

Similar Documents

Publication Publication Date Title
USRE43203E1 (en) Computation time reduction for the three-dimensional displays
Maimone et al. Holographic near-eye displays for virtual and augmented reality
US10884377B2 (en) Method of computing a hologram for reconstructing an object using a display device
EP2160905B1 (en) Multi-user autostereoscopic display
KR100311066B1 (en) Method and apparatus for generating high resolution 3d images in a head tracked stereo display system
US7224526B2 (en) Three-dimensional free space image projection employing Fresnel lenses
EP1183558B1 (en) Holographic displays
Yamaguchi Full-parallax holographic light-field 3-D displays and interactive 3-D touch
KR20080096505A (en) Method for the multimodal representation of image contents on a display unit for video holograms, and multimodal display unit
US20180364643A1 (en) Method for computing holograms for holographic reconstruction of two-dimensional and/or three-dimensional scenes
EP0862096A2 (en) Method and system for producing computer generated holograms realizing real time holographic video production and display
US20030122828A1 (en) Projection of three-dimensional images
JP2005520184A (en) Radiation conditioning system
Brar et al. Laser-based head-tracked 3D display research
US8390676B2 (en) Device and method for tracking a viewer window
US20160033934A1 (en) Hologram generation apparatus and method
US11835721B2 (en) Display device and method for producing a large field of vision
CN104394394A (en) Three-dimensional display method, device and system for realizing holographic image display
Surman et al. Solving the 3D Problem—The History and Development of Viable Domestic
Surman et al. Head tracked single and multi-user autostereoscopic displays
CN104394395A (en) Holographic image display method, device and system capable of changing scanning frequency
Barabas Holographic television: measuring visual performance with holographic and other 3D television technologies
Kreis Referenceless Phase Holography for 3D Imaging
Chatterjee et al. Autostereoscopic, partial pixel, spatially multiplexed, and other 3D display technologies
Byoungho Lee et al. THREE-DIMENSIONAL DISPLAY AND INFORMATION PROCESSING BASED ON INTEGRAL IMAGING

Legal Events

Date Code Title Description
AS Assignment

Owner name: F. POSZAT HU, LLC, DELAWARE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:QINETIQ LIMITED COMPANY;REEL/FRAME:026815/0372

Effective date: 20070327

Owner name: HOLOGRAPHIC IMAGING LLC, MICHIGAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PAYNE, DOUGLAS ALAN;BLANCHARD, PAUL MICHAEL;SIGNING DATES FROM 20030103 TO 20030106;REEL/FRAME:026815/0277

Owner name: QINETIQ LIMITED, UNITED KINGDOM

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HOLOGRAPHIC IMAGING LLC;REEL/FRAME:026815/0338

Effective date: 20051220

AS Assignment

Owner name: HOLOGRAPHIC IMAGING LLC, MICHIGAN

Free format text: DECLARATION TO CORRECT THE INVENTOR NAME CHANGE PREVIOUSLY RECORDED ON REEL 026815 FRAME 0277. ASSIGNORS HEREBY CONFIRMS THE ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:LEVENETS, DOUGLAS ALAN;REEL/FRAME:027131/0980

Effective date: 20030103

CC Certificate of correction
REMI Maintenance fee reminder mailed
LAPS Lapse for failure to pay maintenance fees
AS Assignment

Owner name: INTELLECTUAL VENTURES ASSETS 191 LLC, DELAWARE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:F. POSZAT HU, L.L.C.;REEL/FRAME:062666/0463

Effective date: 20221222

AS Assignment

Owner name: INTELLECTUAL VENTURES ASSETS 186 LLC, DELAWARE

Free format text: SECURITY INTEREST;ASSIGNOR:MIND FUSION, LLC;REEL/FRAME:063295/0001

Effective date: 20230214

Owner name: INTELLECTUAL VENTURES ASSETS 191 LLC, DELAWARE

Free format text: SECURITY INTEREST;ASSIGNOR:MIND FUSION, LLC;REEL/FRAME:063295/0001

Effective date: 20230214

AS Assignment

Owner name: MIND FUSION, LLC, WASHINGTON

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:INTELLECTUAL VENTURES ASSETS 191 LLC;REEL/FRAME:064270/0685

Effective date: 20230214