US20070047040A1 - Apparatus and method for controlling depth of three-dimensional image - Google Patents

Apparatus and method for controlling depth of three-dimensional image Download PDF

Info

Publication number
US20070047040A1
US20070047040A1 US11/492,970 US49297006A US2007047040A1 US 20070047040 A1 US20070047040 A1 US 20070047040A1 US 49297006 A US49297006 A US 49297006A US 2007047040 A1 US2007047040 A1 US 2007047040A1
Authority
US
United States
Prior art keywords
depth
eye image
physical distance
value
controlling
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US11/492,970
Other versions
US8290244B2 (en
Inventor
Tae-Hyeun Ha
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Electronics Co Ltd
Original Assignee
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Samsung Electronics Co Ltd filed Critical Samsung Electronics Co Ltd
Assigned to SAMSUNG ELECTRONICS CO., LTD. reassignment SAMSUNG ELECTRONICS CO., LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HA, TAE-HYEUN
Publication of US20070047040A1 publication Critical patent/US20070047040A1/en
Application granted granted Critical
Publication of US8290244B2 publication Critical patent/US8290244B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/111Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B30/00Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images
    • G02B30/50Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images the image being built up from image elements distributed over a 3D volume, e.g. voxels
    • G02B30/52Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images the image being built up from image elements distributed over a 3D volume, e.g. voxels the 3D volume being constructed from a stack or sequence of 2D planes, e.g. depth sampling systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/122Improving the 3D impression of stereoscopic images by modifying image signal contents, e.g. by filtering or adding monoscopic depth cues
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/128Adjusting depth or disparity
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/139Format conversion, e.g. of frame-rate or size
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/144Processing image signals for flicker reduction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/398Synchronisation thereof; Control thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N2213/00Details of stereoscopic systems
    • H04N2213/002Eyestrain reduction by processing stereoscopic signals or controlling stereoscopic devices

Definitions

  • the present invention relates to an apparatus and method for controlling the depth of a three-dimensional image, and more particularly, to an apparatus and method for adaptively controlling disparity to control the image depth when a user uses a stereoscopic three-dimensional display of a different size than a display used in a manufacturing environment.
  • Current three dimensional displays generally use a method of projecting images having different disparities to the left and right eyes of a user and applying a three-dimensional effect to the displayed image by adding a film-shaped micro-polarizer, a barrier, or a lenticular lens to a two-dimensional flat display.
  • Three-dimensional displays using this method are classified into stereoscopic three-dimensional displays and multi-view three-dimensional displays in accordance with the number of views of the display.
  • the stereoscopic three-dimensional display displays images of two views and the multi-view three-dimensional display displays images of three or more views.
  • a realistic image is created using a stereoscopic camera having two incident lenses or a multi-view camera having several incident lenses, and stereo or multi-view contents are created by computer graphic technology.
  • the image or the image contents are created in consideration of human visual characteristics. However, if the characteristics such as size and resolution of the display used in the manufacturing environment are different from those of the display used by the user, the desired three-dimensional image cannot be displayed.
  • the present invention provides an apparatus and method for adaptively controlling disparity to control the image depth when a user uses a stereoscopic three-dimensional display having a different size than a display used in a manufacturing environment.
  • an apparatus for controlling the depth of a three-dimensional image including: a disparity measuring unit which measures the disparity between a left eye image and a right eye image; a physical distance calculating unit which calculates the physical distance between the left eye image and the right eye image based on the measured disparity and the physical characteristics of a display; and a depth controlling unit which controls the depth depending on the calculated physical distance.
  • the disparity measuring unit may estimate the disparity between the left eye image and the right eye image and determine any one of a maximum value, a minimum value, and a mean value of the disparity as a reference value.
  • the physical characteristics of the display may include the physical size of one pixel in the display, and the physical distance calculating unit may read the physical size of one pixel which has been previously stored and multiply the reference value by the physical size of that pixel to calculate the physical distance of the disparity.
  • the depth controlling unit may control the physical distance based on a threshold value determined depending on the physical characteristics of the display.
  • the depth controlling unit may compare a maximum value of the calculated physical distance with a maximum threshold value of the physical distance, control the maximum value to less than the maximum threshold value if the calculated maximum value is greater than the maximum threshold value, and control the physical distance between objects in the images in proportion to the controlled maximum value.
  • the depth controlling unit may compare a mean value of the calculated physical distance with a mean threshold value of the physical distance, control the mean value to less than the mean threshold value if the calculated mean value is greater than the mean threshold value, and control the physical distance between objects in the images in proportion to the controlled mean value.
  • the depth controlling unit may compare a maximum value of the calculated physical distance with a minimum threshold value of the physical distance, control the maximum value to greater than the minimum threshold value if the calculated maximum value is less than the minimum threshold value, and control the physical distance between objects in the images in proportion to the controlled maximum value.
  • the depth controlling unit may reconstruct an intermediate image using an intermediate view reconstruction method based on the controlled physical distance and replace at least one of the left eye image and the right eye image with the reconstructed intermediate image to control the depth.
  • the depth controlling unit may horizontally move the left eye image and the right eye image based on the controlled physical distance and synthesize the horizontally moved left eye image and right eye image to control the depth.
  • the apparatus for controlling the depth may further include a display unit which displays the image having the controlled depth, output from the depth controlling unit.
  • a method for controlling the depth of a three-dimensional image including: measuring the disparity between a left eye image and a right eye image; calculating the physical distance between the left eye image and the right eye image based on the measured disparity and the physical characteristics of a display; and controlling the depth depending on the calculated physical distance.
  • FIGS. 1A-1C illustrate a disparity variation between a left eye image and a right eye image in accordance with the variation of the screen size of a display
  • FIG. 2 illustrates the relationship between the depth and the size of an object when a user views a large object and a small object at the same position
  • FIGS. 3A-3C illustrate the variation of the depth of an object depending on the screen size of a display and the position of a user
  • FIG. 4 illustrates the structure of an apparatus for adaptively controlling the depth of a three-dimensional image according to an embodiment of the present invention
  • FIG. 5 illustrates a process of controlling the depth using an intermediate view reconstruction method in a depth controlling unit illustrated in FIG. 4 ;
  • FIG. 6 illustrates a process of horizontally moving a left eye image and a right eye image to control image depth in the depth controlling unit illustrated in FIG. 4 ;
  • FIG. 7 is a flowchart illustrating a method for adaptively controlling the depth of a three-dimensional image according to an embodiment of the present invention.
  • FIGS. 1A-1C a disparity variation between a left eye image and a right eye image in accordance with the variation of the screen size of a display.
  • the content When content is created for a display screen illustrated in FIG. 1B , the content has an optimal depth D 2 when reproduced on the same display screen. Accordingly, if the content is intended for a three-dimensional display having a screen size of N2 ⁇ M2 (inch 2 ) illustrated in FIG. 1B but a user watches the content on a display having a smaller screen size of N1 ⁇ M1 (inch 2 ) illustrated in FIG. 1A , the depth of the three-dimensional image is reduced and thus the user feels as if the object on the screen is distant. That is, the three-dimensional effect deteriorates.
  • the user When the user watches the three-dimensional content on a display having a larger screen size of N3 ⁇ M3 (inch 2 ), the user feels as if the object on the screen is very close, which can cause fatigue or eye strain. In some cases, the object may have a depth which cannot be perceived by a human.
  • FIG. 2 illustrates the relationship between the depth and the size of an object when a user views a large object and a small object at the same position.
  • the depth does not actually vary even if the size of the object varies.
  • the disparity between a left eye image and a right eye image is proportional to the depth perceived by the user and varies depending on the screen size of the display, the same problem as described with reference to FIG. 1 occurs.
  • FIGS. 3A-3C the variation of the depth of an object depending on the screen size of a display and the position of a user.
  • the depth when the user watches the image content on a display having a different screen size from the display used when making the content, the depth must be controlled such that the image is suitably displayed on the user's display.
  • FIG. 4 illustrates the structure of an apparatus for adaptively controlling the depth of a three-dimensional image according to an embodiment of the present invention.
  • the apparatus 400 for controlling the depth includes a disparity measuring unit 410 , a physical distance calculating unit 420 , and a depth controlling unit 430 .
  • the disparity measuring unit 410 divides an input three-dimensional image into a left eye image and a right eye image and measures the disparity between the left eye image and the right eye image.
  • the disparity measuring unit 410 estimates the disparity between the left eye image and the right eye image.
  • the method of estimating the disparity may use block-based movement estimation as used by MPEG encoding. First, the left eye image is divided into N ⁇ N uniform blocks. Subsequently, for each block in the left eye image, the most similar block is estimated in the right eye image using a sum of absolute difference (SAD) or a mean of absolute difference (MAD) calculation.
  • SAD sum of absolute difference
  • MAD mean of absolute difference
  • the distance between a reference block and its estimated similar block forms a disparity vector, measured in pixels. That is, the disparity between the left eye image and the right eye image at a specific position is represented by the distance in pixels between the same objects in the left eye image and the right eye image at that point.
  • the disparity measuring unit 410 determines any one of a maximum value d(max), a minimum value d(min), and a mean value d(mean) of the estimated disparities as a reference value and sends it to the physical distance calculating unit 420 .
  • the physical distance calculating unit 420 calculates the physical distance between the left eye image and the right eye image based on the measured disparity and the physical characteristics of the display.
  • the physical distance calculating unit 420 reads the physical size P of one pixel displayed on the screen and multiplies the physical size P of the pixel by the reference value determined by the disparity measuring unit 410 , to calculate the physical distance between the objects in the left eye image and the right eye image.
  • the physical size P of the pixel may be stored in a storage unit such as a ROM of the display.
  • the depth controlling unit 430 controls the calculated physical distance based on a threshold value determined in accordance with the physical characteristics of the display, and controls the depth in accordance with the controlled distance.
  • the depth controlling unit 430 reconstructs the image using an intermediate view reconstruction method, based on the controlled distance, and replaces either the left eye image or the right eye image with the reconstructed image, and outputs the replacement image and the unchanged other eye image. When these are displayed, an image having controlled depth is viewed.
  • the depth control unit 430 horizontally moves the left eye image and the right eye image based on the controlled distance, synthesizes the horizontally moved left eye image and right eye image, and outputs the image having the controlled depth.
  • the depth controlling apparatus 400 may include a display unit (not shown) for displaying the three-dimensional image output from depth controlling unit 430 to embody a display system for outputting the three-dimensional image.
  • FIG. 5 illustrates a process of controlling the depth using the intermediate view reconstruction method in the depth controlling unit 430 illustrated in FIG. 4 .
  • the depth controlling unit 430 controls the depth based on the physical distance calculated by the physical distance calculating unit 420 . If the physical distance must be controlled, the depth controlling unit 430 performs the following operations.
  • the depth controlling unit 430 receives the maximum value of the physical distance calculated by the physical distance calculating unit 420 and compares the received maximum value with a maximum threshold value of the physical distance. If the received maximum value is greater than the maximum threshold value, the depth controlling unit 430 controls the maximum value to be less than the maximum threshold value, and controls the physical distance between the objects in the images in proportion to the controlled value. That is, the value obtained by multiplying the maximum value of the disparity between the objects in the images by the physical size of one pixel of the display is controlled to less than the predetermined threshold value and the total disparity of the image is controlled in proportion to the controlled value.
  • the maximum value of the physical distance, D(max), is obtained by multiplying the maximum value d(max) of the measured disparities by the physical size P of one pixel.
  • the maximum threshold value th(max) is the maximum value of the physical distance between the left eye image 51 and the right eye image 53 in the display as used by a user within a range which does not cause eyestrain or other fatigue, as determined experimentally.
  • the maximum threshold value can be obtained from the depth having the maximum value within a range wherein fatigue is not induced.
  • d 1 denotes the distance from the left edge to the object in the left eye image 51
  • d 2 denotes the distance from the left edge to the object in the right eye image 53
  • d 3 denotes the disparity between the objects in the left eye image 51 and the right eye image 53 .
  • d 4 denotes the disparity which is controlled to prevent user fatigue. That is, if D(max)>th(max), a new image having the disparity d 4 is generated using Equation 1, and the existing right eye image 53 is replaced with the new image, thereby reducing user fatigue.
  • the depth controlling unit 430 receives the mean value of the physical distance calculated by the physical distance calculating unit 420 and compares it with a mean threshold value of the physical distance. Then, if the calculated mean value is greater than the mean threshold value, the depth controlling unit 430 controls the calculated mean value to less than the mean threshold value, and controls the physical distance between the objects in the images in proportion to the controlled mean value.
  • the mean value of the physical distance, D(mean) is obtained by multiplying the mean value d(mean) of the disparities calculated by the disparity measuring unit 410 by the physical size P of one pixel.
  • the mean threshold value th(mean) is the average of the physical distances between the left eye image 51 and the right eye image 53 in the display as used by the user within the range which does not cause eyestrain or other fatigue, as determined experimentally.
  • the depth controlling unit 430 compares the maximum value of the physical distance calculated by the physical distance calculating unit 420 with a minimum threshold value of the physical distance. If the calculated maximum value is less than the minimum threshold value, the depth controlling unit 430 controls the calculated maximum value to greater than the minimum threshold value, and controls the physical distance between the objects in the images in proportion to the controlled maximum value.
  • the minimum threshold value th(min) is the minimum value of the physical distance between the left eye image 51 and the right eye image 53 in the display as used by the user, within the range which does not cause eyestrain or other fatigue, as determined experimentally.
  • the image having the physical distance of d 4 ⁇ P is reconstructed by the intermediate view reconstruction method, the right eye image 53 is replaced with the reconstructed intermediate image 55 , and the left eye image 51 and the intermediate image 55 overlap each other and are output.
  • the proper disparity d 4 or the proper physical distance d 4 ⁇ P is determined in order to control the depth, any one or at least two of Equations 1, 2, and 3 may be used in accordance with the characteristics of the input three-dimensional image.
  • the right eye image 53 was replaced with the reconstructed image using a method of determining proper disparity based on the distance from the left edge to the object.
  • the proper disparity may be calculated based on the distance from the right edge to the object.
  • the left eye image 51 is replaced with the reconstructed image using the intermediate view reconstruction method. That is, the image 53 of FIG. 5 and the reconstructed image reconstructed by the intermediate view reconstruction method overlap each other and are output.
  • FIG. 6 illustrates a process of horizontally moving the left eye image 51 and the right eye image 53 to control the depth in the depth controlling unit 430 illustrated in FIG. 4 .
  • Left eye image 61 is the image obtained by horizontally moving the left eye image 51 of FIG. 5 to the right side, and a hatched region indicates a portion which is lost in the process.
  • Right eye image 63 is the image obtained by horizontally moving the right eye image 53 of FIG. 5 to the left side, and a hatched region indicates a portion which is lost in the process.
  • An image 65 is obtained by synthesizing the horizontally moved left eye image 61 and right eye image 63 . Accordingly, the left eye image 51 and the right eye image 53 must be moved horizontally such that the disparity between the left and right eye images is controlled to the disparity d 4 as determined above.
  • the left eye image 51 is horizontally moved by d 3 -d 4 /2 to the right side and the right eye image 53 is horizontally moved by d 3 -d 4 /2 to the left side.
  • the three-dimensional image 65 having the disparity d 4 can be generated. Similar to the method described with reference to FIG. 5 , when the proper disparity d 4 or the proper physical distance d 4 ⁇ P is determined in order to control the depth, any one or at least two of Equations 1, 2, and 3 may be used in accordance with the characteristics of the input three-dimensional image.
  • FIG. 7 is a flowchart illustrating a method of adaptively controlling the depth of a three-dimensional image according to an embodiment of the present invention.
  • the disparity between the left eye image and the right eye image is measured (S 710 ).
  • the physical distance between the left eye image and the right eye image is calculated (S 720 ).
  • the depth is controlled in accordance with the calculated physical distance (S 730 ).
  • the image having the controlled depth is displayed on the display to provide a user with an image having an adaptively controlled three-dimensional effect.
  • the invention can also be embodied as computer readable code on a computer readable recording medium.
  • the computer readable recording medium is any data storage device that can store data which can be thereafter read by a computer system. Examples of the computer readable recording medium include read-only memory (ROM), random-access memory (RAM), CD-ROMs, magnetic tapes, floppy disks, optical data storage devices, and carrier waves (such as data transmission through the internet).
  • ROM read-only memory
  • RAM random-access memory
  • CD-ROMs compact discs
  • magnetic tapes magnetic tapes
  • floppy disks optical data storage devices
  • carrier waves such as data transmission through the internet
  • carrier waves such as data transmission through the internet.
  • the computer readable recording medium can also be distributed over network coupled computer systems so that the computer readable code is stored and executed in a distributed fashion. Also, functional programs, code, and code segments for accomplishing the present invention can be easily constructed by programmers skilled in the art to which the present invention pertains.
  • the present invention it is possible to provide an apparatus and method for adaptively controlling disparity to control the image depth when a user uses a stereoscopic three-dimensional display having a different screen size than a display used in a manufacturing environment.

Abstract

An apparatus and method for controlling the depth of a three-dimensional image. The apparatus includes: a disparity measuring unit which measures the disparity between a left eye image and a right eye image; a physical distance calculating unit which calculates the physical distance between the left eye image and the right eye image based on the measured disparity and the physical characteristics of a display; and a depth controlling unit which controls the depth depending on the calculated physical distance. It is possible to adaptively control the disparity to control the depth when a user uses a stereoscopic three-dimensional display having a different screen size than a display used in a manufacturing environment.

Description

    CROSS-REFERENCE TO RELATED PATENT APPLICATION
  • This application claims the benefit of Korean Patent Application No. 10-2005-0080611, filed on Aug. 31, 2005, in the Korean Intellectual Property Office, the disclosure of which is incorporated herein in its entirety by reference.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The present invention relates to an apparatus and method for controlling the depth of a three-dimensional image, and more particularly, to an apparatus and method for adaptively controlling disparity to control the image depth when a user uses a stereoscopic three-dimensional display of a different size than a display used in a manufacturing environment.
  • 2. Description of the Related Art
  • Current three dimensional displays generally use a method of projecting images having different disparities to the left and right eyes of a user and applying a three-dimensional effect to the displayed image by adding a film-shaped micro-polarizer, a barrier, or a lenticular lens to a two-dimensional flat display. Three-dimensional displays using this method are classified into stereoscopic three-dimensional displays and multi-view three-dimensional displays in accordance with the number of views of the display. The stereoscopic three-dimensional display displays images of two views and the multi-view three-dimensional display displays images of three or more views.
  • A realistic image is created using a stereoscopic camera having two incident lenses or a multi-view camera having several incident lenses, and stereo or multi-view contents are created by computer graphic technology. Here, the image or the image contents are created in consideration of human visual characteristics. However, if the characteristics such as size and resolution of the display used in the manufacturing environment are different from those of the display used by the user, the desired three-dimensional image cannot be displayed.
  • SUMMARY OF THE INVENTION
  • The present invention provides an apparatus and method for adaptively controlling disparity to control the image depth when a user uses a stereoscopic three-dimensional display having a different size than a display used in a manufacturing environment.
  • According to an aspect of the present invention, there is provided an apparatus for controlling the depth of a three-dimensional image, including: a disparity measuring unit which measures the disparity between a left eye image and a right eye image; a physical distance calculating unit which calculates the physical distance between the left eye image and the right eye image based on the measured disparity and the physical characteristics of a display; and a depth controlling unit which controls the depth depending on the calculated physical distance.
  • The disparity measuring unit may estimate the disparity between the left eye image and the right eye image and determine any one of a maximum value, a minimum value, and a mean value of the disparity as a reference value.
  • The physical characteristics of the display may include the physical size of one pixel in the display, and the physical distance calculating unit may read the physical size of one pixel which has been previously stored and multiply the reference value by the physical size of that pixel to calculate the physical distance of the disparity.
  • The depth controlling unit may control the physical distance based on a threshold value determined depending on the physical characteristics of the display.
  • The depth controlling unit may compare a maximum value of the calculated physical distance with a maximum threshold value of the physical distance, control the maximum value to less than the maximum threshold value if the calculated maximum value is greater than the maximum threshold value, and control the physical distance between objects in the images in proportion to the controlled maximum value.
  • The depth controlling unit may compare a mean value of the calculated physical distance with a mean threshold value of the physical distance, control the mean value to less than the mean threshold value if the calculated mean value is greater than the mean threshold value, and control the physical distance between objects in the images in proportion to the controlled mean value.
  • The depth controlling unit may compare a maximum value of the calculated physical distance with a minimum threshold value of the physical distance, control the maximum value to greater than the minimum threshold value if the calculated maximum value is less than the minimum threshold value, and control the physical distance between objects in the images in proportion to the controlled maximum value.
  • The depth controlling unit may reconstruct an intermediate image using an intermediate view reconstruction method based on the controlled physical distance and replace at least one of the left eye image and the right eye image with the reconstructed intermediate image to control the depth. The depth controlling unit may horizontally move the left eye image and the right eye image based on the controlled physical distance and synthesize the horizontally moved left eye image and right eye image to control the depth.
  • The apparatus for controlling the depth may further include a display unit which displays the image having the controlled depth, output from the depth controlling unit.
  • According to another aspect of the present invention, there is provided a method for controlling the depth of a three-dimensional image, including: measuring the disparity between a left eye image and a right eye image; calculating the physical distance between the left eye image and the right eye image based on the measured disparity and the physical characteristics of a display; and controlling the depth depending on the calculated physical distance.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The above and other features and advantages of the present invention will become more apparent by describing in detail exemplary embodiments thereof with reference to the attached drawings in which:
  • FIGS. 1A-1C illustrate a disparity variation between a left eye image and a right eye image in accordance with the variation of the screen size of a display;
  • FIG. 2 illustrates the relationship between the depth and the size of an object when a user views a large object and a small object at the same position;
  • FIGS. 3A-3C illustrate the variation of the depth of an object depending on the screen size of a display and the position of a user;
  • FIG. 4 illustrates the structure of an apparatus for adaptively controlling the depth of a three-dimensional image according to an embodiment of the present invention;
  • FIG. 5 illustrates a process of controlling the depth using an intermediate view reconstruction method in a depth controlling unit illustrated in FIG. 4;
  • FIG. 6 illustrates a process of horizontally moving a left eye image and a right eye image to control image depth in the depth controlling unit illustrated in FIG. 4; and
  • FIG. 7 is a flowchart illustrating a method for adaptively controlling the depth of a three-dimensional image according to an embodiment of the present invention.
  • DETAILED DESCRIPTION OF THE INVENTION
  • The present invention will now be described more fully with reference to the accompanying drawings, in which exemplary embodiments of the invention are shown.
  • FIGS. 1A-1C a disparity variation between a left eye image and a right eye image in accordance with the variation of the screen size of a display.
  • When content is created for a display screen illustrated in FIG. 1B, the content has an optimal depth D2 when reproduced on the same display screen. Accordingly, if the content is intended for a three-dimensional display having a screen size of N2×M2 (inch2) illustrated in FIG. 1B but a user watches the content on a display having a smaller screen size of N1×M1 (inch2) illustrated in FIG. 1A, the depth of the three-dimensional image is reduced and thus the user feels as if the object on the screen is distant. That is, the three-dimensional effect deteriorates.
  • When the user watches the three-dimensional content on a display having a larger screen size of N3×M3 (inch2), the user feels as if the object on the screen is very close, which can cause fatigue or eye strain. In some cases, the object may have a depth which cannot be perceived by a human.
  • FIG. 2 illustrates the relationship between the depth and the size of an object when a user views a large object and a small object at the same position.
  • If the distance between an object and a user who watches the object is constant, the depth does not actually vary even if the size of the object varies. However, in a three-dimensional display using stereoscopic disparity, since the disparity between a left eye image and a right eye image is proportional to the depth perceived by the user and varies depending on the screen size of the display, the same problem as described with reference to FIG. 1 occurs.
  • FIGS. 3A-3C the variation of the depth of an object depending on the screen size of a display and the position of a user.
  • As illustrated in FIGS. 3A and 3B, if the screen size increases while the position of the user is constant, the disparity increases and thus the depth increases. As illustrated in FIGS. 3B and 3C, if the distance between the display and the user doubles, the depth increases.
  • Accordingly, when the user watches the image content on a display having a different screen size from the display used when making the content, the depth must be controlled such that the image is suitably displayed on the user's display.
  • FIG. 4 illustrates the structure of an apparatus for adaptively controlling the depth of a three-dimensional image according to an embodiment of the present invention.
  • The apparatus 400 for controlling the depth according to the present embodiment includes a disparity measuring unit 410, a physical distance calculating unit 420, and a depth controlling unit 430.
  • The disparity measuring unit 410 divides an input three-dimensional image into a left eye image and a right eye image and measures the disparity between the left eye image and the right eye image. The disparity measuring unit 410 estimates the disparity between the left eye image and the right eye image. The method of estimating the disparity may use block-based movement estimation as used by MPEG encoding. First, the left eye image is divided into N×N uniform blocks. Subsequently, for each block in the left eye image, the most similar block is estimated in the right eye image using a sum of absolute difference (SAD) or a mean of absolute difference (MAD) calculation. The distance between a reference block and its estimated similar block forms a disparity vector, measured in pixels. That is, the disparity between the left eye image and the right eye image at a specific position is represented by the distance in pixels between the same objects in the left eye image and the right eye image at that point.
  • The disparity measuring unit 410 determines any one of a maximum value d(max), a minimum value d(min), and a mean value d(mean) of the estimated disparities as a reference value and sends it to the physical distance calculating unit 420.
  • The physical distance calculating unit 420 calculates the physical distance between the left eye image and the right eye image based on the measured disparity and the physical characteristics of the display. The physical distance calculating unit 420 reads the physical size P of one pixel displayed on the screen and multiplies the physical size P of the pixel by the reference value determined by the disparity measuring unit 410, to calculate the physical distance between the objects in the left eye image and the right eye image. The physical size P of the pixel may be stored in a storage unit such as a ROM of the display.
  • The depth controlling unit 430 controls the calculated physical distance based on a threshold value determined in accordance with the physical characteristics of the display, and controls the depth in accordance with the controlled distance. The depth controlling unit 430 reconstructs the image using an intermediate view reconstruction method, based on the controlled distance, and replaces either the left eye image or the right eye image with the reconstructed image, and outputs the replacement image and the unchanged other eye image. When these are displayed, an image having controlled depth is viewed. Alternatively, the depth control unit 430 horizontally moves the left eye image and the right eye image based on the controlled distance, synthesizes the horizontally moved left eye image and right eye image, and outputs the image having the controlled depth. The depth controlling apparatus 400 may include a display unit (not shown) for displaying the three-dimensional image output from depth controlling unit 430 to embody a display system for outputting the three-dimensional image.
  • The process of controlling the depth with the depth controlling unit 430 will be described in detail with reference to FIGS. 5 and 6.
  • FIG. 5 illustrates a process of controlling the depth using the intermediate view reconstruction method in the depth controlling unit 430 illustrated in FIG. 4.
  • The depth controlling unit 430 controls the depth based on the physical distance calculated by the physical distance calculating unit 420. If the physical distance must be controlled, the depth controlling unit 430 performs the following operations.
  • The depth controlling unit 430 receives the maximum value of the physical distance calculated by the physical distance calculating unit 420 and compares the received maximum value with a maximum threshold value of the physical distance. If the received maximum value is greater than the maximum threshold value, the depth controlling unit 430 controls the maximum value to be less than the maximum threshold value, and controls the physical distance between the objects in the images in proportion to the controlled value. That is, the value obtained by multiplying the maximum value of the disparity between the objects in the images by the physical size of one pixel of the display is controlled to less than the predetermined threshold value and the total disparity of the image is controlled in proportion to the controlled value.
  • The maximum value of the physical distance, D(max), is obtained by multiplying the maximum value d(max) of the measured disparities by the physical size P of one pixel. The maximum threshold value th(max) is the maximum value of the physical distance between the left eye image 51 and the right eye image 53 in the display as used by a user within a range which does not cause eyestrain or other fatigue, as determined experimentally. The maximum threshold value can be obtained from the depth having the maximum value within a range wherein fatigue is not induced.
  • Referring to FIG. 5, d1 denotes the distance from the left edge to the object in the left eye image 51, d2 denotes the distance from the left edge to the object in the right eye image 53, and d3 denotes the disparity between the objects in the left eye image 51 and the right eye image 53. In the reconstructed image 55, d4 denotes the disparity which is controlled to prevent user fatigue. That is, if D(max)>th(max), a new image having the disparity d4 is generated using Equation 1, and the existing right eye image 53 is replaced with the new image, thereby reducing user fatigue.
    dP=th(max)×dP/D(max)=th(max)×dP/(d(max)×P)  Equation 1
  • Although the maximum value of the physical distance is less than the maximum threshold value, if the total disparity of the image is large, the user may feel fatigued. In this case, the depth controlling unit 430 receives the mean value of the physical distance calculated by the physical distance calculating unit 420 and compares it with a mean threshold value of the physical distance. Then, if the calculated mean value is greater than the mean threshold value, the depth controlling unit 430 controls the calculated mean value to less than the mean threshold value, and controls the physical distance between the objects in the images in proportion to the controlled mean value.
  • The mean value of the physical distance, D(mean), is obtained by multiplying the mean value d(mean) of the disparities calculated by the disparity measuring unit 410 by the physical size P of one pixel. The mean threshold value th(mean) is the average of the physical distances between the left eye image 51 and the right eye image 53 in the display as used by the user within the range which does not cause eyestrain or other fatigue, as determined experimentally.
  • If D(mean)>th(mean), a new image having the disparity d4 is generated using Equation 2 and the existing right eye image 53 is replaced with the new image, thereby reducing the fatigue of the user.
    dP=th(mean)×dP/D(mean)  Equation 2
  • Up to now, the case of using a display having a larger screen size than the display which can optimally display the three-dimensional image has been described. When using a display having a smaller screen size than the optimal display, the physical depth or the disparity decreases and thus eyestrain is reduced. However, the three-dimensional effect may reduced by the decreasing depth. In this case, the three-dimensional effect must be increased by increasing the total disparity.
  • Accordingly, the depth controlling unit 430 compares the maximum value of the physical distance calculated by the physical distance calculating unit 420 with a minimum threshold value of the physical distance. If the calculated maximum value is less than the minimum threshold value, the depth controlling unit 430 controls the calculated maximum value to greater than the minimum threshold value, and controls the physical distance between the objects in the images in proportion to the controlled maximum value.
  • The minimum threshold value th(min) is the minimum value of the physical distance between the left eye image 51 and the right eye image 53 in the display as used by the user, within the range which does not cause eyestrain or other fatigue, as determined experimentally.
  • If D(max)<th(min), a new image having the disparity d4 is generated using Equation 3 and the existing right eye image 53 is replaced with the new image, thereby increasing three dimensional effect.
    dP=th(min)×dP/D(max)  Equation 3
  • The image having the physical distance of d4×P is reconstructed by the intermediate view reconstruction method, the right eye image 53 is replaced with the reconstructed intermediate image 55, and the left eye image 51 and the intermediate image 55 overlap each other and are output. When the proper disparity d4 or the proper physical distance d4×P is determined in order to control the depth, any one or at least two of Equations 1, 2, and 3 may be used in accordance with the characteristics of the input three-dimensional image.
  • In FIG. 5, the right eye image 53 was replaced with the reconstructed image using a method of determining proper disparity based on the distance from the left edge to the object. However, the proper disparity may be calculated based on the distance from the right edge to the object. In this case, the left eye image 51 is replaced with the reconstructed image using the intermediate view reconstruction method. That is, the image 53 of FIG. 5 and the reconstructed image reconstructed by the intermediate view reconstruction method overlap each other and are output.
  • FIG. 6 illustrates a process of horizontally moving the left eye image 51 and the right eye image 53 to control the depth in the depth controlling unit 430 illustrated in FIG. 4.
  • Left eye image 61 is the image obtained by horizontally moving the left eye image 51 of FIG. 5 to the right side, and a hatched region indicates a portion which is lost in the process. Right eye image 63 is the image obtained by horizontally moving the right eye image 53 of FIG. 5 to the left side, and a hatched region indicates a portion which is lost in the process. An image 65 is obtained by synthesizing the horizontally moved left eye image 61 and right eye image 63. Accordingly, the left eye image 51 and the right eye image 53 must be moved horizontally such that the disparity between the left and right eye images is controlled to the disparity d4 as determined above.
  • Referring to FIG. 6, in order to output the image having the disparity d4, the left eye image 51 is horizontally moved by d3-d4/2 to the right side and the right eye image 53 is horizontally moved by d3-d4/2 to the left side. When the left eye image 61 and the right eye image 63 are synthesized, the three-dimensional image 65 having the disparity d4 can be generated. Similar to the method described with reference to FIG. 5, when the proper disparity d4 or the proper physical distance d4×P is determined in order to control the depth, any one or at least two of Equations 1, 2, and 3 may be used in accordance with the characteristics of the input three-dimensional image.
  • FIG. 7 is a flowchart illustrating a method of adaptively controlling the depth of a three-dimensional image according to an embodiment of the present invention.
  • In order to control the depth of the three-dimensional image, the disparity between the left eye image and the right eye image is measured (S710).
  • Based on the measured disparity and the physical characteristics of the display, the physical distance between the left eye image and the right eye image is calculated (S720).
  • The depth is controlled in accordance with the calculated physical distance (S730). The image having the controlled depth is displayed on the display to provide a user with an image having an adaptively controlled three-dimensional effect.
  • The invention can also be embodied as computer readable code on a computer readable recording medium. The computer readable recording medium is any data storage device that can store data which can be thereafter read by a computer system. Examples of the computer readable recording medium include read-only memory (ROM), random-access memory (RAM), CD-ROMs, magnetic tapes, floppy disks, optical data storage devices, and carrier waves (such as data transmission through the internet). The computer readable recording medium can also be distributed over network coupled computer systems so that the computer readable code is stored and executed in a distributed fashion. Also, functional programs, code, and code segments for accomplishing the present invention can be easily constructed by programmers skilled in the art to which the present invention pertains.
  • According to the present invention, it is possible to provide an apparatus and method for adaptively controlling disparity to control the image depth when a user uses a stereoscopic three-dimensional display having a different screen size than a display used in a manufacturing environment.
  • While the present invention has been particularly shown and described with reference to exemplary embodiments thereof, it will be understood by those of ordinary skill in the art that various changes in form and detail may be made therein without departing from the spirit and scope of the present invention as defined by the following claims.

Claims (21)

1. An apparatus for controlling the depth of a three-dimensional image, comprising:
a disparity measuring unit which measures the disparity between a left eye image and a right eye image;
a physical distance calculating unit which calculates the physical distance between the left eye image and the right eye image based on the measured disparity and the physical characteristics of a display; and
a depth controlling unit which controls the depth depending on the calculated physical distance.
2. The apparatus according to claim 1, wherein the disparity measuring unit estimates the disparity between the left eye image and the right eye image and determines any one of a maximum value, a minimum value, and a mean value of the disparity as a reference value.
3. The apparatus according to claim 2, wherein the physical characteristics of the display include the physical size of one pixel in the display, and
the physical distance calculating unit reads the physical size of one pixel which has been previously stored and multiplies the reference value by the physical size of that pixel to calculate the physical distance of the disparity.
4. The apparatus according to claim 3, wherein the depth controlling unit controls the physical distance based on a threshold value determined depending on the physical characteristics of the display.
5. The apparatus according to claim 4, wherein the depth controlling unit compares a maximum value of the calculated physical distance with a maximum threshold value of the physical distance, controls the maximum value to less than the maximum threshold value if the calculated maximum value is greater than the maximum threshold value, and controls the physical distance between objects in the images in proportion to the controlled maximum value.
6. The apparatus according to claim 4, wherein the depth controlling unit compares a mean value of the calculated physical distance with a mean threshold value of the physical distance, controls the mean value to less than the mean threshold value if the calculated mean value is greater than the mean threshold value, and controls the physical distance between objects in the images in proportion to the controlled mean value.
7. The apparatus according to claim 4, wherein the depth controlling unit compares a maximum value of the calculated physical distance with a minimum threshold value of the physical distance, controls the maximum value to greater than the minimum threshold value if the calculated maximum value is less than the minimum threshold value, and controls the physical distance between objects in the images in proportion to the controlled maximum value.
8. The apparatus according to claim 4, wherein the depth controlling unit reconstructs an intermediate image using an intermediate view reconstruction method based on the controlled physical distance, and replaces at least one of the left eye image and the right eye image with the reconstructed intermediate image to control the depth.
9. The apparatus according to claim 4, wherein the depth controlling unit horizontally moves the left eye image and the right eye image based on the controlled physical distance and synthesizes the horizontally moved left eye image and right eye image to control the depth.
10. The apparatus according to claim 1, further comprising a display unit which displays the image having the controlled depth, output from the depth controlling unit.
11. A method for controlling the depth of a three-dimensional image, comprising:
measuring the disparity between a left eye image and a right eye image;
calculating the physical distance between the left eye image and the right eye image based on the measured disparity and the physical characteristics of a display; and
controlling the depth depending on the calculated physical distance.
12. The method according to claim 11, wherein the measuring of the disparity includes estimating the disparity between the left eye image and the right eye image and determining any one of a maximum value, a minimum value, and a mean value of the disparity as a reference value.
13. The method according to claim 12, wherein the physical characteristics of the display include the physical size of one pixel in the display, and
the calculating of the physical distance includes reading the physical size of one pixel which is previously stored and multiplying the reference value by the physical size of that pixel to calculate the physical distance of the disparity.
14. The method according to claim 13, wherein the controlling of the depth includes controlling the physical distance based on a threshold value determined depending on the physical characteristics of the display.
15. The method according to claim 14, wherein the controlling of the depth includes comparing a maximum value of the calculated physical distance with a maximum threshold value of the physical distance, controlling the maximum value to less than the maximum threshold value if the calculated maximum value is greater than the maximum threshold value, and controlling the physical distance between objects in the images in proportion to the controlled maximum value.
16. The method according to claim 14, wherein the controlling of the depth includes comparing a mean value of the calculated physical distance with a mean threshold value of the physical distance, controlling the mean value to less than the mean threshold value if the calculated mean value is greater than the mean threshold value, and controlling the physical distance between objects in the images in proportion to the controlled mean value.
17. The method according to claim 14, wherein the controlling of the depth includes comparing a maximum value of the calculated physical distance with a minimum threshold value of the physical distance, controlling the maximum value to greater than the minimum threshold value if the calculated maximum value is less than the minimum threshold value, and controlling the physical distance between objects in the images in proportion to the controlled maximum value.
18. The method according to claim 14, wherein the controlling of the depth includes reconstructing an intermediate image using an intermediate view reconstruction method based on the controlled physical distance and replacing at least one of the left eye image and the right eye image with the reconstructed intermediate image to control the depth.
19. The method according to claim 14, wherein the controlling of the depth includes horizontally moving the left eye image and the right eye image based on the controlled physical distance and synthesizing the horizontally moved left eye image and right eye image to control the depth.
20. The method according to claim 11, further comprising displaying the image having the controlled depth.
21. A computer-readable medium having embodied thereon a computer program for the method of claim 11.
US11/492,970 2005-08-31 2006-07-26 Apparatus and method for controlling depth of three-dimensional image Active 2030-04-09 US8290244B2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR1020050080611A KR100667810B1 (en) 2005-08-31 2005-08-31 Apparatus for controlling depth of 3d picture and method therefor
KR10-2005-0080611 2005-08-31

Publications (2)

Publication Number Publication Date
US20070047040A1 true US20070047040A1 (en) 2007-03-01
US8290244B2 US8290244B2 (en) 2012-10-16

Family

ID=37803684

Family Applications (1)

Application Number Title Priority Date Filing Date
US11/492,970 Active 2030-04-09 US8290244B2 (en) 2005-08-31 2006-07-26 Apparatus and method for controlling depth of three-dimensional image

Country Status (4)

Country Link
US (1) US8290244B2 (en)
KR (1) KR100667810B1 (en)
CN (1) CN100574462C (en)
NL (1) NL1032382C2 (en)

Cited By (73)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1978755A2 (en) * 2007-04-04 2008-10-08 Mitsubishi Electric Corporation Method and system for acquiring, encoding, decoding and displaying 3D light fields
WO2009020277A1 (en) * 2007-08-06 2009-02-12 Samsung Electronics Co., Ltd. Method and apparatus for reproducing stereoscopic image using depth control
US20090324114A1 (en) * 2008-06-26 2009-12-31 Sony Corporation Image compression apparatus and image compression method
WO2010087575A2 (en) * 2009-02-01 2010-08-05 Lg Electronics Inc. Broadcast receiver and 3d video data processing method
US20110032341A1 (en) * 2009-08-04 2011-02-10 Ignatov Artem Konstantinovich Method and system to transform stereo content
US20110122131A1 (en) * 2008-07-24 2011-05-26 Koninklijke Philips Electronics N.V. Versatile 3-d picture format
US20110187831A1 (en) * 2010-02-03 2011-08-04 Korea Institute Of Science And Technology Apparatus and method for displaying three-dimensional images
US20110216163A1 (en) * 2010-03-08 2011-09-08 Dolby Laboratories Licensing Corporation Methods For Carrying And Transmitting 3D Z-Norm Attributes In Digital TV Closed Captioning
WO2011107550A1 (en) * 2010-03-05 2011-09-09 Sony Corporation Disparity distribution estimation for 3d tv
JP2011188108A (en) * 2010-03-05 2011-09-22 Mitsubishi Electric Corp Image processing apparatus, three-dimensional image display device, and image processing method
WO2011121397A1 (en) * 2010-04-01 2011-10-06 Nokia Corporation Method, apparatus and computer program for selecting a stereoscopic imaging viewpoint pair
WO2011121437A1 (en) * 2010-03-31 2011-10-06 Thomson Licensing 3d disparity maps
US20110249017A1 (en) * 2010-04-07 2011-10-13 Sony Corporation Image signal processing device, display device, display method and program product
US20110267338A1 (en) * 2010-05-03 2011-11-03 Kwangwoon University Industry-Academic Collaboration Foundation Apparatus and method for reducing three-dimensional visual fatigue
US20110285819A1 (en) * 2009-09-17 2011-11-24 Panasonic Corporation Video signal processing apparatus and video signal processing method
US20110292178A1 (en) * 2010-05-28 2011-12-01 Qualcomm Incorporated Three-dimensional image processing
US20110304708A1 (en) * 2010-06-10 2011-12-15 Samsung Electronics Co., Ltd. System and method of generating stereo-view and multi-view images for rendering perception of depth of stereoscopic image
US20120013604A1 (en) * 2010-07-14 2012-01-19 Samsung Electronics Co., Ltd. Display apparatus and method for setting sense of depth thereof
US20120038754A1 (en) * 2010-08-16 2012-02-16 Na Sungbin Method for processing images in display device outputting 3-dimensional contents and display device using the same
US20120075419A1 (en) * 2010-09-23 2012-03-29 Thomson Licensing Adaptation of 3D video content
EP2442577A3 (en) * 2010-10-15 2012-05-16 Sony Corporation Information processing apparatus, informaton processing method and program
US20120120202A1 (en) * 2010-11-12 2012-05-17 Gwangju Institute Of Science And Technology Method for improving 3 dimensional effect and reducing visual fatigue and apparatus enabling the same
US8200045B2 (en) * 2007-02-07 2012-06-12 Thomson Licensing Image processing method
US20120206453A1 (en) * 2009-09-16 2012-08-16 Koninklijke Philips Electronics N.V. 3d screen size compensation
FR2971875A1 (en) * 2011-02-23 2012-08-24 Mobiclip DEVICE AND METHOD FOR MANAGING THE POSITION OF THE FOCAL PLANE IN A STEREOSCOPIC SCENE
US20120249747A1 (en) * 2011-03-30 2012-10-04 Ziv Aviv Real-time depth extraction using stereo correspondence
EP2408213A3 (en) * 2010-07-14 2012-10-24 LG Electronics Inc. Mobile terminal and controlling method thereof
EP2525579A2 (en) * 2010-01-11 2012-11-21 LG Electronics Inc. Broadcasting receiver and method for displaying 3d images
WO2012162096A1 (en) * 2011-05-23 2012-11-29 Qualcomm Incorporated Interactive user interface for stereoscopic effect adjustment
US20120327191A1 (en) * 2010-03-05 2012-12-27 Panasonic Corporation 3d imaging device and 3d imaging method
CN103024591A (en) * 2012-12-26 2013-04-03 新奥特(北京)视频技术有限公司 Method and device for adjusting 3D (three-dimensional) parallax
EP2375766A3 (en) * 2010-04-09 2013-05-29 Tektronix International Sales GmbH Method and apparatus for measuring an audiovisual parameter
US20130141423A1 (en) * 2011-12-06 2013-06-06 Jung-hyun Cho Three-dimensional image display apparatus
US20130278631A1 (en) * 2010-02-28 2013-10-24 Osterhout Group, Inc. 3d positioning of augmented reality information
WO2013158784A1 (en) * 2012-04-17 2013-10-24 3Dmedia Corporation Systems and methods for improving overall quality of three-dimensional content by altering parallax budget or compensating for moving objects
US20140063206A1 (en) * 2012-08-28 2014-03-06 Himax Technologies Limited System and method of viewer centric depth adjustment
US8810635B2 (en) 2009-07-31 2014-08-19 3Dmedia Corporation Methods, systems, and computer-readable storage media for selecting image capture positions to generate three-dimensional images
US20140292748A1 (en) * 2013-04-01 2014-10-02 Electronics And Telecommunications Research Institute System and method for providing stereoscopic image by adjusting depth value
US20150015666A1 (en) * 2013-07-09 2015-01-15 Electronics And Telecommunications Research Institute Method and apparatus for providing 3d video streaming service
US9049434B2 (en) 2010-03-05 2015-06-02 Panasonic Intellectual Property Management Co., Ltd. 3D imaging device and 3D imaging method
US20150156472A1 (en) * 2012-07-06 2015-06-04 Lg Electronics Inc. Terminal for increasing visual comfort sensation of 3d object and control method thereof
US9082210B2 (en) 2012-09-19 2015-07-14 Ali (Zhuhai) Corporation Method and apparatus for adjusting image depth
US9091851B2 (en) 2010-02-28 2015-07-28 Microsoft Technology Licensing, Llc Light control in head mounted displays
US9097890B2 (en) 2010-02-28 2015-08-04 Microsoft Technology Licensing, Llc Grating in a light transmissive illumination system for see-through near-eye display glasses
US9097891B2 (en) 2010-02-28 2015-08-04 Microsoft Technology Licensing, Llc See-through near-eye display glasses including an auto-brightness control for the display brightness based on the brightness in the environment
US9128281B2 (en) 2010-09-14 2015-09-08 Microsoft Technology Licensing, Llc Eyepiece with uniformly illuminated reflective display
US9129295B2 (en) 2010-02-28 2015-09-08 Microsoft Technology Licensing, Llc See-through near-eye display glasses with a fast response photochromic film system for quick transition from dark to clear
US9134534B2 (en) 2010-02-28 2015-09-15 Microsoft Technology Licensing, Llc See-through near-eye display glasses including a modular image source
EP2432228A3 (en) * 2010-09-17 2015-10-07 Samsung Electronics Co., Ltd. Display apparatus and control method thereof
US9182596B2 (en) 2010-02-28 2015-11-10 Microsoft Technology Licensing, Llc See-through near-eye display glasses with the optical assembly including absorptive polarizers or anti-reflective coatings to reduce stray light
US9185388B2 (en) 2010-11-03 2015-11-10 3Dmedia Corporation Methods, systems, and computer program products for creating three-dimensional video sequences
US20150324951A1 (en) * 2014-05-08 2015-11-12 Glasses.Com Systems and methods for scaling an object
US9188849B2 (en) 2010-03-05 2015-11-17 Panasonic Intellectual Property Management Co., Ltd. 3D imaging device and 3D imaging method
US9215436B2 (en) 2009-06-24 2015-12-15 Dolby Laboratories Licensing Corporation Insertion of 3D objects in a stereoscopic image at relative depth
US9223134B2 (en) 2010-02-28 2015-12-29 Microsoft Technology Licensing, Llc Optical imperfections in a light transmissive illumination system for see-through near-eye display glasses
US9229227B2 (en) 2010-02-28 2016-01-05 Microsoft Technology Licensing, Llc See-through near-eye display glasses with a light transmissive wedge shaped illumination system
EP2757789A4 (en) * 2011-09-16 2016-01-20 Konica Minolta Inc Image processing system, image processing method, and image processing program
US9285589B2 (en) 2010-02-28 2016-03-15 Microsoft Technology Licensing, Llc AR glasses with event and sensor triggered control of AR eyepiece applications
US9329689B2 (en) 2010-02-28 2016-05-03 Microsoft Technology Licensing, Llc Method and apparatus for biometric data capture
US9344701B2 (en) 2010-07-23 2016-05-17 3Dmedia Corporation Methods, systems, and computer-readable storage media for identifying a rough depth map in a scene and for determining a stereo-base distance for three-dimensional (3D) content creation
US9341843B2 (en) 2010-02-28 2016-05-17 Microsoft Technology Licensing, Llc See-through near-eye display glasses with a small scale image source
US9366862B2 (en) 2010-02-28 2016-06-14 Microsoft Technology Licensing, Llc System and method for delivering content to a group of see-through near eye display eyepieces
US9380292B2 (en) 2009-07-31 2016-06-28 3Dmedia Corporation Methods, systems, and computer-readable storage media for generating three-dimensional (3D) images of a scene
US9519994B2 (en) 2011-04-15 2016-12-13 Dolby Laboratories Licensing Corporation Systems and methods for rendering 3D image independent of display size and viewing distance
US9600923B2 (en) 2011-05-26 2017-03-21 Thomson Licensing Scale-independent maps
US9759917B2 (en) 2010-02-28 2017-09-12 Microsoft Technology Licensing, Llc AR glasses with event and sensor triggered AR eyepiece interface to external devices
US10154243B2 (en) 2010-06-28 2018-12-11 Interdigital Madison Patent Holdings Method and apparatus for customizing 3-dimensional effects of stereo content
US10180572B2 (en) 2010-02-28 2019-01-15 Microsoft Technology Licensing, Llc AR glasses with event and user action control of external applications
US10200671B2 (en) 2010-12-27 2019-02-05 3Dmedia Corporation Primary and auxiliary image capture devices for image processing and related methods
US10539787B2 (en) 2010-02-28 2020-01-21 Microsoft Technology Licensing, Llc Head-worn adaptive display
US10860100B2 (en) 2010-02-28 2020-12-08 Microsoft Technology Licensing, Llc AR glasses with predictive control of external device based on event input
US10895868B2 (en) * 2015-04-17 2021-01-19 Tulip Interfaces, Inc. Augmented interface authoring
EP4201060A4 (en) * 2021-01-26 2024-02-14 Samsung Electronics Co Ltd Display apparatus and control method thereof

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100824942B1 (en) 2007-05-31 2008-04-28 한국과학기술원 Method of generating lenticular display image and recording medium thereof
DE102008001451A1 (en) * 2008-04-29 2009-11-05 Robert Bosch Gmbh Camera and method for controlling a camera
JP4737573B2 (en) 2009-02-05 2011-08-03 富士フイルム株式会社 3D image output apparatus and method
JP5501458B2 (en) * 2010-06-11 2014-05-21 富士フイルム株式会社 Stereoscopic image display apparatus, stereoscopic imaging apparatus, and method
JP2012047995A (en) * 2010-08-27 2012-03-08 Fujitsu Ltd Information display device
KR101176500B1 (en) * 2010-10-06 2012-08-22 엘지전자 주식회사 Image display apparatus, and method for operating the same
KR101153120B1 (en) * 2010-10-06 2012-06-04 차형경 Method For Depth Adaptation of Autostereoscopic Image and User Equipment For the Same
JP5876983B2 (en) * 2010-12-29 2016-03-02 任天堂株式会社 Display control program, display control device, display control method, and display control system
JP2012204852A (en) * 2011-03-23 2012-10-22 Sony Corp Image processing apparatus and method, and program
CN102724521A (en) * 2011-03-29 2012-10-10 青岛海信电器股份有限公司 Method and apparatus for stereoscopic display
KR101165810B1 (en) 2011-04-27 2012-07-16 주식회사 아이브이넷 A method and an apparatus extracting a depth image information using a stereo camera
JP6021063B2 (en) * 2011-07-26 2016-11-02 パナソニックIpマネジメント株式会社 Movie editing apparatus, movie editing method, program, and integrated circuit
KR101888672B1 (en) * 2011-07-27 2018-08-16 엘지디스플레이 주식회사 Streoscopic image display device and method for driving thereof
KR101247501B1 (en) 2011-08-30 2013-03-25 김재도 Stereopic 3 Dimension Display Device
KR101246846B1 (en) 2011-08-30 2013-03-25 김재도 Method for displaying of Stereopic 3 Dimension
KR20160073787A (en) 2014-12-17 2016-06-27 삼성전자주식회사 Method and apparatus for generating 3d image on curved display

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5801760A (en) * 1993-08-26 1998-09-01 Matsushita Electric Industrial Co., Ltd. Stereoscopic image pickup and display apparatus
US6160909A (en) * 1998-04-01 2000-12-12 Canon Kabushiki Kaisha Depth control for stereoscopic images
US6798406B1 (en) * 1999-09-15 2004-09-28 Sharp Kabushiki Kaisha Stereo images with comfortable perceived depth
US20050201612A1 (en) * 2004-03-04 2005-09-15 Samsung Electronics Co.,Ltd. Method and apparatus for detecting people using stereo camera
US20060203085A1 (en) * 2002-11-28 2006-09-14 Seijiro Tomita There dimensional image signal producing circuit and three-dimensional image display apparatus
US20090022393A1 (en) * 2005-04-07 2009-01-22 Visionsense Ltd. Method for reconstructing a three-dimensional surface of an object

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6005607A (en) 1995-06-29 1999-12-21 Matsushita Electric Industrial Co., Ltd. Stereoscopic computer graphics image generating apparatus and stereoscopic TV apparatus
KR100294925B1 (en) 1999-06-03 2001-07-12 윤종용 3-D graphic image manufacturing method and binocular visual disparity adjustment method therefor
EP1085769B1 (en) 1999-09-15 2012-02-01 Sharp Kabushiki Kaisha Stereoscopic image pickup apparatus
JP2002228976A (en) 2001-02-01 2002-08-14 Mixed Reality Systems Laboratory Inc Stereoscopic image display device and stereoscopic image display method
JP2003107603A (en) 2001-09-28 2003-04-09 Namco Ltd Stereophonic image generating device, stereophonic image generation information and information storage medium
KR100439341B1 (en) 2002-08-27 2004-07-07 한국전자통신연구원 Depth of field adjustment apparatus and method of stereo image for reduction of visual fatigue
JP4228646B2 (en) 2002-10-02 2009-02-25 株式会社セガ Stereoscopic image generation method and stereoscopic image generation apparatus
JP3978392B2 (en) 2002-11-28 2007-09-19 誠次郎 富田 3D image signal generation circuit and 3D image display device
JP2005073049A (en) 2003-08-26 2005-03-17 Sharp Corp Device and method for reproducing stereoscopic image

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5801760A (en) * 1993-08-26 1998-09-01 Matsushita Electric Industrial Co., Ltd. Stereoscopic image pickup and display apparatus
US6160909A (en) * 1998-04-01 2000-12-12 Canon Kabushiki Kaisha Depth control for stereoscopic images
US6798406B1 (en) * 1999-09-15 2004-09-28 Sharp Kabushiki Kaisha Stereo images with comfortable perceived depth
US20060203085A1 (en) * 2002-11-28 2006-09-14 Seijiro Tomita There dimensional image signal producing circuit and three-dimensional image display apparatus
US20050201612A1 (en) * 2004-03-04 2005-09-15 Samsung Electronics Co.,Ltd. Method and apparatus for detecting people using stereo camera
US20090022393A1 (en) * 2005-04-07 2009-01-22 Visionsense Ltd. Method for reconstructing a three-dimensional surface of an object

Cited By (115)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8200045B2 (en) * 2007-02-07 2012-06-12 Thomson Licensing Image processing method
EP1978755A2 (en) * 2007-04-04 2008-10-08 Mitsubishi Electric Corporation Method and system for acquiring, encoding, decoding and displaying 3D light fields
WO2009020277A1 (en) * 2007-08-06 2009-02-12 Samsung Electronics Co., Ltd. Method and apparatus for reproducing stereoscopic image using depth control
US20090324114A1 (en) * 2008-06-26 2009-12-31 Sony Corporation Image compression apparatus and image compression method
US8204319B2 (en) * 2008-06-26 2012-06-19 Sony Corporation Image compression apparatus and image compression method
US20110122131A1 (en) * 2008-07-24 2011-05-26 Koninklijke Philips Electronics N.V. Versatile 3-d picture format
US10567728B2 (en) 2008-07-24 2020-02-18 Koninklijke Philips N.V. Versatile 3-D picture format
US9432651B2 (en) 2008-07-24 2016-08-30 Koninklijke Philips N.V. Versatile 3-D picture format
WO2010087575A2 (en) * 2009-02-01 2010-08-05 Lg Electronics Inc. Broadcast receiver and 3d video data processing method
WO2010087575A3 (en) * 2009-02-01 2011-09-15 Lg Electronics Inc. Broadcast receiver and 3d video data processing method
US9756309B2 (en) 2009-02-01 2017-09-05 Lg Electronics Inc. Broadcast receiver and 3D video data processing method
US8803945B2 (en) 2009-02-01 2014-08-12 Lg Electronics Inc. Broadcast receiver and 3D video data processing method
EP2384585A4 (en) * 2009-02-01 2017-03-15 LG Electronics Inc. Broadcast receiver and 3d video data processing method
US9215436B2 (en) 2009-06-24 2015-12-15 Dolby Laboratories Licensing Corporation Insertion of 3D objects in a stereoscopic image at relative depth
US9380292B2 (en) 2009-07-31 2016-06-28 3Dmedia Corporation Methods, systems, and computer-readable storage media for generating three-dimensional (3D) images of a scene
US8810635B2 (en) 2009-07-31 2014-08-19 3Dmedia Corporation Methods, systems, and computer-readable storage media for selecting image capture positions to generate three-dimensional images
US11044458B2 (en) 2009-07-31 2021-06-22 3Dmedia Corporation Methods, systems, and computer-readable storage media for generating three-dimensional (3D) images of a scene
US20110032341A1 (en) * 2009-08-04 2011-02-10 Ignatov Artem Konstantinovich Method and system to transform stereo content
US20120206453A1 (en) * 2009-09-16 2012-08-16 Koninklijke Philips Electronics N.V. 3d screen size compensation
US20110285819A1 (en) * 2009-09-17 2011-11-24 Panasonic Corporation Video signal processing apparatus and video signal processing method
EP2525579A4 (en) * 2010-01-11 2014-11-12 Lg Electronics Inc Broadcasting receiver and method for displaying 3d images
US9485489B2 (en) 2010-01-11 2016-11-01 Lg Electronics Inc. Broadcasting receiver and method for displaying 3D images
EP2525579A2 (en) * 2010-01-11 2012-11-21 LG Electronics Inc. Broadcasting receiver and method for displaying 3d images
US9215444B2 (en) 2010-01-11 2015-12-15 Lg Electronics Inc. Broadcasting receiver and method for displaying 3D images
US20110187831A1 (en) * 2010-02-03 2011-08-04 Korea Institute Of Science And Technology Apparatus and method for displaying three-dimensional images
US8994791B2 (en) 2010-02-03 2015-03-31 Korea Institute Of Science And Technology Apparatus and method for displaying three-dimensional images
US9366862B2 (en) 2010-02-28 2016-06-14 Microsoft Technology Licensing, Llc System and method for delivering content to a group of see-through near eye display eyepieces
US9285589B2 (en) 2010-02-28 2016-03-15 Microsoft Technology Licensing, Llc AR glasses with event and sensor triggered control of AR eyepiece applications
US10539787B2 (en) 2010-02-28 2020-01-21 Microsoft Technology Licensing, Llc Head-worn adaptive display
US10268888B2 (en) 2010-02-28 2019-04-23 Microsoft Technology Licensing, Llc Method and apparatus for biometric data capture
US9229227B2 (en) 2010-02-28 2016-01-05 Microsoft Technology Licensing, Llc See-through near-eye display glasses with a light transmissive wedge shaped illumination system
US10860100B2 (en) 2010-02-28 2020-12-08 Microsoft Technology Licensing, Llc AR glasses with predictive control of external device based on event input
US10180572B2 (en) 2010-02-28 2019-01-15 Microsoft Technology Licensing, Llc AR glasses with event and user action control of external applications
US9329689B2 (en) 2010-02-28 2016-05-03 Microsoft Technology Licensing, Llc Method and apparatus for biometric data capture
US9134534B2 (en) 2010-02-28 2015-09-15 Microsoft Technology Licensing, Llc See-through near-eye display glasses including a modular image source
US9875406B2 (en) 2010-02-28 2018-01-23 Microsoft Technology Licensing, Llc Adjustable extension for temple arm
US9129295B2 (en) 2010-02-28 2015-09-08 Microsoft Technology Licensing, Llc See-through near-eye display glasses with a fast response photochromic film system for quick transition from dark to clear
US9759917B2 (en) 2010-02-28 2017-09-12 Microsoft Technology Licensing, Llc AR glasses with event and sensor triggered AR eyepiece interface to external devices
US9097891B2 (en) 2010-02-28 2015-08-04 Microsoft Technology Licensing, Llc See-through near-eye display glasses including an auto-brightness control for the display brightness based on the brightness in the environment
US20130278631A1 (en) * 2010-02-28 2013-10-24 Osterhout Group, Inc. 3d positioning of augmented reality information
US9223134B2 (en) 2010-02-28 2015-12-29 Microsoft Technology Licensing, Llc Optical imperfections in a light transmissive illumination system for see-through near-eye display glasses
US9097890B2 (en) 2010-02-28 2015-08-04 Microsoft Technology Licensing, Llc Grating in a light transmissive illumination system for see-through near-eye display glasses
US9091851B2 (en) 2010-02-28 2015-07-28 Microsoft Technology Licensing, Llc Light control in head mounted displays
US9341843B2 (en) 2010-02-28 2016-05-17 Microsoft Technology Licensing, Llc See-through near-eye display glasses with a small scale image source
US9182596B2 (en) 2010-02-28 2015-11-10 Microsoft Technology Licensing, Llc See-through near-eye display glasses with the optical assembly including absorptive polarizers or anti-reflective coatings to reduce stray light
JP2011188108A (en) * 2010-03-05 2011-09-22 Mitsubishi Electric Corp Image processing apparatus, three-dimensional image display device, and image processing method
US20120327191A1 (en) * 2010-03-05 2012-12-27 Panasonic Corporation 3d imaging device and 3d imaging method
US9188849B2 (en) 2010-03-05 2015-11-17 Panasonic Intellectual Property Management Co., Ltd. 3D imaging device and 3D imaging method
US9128367B2 (en) * 2010-03-05 2015-09-08 Panasonic Intellectual Property Management Co., Ltd. 3D imaging device and 3D imaging method
US9049434B2 (en) 2010-03-05 2015-06-02 Panasonic Intellectual Property Management Co., Ltd. 3D imaging device and 3D imaging method
WO2011107550A1 (en) * 2010-03-05 2011-09-09 Sony Corporation Disparity distribution estimation for 3d tv
US9426441B2 (en) 2010-03-08 2016-08-23 Dolby Laboratories Licensing Corporation Methods for carrying and transmitting 3D z-norm attributes in digital TV closed captioning
US20110216163A1 (en) * 2010-03-08 2011-09-08 Dolby Laboratories Licensing Corporation Methods For Carrying And Transmitting 3D Z-Norm Attributes In Digital TV Closed Captioning
CN106131531A (en) * 2010-03-31 2016-11-16 汤姆森特许公司 Method for processing video frequency and device
US10791314B2 (en) 2010-03-31 2020-09-29 Interdigital Ce Patent Holdings, Sas 3D disparity maps
WO2011121437A1 (en) * 2010-03-31 2011-10-06 Thomson Licensing 3d disparity maps
WO2011121397A1 (en) * 2010-04-01 2011-10-06 Nokia Corporation Method, apparatus and computer program for selecting a stereoscopic imaging viewpoint pair
US20110249017A1 (en) * 2010-04-07 2011-10-13 Sony Corporation Image signal processing device, display device, display method and program product
US8542909B2 (en) 2010-04-09 2013-09-24 Tektronix, Inc. Method and apparatus for measuring an audiovisual parameter
EP2375766A3 (en) * 2010-04-09 2013-05-29 Tektronix International Sales GmbH Method and apparatus for measuring an audiovisual parameter
US8810564B2 (en) * 2010-05-03 2014-08-19 Samsung Electronics Co., Ltd. Apparatus and method for reducing three-dimensional visual fatigue
US20110267338A1 (en) * 2010-05-03 2011-11-03 Kwangwoon University Industry-Academic Collaboration Foundation Apparatus and method for reducing three-dimensional visual fatigue
WO2011149967A3 (en) * 2010-05-28 2012-01-12 Qualcomm Incorporated Three-dimensional image processing
US8970672B2 (en) * 2010-05-28 2015-03-03 Qualcomm Incorporated Three-dimensional image processing
US20110292178A1 (en) * 2010-05-28 2011-12-01 Qualcomm Incorporated Three-dimensional image processing
US20110304708A1 (en) * 2010-06-10 2011-12-15 Samsung Electronics Co., Ltd. System and method of generating stereo-view and multi-view images for rendering perception of depth of stereoscopic image
US10154243B2 (en) 2010-06-28 2018-12-11 Interdigital Madison Patent Holdings Method and apparatus for customizing 3-dimensional effects of stereo content
US9420257B2 (en) 2010-07-14 2016-08-16 Lg Electronics Inc. Mobile terminal and method for adjusting and displaying a stereoscopic image
US20120013604A1 (en) * 2010-07-14 2012-01-19 Samsung Electronics Co., Ltd. Display apparatus and method for setting sense of depth thereof
EP2408213A3 (en) * 2010-07-14 2012-10-24 LG Electronics Inc. Mobile terminal and controlling method thereof
US9344701B2 (en) 2010-07-23 2016-05-17 3Dmedia Corporation Methods, systems, and computer-readable storage media for identifying a rough depth map in a scene and for determining a stereo-base distance for three-dimensional (3D) content creation
US9288482B2 (en) * 2010-08-16 2016-03-15 Lg Electronics Inc. Method for processing images in display device outputting 3-dimensional contents and display device using the same
KR101763592B1 (en) * 2010-08-16 2017-08-01 엘지전자 주식회사 Method for processing image of display system outputting 3 dimensional contents and display system enabling of the method
US20120038754A1 (en) * 2010-08-16 2012-02-16 Na Sungbin Method for processing images in display device outputting 3-dimensional contents and display device using the same
US9128281B2 (en) 2010-09-14 2015-09-08 Microsoft Technology Licensing, Llc Eyepiece with uniformly illuminated reflective display
EP2432228A3 (en) * 2010-09-17 2015-10-07 Samsung Electronics Co., Ltd. Display apparatus and control method thereof
TWI491216B (en) * 2010-09-23 2015-07-01 Thomson Licensing Adaptation of 3d video content
US9204122B2 (en) * 2010-09-23 2015-12-01 Thomson Licensing Adaptation of 3D video content
US20120075419A1 (en) * 2010-09-23 2012-03-29 Thomson Licensing Adaptation of 3D video content
EP2442577A3 (en) * 2010-10-15 2012-05-16 Sony Corporation Information processing apparatus, informaton processing method and program
US9001188B2 (en) 2010-10-15 2015-04-07 Sony Corporation Information processing apparatus, information processing method and program
US9185388B2 (en) 2010-11-03 2015-11-10 3Dmedia Corporation Methods, systems, and computer program products for creating three-dimensional video sequences
US20120120202A1 (en) * 2010-11-12 2012-05-17 Gwangju Institute Of Science And Technology Method for improving 3 dimensional effect and reducing visual fatigue and apparatus enabling the same
US8760502B2 (en) * 2010-11-12 2014-06-24 Samsung Electronics Co., Ltd. Method for improving 3 dimensional effect and reducing visual fatigue and apparatus enabling the same
US10911737B2 (en) 2010-12-27 2021-02-02 3Dmedia Corporation Primary and auxiliary image capture devices for image processing and related methods
US11388385B2 (en) 2010-12-27 2022-07-12 3Dmedia Corporation Primary and auxiliary image capture devices for image processing and related methods
US10200671B2 (en) 2010-12-27 2019-02-05 3Dmedia Corporation Primary and auxiliary image capture devices for image processing and related methods
US9800865B2 (en) 2011-02-23 2017-10-24 Nintendo European Research And Development Sas Device and method for managing the position of the focal plane in a stereoscopic scene
FR2971875A1 (en) * 2011-02-23 2012-08-24 Mobiclip DEVICE AND METHOD FOR MANAGING THE POSITION OF THE FOCAL PLANE IN A STEREOSCOPIC SCENE
WO2012113797A1 (en) * 2011-02-23 2012-08-30 Mobiclip Device and method for managing the position of the focal plane in a stereoscopic scene
US8823777B2 (en) * 2011-03-30 2014-09-02 Intel Corporation Real-time depth extraction using stereo correspondence
US20120249747A1 (en) * 2011-03-30 2012-10-04 Ziv Aviv Real-time depth extraction using stereo correspondence
WO2012135220A3 (en) * 2011-03-30 2013-01-03 Intel Corporation Real-time depth extraction using stereo correspondence
US9519994B2 (en) 2011-04-15 2016-12-13 Dolby Laboratories Licensing Corporation Systems and methods for rendering 3D image independent of display size and viewing distance
CN103609104A (en) * 2011-05-23 2014-02-26 高通股份有限公司 Interactive user interface for stereoscopic effect adjustment
WO2012162096A1 (en) * 2011-05-23 2012-11-29 Qualcomm Incorporated Interactive user interface for stereoscopic effect adjustment
JP2016192773A (en) * 2011-05-23 2016-11-10 クゥアルコム・インコーポレイテッドQualcomm Incorporated Interactive user interface for stereoscopic effect adjustment
JP2014517619A (en) * 2011-05-23 2014-07-17 クゥアルコム・インコーポレイテッド Interactive user interface for adjusting stereoscopic effects
US9600923B2 (en) 2011-05-26 2017-03-21 Thomson Licensing Scale-independent maps
EP2757789A4 (en) * 2011-09-16 2016-01-20 Konica Minolta Inc Image processing system, image processing method, and image processing program
US20130141423A1 (en) * 2011-12-06 2013-06-06 Jung-hyun Cho Three-dimensional image display apparatus
US9300948B2 (en) * 2011-12-06 2016-03-29 Samsung Display Co., Ltd. Three-dimensional image display apparatus
WO2013158784A1 (en) * 2012-04-17 2013-10-24 3Dmedia Corporation Systems and methods for improving overall quality of three-dimensional content by altering parallax budget or compensating for moving objects
US20150156472A1 (en) * 2012-07-06 2015-06-04 Lg Electronics Inc. Terminal for increasing visual comfort sensation of 3d object and control method thereof
US9674501B2 (en) * 2012-07-06 2017-06-06 Lg Electronics Inc. Terminal for increasing visual comfort sensation of 3D object and control method thereof
US20140063206A1 (en) * 2012-08-28 2014-03-06 Himax Technologies Limited System and method of viewer centric depth adjustment
US9082210B2 (en) 2012-09-19 2015-07-14 Ali (Zhuhai) Corporation Method and apparatus for adjusting image depth
CN103024591A (en) * 2012-12-26 2013-04-03 新奥特(北京)视频技术有限公司 Method and device for adjusting 3D (three-dimensional) parallax
US20140292748A1 (en) * 2013-04-01 2014-10-02 Electronics And Telecommunications Research Institute System and method for providing stereoscopic image by adjusting depth value
US20150015666A1 (en) * 2013-07-09 2015-01-15 Electronics And Telecommunications Research Institute Method and apparatus for providing 3d video streaming service
US9996899B2 (en) * 2014-05-08 2018-06-12 Glasses.Com Inc. Systems and methods for scaling an object
US20150324951A1 (en) * 2014-05-08 2015-11-12 Glasses.Com Systems and methods for scaling an object
US10895868B2 (en) * 2015-04-17 2021-01-19 Tulip Interfaces, Inc. Augmented interface authoring
US10996660B2 (en) 2015-04-17 2021-05-04 Tulip Interfaces, Ine. Augmented manufacturing system
EP4201060A4 (en) * 2021-01-26 2024-02-14 Samsung Electronics Co Ltd Display apparatus and control method thereof

Also Published As

Publication number Publication date
NL1032382A1 (en) 2007-03-01
US8290244B2 (en) 2012-10-16
NL1032382C2 (en) 2011-02-10
CN1925627A (en) 2007-03-07
CN100574462C (en) 2009-12-23
KR100667810B1 (en) 2007-01-11

Similar Documents

Publication Publication Date Title
US8290244B2 (en) Apparatus and method for controlling depth of three-dimensional image
CN102474644B (en) Stereo image display system, parallax conversion equipment, parallax conversion method
CA2723627C (en) System and method for measuring potential eyestrain of stereoscopic motion pictures
US8116557B2 (en) 3D image processing apparatus and method
JP6147275B2 (en) Stereoscopic image processing apparatus, stereoscopic image processing method, and program
JP3826236B2 (en) Intermediate image generation method, intermediate image generation device, parallax estimation method, and image transmission display device
US7944444B2 (en) 3D image processing apparatus and method
US9445071B2 (en) Method and apparatus generating multi-view images for three-dimensional display
TWI493505B (en) Image processing method and image processing apparatus thereof
US20110304708A1 (en) System and method of generating stereo-view and multi-view images for rendering perception of depth of stereoscopic image
KR20100040593A (en) Apparatus and method for image processing
US20150022631A1 (en) Content-aware display adaptation methods and editing interfaces and methods for stereoscopic images
KR20150121127A (en) Binocular fixation imaging method and apparatus
KR101105359B1 (en) Fatigue evaluation method and apparatus for 3D video based on depth map
US8766974B2 (en) Display apparatus and method
WO2011030399A1 (en) Image processing method and apparatus
US8884951B2 (en) Depth estimation data generating apparatus, depth estimation data generating method, and depth estimation data generating program, and pseudo three-dimensional image generating apparatus, pseudo three-dimensional image generating method, and pseudo three-dimensional image generating program
US20170309055A1 (en) Adjusting parallax of three-dimensional display material
US8976171B2 (en) Depth estimation data generating apparatus, depth estimation data generating method, and depth estimation data generating program, and pseudo three-dimensional image generating apparatus, pseudo three-dimensional image generating method, and pseudo three-dimensional image generating program
JP2013090129A (en) Image processing apparatus, image processing method and program
JP4815004B2 (en) Multi-view image encoding device
KR101521213B1 (en) Apparatus for correcting stereoscopic image using visual discomfort model and method thereof
KR101326279B1 (en) Motion estimation system for stereoscopic videos and method thereof
KR20130051119A (en) Image display apparatus and method for operating the same

Legal Events

Date Code Title Description
AS Assignment

Owner name: SAMSUNG ELECTRONICS CO., LTD., KOREA, REPUBLIC OF

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HA, TAE-HYEUN;REEL/FRAME:018135/0150

Effective date: 20060517

STCF Information on status: patent grant

Free format text: PATENTED CASE

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

FPAY Fee payment

Year of fee payment: 4

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 12