US20070047040A1 - Apparatus and method for controlling depth of three-dimensional image - Google Patents
Apparatus and method for controlling depth of three-dimensional image Download PDFInfo
- Publication number
- US20070047040A1 US20070047040A1 US11/492,970 US49297006A US2007047040A1 US 20070047040 A1 US20070047040 A1 US 20070047040A1 US 49297006 A US49297006 A US 49297006A US 2007047040 A1 US2007047040 A1 US 2007047040A1
- Authority
- US
- United States
- Prior art keywords
- depth
- eye image
- physical distance
- value
- controlling
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 41
- 230000002194 synthesizing effect Effects 0.000 claims description 2
- 238000004590 computer program Methods 0.000 claims 1
- 238000004519 manufacturing process Methods 0.000 abstract description 5
- 230000000694 effects Effects 0.000 description 6
- 208000003464 asthenopia Diseases 0.000 description 5
- 230000008901 benefit Effects 0.000 description 2
- 238000013500 data storage Methods 0.000 description 2
- 230000004888 barrier function Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/111—Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B30/00—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images
- G02B30/50—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images the image being built up from image elements distributed over a 3D volume, e.g. voxels
- G02B30/52—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images the image being built up from image elements distributed over a 3D volume, e.g. voxels the 3D volume being constructed from a stack or sequence of 2D planes, e.g. depth sampling systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/122—Improving the 3D impression of stereoscopic images by modifying image signal contents, e.g. by filtering or adding monoscopic depth cues
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/128—Adjusting depth or disparity
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/139—Format conversion, e.g. of frame-rate or size
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/144—Processing image signals for flicker reduction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/398—Synchronisation thereof; Control thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N2213/00—Details of stereoscopic systems
- H04N2213/002—Eyestrain reduction by processing stereoscopic signals or controlling stereoscopic devices
Definitions
- the present invention relates to an apparatus and method for controlling the depth of a three-dimensional image, and more particularly, to an apparatus and method for adaptively controlling disparity to control the image depth when a user uses a stereoscopic three-dimensional display of a different size than a display used in a manufacturing environment.
- Current three dimensional displays generally use a method of projecting images having different disparities to the left and right eyes of a user and applying a three-dimensional effect to the displayed image by adding a film-shaped micro-polarizer, a barrier, or a lenticular lens to a two-dimensional flat display.
- Three-dimensional displays using this method are classified into stereoscopic three-dimensional displays and multi-view three-dimensional displays in accordance with the number of views of the display.
- the stereoscopic three-dimensional display displays images of two views and the multi-view three-dimensional display displays images of three or more views.
- a realistic image is created using a stereoscopic camera having two incident lenses or a multi-view camera having several incident lenses, and stereo or multi-view contents are created by computer graphic technology.
- the image or the image contents are created in consideration of human visual characteristics. However, if the characteristics such as size and resolution of the display used in the manufacturing environment are different from those of the display used by the user, the desired three-dimensional image cannot be displayed.
- the present invention provides an apparatus and method for adaptively controlling disparity to control the image depth when a user uses a stereoscopic three-dimensional display having a different size than a display used in a manufacturing environment.
- an apparatus for controlling the depth of a three-dimensional image including: a disparity measuring unit which measures the disparity between a left eye image and a right eye image; a physical distance calculating unit which calculates the physical distance between the left eye image and the right eye image based on the measured disparity and the physical characteristics of a display; and a depth controlling unit which controls the depth depending on the calculated physical distance.
- the disparity measuring unit may estimate the disparity between the left eye image and the right eye image and determine any one of a maximum value, a minimum value, and a mean value of the disparity as a reference value.
- the physical characteristics of the display may include the physical size of one pixel in the display, and the physical distance calculating unit may read the physical size of one pixel which has been previously stored and multiply the reference value by the physical size of that pixel to calculate the physical distance of the disparity.
- the depth controlling unit may control the physical distance based on a threshold value determined depending on the physical characteristics of the display.
- the depth controlling unit may compare a maximum value of the calculated physical distance with a maximum threshold value of the physical distance, control the maximum value to less than the maximum threshold value if the calculated maximum value is greater than the maximum threshold value, and control the physical distance between objects in the images in proportion to the controlled maximum value.
- the depth controlling unit may compare a mean value of the calculated physical distance with a mean threshold value of the physical distance, control the mean value to less than the mean threshold value if the calculated mean value is greater than the mean threshold value, and control the physical distance between objects in the images in proportion to the controlled mean value.
- the depth controlling unit may compare a maximum value of the calculated physical distance with a minimum threshold value of the physical distance, control the maximum value to greater than the minimum threshold value if the calculated maximum value is less than the minimum threshold value, and control the physical distance between objects in the images in proportion to the controlled maximum value.
- the depth controlling unit may reconstruct an intermediate image using an intermediate view reconstruction method based on the controlled physical distance and replace at least one of the left eye image and the right eye image with the reconstructed intermediate image to control the depth.
- the depth controlling unit may horizontally move the left eye image and the right eye image based on the controlled physical distance and synthesize the horizontally moved left eye image and right eye image to control the depth.
- the apparatus for controlling the depth may further include a display unit which displays the image having the controlled depth, output from the depth controlling unit.
- a method for controlling the depth of a three-dimensional image including: measuring the disparity between a left eye image and a right eye image; calculating the physical distance between the left eye image and the right eye image based on the measured disparity and the physical characteristics of a display; and controlling the depth depending on the calculated physical distance.
- FIGS. 1A-1C illustrate a disparity variation between a left eye image and a right eye image in accordance with the variation of the screen size of a display
- FIG. 2 illustrates the relationship between the depth and the size of an object when a user views a large object and a small object at the same position
- FIGS. 3A-3C illustrate the variation of the depth of an object depending on the screen size of a display and the position of a user
- FIG. 4 illustrates the structure of an apparatus for adaptively controlling the depth of a three-dimensional image according to an embodiment of the present invention
- FIG. 5 illustrates a process of controlling the depth using an intermediate view reconstruction method in a depth controlling unit illustrated in FIG. 4 ;
- FIG. 6 illustrates a process of horizontally moving a left eye image and a right eye image to control image depth in the depth controlling unit illustrated in FIG. 4 ;
- FIG. 7 is a flowchart illustrating a method for adaptively controlling the depth of a three-dimensional image according to an embodiment of the present invention.
- FIGS. 1A-1C a disparity variation between a left eye image and a right eye image in accordance with the variation of the screen size of a display.
- the content When content is created for a display screen illustrated in FIG. 1B , the content has an optimal depth D 2 when reproduced on the same display screen. Accordingly, if the content is intended for a three-dimensional display having a screen size of N2 ⁇ M2 (inch 2 ) illustrated in FIG. 1B but a user watches the content on a display having a smaller screen size of N1 ⁇ M1 (inch 2 ) illustrated in FIG. 1A , the depth of the three-dimensional image is reduced and thus the user feels as if the object on the screen is distant. That is, the three-dimensional effect deteriorates.
- the user When the user watches the three-dimensional content on a display having a larger screen size of N3 ⁇ M3 (inch 2 ), the user feels as if the object on the screen is very close, which can cause fatigue or eye strain. In some cases, the object may have a depth which cannot be perceived by a human.
- FIG. 2 illustrates the relationship between the depth and the size of an object when a user views a large object and a small object at the same position.
- the depth does not actually vary even if the size of the object varies.
- the disparity between a left eye image and a right eye image is proportional to the depth perceived by the user and varies depending on the screen size of the display, the same problem as described with reference to FIG. 1 occurs.
- FIGS. 3A-3C the variation of the depth of an object depending on the screen size of a display and the position of a user.
- the depth when the user watches the image content on a display having a different screen size from the display used when making the content, the depth must be controlled such that the image is suitably displayed on the user's display.
- FIG. 4 illustrates the structure of an apparatus for adaptively controlling the depth of a three-dimensional image according to an embodiment of the present invention.
- the apparatus 400 for controlling the depth includes a disparity measuring unit 410 , a physical distance calculating unit 420 , and a depth controlling unit 430 .
- the disparity measuring unit 410 divides an input three-dimensional image into a left eye image and a right eye image and measures the disparity between the left eye image and the right eye image.
- the disparity measuring unit 410 estimates the disparity between the left eye image and the right eye image.
- the method of estimating the disparity may use block-based movement estimation as used by MPEG encoding. First, the left eye image is divided into N ⁇ N uniform blocks. Subsequently, for each block in the left eye image, the most similar block is estimated in the right eye image using a sum of absolute difference (SAD) or a mean of absolute difference (MAD) calculation.
- SAD sum of absolute difference
- MAD mean of absolute difference
- the distance between a reference block and its estimated similar block forms a disparity vector, measured in pixels. That is, the disparity between the left eye image and the right eye image at a specific position is represented by the distance in pixels between the same objects in the left eye image and the right eye image at that point.
- the disparity measuring unit 410 determines any one of a maximum value d(max), a minimum value d(min), and a mean value d(mean) of the estimated disparities as a reference value and sends it to the physical distance calculating unit 420 .
- the physical distance calculating unit 420 calculates the physical distance between the left eye image and the right eye image based on the measured disparity and the physical characteristics of the display.
- the physical distance calculating unit 420 reads the physical size P of one pixel displayed on the screen and multiplies the physical size P of the pixel by the reference value determined by the disparity measuring unit 410 , to calculate the physical distance between the objects in the left eye image and the right eye image.
- the physical size P of the pixel may be stored in a storage unit such as a ROM of the display.
- the depth controlling unit 430 controls the calculated physical distance based on a threshold value determined in accordance with the physical characteristics of the display, and controls the depth in accordance with the controlled distance.
- the depth controlling unit 430 reconstructs the image using an intermediate view reconstruction method, based on the controlled distance, and replaces either the left eye image or the right eye image with the reconstructed image, and outputs the replacement image and the unchanged other eye image. When these are displayed, an image having controlled depth is viewed.
- the depth control unit 430 horizontally moves the left eye image and the right eye image based on the controlled distance, synthesizes the horizontally moved left eye image and right eye image, and outputs the image having the controlled depth.
- the depth controlling apparatus 400 may include a display unit (not shown) for displaying the three-dimensional image output from depth controlling unit 430 to embody a display system for outputting the three-dimensional image.
- FIG. 5 illustrates a process of controlling the depth using the intermediate view reconstruction method in the depth controlling unit 430 illustrated in FIG. 4 .
- the depth controlling unit 430 controls the depth based on the physical distance calculated by the physical distance calculating unit 420 . If the physical distance must be controlled, the depth controlling unit 430 performs the following operations.
- the depth controlling unit 430 receives the maximum value of the physical distance calculated by the physical distance calculating unit 420 and compares the received maximum value with a maximum threshold value of the physical distance. If the received maximum value is greater than the maximum threshold value, the depth controlling unit 430 controls the maximum value to be less than the maximum threshold value, and controls the physical distance between the objects in the images in proportion to the controlled value. That is, the value obtained by multiplying the maximum value of the disparity between the objects in the images by the physical size of one pixel of the display is controlled to less than the predetermined threshold value and the total disparity of the image is controlled in proportion to the controlled value.
- the maximum value of the physical distance, D(max), is obtained by multiplying the maximum value d(max) of the measured disparities by the physical size P of one pixel.
- the maximum threshold value th(max) is the maximum value of the physical distance between the left eye image 51 and the right eye image 53 in the display as used by a user within a range which does not cause eyestrain or other fatigue, as determined experimentally.
- the maximum threshold value can be obtained from the depth having the maximum value within a range wherein fatigue is not induced.
- d 1 denotes the distance from the left edge to the object in the left eye image 51
- d 2 denotes the distance from the left edge to the object in the right eye image 53
- d 3 denotes the disparity between the objects in the left eye image 51 and the right eye image 53 .
- d 4 denotes the disparity which is controlled to prevent user fatigue. That is, if D(max)>th(max), a new image having the disparity d 4 is generated using Equation 1, and the existing right eye image 53 is replaced with the new image, thereby reducing user fatigue.
- the depth controlling unit 430 receives the mean value of the physical distance calculated by the physical distance calculating unit 420 and compares it with a mean threshold value of the physical distance. Then, if the calculated mean value is greater than the mean threshold value, the depth controlling unit 430 controls the calculated mean value to less than the mean threshold value, and controls the physical distance between the objects in the images in proportion to the controlled mean value.
- the mean value of the physical distance, D(mean) is obtained by multiplying the mean value d(mean) of the disparities calculated by the disparity measuring unit 410 by the physical size P of one pixel.
- the mean threshold value th(mean) is the average of the physical distances between the left eye image 51 and the right eye image 53 in the display as used by the user within the range which does not cause eyestrain or other fatigue, as determined experimentally.
- the depth controlling unit 430 compares the maximum value of the physical distance calculated by the physical distance calculating unit 420 with a minimum threshold value of the physical distance. If the calculated maximum value is less than the minimum threshold value, the depth controlling unit 430 controls the calculated maximum value to greater than the minimum threshold value, and controls the physical distance between the objects in the images in proportion to the controlled maximum value.
- the minimum threshold value th(min) is the minimum value of the physical distance between the left eye image 51 and the right eye image 53 in the display as used by the user, within the range which does not cause eyestrain or other fatigue, as determined experimentally.
- the image having the physical distance of d 4 ⁇ P is reconstructed by the intermediate view reconstruction method, the right eye image 53 is replaced with the reconstructed intermediate image 55 , and the left eye image 51 and the intermediate image 55 overlap each other and are output.
- the proper disparity d 4 or the proper physical distance d 4 ⁇ P is determined in order to control the depth, any one or at least two of Equations 1, 2, and 3 may be used in accordance with the characteristics of the input three-dimensional image.
- the right eye image 53 was replaced with the reconstructed image using a method of determining proper disparity based on the distance from the left edge to the object.
- the proper disparity may be calculated based on the distance from the right edge to the object.
- the left eye image 51 is replaced with the reconstructed image using the intermediate view reconstruction method. That is, the image 53 of FIG. 5 and the reconstructed image reconstructed by the intermediate view reconstruction method overlap each other and are output.
- FIG. 6 illustrates a process of horizontally moving the left eye image 51 and the right eye image 53 to control the depth in the depth controlling unit 430 illustrated in FIG. 4 .
- Left eye image 61 is the image obtained by horizontally moving the left eye image 51 of FIG. 5 to the right side, and a hatched region indicates a portion which is lost in the process.
- Right eye image 63 is the image obtained by horizontally moving the right eye image 53 of FIG. 5 to the left side, and a hatched region indicates a portion which is lost in the process.
- An image 65 is obtained by synthesizing the horizontally moved left eye image 61 and right eye image 63 . Accordingly, the left eye image 51 and the right eye image 53 must be moved horizontally such that the disparity between the left and right eye images is controlled to the disparity d 4 as determined above.
- the left eye image 51 is horizontally moved by d 3 -d 4 /2 to the right side and the right eye image 53 is horizontally moved by d 3 -d 4 /2 to the left side.
- the three-dimensional image 65 having the disparity d 4 can be generated. Similar to the method described with reference to FIG. 5 , when the proper disparity d 4 or the proper physical distance d 4 ⁇ P is determined in order to control the depth, any one or at least two of Equations 1, 2, and 3 may be used in accordance with the characteristics of the input three-dimensional image.
- FIG. 7 is a flowchart illustrating a method of adaptively controlling the depth of a three-dimensional image according to an embodiment of the present invention.
- the disparity between the left eye image and the right eye image is measured (S 710 ).
- the physical distance between the left eye image and the right eye image is calculated (S 720 ).
- the depth is controlled in accordance with the calculated physical distance (S 730 ).
- the image having the controlled depth is displayed on the display to provide a user with an image having an adaptively controlled three-dimensional effect.
- the invention can also be embodied as computer readable code on a computer readable recording medium.
- the computer readable recording medium is any data storage device that can store data which can be thereafter read by a computer system. Examples of the computer readable recording medium include read-only memory (ROM), random-access memory (RAM), CD-ROMs, magnetic tapes, floppy disks, optical data storage devices, and carrier waves (such as data transmission through the internet).
- ROM read-only memory
- RAM random-access memory
- CD-ROMs compact discs
- magnetic tapes magnetic tapes
- floppy disks optical data storage devices
- carrier waves such as data transmission through the internet
- carrier waves such as data transmission through the internet.
- the computer readable recording medium can also be distributed over network coupled computer systems so that the computer readable code is stored and executed in a distributed fashion. Also, functional programs, code, and code segments for accomplishing the present invention can be easily constructed by programmers skilled in the art to which the present invention pertains.
- the present invention it is possible to provide an apparatus and method for adaptively controlling disparity to control the image depth when a user uses a stereoscopic three-dimensional display having a different screen size than a display used in a manufacturing environment.
Abstract
Description
- This application claims the benefit of Korean Patent Application No. 10-2005-0080611, filed on Aug. 31, 2005, in the Korean Intellectual Property Office, the disclosure of which is incorporated herein in its entirety by reference.
- 1. Field of the Invention
- The present invention relates to an apparatus and method for controlling the depth of a three-dimensional image, and more particularly, to an apparatus and method for adaptively controlling disparity to control the image depth when a user uses a stereoscopic three-dimensional display of a different size than a display used in a manufacturing environment.
- 2. Description of the Related Art
- Current three dimensional displays generally use a method of projecting images having different disparities to the left and right eyes of a user and applying a three-dimensional effect to the displayed image by adding a film-shaped micro-polarizer, a barrier, or a lenticular lens to a two-dimensional flat display. Three-dimensional displays using this method are classified into stereoscopic three-dimensional displays and multi-view three-dimensional displays in accordance with the number of views of the display. The stereoscopic three-dimensional display displays images of two views and the multi-view three-dimensional display displays images of three or more views.
- A realistic image is created using a stereoscopic camera having two incident lenses or a multi-view camera having several incident lenses, and stereo or multi-view contents are created by computer graphic technology. Here, the image or the image contents are created in consideration of human visual characteristics. However, if the characteristics such as size and resolution of the display used in the manufacturing environment are different from those of the display used by the user, the desired three-dimensional image cannot be displayed.
- The present invention provides an apparatus and method for adaptively controlling disparity to control the image depth when a user uses a stereoscopic three-dimensional display having a different size than a display used in a manufacturing environment.
- According to an aspect of the present invention, there is provided an apparatus for controlling the depth of a three-dimensional image, including: a disparity measuring unit which measures the disparity between a left eye image and a right eye image; a physical distance calculating unit which calculates the physical distance between the left eye image and the right eye image based on the measured disparity and the physical characteristics of a display; and a depth controlling unit which controls the depth depending on the calculated physical distance.
- The disparity measuring unit may estimate the disparity between the left eye image and the right eye image and determine any one of a maximum value, a minimum value, and a mean value of the disparity as a reference value.
- The physical characteristics of the display may include the physical size of one pixel in the display, and the physical distance calculating unit may read the physical size of one pixel which has been previously stored and multiply the reference value by the physical size of that pixel to calculate the physical distance of the disparity.
- The depth controlling unit may control the physical distance based on a threshold value determined depending on the physical characteristics of the display.
- The depth controlling unit may compare a maximum value of the calculated physical distance with a maximum threshold value of the physical distance, control the maximum value to less than the maximum threshold value if the calculated maximum value is greater than the maximum threshold value, and control the physical distance between objects in the images in proportion to the controlled maximum value.
- The depth controlling unit may compare a mean value of the calculated physical distance with a mean threshold value of the physical distance, control the mean value to less than the mean threshold value if the calculated mean value is greater than the mean threshold value, and control the physical distance between objects in the images in proportion to the controlled mean value.
- The depth controlling unit may compare a maximum value of the calculated physical distance with a minimum threshold value of the physical distance, control the maximum value to greater than the minimum threshold value if the calculated maximum value is less than the minimum threshold value, and control the physical distance between objects in the images in proportion to the controlled maximum value.
- The depth controlling unit may reconstruct an intermediate image using an intermediate view reconstruction method based on the controlled physical distance and replace at least one of the left eye image and the right eye image with the reconstructed intermediate image to control the depth. The depth controlling unit may horizontally move the left eye image and the right eye image based on the controlled physical distance and synthesize the horizontally moved left eye image and right eye image to control the depth.
- The apparatus for controlling the depth may further include a display unit which displays the image having the controlled depth, output from the depth controlling unit.
- According to another aspect of the present invention, there is provided a method for controlling the depth of a three-dimensional image, including: measuring the disparity between a left eye image and a right eye image; calculating the physical distance between the left eye image and the right eye image based on the measured disparity and the physical characteristics of a display; and controlling the depth depending on the calculated physical distance.
- The above and other features and advantages of the present invention will become more apparent by describing in detail exemplary embodiments thereof with reference to the attached drawings in which:
-
FIGS. 1A-1C illustrate a disparity variation between a left eye image and a right eye image in accordance with the variation of the screen size of a display; -
FIG. 2 illustrates the relationship between the depth and the size of an object when a user views a large object and a small object at the same position; -
FIGS. 3A-3C illustrate the variation of the depth of an object depending on the screen size of a display and the position of a user; -
FIG. 4 illustrates the structure of an apparatus for adaptively controlling the depth of a three-dimensional image according to an embodiment of the present invention; -
FIG. 5 illustrates a process of controlling the depth using an intermediate view reconstruction method in a depth controlling unit illustrated inFIG. 4 ; -
FIG. 6 illustrates a process of horizontally moving a left eye image and a right eye image to control image depth in the depth controlling unit illustrated inFIG. 4 ; and -
FIG. 7 is a flowchart illustrating a method for adaptively controlling the depth of a three-dimensional image according to an embodiment of the present invention. - The present invention will now be described more fully with reference to the accompanying drawings, in which exemplary embodiments of the invention are shown.
-
FIGS. 1A-1C a disparity variation between a left eye image and a right eye image in accordance with the variation of the screen size of a display. - When content is created for a display screen illustrated in
FIG. 1B , the content has an optimal depth D2 when reproduced on the same display screen. Accordingly, if the content is intended for a three-dimensional display having a screen size of N2×M2 (inch2) illustrated inFIG. 1B but a user watches the content on a display having a smaller screen size of N1×M1 (inch2) illustrated inFIG. 1A , the depth of the three-dimensional image is reduced and thus the user feels as if the object on the screen is distant. That is, the three-dimensional effect deteriorates. - When the user watches the three-dimensional content on a display having a larger screen size of N3×M3 (inch2), the user feels as if the object on the screen is very close, which can cause fatigue or eye strain. In some cases, the object may have a depth which cannot be perceived by a human.
-
FIG. 2 illustrates the relationship between the depth and the size of an object when a user views a large object and a small object at the same position. - If the distance between an object and a user who watches the object is constant, the depth does not actually vary even if the size of the object varies. However, in a three-dimensional display using stereoscopic disparity, since the disparity between a left eye image and a right eye image is proportional to the depth perceived by the user and varies depending on the screen size of the display, the same problem as described with reference to
FIG. 1 occurs. -
FIGS. 3A-3C the variation of the depth of an object depending on the screen size of a display and the position of a user. - As illustrated in
FIGS. 3A and 3B , if the screen size increases while the position of the user is constant, the disparity increases and thus the depth increases. As illustrated inFIGS. 3B and 3C , if the distance between the display and the user doubles, the depth increases. - Accordingly, when the user watches the image content on a display having a different screen size from the display used when making the content, the depth must be controlled such that the image is suitably displayed on the user's display.
-
FIG. 4 illustrates the structure of an apparatus for adaptively controlling the depth of a three-dimensional image according to an embodiment of the present invention. - The
apparatus 400 for controlling the depth according to the present embodiment includes adisparity measuring unit 410, a physicaldistance calculating unit 420, and adepth controlling unit 430. - The
disparity measuring unit 410 divides an input three-dimensional image into a left eye image and a right eye image and measures the disparity between the left eye image and the right eye image. Thedisparity measuring unit 410 estimates the disparity between the left eye image and the right eye image. The method of estimating the disparity may use block-based movement estimation as used by MPEG encoding. First, the left eye image is divided into N×N uniform blocks. Subsequently, for each block in the left eye image, the most similar block is estimated in the right eye image using a sum of absolute difference (SAD) or a mean of absolute difference (MAD) calculation. The distance between a reference block and its estimated similar block forms a disparity vector, measured in pixels. That is, the disparity between the left eye image and the right eye image at a specific position is represented by the distance in pixels between the same objects in the left eye image and the right eye image at that point. - The
disparity measuring unit 410 determines any one of a maximum value d(max), a minimum value d(min), and a mean value d(mean) of the estimated disparities as a reference value and sends it to the physicaldistance calculating unit 420. - The physical
distance calculating unit 420 calculates the physical distance between the left eye image and the right eye image based on the measured disparity and the physical characteristics of the display. The physicaldistance calculating unit 420 reads the physical size P of one pixel displayed on the screen and multiplies the physical size P of the pixel by the reference value determined by thedisparity measuring unit 410, to calculate the physical distance between the objects in the left eye image and the right eye image. The physical size P of the pixel may be stored in a storage unit such as a ROM of the display. - The
depth controlling unit 430 controls the calculated physical distance based on a threshold value determined in accordance with the physical characteristics of the display, and controls the depth in accordance with the controlled distance. Thedepth controlling unit 430 reconstructs the image using an intermediate view reconstruction method, based on the controlled distance, and replaces either the left eye image or the right eye image with the reconstructed image, and outputs the replacement image and the unchanged other eye image. When these are displayed, an image having controlled depth is viewed. Alternatively, thedepth control unit 430 horizontally moves the left eye image and the right eye image based on the controlled distance, synthesizes the horizontally moved left eye image and right eye image, and outputs the image having the controlled depth. Thedepth controlling apparatus 400 may include a display unit (not shown) for displaying the three-dimensional image output fromdepth controlling unit 430 to embody a display system for outputting the three-dimensional image. - The process of controlling the depth with the
depth controlling unit 430 will be described in detail with reference toFIGS. 5 and 6 . -
FIG. 5 illustrates a process of controlling the depth using the intermediate view reconstruction method in thedepth controlling unit 430 illustrated inFIG. 4 . - The
depth controlling unit 430 controls the depth based on the physical distance calculated by the physicaldistance calculating unit 420. If the physical distance must be controlled, thedepth controlling unit 430 performs the following operations. - The
depth controlling unit 430 receives the maximum value of the physical distance calculated by the physicaldistance calculating unit 420 and compares the received maximum value with a maximum threshold value of the physical distance. If the received maximum value is greater than the maximum threshold value, thedepth controlling unit 430 controls the maximum value to be less than the maximum threshold value, and controls the physical distance between the objects in the images in proportion to the controlled value. That is, the value obtained by multiplying the maximum value of the disparity between the objects in the images by the physical size of one pixel of the display is controlled to less than the predetermined threshold value and the total disparity of the image is controlled in proportion to the controlled value. - The maximum value of the physical distance, D(max), is obtained by multiplying the maximum value d(max) of the measured disparities by the physical size P of one pixel. The maximum threshold value th(max) is the maximum value of the physical distance between the
left eye image 51 and theright eye image 53 in the display as used by a user within a range which does not cause eyestrain or other fatigue, as determined experimentally. The maximum threshold value can be obtained from the depth having the maximum value within a range wherein fatigue is not induced. - Referring to
FIG. 5 , d1 denotes the distance from the left edge to the object in theleft eye image 51, d2 denotes the distance from the left edge to the object in theright eye image 53, and d3 denotes the disparity between the objects in theleft eye image 51 and theright eye image 53. In thereconstructed image 55, d4 denotes the disparity which is controlled to prevent user fatigue. That is, if D(max)>th(max), a new image having the disparity d4 is generated using Equation 1, and the existingright eye image 53 is replaced with the new image, thereby reducing user fatigue.
d4×P=th(max)×d3×P/D(max)=th(max)×d3×P/(d(max)×P) Equation 1 - Although the maximum value of the physical distance is less than the maximum threshold value, if the total disparity of the image is large, the user may feel fatigued. In this case, the
depth controlling unit 430 receives the mean value of the physical distance calculated by the physicaldistance calculating unit 420 and compares it with a mean threshold value of the physical distance. Then, if the calculated mean value is greater than the mean threshold value, thedepth controlling unit 430 controls the calculated mean value to less than the mean threshold value, and controls the physical distance between the objects in the images in proportion to the controlled mean value. - The mean value of the physical distance, D(mean), is obtained by multiplying the mean value d(mean) of the disparities calculated by the
disparity measuring unit 410 by the physical size P of one pixel. The mean threshold value th(mean) is the average of the physical distances between theleft eye image 51 and theright eye image 53 in the display as used by the user within the range which does not cause eyestrain or other fatigue, as determined experimentally. - If D(mean)>th(mean), a new image having the disparity d4 is generated using Equation 2 and the existing
right eye image 53 is replaced with the new image, thereby reducing the fatigue of the user.
d4×P=th(mean)×d3×P/D(mean) Equation 2 - Up to now, the case of using a display having a larger screen size than the display which can optimally display the three-dimensional image has been described. When using a display having a smaller screen size than the optimal display, the physical depth or the disparity decreases and thus eyestrain is reduced. However, the three-dimensional effect may reduced by the decreasing depth. In this case, the three-dimensional effect must be increased by increasing the total disparity.
- Accordingly, the
depth controlling unit 430 compares the maximum value of the physical distance calculated by the physicaldistance calculating unit 420 with a minimum threshold value of the physical distance. If the calculated maximum value is less than the minimum threshold value, thedepth controlling unit 430 controls the calculated maximum value to greater than the minimum threshold value, and controls the physical distance between the objects in the images in proportion to the controlled maximum value. - The minimum threshold value th(min) is the minimum value of the physical distance between the
left eye image 51 and theright eye image 53 in the display as used by the user, within the range which does not cause eyestrain or other fatigue, as determined experimentally. - If D(max)<th(min), a new image having the disparity d4 is generated using Equation 3 and the existing
right eye image 53 is replaced with the new image, thereby increasing three dimensional effect.
d4×P=th(min)×d3×P/D(max) Equation 3 - The image having the physical distance of d4×P is reconstructed by the intermediate view reconstruction method, the
right eye image 53 is replaced with the reconstructedintermediate image 55, and theleft eye image 51 and theintermediate image 55 overlap each other and are output. When the proper disparity d4 or the proper physical distance d4×P is determined in order to control the depth, any one or at least two of Equations 1, 2, and 3 may be used in accordance with the characteristics of the input three-dimensional image. - In
FIG. 5 , theright eye image 53 was replaced with the reconstructed image using a method of determining proper disparity based on the distance from the left edge to the object. However, the proper disparity may be calculated based on the distance from the right edge to the object. In this case, theleft eye image 51 is replaced with the reconstructed image using the intermediate view reconstruction method. That is, theimage 53 ofFIG. 5 and the reconstructed image reconstructed by the intermediate view reconstruction method overlap each other and are output. -
FIG. 6 illustrates a process of horizontally moving theleft eye image 51 and theright eye image 53 to control the depth in thedepth controlling unit 430 illustrated inFIG. 4 . -
Left eye image 61 is the image obtained by horizontally moving theleft eye image 51 ofFIG. 5 to the right side, and a hatched region indicates a portion which is lost in the process.Right eye image 63 is the image obtained by horizontally moving theright eye image 53 ofFIG. 5 to the left side, and a hatched region indicates a portion which is lost in the process. Animage 65 is obtained by synthesizing the horizontally moved lefteye image 61 andright eye image 63. Accordingly, theleft eye image 51 and theright eye image 53 must be moved horizontally such that the disparity between the left and right eye images is controlled to the disparity d4 as determined above. - Referring to
FIG. 6 , in order to output the image having the disparity d4, theleft eye image 51 is horizontally moved by d3-d4/2 to the right side and theright eye image 53 is horizontally moved by d3-d4/2 to the left side. When theleft eye image 61 and theright eye image 63 are synthesized, the three-dimensional image 65 having the disparity d4 can be generated. Similar to the method described with reference toFIG. 5 , when the proper disparity d4 or the proper physical distance d4×P is determined in order to control the depth, any one or at least two of Equations 1, 2, and 3 may be used in accordance with the characteristics of the input three-dimensional image. -
FIG. 7 is a flowchart illustrating a method of adaptively controlling the depth of a three-dimensional image according to an embodiment of the present invention. - In order to control the depth of the three-dimensional image, the disparity between the left eye image and the right eye image is measured (S710).
- Based on the measured disparity and the physical characteristics of the display, the physical distance between the left eye image and the right eye image is calculated (S720).
- The depth is controlled in accordance with the calculated physical distance (S730). The image having the controlled depth is displayed on the display to provide a user with an image having an adaptively controlled three-dimensional effect.
- The invention can also be embodied as computer readable code on a computer readable recording medium. The computer readable recording medium is any data storage device that can store data which can be thereafter read by a computer system. Examples of the computer readable recording medium include read-only memory (ROM), random-access memory (RAM), CD-ROMs, magnetic tapes, floppy disks, optical data storage devices, and carrier waves (such as data transmission through the internet). The computer readable recording medium can also be distributed over network coupled computer systems so that the computer readable code is stored and executed in a distributed fashion. Also, functional programs, code, and code segments for accomplishing the present invention can be easily constructed by programmers skilled in the art to which the present invention pertains.
- According to the present invention, it is possible to provide an apparatus and method for adaptively controlling disparity to control the image depth when a user uses a stereoscopic three-dimensional display having a different screen size than a display used in a manufacturing environment.
- While the present invention has been particularly shown and described with reference to exemplary embodiments thereof, it will be understood by those of ordinary skill in the art that various changes in form and detail may be made therein without departing from the spirit and scope of the present invention as defined by the following claims.
Claims (21)
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020050080611A KR100667810B1 (en) | 2005-08-31 | 2005-08-31 | Apparatus for controlling depth of 3d picture and method therefor |
KR10-2005-0080611 | 2005-08-31 |
Publications (2)
Publication Number | Publication Date |
---|---|
US20070047040A1 true US20070047040A1 (en) | 2007-03-01 |
US8290244B2 US8290244B2 (en) | 2012-10-16 |
Family
ID=37803684
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/492,970 Active 2030-04-09 US8290244B2 (en) | 2005-08-31 | 2006-07-26 | Apparatus and method for controlling depth of three-dimensional image |
Country Status (4)
Country | Link |
---|---|
US (1) | US8290244B2 (en) |
KR (1) | KR100667810B1 (en) |
CN (1) | CN100574462C (en) |
NL (1) | NL1032382C2 (en) |
Cited By (73)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1978755A2 (en) * | 2007-04-04 | 2008-10-08 | Mitsubishi Electric Corporation | Method and system for acquiring, encoding, decoding and displaying 3D light fields |
WO2009020277A1 (en) * | 2007-08-06 | 2009-02-12 | Samsung Electronics Co., Ltd. | Method and apparatus for reproducing stereoscopic image using depth control |
US20090324114A1 (en) * | 2008-06-26 | 2009-12-31 | Sony Corporation | Image compression apparatus and image compression method |
WO2010087575A2 (en) * | 2009-02-01 | 2010-08-05 | Lg Electronics Inc. | Broadcast receiver and 3d video data processing method |
US20110032341A1 (en) * | 2009-08-04 | 2011-02-10 | Ignatov Artem Konstantinovich | Method and system to transform stereo content |
US20110122131A1 (en) * | 2008-07-24 | 2011-05-26 | Koninklijke Philips Electronics N.V. | Versatile 3-d picture format |
US20110187831A1 (en) * | 2010-02-03 | 2011-08-04 | Korea Institute Of Science And Technology | Apparatus and method for displaying three-dimensional images |
US20110216163A1 (en) * | 2010-03-08 | 2011-09-08 | Dolby Laboratories Licensing Corporation | Methods For Carrying And Transmitting 3D Z-Norm Attributes In Digital TV Closed Captioning |
WO2011107550A1 (en) * | 2010-03-05 | 2011-09-09 | Sony Corporation | Disparity distribution estimation for 3d tv |
JP2011188108A (en) * | 2010-03-05 | 2011-09-22 | Mitsubishi Electric Corp | Image processing apparatus, three-dimensional image display device, and image processing method |
WO2011121397A1 (en) * | 2010-04-01 | 2011-10-06 | Nokia Corporation | Method, apparatus and computer program for selecting a stereoscopic imaging viewpoint pair |
WO2011121437A1 (en) * | 2010-03-31 | 2011-10-06 | Thomson Licensing | 3d disparity maps |
US20110249017A1 (en) * | 2010-04-07 | 2011-10-13 | Sony Corporation | Image signal processing device, display device, display method and program product |
US20110267338A1 (en) * | 2010-05-03 | 2011-11-03 | Kwangwoon University Industry-Academic Collaboration Foundation | Apparatus and method for reducing three-dimensional visual fatigue |
US20110285819A1 (en) * | 2009-09-17 | 2011-11-24 | Panasonic Corporation | Video signal processing apparatus and video signal processing method |
US20110292178A1 (en) * | 2010-05-28 | 2011-12-01 | Qualcomm Incorporated | Three-dimensional image processing |
US20110304708A1 (en) * | 2010-06-10 | 2011-12-15 | Samsung Electronics Co., Ltd. | System and method of generating stereo-view and multi-view images for rendering perception of depth of stereoscopic image |
US20120013604A1 (en) * | 2010-07-14 | 2012-01-19 | Samsung Electronics Co., Ltd. | Display apparatus and method for setting sense of depth thereof |
US20120038754A1 (en) * | 2010-08-16 | 2012-02-16 | Na Sungbin | Method for processing images in display device outputting 3-dimensional contents and display device using the same |
US20120075419A1 (en) * | 2010-09-23 | 2012-03-29 | Thomson Licensing | Adaptation of 3D video content |
EP2442577A3 (en) * | 2010-10-15 | 2012-05-16 | Sony Corporation | Information processing apparatus, informaton processing method and program |
US20120120202A1 (en) * | 2010-11-12 | 2012-05-17 | Gwangju Institute Of Science And Technology | Method for improving 3 dimensional effect and reducing visual fatigue and apparatus enabling the same |
US8200045B2 (en) * | 2007-02-07 | 2012-06-12 | Thomson Licensing | Image processing method |
US20120206453A1 (en) * | 2009-09-16 | 2012-08-16 | Koninklijke Philips Electronics N.V. | 3d screen size compensation |
FR2971875A1 (en) * | 2011-02-23 | 2012-08-24 | Mobiclip | DEVICE AND METHOD FOR MANAGING THE POSITION OF THE FOCAL PLANE IN A STEREOSCOPIC SCENE |
US20120249747A1 (en) * | 2011-03-30 | 2012-10-04 | Ziv Aviv | Real-time depth extraction using stereo correspondence |
EP2408213A3 (en) * | 2010-07-14 | 2012-10-24 | LG Electronics Inc. | Mobile terminal and controlling method thereof |
EP2525579A2 (en) * | 2010-01-11 | 2012-11-21 | LG Electronics Inc. | Broadcasting receiver and method for displaying 3d images |
WO2012162096A1 (en) * | 2011-05-23 | 2012-11-29 | Qualcomm Incorporated | Interactive user interface for stereoscopic effect adjustment |
US20120327191A1 (en) * | 2010-03-05 | 2012-12-27 | Panasonic Corporation | 3d imaging device and 3d imaging method |
CN103024591A (en) * | 2012-12-26 | 2013-04-03 | 新奥特(北京)视频技术有限公司 | Method and device for adjusting 3D (three-dimensional) parallax |
EP2375766A3 (en) * | 2010-04-09 | 2013-05-29 | Tektronix International Sales GmbH | Method and apparatus for measuring an audiovisual parameter |
US20130141423A1 (en) * | 2011-12-06 | 2013-06-06 | Jung-hyun Cho | Three-dimensional image display apparatus |
US20130278631A1 (en) * | 2010-02-28 | 2013-10-24 | Osterhout Group, Inc. | 3d positioning of augmented reality information |
WO2013158784A1 (en) * | 2012-04-17 | 2013-10-24 | 3Dmedia Corporation | Systems and methods for improving overall quality of three-dimensional content by altering parallax budget or compensating for moving objects |
US20140063206A1 (en) * | 2012-08-28 | 2014-03-06 | Himax Technologies Limited | System and method of viewer centric depth adjustment |
US8810635B2 (en) | 2009-07-31 | 2014-08-19 | 3Dmedia Corporation | Methods, systems, and computer-readable storage media for selecting image capture positions to generate three-dimensional images |
US20140292748A1 (en) * | 2013-04-01 | 2014-10-02 | Electronics And Telecommunications Research Institute | System and method for providing stereoscopic image by adjusting depth value |
US20150015666A1 (en) * | 2013-07-09 | 2015-01-15 | Electronics And Telecommunications Research Institute | Method and apparatus for providing 3d video streaming service |
US9049434B2 (en) | 2010-03-05 | 2015-06-02 | Panasonic Intellectual Property Management Co., Ltd. | 3D imaging device and 3D imaging method |
US20150156472A1 (en) * | 2012-07-06 | 2015-06-04 | Lg Electronics Inc. | Terminal for increasing visual comfort sensation of 3d object and control method thereof |
US9082210B2 (en) | 2012-09-19 | 2015-07-14 | Ali (Zhuhai) Corporation | Method and apparatus for adjusting image depth |
US9091851B2 (en) | 2010-02-28 | 2015-07-28 | Microsoft Technology Licensing, Llc | Light control in head mounted displays |
US9097890B2 (en) | 2010-02-28 | 2015-08-04 | Microsoft Technology Licensing, Llc | Grating in a light transmissive illumination system for see-through near-eye display glasses |
US9097891B2 (en) | 2010-02-28 | 2015-08-04 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses including an auto-brightness control for the display brightness based on the brightness in the environment |
US9128281B2 (en) | 2010-09-14 | 2015-09-08 | Microsoft Technology Licensing, Llc | Eyepiece with uniformly illuminated reflective display |
US9129295B2 (en) | 2010-02-28 | 2015-09-08 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a fast response photochromic film system for quick transition from dark to clear |
US9134534B2 (en) | 2010-02-28 | 2015-09-15 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses including a modular image source |
EP2432228A3 (en) * | 2010-09-17 | 2015-10-07 | Samsung Electronics Co., Ltd. | Display apparatus and control method thereof |
US9182596B2 (en) | 2010-02-28 | 2015-11-10 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with the optical assembly including absorptive polarizers or anti-reflective coatings to reduce stray light |
US9185388B2 (en) | 2010-11-03 | 2015-11-10 | 3Dmedia Corporation | Methods, systems, and computer program products for creating three-dimensional video sequences |
US20150324951A1 (en) * | 2014-05-08 | 2015-11-12 | Glasses.Com | Systems and methods for scaling an object |
US9188849B2 (en) | 2010-03-05 | 2015-11-17 | Panasonic Intellectual Property Management Co., Ltd. | 3D imaging device and 3D imaging method |
US9215436B2 (en) | 2009-06-24 | 2015-12-15 | Dolby Laboratories Licensing Corporation | Insertion of 3D objects in a stereoscopic image at relative depth |
US9223134B2 (en) | 2010-02-28 | 2015-12-29 | Microsoft Technology Licensing, Llc | Optical imperfections in a light transmissive illumination system for see-through near-eye display glasses |
US9229227B2 (en) | 2010-02-28 | 2016-01-05 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a light transmissive wedge shaped illumination system |
EP2757789A4 (en) * | 2011-09-16 | 2016-01-20 | Konica Minolta Inc | Image processing system, image processing method, and image processing program |
US9285589B2 (en) | 2010-02-28 | 2016-03-15 | Microsoft Technology Licensing, Llc | AR glasses with event and sensor triggered control of AR eyepiece applications |
US9329689B2 (en) | 2010-02-28 | 2016-05-03 | Microsoft Technology Licensing, Llc | Method and apparatus for biometric data capture |
US9344701B2 (en) | 2010-07-23 | 2016-05-17 | 3Dmedia Corporation | Methods, systems, and computer-readable storage media for identifying a rough depth map in a scene and for determining a stereo-base distance for three-dimensional (3D) content creation |
US9341843B2 (en) | 2010-02-28 | 2016-05-17 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a small scale image source |
US9366862B2 (en) | 2010-02-28 | 2016-06-14 | Microsoft Technology Licensing, Llc | System and method for delivering content to a group of see-through near eye display eyepieces |
US9380292B2 (en) | 2009-07-31 | 2016-06-28 | 3Dmedia Corporation | Methods, systems, and computer-readable storage media for generating three-dimensional (3D) images of a scene |
US9519994B2 (en) | 2011-04-15 | 2016-12-13 | Dolby Laboratories Licensing Corporation | Systems and methods for rendering 3D image independent of display size and viewing distance |
US9600923B2 (en) | 2011-05-26 | 2017-03-21 | Thomson Licensing | Scale-independent maps |
US9759917B2 (en) | 2010-02-28 | 2017-09-12 | Microsoft Technology Licensing, Llc | AR glasses with event and sensor triggered AR eyepiece interface to external devices |
US10154243B2 (en) | 2010-06-28 | 2018-12-11 | Interdigital Madison Patent Holdings | Method and apparatus for customizing 3-dimensional effects of stereo content |
US10180572B2 (en) | 2010-02-28 | 2019-01-15 | Microsoft Technology Licensing, Llc | AR glasses with event and user action control of external applications |
US10200671B2 (en) | 2010-12-27 | 2019-02-05 | 3Dmedia Corporation | Primary and auxiliary image capture devices for image processing and related methods |
US10539787B2 (en) | 2010-02-28 | 2020-01-21 | Microsoft Technology Licensing, Llc | Head-worn adaptive display |
US10860100B2 (en) | 2010-02-28 | 2020-12-08 | Microsoft Technology Licensing, Llc | AR glasses with predictive control of external device based on event input |
US10895868B2 (en) * | 2015-04-17 | 2021-01-19 | Tulip Interfaces, Inc. | Augmented interface authoring |
EP4201060A4 (en) * | 2021-01-26 | 2024-02-14 | Samsung Electronics Co Ltd | Display apparatus and control method thereof |
Families Citing this family (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR100824942B1 (en) | 2007-05-31 | 2008-04-28 | 한국과학기술원 | Method of generating lenticular display image and recording medium thereof |
DE102008001451A1 (en) * | 2008-04-29 | 2009-11-05 | Robert Bosch Gmbh | Camera and method for controlling a camera |
JP4737573B2 (en) | 2009-02-05 | 2011-08-03 | 富士フイルム株式会社 | 3D image output apparatus and method |
JP5501458B2 (en) * | 2010-06-11 | 2014-05-21 | 富士フイルム株式会社 | Stereoscopic image display apparatus, stereoscopic imaging apparatus, and method |
JP2012047995A (en) * | 2010-08-27 | 2012-03-08 | Fujitsu Ltd | Information display device |
KR101176500B1 (en) * | 2010-10-06 | 2012-08-22 | 엘지전자 주식회사 | Image display apparatus, and method for operating the same |
KR101153120B1 (en) * | 2010-10-06 | 2012-06-04 | 차형경 | Method For Depth Adaptation of Autostereoscopic Image and User Equipment For the Same |
JP5876983B2 (en) * | 2010-12-29 | 2016-03-02 | 任天堂株式会社 | Display control program, display control device, display control method, and display control system |
JP2012204852A (en) * | 2011-03-23 | 2012-10-22 | Sony Corp | Image processing apparatus and method, and program |
CN102724521A (en) * | 2011-03-29 | 2012-10-10 | 青岛海信电器股份有限公司 | Method and apparatus for stereoscopic display |
KR101165810B1 (en) | 2011-04-27 | 2012-07-16 | 주식회사 아이브이넷 | A method and an apparatus extracting a depth image information using a stereo camera |
JP6021063B2 (en) * | 2011-07-26 | 2016-11-02 | パナソニックIpマネジメント株式会社 | Movie editing apparatus, movie editing method, program, and integrated circuit |
KR101888672B1 (en) * | 2011-07-27 | 2018-08-16 | 엘지디스플레이 주식회사 | Streoscopic image display device and method for driving thereof |
KR101247501B1 (en) | 2011-08-30 | 2013-03-25 | 김재도 | Stereopic 3 Dimension Display Device |
KR101246846B1 (en) | 2011-08-30 | 2013-03-25 | 김재도 | Method for displaying of Stereopic 3 Dimension |
KR20160073787A (en) | 2014-12-17 | 2016-06-27 | 삼성전자주식회사 | Method and apparatus for generating 3d image on curved display |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5801760A (en) * | 1993-08-26 | 1998-09-01 | Matsushita Electric Industrial Co., Ltd. | Stereoscopic image pickup and display apparatus |
US6160909A (en) * | 1998-04-01 | 2000-12-12 | Canon Kabushiki Kaisha | Depth control for stereoscopic images |
US6798406B1 (en) * | 1999-09-15 | 2004-09-28 | Sharp Kabushiki Kaisha | Stereo images with comfortable perceived depth |
US20050201612A1 (en) * | 2004-03-04 | 2005-09-15 | Samsung Electronics Co.,Ltd. | Method and apparatus for detecting people using stereo camera |
US20060203085A1 (en) * | 2002-11-28 | 2006-09-14 | Seijiro Tomita | There dimensional image signal producing circuit and three-dimensional image display apparatus |
US20090022393A1 (en) * | 2005-04-07 | 2009-01-22 | Visionsense Ltd. | Method for reconstructing a three-dimensional surface of an object |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6005607A (en) | 1995-06-29 | 1999-12-21 | Matsushita Electric Industrial Co., Ltd. | Stereoscopic computer graphics image generating apparatus and stereoscopic TV apparatus |
KR100294925B1 (en) | 1999-06-03 | 2001-07-12 | 윤종용 | 3-D graphic image manufacturing method and binocular visual disparity adjustment method therefor |
EP1085769B1 (en) | 1999-09-15 | 2012-02-01 | Sharp Kabushiki Kaisha | Stereoscopic image pickup apparatus |
JP2002228976A (en) | 2001-02-01 | 2002-08-14 | Mixed Reality Systems Laboratory Inc | Stereoscopic image display device and stereoscopic image display method |
JP2003107603A (en) | 2001-09-28 | 2003-04-09 | Namco Ltd | Stereophonic image generating device, stereophonic image generation information and information storage medium |
KR100439341B1 (en) | 2002-08-27 | 2004-07-07 | 한국전자통신연구원 | Depth of field adjustment apparatus and method of stereo image for reduction of visual fatigue |
JP4228646B2 (en) | 2002-10-02 | 2009-02-25 | 株式会社セガ | Stereoscopic image generation method and stereoscopic image generation apparatus |
JP3978392B2 (en) | 2002-11-28 | 2007-09-19 | 誠次郎 富田 | 3D image signal generation circuit and 3D image display device |
JP2005073049A (en) | 2003-08-26 | 2005-03-17 | Sharp Corp | Device and method for reproducing stereoscopic image |
-
2005
- 2005-08-31 KR KR1020050080611A patent/KR100667810B1/en not_active IP Right Cessation
-
2006
- 2006-07-26 US US11/492,970 patent/US8290244B2/en active Active
- 2006-08-29 NL NL1032382A patent/NL1032382C2/en not_active IP Right Cessation
- 2006-08-31 CN CNB2006101280132A patent/CN100574462C/en not_active Expired - Fee Related
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5801760A (en) * | 1993-08-26 | 1998-09-01 | Matsushita Electric Industrial Co., Ltd. | Stereoscopic image pickup and display apparatus |
US6160909A (en) * | 1998-04-01 | 2000-12-12 | Canon Kabushiki Kaisha | Depth control for stereoscopic images |
US6798406B1 (en) * | 1999-09-15 | 2004-09-28 | Sharp Kabushiki Kaisha | Stereo images with comfortable perceived depth |
US20060203085A1 (en) * | 2002-11-28 | 2006-09-14 | Seijiro Tomita | There dimensional image signal producing circuit and three-dimensional image display apparatus |
US20050201612A1 (en) * | 2004-03-04 | 2005-09-15 | Samsung Electronics Co.,Ltd. | Method and apparatus for detecting people using stereo camera |
US20090022393A1 (en) * | 2005-04-07 | 2009-01-22 | Visionsense Ltd. | Method for reconstructing a three-dimensional surface of an object |
Cited By (115)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8200045B2 (en) * | 2007-02-07 | 2012-06-12 | Thomson Licensing | Image processing method |
EP1978755A2 (en) * | 2007-04-04 | 2008-10-08 | Mitsubishi Electric Corporation | Method and system for acquiring, encoding, decoding and displaying 3D light fields |
WO2009020277A1 (en) * | 2007-08-06 | 2009-02-12 | Samsung Electronics Co., Ltd. | Method and apparatus for reproducing stereoscopic image using depth control |
US20090324114A1 (en) * | 2008-06-26 | 2009-12-31 | Sony Corporation | Image compression apparatus and image compression method |
US8204319B2 (en) * | 2008-06-26 | 2012-06-19 | Sony Corporation | Image compression apparatus and image compression method |
US20110122131A1 (en) * | 2008-07-24 | 2011-05-26 | Koninklijke Philips Electronics N.V. | Versatile 3-d picture format |
US10567728B2 (en) | 2008-07-24 | 2020-02-18 | Koninklijke Philips N.V. | Versatile 3-D picture format |
US9432651B2 (en) | 2008-07-24 | 2016-08-30 | Koninklijke Philips N.V. | Versatile 3-D picture format |
WO2010087575A2 (en) * | 2009-02-01 | 2010-08-05 | Lg Electronics Inc. | Broadcast receiver and 3d video data processing method |
WO2010087575A3 (en) * | 2009-02-01 | 2011-09-15 | Lg Electronics Inc. | Broadcast receiver and 3d video data processing method |
US9756309B2 (en) | 2009-02-01 | 2017-09-05 | Lg Electronics Inc. | Broadcast receiver and 3D video data processing method |
US8803945B2 (en) | 2009-02-01 | 2014-08-12 | Lg Electronics Inc. | Broadcast receiver and 3D video data processing method |
EP2384585A4 (en) * | 2009-02-01 | 2017-03-15 | LG Electronics Inc. | Broadcast receiver and 3d video data processing method |
US9215436B2 (en) | 2009-06-24 | 2015-12-15 | Dolby Laboratories Licensing Corporation | Insertion of 3D objects in a stereoscopic image at relative depth |
US9380292B2 (en) | 2009-07-31 | 2016-06-28 | 3Dmedia Corporation | Methods, systems, and computer-readable storage media for generating three-dimensional (3D) images of a scene |
US8810635B2 (en) | 2009-07-31 | 2014-08-19 | 3Dmedia Corporation | Methods, systems, and computer-readable storage media for selecting image capture positions to generate three-dimensional images |
US11044458B2 (en) | 2009-07-31 | 2021-06-22 | 3Dmedia Corporation | Methods, systems, and computer-readable storage media for generating three-dimensional (3D) images of a scene |
US20110032341A1 (en) * | 2009-08-04 | 2011-02-10 | Ignatov Artem Konstantinovich | Method and system to transform stereo content |
US20120206453A1 (en) * | 2009-09-16 | 2012-08-16 | Koninklijke Philips Electronics N.V. | 3d screen size compensation |
US20110285819A1 (en) * | 2009-09-17 | 2011-11-24 | Panasonic Corporation | Video signal processing apparatus and video signal processing method |
EP2525579A4 (en) * | 2010-01-11 | 2014-11-12 | Lg Electronics Inc | Broadcasting receiver and method for displaying 3d images |
US9485489B2 (en) | 2010-01-11 | 2016-11-01 | Lg Electronics Inc. | Broadcasting receiver and method for displaying 3D images |
EP2525579A2 (en) * | 2010-01-11 | 2012-11-21 | LG Electronics Inc. | Broadcasting receiver and method for displaying 3d images |
US9215444B2 (en) | 2010-01-11 | 2015-12-15 | Lg Electronics Inc. | Broadcasting receiver and method for displaying 3D images |
US20110187831A1 (en) * | 2010-02-03 | 2011-08-04 | Korea Institute Of Science And Technology | Apparatus and method for displaying three-dimensional images |
US8994791B2 (en) | 2010-02-03 | 2015-03-31 | Korea Institute Of Science And Technology | Apparatus and method for displaying three-dimensional images |
US9366862B2 (en) | 2010-02-28 | 2016-06-14 | Microsoft Technology Licensing, Llc | System and method for delivering content to a group of see-through near eye display eyepieces |
US9285589B2 (en) | 2010-02-28 | 2016-03-15 | Microsoft Technology Licensing, Llc | AR glasses with event and sensor triggered control of AR eyepiece applications |
US10539787B2 (en) | 2010-02-28 | 2020-01-21 | Microsoft Technology Licensing, Llc | Head-worn adaptive display |
US10268888B2 (en) | 2010-02-28 | 2019-04-23 | Microsoft Technology Licensing, Llc | Method and apparatus for biometric data capture |
US9229227B2 (en) | 2010-02-28 | 2016-01-05 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a light transmissive wedge shaped illumination system |
US10860100B2 (en) | 2010-02-28 | 2020-12-08 | Microsoft Technology Licensing, Llc | AR glasses with predictive control of external device based on event input |
US10180572B2 (en) | 2010-02-28 | 2019-01-15 | Microsoft Technology Licensing, Llc | AR glasses with event and user action control of external applications |
US9329689B2 (en) | 2010-02-28 | 2016-05-03 | Microsoft Technology Licensing, Llc | Method and apparatus for biometric data capture |
US9134534B2 (en) | 2010-02-28 | 2015-09-15 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses including a modular image source |
US9875406B2 (en) | 2010-02-28 | 2018-01-23 | Microsoft Technology Licensing, Llc | Adjustable extension for temple arm |
US9129295B2 (en) | 2010-02-28 | 2015-09-08 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a fast response photochromic film system for quick transition from dark to clear |
US9759917B2 (en) | 2010-02-28 | 2017-09-12 | Microsoft Technology Licensing, Llc | AR glasses with event and sensor triggered AR eyepiece interface to external devices |
US9097891B2 (en) | 2010-02-28 | 2015-08-04 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses including an auto-brightness control for the display brightness based on the brightness in the environment |
US20130278631A1 (en) * | 2010-02-28 | 2013-10-24 | Osterhout Group, Inc. | 3d positioning of augmented reality information |
US9223134B2 (en) | 2010-02-28 | 2015-12-29 | Microsoft Technology Licensing, Llc | Optical imperfections in a light transmissive illumination system for see-through near-eye display glasses |
US9097890B2 (en) | 2010-02-28 | 2015-08-04 | Microsoft Technology Licensing, Llc | Grating in a light transmissive illumination system for see-through near-eye display glasses |
US9091851B2 (en) | 2010-02-28 | 2015-07-28 | Microsoft Technology Licensing, Llc | Light control in head mounted displays |
US9341843B2 (en) | 2010-02-28 | 2016-05-17 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a small scale image source |
US9182596B2 (en) | 2010-02-28 | 2015-11-10 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with the optical assembly including absorptive polarizers or anti-reflective coatings to reduce stray light |
JP2011188108A (en) * | 2010-03-05 | 2011-09-22 | Mitsubishi Electric Corp | Image processing apparatus, three-dimensional image display device, and image processing method |
US20120327191A1 (en) * | 2010-03-05 | 2012-12-27 | Panasonic Corporation | 3d imaging device and 3d imaging method |
US9188849B2 (en) | 2010-03-05 | 2015-11-17 | Panasonic Intellectual Property Management Co., Ltd. | 3D imaging device and 3D imaging method |
US9128367B2 (en) * | 2010-03-05 | 2015-09-08 | Panasonic Intellectual Property Management Co., Ltd. | 3D imaging device and 3D imaging method |
US9049434B2 (en) | 2010-03-05 | 2015-06-02 | Panasonic Intellectual Property Management Co., Ltd. | 3D imaging device and 3D imaging method |
WO2011107550A1 (en) * | 2010-03-05 | 2011-09-09 | Sony Corporation | Disparity distribution estimation for 3d tv |
US9426441B2 (en) | 2010-03-08 | 2016-08-23 | Dolby Laboratories Licensing Corporation | Methods for carrying and transmitting 3D z-norm attributes in digital TV closed captioning |
US20110216163A1 (en) * | 2010-03-08 | 2011-09-08 | Dolby Laboratories Licensing Corporation | Methods For Carrying And Transmitting 3D Z-Norm Attributes In Digital TV Closed Captioning |
CN106131531A (en) * | 2010-03-31 | 2016-11-16 | 汤姆森特许公司 | Method for processing video frequency and device |
US10791314B2 (en) | 2010-03-31 | 2020-09-29 | Interdigital Ce Patent Holdings, Sas | 3D disparity maps |
WO2011121437A1 (en) * | 2010-03-31 | 2011-10-06 | Thomson Licensing | 3d disparity maps |
WO2011121397A1 (en) * | 2010-04-01 | 2011-10-06 | Nokia Corporation | Method, apparatus and computer program for selecting a stereoscopic imaging viewpoint pair |
US20110249017A1 (en) * | 2010-04-07 | 2011-10-13 | Sony Corporation | Image signal processing device, display device, display method and program product |
US8542909B2 (en) | 2010-04-09 | 2013-09-24 | Tektronix, Inc. | Method and apparatus for measuring an audiovisual parameter |
EP2375766A3 (en) * | 2010-04-09 | 2013-05-29 | Tektronix International Sales GmbH | Method and apparatus for measuring an audiovisual parameter |
US8810564B2 (en) * | 2010-05-03 | 2014-08-19 | Samsung Electronics Co., Ltd. | Apparatus and method for reducing three-dimensional visual fatigue |
US20110267338A1 (en) * | 2010-05-03 | 2011-11-03 | Kwangwoon University Industry-Academic Collaboration Foundation | Apparatus and method for reducing three-dimensional visual fatigue |
WO2011149967A3 (en) * | 2010-05-28 | 2012-01-12 | Qualcomm Incorporated | Three-dimensional image processing |
US8970672B2 (en) * | 2010-05-28 | 2015-03-03 | Qualcomm Incorporated | Three-dimensional image processing |
US20110292178A1 (en) * | 2010-05-28 | 2011-12-01 | Qualcomm Incorporated | Three-dimensional image processing |
US20110304708A1 (en) * | 2010-06-10 | 2011-12-15 | Samsung Electronics Co., Ltd. | System and method of generating stereo-view and multi-view images for rendering perception of depth of stereoscopic image |
US10154243B2 (en) | 2010-06-28 | 2018-12-11 | Interdigital Madison Patent Holdings | Method and apparatus for customizing 3-dimensional effects of stereo content |
US9420257B2 (en) | 2010-07-14 | 2016-08-16 | Lg Electronics Inc. | Mobile terminal and method for adjusting and displaying a stereoscopic image |
US20120013604A1 (en) * | 2010-07-14 | 2012-01-19 | Samsung Electronics Co., Ltd. | Display apparatus and method for setting sense of depth thereof |
EP2408213A3 (en) * | 2010-07-14 | 2012-10-24 | LG Electronics Inc. | Mobile terminal and controlling method thereof |
US9344701B2 (en) | 2010-07-23 | 2016-05-17 | 3Dmedia Corporation | Methods, systems, and computer-readable storage media for identifying a rough depth map in a scene and for determining a stereo-base distance for three-dimensional (3D) content creation |
US9288482B2 (en) * | 2010-08-16 | 2016-03-15 | Lg Electronics Inc. | Method for processing images in display device outputting 3-dimensional contents and display device using the same |
KR101763592B1 (en) * | 2010-08-16 | 2017-08-01 | 엘지전자 주식회사 | Method for processing image of display system outputting 3 dimensional contents and display system enabling of the method |
US20120038754A1 (en) * | 2010-08-16 | 2012-02-16 | Na Sungbin | Method for processing images in display device outputting 3-dimensional contents and display device using the same |
US9128281B2 (en) | 2010-09-14 | 2015-09-08 | Microsoft Technology Licensing, Llc | Eyepiece with uniformly illuminated reflective display |
EP2432228A3 (en) * | 2010-09-17 | 2015-10-07 | Samsung Electronics Co., Ltd. | Display apparatus and control method thereof |
TWI491216B (en) * | 2010-09-23 | 2015-07-01 | Thomson Licensing | Adaptation of 3d video content |
US9204122B2 (en) * | 2010-09-23 | 2015-12-01 | Thomson Licensing | Adaptation of 3D video content |
US20120075419A1 (en) * | 2010-09-23 | 2012-03-29 | Thomson Licensing | Adaptation of 3D video content |
EP2442577A3 (en) * | 2010-10-15 | 2012-05-16 | Sony Corporation | Information processing apparatus, informaton processing method and program |
US9001188B2 (en) | 2010-10-15 | 2015-04-07 | Sony Corporation | Information processing apparatus, information processing method and program |
US9185388B2 (en) | 2010-11-03 | 2015-11-10 | 3Dmedia Corporation | Methods, systems, and computer program products for creating three-dimensional video sequences |
US20120120202A1 (en) * | 2010-11-12 | 2012-05-17 | Gwangju Institute Of Science And Technology | Method for improving 3 dimensional effect and reducing visual fatigue and apparatus enabling the same |
US8760502B2 (en) * | 2010-11-12 | 2014-06-24 | Samsung Electronics Co., Ltd. | Method for improving 3 dimensional effect and reducing visual fatigue and apparatus enabling the same |
US10911737B2 (en) | 2010-12-27 | 2021-02-02 | 3Dmedia Corporation | Primary and auxiliary image capture devices for image processing and related methods |
US11388385B2 (en) | 2010-12-27 | 2022-07-12 | 3Dmedia Corporation | Primary and auxiliary image capture devices for image processing and related methods |
US10200671B2 (en) | 2010-12-27 | 2019-02-05 | 3Dmedia Corporation | Primary and auxiliary image capture devices for image processing and related methods |
US9800865B2 (en) | 2011-02-23 | 2017-10-24 | Nintendo European Research And Development Sas | Device and method for managing the position of the focal plane in a stereoscopic scene |
FR2971875A1 (en) * | 2011-02-23 | 2012-08-24 | Mobiclip | DEVICE AND METHOD FOR MANAGING THE POSITION OF THE FOCAL PLANE IN A STEREOSCOPIC SCENE |
WO2012113797A1 (en) * | 2011-02-23 | 2012-08-30 | Mobiclip | Device and method for managing the position of the focal plane in a stereoscopic scene |
US8823777B2 (en) * | 2011-03-30 | 2014-09-02 | Intel Corporation | Real-time depth extraction using stereo correspondence |
US20120249747A1 (en) * | 2011-03-30 | 2012-10-04 | Ziv Aviv | Real-time depth extraction using stereo correspondence |
WO2012135220A3 (en) * | 2011-03-30 | 2013-01-03 | Intel Corporation | Real-time depth extraction using stereo correspondence |
US9519994B2 (en) | 2011-04-15 | 2016-12-13 | Dolby Laboratories Licensing Corporation | Systems and methods for rendering 3D image independent of display size and viewing distance |
CN103609104A (en) * | 2011-05-23 | 2014-02-26 | 高通股份有限公司 | Interactive user interface for stereoscopic effect adjustment |
WO2012162096A1 (en) * | 2011-05-23 | 2012-11-29 | Qualcomm Incorporated | Interactive user interface for stereoscopic effect adjustment |
JP2016192773A (en) * | 2011-05-23 | 2016-11-10 | クゥアルコム・インコーポレイテッドQualcomm Incorporated | Interactive user interface for stereoscopic effect adjustment |
JP2014517619A (en) * | 2011-05-23 | 2014-07-17 | クゥアルコム・インコーポレイテッド | Interactive user interface for adjusting stereoscopic effects |
US9600923B2 (en) | 2011-05-26 | 2017-03-21 | Thomson Licensing | Scale-independent maps |
EP2757789A4 (en) * | 2011-09-16 | 2016-01-20 | Konica Minolta Inc | Image processing system, image processing method, and image processing program |
US20130141423A1 (en) * | 2011-12-06 | 2013-06-06 | Jung-hyun Cho | Three-dimensional image display apparatus |
US9300948B2 (en) * | 2011-12-06 | 2016-03-29 | Samsung Display Co., Ltd. | Three-dimensional image display apparatus |
WO2013158784A1 (en) * | 2012-04-17 | 2013-10-24 | 3Dmedia Corporation | Systems and methods for improving overall quality of three-dimensional content by altering parallax budget or compensating for moving objects |
US20150156472A1 (en) * | 2012-07-06 | 2015-06-04 | Lg Electronics Inc. | Terminal for increasing visual comfort sensation of 3d object and control method thereof |
US9674501B2 (en) * | 2012-07-06 | 2017-06-06 | Lg Electronics Inc. | Terminal for increasing visual comfort sensation of 3D object and control method thereof |
US20140063206A1 (en) * | 2012-08-28 | 2014-03-06 | Himax Technologies Limited | System and method of viewer centric depth adjustment |
US9082210B2 (en) | 2012-09-19 | 2015-07-14 | Ali (Zhuhai) Corporation | Method and apparatus for adjusting image depth |
CN103024591A (en) * | 2012-12-26 | 2013-04-03 | 新奥特(北京)视频技术有限公司 | Method and device for adjusting 3D (three-dimensional) parallax |
US20140292748A1 (en) * | 2013-04-01 | 2014-10-02 | Electronics And Telecommunications Research Institute | System and method for providing stereoscopic image by adjusting depth value |
US20150015666A1 (en) * | 2013-07-09 | 2015-01-15 | Electronics And Telecommunications Research Institute | Method and apparatus for providing 3d video streaming service |
US9996899B2 (en) * | 2014-05-08 | 2018-06-12 | Glasses.Com Inc. | Systems and methods for scaling an object |
US20150324951A1 (en) * | 2014-05-08 | 2015-11-12 | Glasses.Com | Systems and methods for scaling an object |
US10895868B2 (en) * | 2015-04-17 | 2021-01-19 | Tulip Interfaces, Inc. | Augmented interface authoring |
US10996660B2 (en) | 2015-04-17 | 2021-05-04 | Tulip Interfaces, Ine. | Augmented manufacturing system |
EP4201060A4 (en) * | 2021-01-26 | 2024-02-14 | Samsung Electronics Co Ltd | Display apparatus and control method thereof |
Also Published As
Publication number | Publication date |
---|---|
NL1032382A1 (en) | 2007-03-01 |
US8290244B2 (en) | 2012-10-16 |
NL1032382C2 (en) | 2011-02-10 |
CN1925627A (en) | 2007-03-07 |
CN100574462C (en) | 2009-12-23 |
KR100667810B1 (en) | 2007-01-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8290244B2 (en) | Apparatus and method for controlling depth of three-dimensional image | |
CN102474644B (en) | Stereo image display system, parallax conversion equipment, parallax conversion method | |
CA2723627C (en) | System and method for measuring potential eyestrain of stereoscopic motion pictures | |
US8116557B2 (en) | 3D image processing apparatus and method | |
JP6147275B2 (en) | Stereoscopic image processing apparatus, stereoscopic image processing method, and program | |
JP3826236B2 (en) | Intermediate image generation method, intermediate image generation device, parallax estimation method, and image transmission display device | |
US7944444B2 (en) | 3D image processing apparatus and method | |
US9445071B2 (en) | Method and apparatus generating multi-view images for three-dimensional display | |
TWI493505B (en) | Image processing method and image processing apparatus thereof | |
US20110304708A1 (en) | System and method of generating stereo-view and multi-view images for rendering perception of depth of stereoscopic image | |
KR20100040593A (en) | Apparatus and method for image processing | |
US20150022631A1 (en) | Content-aware display adaptation methods and editing interfaces and methods for stereoscopic images | |
KR20150121127A (en) | Binocular fixation imaging method and apparatus | |
KR101105359B1 (en) | Fatigue evaluation method and apparatus for 3D video based on depth map | |
US8766974B2 (en) | Display apparatus and method | |
WO2011030399A1 (en) | Image processing method and apparatus | |
US8884951B2 (en) | Depth estimation data generating apparatus, depth estimation data generating method, and depth estimation data generating program, and pseudo three-dimensional image generating apparatus, pseudo three-dimensional image generating method, and pseudo three-dimensional image generating program | |
US20170309055A1 (en) | Adjusting parallax of three-dimensional display material | |
US8976171B2 (en) | Depth estimation data generating apparatus, depth estimation data generating method, and depth estimation data generating program, and pseudo three-dimensional image generating apparatus, pseudo three-dimensional image generating method, and pseudo three-dimensional image generating program | |
JP2013090129A (en) | Image processing apparatus, image processing method and program | |
JP4815004B2 (en) | Multi-view image encoding device | |
KR101521213B1 (en) | Apparatus for correcting stereoscopic image using visual discomfort model and method thereof | |
KR101326279B1 (en) | Motion estimation system for stereoscopic videos and method thereof | |
KR20130051119A (en) | Image display apparatus and method for operating the same |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SAMSUNG ELECTRONICS CO., LTD., KOREA, REPUBLIC OF Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HA, TAE-HYEUN;REEL/FRAME:018135/0150 Effective date: 20060517 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 12 |