US9667944B2 - Imaging optical system and 3D image acquisition apparatus including the imaging optical system - Google Patents

Imaging optical system and 3D image acquisition apparatus including the imaging optical system Download PDF

Info

Publication number
US9667944B2
US9667944B2 US14/028,961 US201314028961A US9667944B2 US 9667944 B2 US9667944 B2 US 9667944B2 US 201314028961 A US201314028961 A US 201314028961A US 9667944 B2 US9667944 B2 US 9667944B2
Authority
US
United States
Prior art keywords
light
wavelength band
sensing layer
sense
nir
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US14/028,961
Other versions
US20140098192A1 (en
Inventor
Yong-hwa Park
Jang-woo YOU
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Electronics Co Ltd
Original Assignee
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Samsung Electronics Co Ltd filed Critical Samsung Electronics Co Ltd
Assigned to SAMSUNG ELECTRONICS CO., LTD. reassignment SAMSUNG ELECTRONICS CO., LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: PARK, YONG-HWA, YOU, JANG-WOO
Publication of US20140098192A1 publication Critical patent/US20140098192A1/en
Priority to US15/589,506 priority Critical patent/US9998730B2/en
Application granted granted Critical
Publication of US9667944B2 publication Critical patent/US9667944B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03BAPPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
    • G03B35/00Stereoscopic photography
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/239Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
    • H04N13/0239
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/02Systems using the reflection of electromagnetic waves other than radio waves
    • G01S17/06Systems determining position data of a target
    • G01S17/08Systems determining position data of a target for measuring distance only
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/88Lidar systems specially adapted for specific applications
    • G01S17/89Lidar systems specially adapted for specific applications for mapping or imaging
    • H04N13/0253
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/207Image signal generators using stereoscopic image cameras using a single 2D image sensor
    • H04N13/211Image signal generators using stereoscopic image cameras using a single 2D image sensor using temporal multiplexing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/207Image signal generators using stereoscopic image cameras using a single 2D image sensor
    • H04N13/214Image signal generators using stereoscopic image cameras using a single 2D image sensor using spectral multiplexing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/254Image signal generators using stereoscopic image cameras in combination with electromagnetic radiation sources for illuminating objects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/257Colour aspects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/10Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths
    • H04N23/11Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths for generating image signals from visible and infrared light wavelengths
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N2213/00Details of stereoscopic systems
    • H04N2213/001Constructional or mechanical details

Definitions

  • the present disclosure relates to imaging optical systems and 3-dimensional (3D) image acquisition apparatuses, and more particularly, to imaging optical systems using one lens and one image sensor and 3D image acquisition apparatuses which are capable of acquiring a color image and a depth image at the same time by using the imaging optical system.
  • 3D image acquisition apparatuses such as 3D cameras which are capable of directly producing 3D content based on user input are being researched.
  • Such a 3D camera may acquire depth information together with existing 2D color image information through one photograph.
  • Depth information which may be information related to a distance between surfaces of an object and a 3D camera, may be acquired by either a stereo vision method using two cameras or a triangulation method using structured light and a camera.
  • a stereo vision method using two cameras
  • a triangulation method using structured light and a camera.
  • the depth information depends upon the surface state of the object.
  • the TOF technique is a method of measuring a TOF taken from a time period during which, after illumination light is projected onto an object, the illumination light reflected from the object is received by a light receiving unit.
  • light having a specific wavelength for example, a near-infrared ray having a wavelength of about 850 nm
  • an illumination optical system including a light emitting diode (LED) or a laser diode (LD).
  • the projected light is reflected from the object at the same specific wavelength and is received by a light receiving unit.
  • a series of processes for extracting depth information such as the modulation of the received light using a modulator having a previously known gain waveform, is performed.
  • Various TOF techniques have been introduced according to the series of optical processing processes.
  • a 3D camera adopting the TOF technique includes an illumination optical system that emits illumination light for acquiring depth information and an imaging optical system for acquiring images of an object.
  • the imaging optical system detects visible light reflected from the object to produce a general color image and simultaneously detects illumination light reflected from the object to produce a depth image having only depth information.
  • the imaging optical system may separately include an objective lens and image sensor for detecting visible light and another objective lens and another image sensor for detecting illumination light (i.e., a 2-lens 2-sensor structure).
  • a 2-lens 2-sensor structure a color image and a depth image have viewing angles different from each other.
  • a separate process and applicable processing circuitry for matching the two images is required.
  • the 3D camera may be increased in size, and also, manufacturing costs thereof may be increased.
  • a 3D camera including one common objective lens and two image sensors i.e., a 1-lens 2-sensor structure
  • image sensors i.e., a 1-lens 2-sensor structure
  • visible light and illumination light are separated from each other, and thus are respectively transmitted to separate image sensors.
  • an image optical system using one lens and one image sensor.
  • a 3D image acquisition apparatus which is capable of photographing a color image and a depth image at the same time by using the image optical system.
  • an imaging optical system includes: an objective lens configured to focus light having a first wavelength band and light having a second wavelength band, wherein the light having the first wavelength band and the light having the second wavelength band are reflected from an external object; an optical shutter module configured to reflect the light having the first wavelength band, which is focused by the objective lens, without modulating the light having the first wavelength band and to modulate the light having the second wavelength band, which is focused by the objective lens, and to reflect the modulated light having the second wavelength band; and an image sensor configured to respectively sense the light having the first wavelength band and the modulated light having the second wavelength band, which are reflected by the optical shutter module, and to output a first image signal with respect to the light having the first wavelength band and a second image signal with respect to the modulated light having the second wavelength band.
  • the optical shutter module may include an optical shutter configured to modulate the light having the second wavelength band with a predetermined gain waveform and a wavelength-selective mirror disposed on a light incident surface of the optical shutter to reflect the light having the first wavelength band and transmit the light having the second wavelength band.
  • the optical shutter module may maintain a constant reflectance with respect to the light having the first wavelength band and has a reflectance varying according to a modulation signal with respect to the light having the second wavelength band.
  • the light having the first wavelength band may be visible light and the light having the second wavelength band may be NIR light
  • the image sensor may have an array of a plurality of pixels, and each of the pixels may include a red sub-pixel configured to sense red light, a green sub-pixel configured to sense green light, a blue sub-pixel configured to sense blue light, and an NIR sub-pixel configured to sense the NIR light.
  • the imaging optical system may further include a time division filter disposed between the optical shutter module and the image sensor to selectively transmit one of the lights having the first wavelength band and the modulated light having the second wavelength band.
  • the time division filter may be configured to transmit the light having the first wavelength band during a first time period and the modulated light having the second wavelength band during a second time period.
  • the time division filter may be configured to switch between a first color filter configured to transmit the light having the first wavelength band on an optical path between the optical shutter module and the image sensor, and a second color filter configured to transmit the modulated light having the second wavelength band on the optical path between the optical shutter module and the image sensor.
  • the light having the first wavelength band may be visible light
  • the light having the second wavelength band may be NIR light
  • the image sensor may have an array of a plurality of pixels, and each of the pixels may include a red sub-pixel configured to sense red light and the NIR light, a green sub-pixel configured to sense green light and the NIR light, and a blue sub-pixel configured to sense blue light and the NIR light.
  • the image sensor may be configured to sense the light having the first wavelength band during the first time period and the modulated light having the second wavelength band during the second time period.
  • the image sensor may include a first sensing layer configured to sense the light having the first wavelength band and a second sensing layer disposed under the first sensing layer such that the first sensing layer is disposed between the second sensing layer and the optical shutter module, the second sensing layer being configured to sense the light having the second wavelength band.
  • the light having the first wavelength band may be visible light
  • the light having the second wavelength band is light having an NIR light
  • the first sensing layer may include a red sub-pixel configured to sense the visible light having red light, a green sub-pixel configured to sense the visible light having green light, and a blue sub-pixel configured to sense the visible light having blue light.
  • the first sensing layer may further include a first color filter disposed on the red sub-pixel to transmit the red light and the NIR light, a second color filter disposed on the green sub-pixel to transmit the green light and the NIR light, and a third color filter disposed on the blue sub-pixel to transmit the blue light and the NIR light.
  • the second sensing layer may include an NIR sub-pixel configured to sense the NIR light.
  • the imaging optical system may further include a filter disposed between the first sensing layer and the second sensing layer to block the light having the first wavelength band.
  • the image sensor may include four sensing layers which are vertically stacked to respectively sense visible light having one of four light colors.
  • the light having the first wavelength band may be visible light
  • the light having the second wavelength band may be NIR light
  • the four sensing layers may include a first sensing layer configured to sense visible light having a blue light, a second sensing layer configured to sense visible light having green light, a third sensing layer configured to sense visible light having red light, and a fourth sensing layer configured to sense the NIR light, which are sequentially disposed from an upper side closest to the optical shutter module to a lower side farthest from the optical shutter module.
  • the imaging optical system may further include a first filter disposed between the first sensing layer and the second sensing layer to block the blue light; a second filter disposed between the second sensing layer and the third sensing layer to block the green light; and a third filter disposed between the third sensing layer and the fourth sensing layer to block the red light.
  • a 3D image acquisition apparatus includes: an objective lens configured to focus light having a first wavelength band and light having a second wavelength band, wherein the light having the first wavelength band and the light having the second wavelength band are reflected from an external object; an optical shutter module configured to reflect the light having the first wavelength band, which is focused by the objective lens, without modulating the light having the first wavelength band and to modulate the light having the second wavelength band, which is focused by the objective lens, to reflect the modulated light having the second wavelength band; and an image sensor configured to respectively sense the light having the first wavelength band and the modulated light having the second wavelength band, which are reflected by the optical shutter module, and to output a first image signal with respect to the modulated light having the first wavelength band and a second image signal with respect to the modulated light having the second wavelength band; a light source configured to emit light having the second wavelength band onto the external object; an image signal processor configured to generate a 3D image by using the first and second image signals outputted from the image sensor; and
  • the light source may be configured to emit the light having the second wavelength band with a predetermined period and waveform onto the external object under the control of the control unit.
  • the light having the first wavelength band may be visible light
  • the light having the second wavelength band may be NIR light
  • the first image signal may be a color image signal having red, green, and blue color components
  • the second image signal may be a depth image signal related to a distance between the 3D image acquisition apparatus and the external object.
  • the image signal processor may calculate the distance between the 3D image acquisition apparatus and the external object for each of a plurality of pixels by using the depth image signal, and the calculated distance may be combined with the color image signal to produce a final 3D image.
  • control unit may control the optical shutter module to modulate the light having the second wavelength band into a predetermined gain waveform having the same period as the light having the second wavelength band.
  • FIG. 1 is a conceptual view of an imaging optical system and a 3D image acquisition apparatus including the imaging optical system according to an exemplary embodiment
  • FIG. 2 is an enlarged view illustrating a structure of an optical shutter module of FIG. 1 ;
  • FIG. 3 is a graph for explaining operation characteristics of the optical shutter module of FIG. 1 ;
  • FIG. 4 is a schematic plan view illustrating a pixel structure of an image sensor of FIG. 1 ;
  • FIGS. 5A through 5D are graphs illustrating transmittance characteristics of color filters arranged within the image sensor of FIG. 4 ;
  • FIG. 6 is a schematic conceptual view illustrating a structure of an imaging optical system according to another exemplary embodiment
  • FIG. 7 is a view illustrating an example of an operation of a time-division filter and an image sensor of FIG. 6 ;
  • FIGS. 8A and 8B are graphs illustrating transmittance characteristics for explaining an operation of the time-division filter of FIG. 6 ;
  • FIGS. 9A through 9C are graphs illustrating transmittance characteristics of color filters arranged within the image sensor of FIG. 6 ;
  • FIG. 10 is a schematic conceptual view of a structure of an imaging optical system according to another exemplary embodiment.
  • FIG. 11 is a schematic plan view illustrating a structure of each of a plurality of sensing layers within an image sensor of FIG. 10 ;
  • FIGS. 12A through 12C are graphs illustrating an example of sensitivities according to wavelengths for each pixel within a color image sensing unit according to exemplary embodiments
  • FIG. 13 is a graph illustrating an example of sensitivities according to wavelengths for each pixel within an IR image sensing unit according to exemplary embodiments
  • FIG. 14 is a schematic conceptual view illustrating a structure of an imaging optical system according to another exemplary embodiment
  • FIG. 15 is a schematic plan view illustrating a structure of each of a plurality of sensing layers within an image sensor of FIGS. 14 ;
  • FIGS. 16A through 16D are graphs illustrating an example of sensitivities according to wavelengths for each of a plurality of sensing layers of an image sensor of FIG. 15 .
  • FIG. 1 is a conceptual view of an imaging optical system and a 3D image acquisition apparatus including the imaging optical system according to an exemplary embodiment.
  • a 3D image acquisition apparatus 100 includes a light source 101 configured to generate illumination light having a predetermined wavelength band, an imaging optical system 110 configured to generate a color image signal and a depth image signal by using visible light and the illumination light which are reflected from an external object (not shown), an image signal processing unit 105 (e.g., image signal processor) configured to produce a 3D image by using the color image signal and the depth image signal, and a control unit 102 configured to control operations of the light source 101 , the imaging optical system 110 , and the image signal processing unit 105 .
  • the 3D image acquisition apparatus 100 may further include a memory 106 configured to store a final 3D image and a display panel 104 configured to display the final 3D image.
  • the imaging optical system 110 may include an objective lens 111 configured to focus the visible light and the illumination light which are reflected from the external object, an optical shutter module 112 configured to reflect the visible light focused by the objective lens 111 without modulating the visible light and configured to modulate the illumination light focused by the objective lens 111 to reflect the modulated illumination light, and an image sensor 113 configured to sense the visible light and the illumination light, which are reflected by the optical shutter module 112 , to generate the color image signal and the depth image signal.
  • an objective lens 111 configured to focus the visible light and the illumination light which are reflected from the external object
  • an optical shutter module 112 configured to reflect the visible light focused by the objective lens 111 without modulating the visible light and configured to modulate the illumination light focused by the objective lens 111 to reflect the modulated illumination light
  • an image sensor 113 configured to sense the visible light and the illumination light, which are reflected by the optical shutter module 112 , to generate the color image signal and the depth image signal.
  • a light emitting diode (LED) or a laser diode (LD) which emit illumination light having a near-infrared (NIR) wavelength of about 850 nm, which is invisible to the human eye for safety may be used as the light source 101 .
  • LED light emitting diode
  • LD laser diode
  • NIR near-infrared
  • the light source 101 may emit illumination light having a specific waveform, for example, a sine wave, a ramp wave, a square wave, and the like according to a control signal received from the control unit 102 .
  • FIG. 2 is an enlarged view illustrating a structure of the optical shutter module 112 of FIG. 1 .
  • the optical shutter module 112 includes an optical shutter 112 b configured to modulate illumination light with a predetermined gain waveform according to a TOF method and a wavelength-selective mirror 112 a disposed on a light incident surface of the optical shutter 112 b .
  • the wavelength-selective mirror 112 a may reflect light RGB having a wavelength band corresponding to that of the visible light and transmit light IR having a wavelength band corresponding to that of the illumination light.
  • the visible light RGB is reflected by the wavelength-selective mirror 112 a toward the image sensor 113 , and the illumination light IR is transmitted through the wavelength-selective mirror 112 a to reach the optical shutter 112 b .
  • the wavelength-selective mirror 112 a may be, for example, coated in a thin film form on a surface of the optical shutter 112 b.
  • the optical shutter 112 b modulates the illumination light IR according to a modulation signal applied from the control unit 102 and reflects the modulated illumination light IR′ toward the image sensor 113 .
  • the optical shutter 112 may be, for example, a GaAs-based reflective modulator which is operable at an ultra-high speed of about several tens of MHz to about several hundreds of MHz and has a reflectance varying according to the modulation signal.
  • FIG. 3 is a graph for explaining operation characteristics of the optical shutter module 112 having the above-described structure.
  • the optical shutter module 112 since the wavelength-selective mirror 112 a of the optical shutter module 112 reflects the light RGB having the wavelength band corresponding to that of the visible light, the optical shutter module 112 may maintain a constant high reflectance at all times with respect to the visible light having a wavelength band of about 400 nm to about 700 nm.
  • the optical shutter module 112 may have a reflectance varying according to the modulation signal with respect to the illumination light IR having a wavelength band of about 845 nm to about 855 nm.
  • the visible light RGB reflected by the optical shutter module 112 and the illumination light IR modulated and reflected by the optical shutter module 112 may be sensed by one image sensor 113 .
  • the image sensor 113 may be implemented as a semiconductor imaging device such as a charge-coupled device (CCD) or a complementary metal oxide semiconductor (CMOS).
  • CMOS complementary metal oxide semiconductor
  • the image sensor 113 has an array of a plurality of pixels.
  • FIG. 4 is a schematic plan view illustrating a pixel structure of the image sensor 113 of FIG. 1 .
  • Each of the pixels of the image sensor 113 may include a plurality of sub-pixels each having different sensitivities to wavelengths. Referring to FIG.
  • the image sensor 113 may include a red sub-pixel 113 R that senses red light, a green sub-pixel 113 G that senses green light, a blue sub-pixel 113 B that senses blue light, and an NIR sub-pixel 113 IR that senses illumination light having an NIR light. Since the image sensor 113 further includes the NIR sub-pixel 113 IR capable of sensing the illumination light having the NIR light in addition to the general red, green, and blue sub-pixels 113 R, 113 G, and 113 B, the image sensor 113 may sense the visible light and the illumination light at the same time.
  • the image sensor 113 may be realized by respectively disposing color filters having different transmittance characteristics on light incident surfaces of the sub-pixels.
  • FIGS. 5A to 5D are graphs illustrating transmittance characteristics of color filters arranged within the image sensor 113 of FIG. 4 .
  • a color filter transmitting only blue light having a wavelength of about 400 nm to about 500 nm may be disposed on the blue sub-pixel 113 B
  • a color filer transmitting only green light having a wavelength of about 500 nm to about 600 nm may be disposed on the green sub-pixel 113 G.
  • a color filter transmitting only red light having a wavelength of about 600 nm to about 700 nm may be disposed on the red sub-pixel 1 13R, and a color filter transmitting only an NIR ray having a wavelength of about 800 nm to about 900 nm may be disposed on the NIR sub-pixel 113 IR.
  • a photosensitive material having different sensitivities to each wavelength may be used for each sub-pixel, instead of using the color filters.
  • the image sensor 113 converts the intensity of the incident light for each sub-pixel into an electrical image signal and outputs the electrical image signal for each sub-pixel to the image signal processing unit 105 .
  • the image signal processing unit 105 produces a color image using electrical signals provided from the red, green, and blue sub-pixels 113 R, 113 G, and 113 B of the image sensor 113 and produces a depth image using an electrical signal provided from the NIR sub-pixel 113 IR.
  • the light source 101 projects, for example, the NIR illumination light onto an object, the light source 101 being operated under the control of the control unit 102 .
  • the light source 101 may project at least three illumination lights having the same period and different phases onto the object in a sequential order of time according to the TOF method.
  • the NIR illumination light reflected by the object is focused by the objective lens 111 .
  • the general visible light reflected from the object may also be focused by the objective lens 111 .
  • the visible light may be reflected by the wavelength-selective mirror 112 a to be incident onto the image sensor 113 .
  • the illumination light may pass through the wavelength-selective mirror 112 a to be incident onto the optical shutter 112 b .
  • the light incident onto the optical shutter 112 b may be modulated and reflected by the optical shutter 112 b and then be incident onto the image sensor 113 .
  • the optical shutter 112 b may sequentially modulate the at least three illumination lights by using a predetermined gain waveform having the same period as each of the at least three illumination lights according to the TOF method.
  • the image sensor 113 converts the intensity of the visible light incident onto each of the red, green, and blue sub-pixels into an electrical color image signal and provides an electrical color image signal for each of the red, green, and blue sub-pixels to the image signal processing unit 105 . Also, the image sensor 113 sequentially converts the intensities of each of the at least three modulated illumination lights incident onto the NIR sub-pixel 113 IR to successively provide the produced at least three electrical depth image signals to the image signal processing unit 105 . The image signal processing unit 105 produces a color image by using the electrical color image signals provided from the image sensor 113 .
  • the image signal processing unit 105 calculates, for each pixel, a distance between the object and the 3D image acquisition apparatus 100 by using the at least three electrical depth image signals provided from the image sensor 113 .
  • the image signal processing unit 105 produces a depth image from the calculated results.
  • the color image and the depth image may be combined with each other to produce a 3D image.
  • the produced 3D image may be, for example, stored in the memory 106 or displayed on the display panel 104 according to a user selection. Since an algorithm used to calculate the distance between the object and the 3D image acquisition apparatus 100 by using the at least three electrical depth image signals may be performed according to the normal TOF method, a detailed description thereof will be omitted.
  • the imaging optical system 110 may obtain the color image and the depth image at the same time by using only one common objective lens 111 and only one image sensor 113 .
  • the number of components may be decreased to reduce production costs.
  • the imaging optical system 110 and the 3D image acquisition apparatus 100 including the imaging optical system 110 may be reduced in size and weight. Also, since the color image and the depth image are photographed using the same objective lens 111 and have the same viewing angle, it is unnecessary to perform a process of matching their respective viewing angles. Also, since the color image and the depth image are obtained through the same image sensor 113 , image matching between the pixels within the color and depth images may be easily performed.
  • FIG. 6 is a schematic conceptual view illustrating a structure of an image optical system 120 according to another exemplary embodiment which implements the time division method.
  • the imaging optical system 120 may include an objective lens 111 configured to focus visible light and illumination light which are reflected from an external object, an optical shutter module 112 configured to reflect the visible light focused by the objective lens 111 without modulating the visible light and modulate the illumination light focused by the objective lens 111 to reflect the modulated illumination light, an image sensor 123 configured to respectively sense the visible light and the illumination light, which are reflected by the optical shutter module 112 , and to generate a color image signal and a depth image signal, and a time division filter 114 disposed between the optical shutter module 112 and the image sensor 123 to selectively transmit only one of the visible light and the illumination light.
  • the time division filter 114 may be configured to transmit only the visible light at a time 1 so that the image sensor 123 senses only the visible light at time 1 and transmit only the illumination light at a time 2 so that the image sensor 123 senses only the illumination light at time 2 .
  • the time division filter 114 may transmit visible light having a wavelength band of about 350 nm to about 700 nm at the time 1 as shown in FIG. 8A .
  • the time division filter 114 may transmit illumination light having a wavelength band in a range of about 800 nm to about 900 nm at the time 2 as shown in FIG. 8B .
  • the time division filter 114 may be realized using a disc-shaped color wheel in which a semicircular color filter configured to transmit the visible light and a semicircular color filter configured to transmit the NIR ray are coupled to each other.
  • the color wheel may be rotated using a motor such that, at the time 1 , the color filter configured to transmit the visible light is located on an optical path.
  • the color wheel may be rotated using the motor such that, at the time 2 , the color filter configured to transmit the NIR ray is located on the optical path.
  • any mechanical or electrical structure which switches a location of the color filter configured to transmit the visible light and the color filter configured to transmit the NIR ray on an optical path between the optical shutter module 112 and the image sensor 123 may be implemented.
  • the image sensor 123 may only include three kinds of sub-pixels. For example, as shown in an upper side of FIG. 7 , the image sensor 123 may include a red sub-pixel 123 R that senses red light, a green sub-pixel 123 G that senses green light, and a blue sub-pixel 123 B that senses blue light. Also, according to the current exemplary embodiment, the sub-pixels 123 R, 123 G, and 123 B may all be configured to sense the illumination light having the NIR light at the time 2 .
  • the red sub-pixel 123 R may sense the red light and the NIR ray
  • the green sub-pixel 123 G may sense the green light and the NIR ray
  • the blue sub-pixel 123 B may sense the blue light and the NIR ray.
  • FIGS. 9A through 9C are graphs illustrating transmittance characteristics of color filters arranged on light incident surfaces of the image sensor 123 according to an exemplary embodiment.
  • a color filter configured to transmit blue light having a wavelength of about 400 nm to about 500 nm and an NIR ray having a wavelength of about 800 nm to about 900 nm may be disposed on the blue sub-pixel 123 B.
  • a color filter configured to transmit green light having a wavelength of about 500 nm to about 600 nm and the NIR ray having the wavelength of about 800 nm to about 900 nm may be disposed on the green sub-pixel 123 G, and a color filter configured to transmit red light having a wavelength of about 600 nm to about 700 nm and the NIR ray having the wavelength of about 800 nm to about 900 nm may be disposed on the red sub-pixel 123 R.
  • the sub-pixels 123 R, 123 G, and 123 B of the image sensor 123 may sense the red, green, and blue light at a time 1 , and the image sensor 123 may sense the illumination light having the NIR light in the entire region of the image sensor 123 at the time 2 .
  • the imaging optical system 120 may be applied to the 3D image acquisition apparatus 100 of FIG. 1 .
  • an example operation of the 3D image acquisition apparatus including the imaging optical system 120 of FIG. 6 will be described.
  • the control unit 102 controls the time division filter 114 to transmit visible light at a time 1 and sense the visible light through the image sensor 123 .
  • the image signal processing unit 105 produces a color image by using an image signal provided from the image sensor 123 .
  • the control unit 102 controls the time division filter 114 to transmit the illumination light and controls the light source 101 to emit illumination light at a time 2 .
  • the image sensor 123 senses the illumination light
  • the image signal processing unit 105 produces a depth image according to the TOF method by using the image signal provided from the image sensor 123 .
  • At least three illumination lights having different phases may be sequentially projected onto the object.
  • the at least three illumination lights may be sensed at the time 2 all at once.
  • the at least three illumination lights may be separately sensed at several different times.
  • the visible light may be sensed at the time 1 , and then, the four illumination lights may be sequentially sensed at different times (e.g., four different times 2 )
  • the light source 101 may sequentially project four illumination lights each having different phases onto the object at the time 2 .
  • the illumination light including the four illumination lights may be sensed over the four times 2 , in the order of: visible light (time 1 ) -> illumination light 1 (time 2 ) -> visible light (time 1 ) -> illumination light 2 (time 2 ) -> visible light (time 1 ) -> illumination light 3 (time 2 ) -> visible light (time 1 ) -> illumination light 4 (time 2 ).
  • the illumination light including the four illumination lights may be sensed over two times 2 , in the order of: visible light (time 1 ) -> illumination light 1 , 2 (time 2 ) -> visible light (time 1 ) -> illumination light 3 , 4 (time 2 ) at the time 2 .
  • the visible light when the visible light is sensed at the time 1 , only three kinds of sub-pixels 123 R, 123 G, and 123 B are used within the image sensor 123 .
  • the color image when compared to a case in which four kinds of sub-pixels are used, the color image may be improved in resolution.
  • the illumination light when the illumination light is sensed at the time 2 , the illumination light may be sensed throughout the entire area of the image sensor 123 . Thus, distance calculation may be increased in accuracy, and light utilization efficiency may be improved.
  • FIG. 10 is a schematic conceptual view of a structure of an imaging optical system 130 according to another exemplary embodiment.
  • the imaging optical system 130 includes an objective lens 111 configured to focus visible light and illumination light which are reflected from an external object, an optical shutter module 112 configured to reflect the visible light focused by the objective lens 111 without modulating the visible light and modulate the illumination light focused by the objective lens 111 to reflect the modulated illumination light, and an image sensor 133 configured to respectively sense the visible light and the illumination light, which are reflected by the optical shutter module 112 , to generate a color image signal and a depth image signal.
  • the exemplary embodiment of FIG. 10 is different from that of FIG. 1 in that the image sensor 133 includes two sensing layers which are vertically stacked. That is, the image sensor 133 includes a first sensing layer 133 a and a second sensing layer 133 b which are vertically stacked.
  • the first sensing layer 133 a may be formed of a light absorption material such as silicon and may be disposed on the second sensing layer 133 b to sense the visible light.
  • the second sensing layer 133 b disposed under the first sensing layer 133 a may be formed as an illumination layer to sense the NIR light.
  • light having a relatively short wavelength may have a relatively shallow penetration depth with respect to a light absorption material such as silicon
  • light having a relatively long wavelength may have a relatively deep penetration depth with respect to the light absorption material.
  • the second sensing layer 133 b configured to sense the illumination light is disposed under the first sensing layer 133 a because the illumination light has a relatively long wavelength and is able to more deeply penetrate through the first sensing layer 133 a and reach the second sensing layer 133 b.
  • FIG. 11 is a schematic plan view illustrating a structure of each of a plurality of sensing layers within the image sensor 133 of FIG. 10 .
  • the first sensing layer 133 a may include only three kinds of sub-pixels.
  • the first sensing layer 133 a may include a red sub-pixel 133 R that senses red light, a green sub-pixel 133 G that senses green light and a blue sub-pixel 133 B that senses blue light.
  • transmittance characteristics of color filters disposed on the sub-pixels 133 R, 133 G, and 133 B may be the same as those of the color filters illustrated in FIGS. 9A to 9C . That is, referring to FIGS.
  • a color filter configured to transmit blue light having a wavelength of about 400 nm to about 500 nm and an NIR ray having a wavelength of about 800 nm to about 900 nm may be disposed on the blue sub-pixel 133 B.
  • a color filter configured to transmit green light having a wavelength of about 500 nm to about 600 nm and the NIR ray having the wavelength of about 800 nm to about 900 nm may be disposed on the green sub-pixel 133 G, and a color filter configured to transmit red light having a wavelength of about 600 nm to about 700 nm and the NIR ray having the wavelength of about 800 nm to about 900 nm may be disposed on the red sub-pixel 133 R.
  • the illumination light having the NIR ray together with the visible light are incident onto each of the sub-pixels 133 R, 133 G, and 133 B of the first sensing layer 133 a .
  • the sub-pixels 133 R, 133 G, and 133 B may be formed of photosensitive materials having different sensitivities based on wavelength.
  • the blue sub-pixel 133 B may be formed of a photosensitive material having high sensitivity with respect to blue light having a wavelength of about 400 nm to about 500 nm.
  • the green sub-pixel 133 G may be formed of a photosensitive material having high sensitivity with respect to green light having a wavelength of about 500 nm to about 600 nm
  • the red sub-pixel 133 R may be formed of a photosensitive material having high sensitivity with respect to red light having a wavelength of about 600 nm to about 700 nm.
  • the second sensing layer 133 b senses the illumination light that is transmitted through the first sensing layer 133 a .
  • the second sensing layer 133 b includes only a single type of sub-pixel 133 IR that senses the illumination light. That is, as shown in FIG. 13 , the second sensing layer 133 b may be configured so that the all pixels in the sensing layer 133 b have high sensitivities with respect to the illumination light.
  • a filter 135 blocking the visible light may be disposed between the first sensing layer 133 a and the second sensing layer 133 b.
  • the sub-pixels 133 R, 133 G, and 133 B may be formed of the same photosensitive material, instead of being formed of different photosensitive materials.
  • the light incident onto the blue sub-pixel 133 B of the first sensing layer 133 a may have the same intensity as the sum of the blue light and the illumination light.
  • the sub-pixel 133 IR of the second sensing layer 133 a may sense only the illumination light.
  • the intensity of the blue light incident onto the blue sub-pixel 133 B of the first sensing layer 133 a may be calculated.
  • intensities of the green and red light incident onto the green and red sub-pixels 133 G and 133 R of the first sensing layer 133 a facing the sub-pixel 133 IR of the second sensing layer 133 b may be calculated.
  • the visible light and the illumination light may be sensed at the same time without performing a time division process, and also, advantages described in the exemplary embodiment of FIG. 6 may be obtained.
  • FIG. 14 is a schematic conceptual view illustrating a structure of an imaging optical system 140 according to another exemplary embodiment.
  • the imaging optical system 140 may include an objective lens 111 configured to focus visible light and illumination light which are reflected from an external object, an optical shutter module 112 configured to reflect the visible light focused by the objective lens 111 without modulating the visible light and modulate the illumination light focused by the objective lens 111 to reflect the modulated illumination light, and an image sensor 143 configured to respectively sense the visible light and the illumination light, which are reflected by the optical shutter module 112 , to generate a color image signal and a depth image signal.
  • an objective lens 111 configured to focus visible light and illumination light which are reflected from an external object
  • an optical shutter module 112 configured to reflect the visible light focused by the objective lens 111 without modulating the visible light and modulate the illumination light focused by the objective lens 111 to reflect the modulated illumination light
  • an image sensor 143 configured to respectively sense the visible light and the illumination light, which are reflected by the optical shutter module 112 , to generate a color image signal and a depth image signal.
  • the exemplary embodiment of FIG. 14 is different from the exemplary embodiment of FIG. 10 in that the image sensor 143 includes four sensing layers which are vertically stacked. That is, although the image sensor 133 includes only the first sensing layer 133 a configured to sense the visible light and the second sensing layer 133 b configured to sense the NIR ray in the exemplary embodiment of FIG. 10 , the image sensor 143 may include four sensing layers 143 a to 143 d respectively configured to sense the red, green and blue light and the NIR ray in the exemplary embodiment of FIG. 14 .
  • the image sensor 143 includes a first sensing layer 143 a that senses the blue light, a second sensing layer 143 b that senses the green light, a third sensing layer 143 c that senses the red light, and a fourth sensing layer 143 d that senses the NIR ray, which are sequentially disposed from an upper side.
  • light having a relatively short wavelength may have a relatively shallow penetration depth with respect to a light absorption material such as silicon, and light having a relatively long wavelength may have a relatively deep penetration depth with respect to the light absorption material.
  • the image sensor 143 may be configured so that the uppermost first sensing layer 143 a senses the blue light (wavelength of about 400 nm to about 500 nm), and the lowermost fourth sensing layer 143 d senses the NIR ray (wavelength of about 800 nm to about 900 nm).
  • the intensity of each of the sensing layers 143 a to 143 d may be selected so that each of the four sensing layers 143 a to 143 d accurately senses light of a corresponding color.
  • the first sensing layer 143 a may be formed of a photosensitive material having high sensitivity with respect to blue light having a wavelength of about 400 nm to about 500 nm
  • the second sensing layer 143 b may be formed of a photosensitive material having high sensitivity with respect to green light having a wavelength of about 500 nm to about 600 nm.
  • the third sensing layer 143 c may be formed of a photosensitive material having high sensitivity with respect to red light having a wavelength of about 600 nm to about 700 nm
  • the fourth sensing layer 143 d may be formed of a photosensitive material having high sensitivity with respect to light having a wavelength of about 800 nm to about 900 nm.
  • the sensing layers 143 a to 143 d may be formed of the same photosensitive material, instead of being formed of different photosensitive materials.
  • a filter 145 a configured to block the blue light may be disposed between the first sensing layer 143 a and the second sensing layer 143 b.
  • a filter 145 b configured to block the green light may be disposed between the second sensing layer 143 b and the third sensing layer 143 c
  • a filter 145 c configured to block the red light may be disposed between the third sensing layer 143 c and the fourth sensing layer 143 d .
  • the filters 145 a , 145 b , and 145 c are added to separate the first through fourth sensing layers, according to other exemplary embodiments, the light absorption materials of each of the first through third sensing layers 143 a - 143 c may also serve as optical filters by adjusting a thickness thereof.
  • the thickness of the light absorption material of the first sensing layer 143 a may be adjusted to completely absorb the blue light within the first sensing layer 143 a
  • the thickness of the light absorption material of the second sensing layer 143 b may be adjusted to completely absorb the green light within the second sensing layer 143 b
  • the thickness of the light absorption material of the third sensing layer 143 c may be adjusted to completely absorb the red light within the third sensing layer 143 c.
  • the first sensing layer 143 a may sense all of the blue, green, red, and illumination light.
  • the second sensing layer 143 b may sense the green, red, and illumination light
  • the third sensing layer 143 c may sense the red and illumination light
  • the fourth sensing layer 143 d may sense only the illumination light.
  • the blue, green, red, and illumination light may be sensed through the separate sensing layers 143 a to 143 d , respectively, a high fill factor may be obtained with respect to all colors.
  • light utilization efficiency may be improved, and high resolution color and depth images may be obtained.
  • the above-described 3D image acquisition apparatus 100 may be applied to various fields.
  • the 3D image acquisition apparatus 100 may be applied to a range-finding device, a gesture recognizer, a game console, and the like.

Abstract

An imaging optical system includes an objective lens configured to focus light having a first wavelength band and light having a second wavelength band, an optical shutter module configured to reflect the light having the first wavelength band, which is focused by the objective lens, without modulating the light having the first wavelength band and to modulate the light having the second wavelength band, which is focused by the objective lens, and reflect the modulated light having the second wavelength band, and an image sensor configured to respectively sense the light having the first wavelength band and the modulated light having the second wavelength band, which are reflected by the optical shutter module, and to output a first image signal with respect to the light having the first wavelength band and a second image signal with respect to the modulated light having the second wavelength band.

Description

CROSS-REFERENCE TO RELATED APPLICATION
This application claims the benefit of Korean Patent Application No. 10-2012-0112660, filed on Oct. 10, 2012, in the Korean Intellectual Property Office, the disclosure of which is incorporated herein in its entirety by reference.
BACKGROUND
1. Field
The present disclosure relates to imaging optical systems and 3-dimensional (3D) image acquisition apparatuses, and more particularly, to imaging optical systems using one lens and one image sensor and 3D image acquisition apparatuses which are capable of acquiring a color image and a depth image at the same time by using the imaging optical system.
2. Description of the Related Art
Recently, with the development and the increased demand of 3D display devices displaying depth images, the importance of 3D content is being emphasized. Thus, 3D image acquisition apparatuses such as 3D cameras which are capable of directly producing 3D content based on user input are being researched. Such a 3D camera may acquire depth information together with existing 2D color image information through one photograph.
Depth information, which may be information related to a distance between surfaces of an object and a 3D camera, may be acquired by either a stereo vision method using two cameras or a triangulation method using structured light and a camera. However, it is difficult to acquire precise depth information through these methods because, as a distance between the object and the 3D camera is increased, the depth information is significantly decreased in accuracy. The depth information depends upon the surface state of the object.
To improve these limitations, a time-of-flight (TOF) technique has been introduced. The TOF technique is a method of measuring a TOF taken from a time period during which, after illumination light is projected onto an object, the illumination light reflected from the object is received by a light receiving unit. According to the TOF technique, light having a specific wavelength (for example, a near-infrared ray having a wavelength of about 850 nm) is projected onto a object by using an illumination optical system including a light emitting diode (LED) or a laser diode (LD). Then, the projected light is reflected from the object at the same specific wavelength and is received by a light receiving unit. Thereafter, a series of processes for extracting depth information, such as the modulation of the received light using a modulator having a previously known gain waveform, is performed. Various TOF techniques have been introduced according to the series of optical processing processes.
A 3D camera adopting the TOF technique includes an illumination optical system that emits illumination light for acquiring depth information and an imaging optical system for acquiring images of an object. The imaging optical system detects visible light reflected from the object to produce a general color image and simultaneously detects illumination light reflected from the object to produce a depth image having only depth information. Conventionally, the imaging optical system may separately include an objective lens and image sensor for detecting visible light and another objective lens and another image sensor for detecting illumination light (i.e., a 2-lens 2-sensor structure). However, in the 2-lens 2-sensor structure, a color image and a depth image have viewing angles different from each other. Thus, a separate process and applicable processing circuitry for matching the two images is required. As a result, the 3D camera may be increased in size, and also, manufacturing costs thereof may be increased.
Thus, a 3D camera including one common objective lens and two image sensors (i.e., a 1-lens 2-sensor structure) has been proposed. However, in the case of the 1-lens 2-sensor structure, visible light and illumination light are separated from each other, and thus are respectively transmitted to separate image sensors.
SUMMARY
According to exemplary embodiments, provided is an image optical system using one lens and one image sensor.
According to exemplary embodiments, provided is a 3D image acquisition apparatus which is capable of photographing a color image and a depth image at the same time by using the image optical system.
Additional aspects will be set forth in part in the description which follows and, in part, will be apparent from the description, or may be learned by practice of the presented exemplary embodiments.
According to an aspect of the exemplary embodiments, an imaging optical system includes: an objective lens configured to focus light having a first wavelength band and light having a second wavelength band, wherein the light having the first wavelength band and the light having the second wavelength band are reflected from an external object; an optical shutter module configured to reflect the light having the first wavelength band, which is focused by the objective lens, without modulating the light having the first wavelength band and to modulate the light having the second wavelength band, which is focused by the objective lens, and to reflect the modulated light having the second wavelength band; and an image sensor configured to respectively sense the light having the first wavelength band and the modulated light having the second wavelength band, which are reflected by the optical shutter module, and to output a first image signal with respect to the light having the first wavelength band and a second image signal with respect to the modulated light having the second wavelength band.
For example, the optical shutter module may include an optical shutter configured to modulate the light having the second wavelength band with a predetermined gain waveform and a wavelength-selective mirror disposed on a light incident surface of the optical shutter to reflect the light having the first wavelength band and transmit the light having the second wavelength band.
The optical shutter module may maintain a constant reflectance with respect to the light having the first wavelength band and has a reflectance varying according to a modulation signal with respect to the light having the second wavelength band.
According to an exemplary embodiment, the light having the first wavelength band may be visible light and the light having the second wavelength band may be NIR light, wherein the image sensor may have an array of a plurality of pixels, and each of the pixels may include a red sub-pixel configured to sense red light, a green sub-pixel configured to sense green light, a blue sub-pixel configured to sense blue light, and an NIR sub-pixel configured to sense the NIR light.
In addition, the imaging optical system may further include a time division filter disposed between the optical shutter module and the image sensor to selectively transmit one of the lights having the first wavelength band and the modulated light having the second wavelength band.
For example, the time division filter may be configured to transmit the light having the first wavelength band during a first time period and the modulated light having the second wavelength band during a second time period.
In addition, the time division filter may be configured to switch between a first color filter configured to transmit the light having the first wavelength band on an optical path between the optical shutter module and the image sensor, and a second color filter configured to transmit the modulated light having the second wavelength band on the optical path between the optical shutter module and the image sensor.
According to another exemplary embodiment, the light having the first wavelength band may be visible light, and the light having the second wavelength band may be NIR light, wherein the image sensor may have an array of a plurality of pixels, and each of the pixels may include a red sub-pixel configured to sense red light and the NIR light, a green sub-pixel configured to sense green light and the NIR light, and a blue sub-pixel configured to sense blue light and the NIR light.
For example, the image sensor may be configured to sense the light having the first wavelength band during the first time period and the modulated light having the second wavelength band during the second time period.
According to another exemplary embodiment, the image sensor may include a first sensing layer configured to sense the light having the first wavelength band and a second sensing layer disposed under the first sensing layer such that the first sensing layer is disposed between the second sensing layer and the optical shutter module, the second sensing layer being configured to sense the light having the second wavelength band.
For example, the light having the first wavelength band may be visible light, and the light having the second wavelength band is light having an NIR light, wherein the first sensing layer may include a red sub-pixel configured to sense the visible light having red light, a green sub-pixel configured to sense the visible light having green light, and a blue sub-pixel configured to sense the visible light having blue light.
In addition, the first sensing layer may further include a first color filter disposed on the red sub-pixel to transmit the red light and the NIR light, a second color filter disposed on the green sub-pixel to transmit the green light and the NIR light, and a third color filter disposed on the blue sub-pixel to transmit the blue light and the NIR light.
Also, the second sensing layer may include an NIR sub-pixel configured to sense the NIR light.
Also, the imaging optical system may further include a filter disposed between the first sensing layer and the second sensing layer to block the light having the first wavelength band.
According to another exemplary embodiment, the image sensor may include four sensing layers which are vertically stacked to respectively sense visible light having one of four light colors.
For example, the light having the first wavelength band may be visible light, and the light having the second wavelength band may be NIR light, wherein the four sensing layers may include a first sensing layer configured to sense visible light having a blue light, a second sensing layer configured to sense visible light having green light, a third sensing layer configured to sense visible light having red light, and a fourth sensing layer configured to sense the NIR light, which are sequentially disposed from an upper side closest to the optical shutter module to a lower side farthest from the optical shutter module.
In addition, the imaging optical system may further include a first filter disposed between the first sensing layer and the second sensing layer to block the blue light; a second filter disposed between the second sensing layer and the third sensing layer to block the green light; and a third filter disposed between the third sensing layer and the fourth sensing layer to block the red light.
Meanwhile, according to another aspect of the exemplary embodiments, a 3D image acquisition apparatus includes: an objective lens configured to focus light having a first wavelength band and light having a second wavelength band, wherein the light having the first wavelength band and the light having the second wavelength band are reflected from an external object; an optical shutter module configured to reflect the light having the first wavelength band, which is focused by the objective lens, without modulating the light having the first wavelength band and to modulate the light having the second wavelength band, which is focused by the objective lens, to reflect the modulated light having the second wavelength band; and an image sensor configured to respectively sense the light having the first wavelength band and the modulated light having the second wavelength band, which are reflected by the optical shutter module, and to output a first image signal with respect to the modulated light having the first wavelength band and a second image signal with respect to the modulated light having the second wavelength band; a light source configured to emit light having the second wavelength band onto the external object; an image signal processor configured to generate a 3D image by using the first and second image signals outputted from the image sensor; and a control unit configured to control operations of the light source and the image signal processing unit.
For example, the light source may be configured to emit the light having the second wavelength band with a predetermined period and waveform onto the external object under the control of the control unit.
For example, the light having the first wavelength band may be visible light, and the light having the second wavelength band may be NIR light, wherein the first image signal may be a color image signal having red, green, and blue color components, and the second image signal may be a depth image signal related to a distance between the 3D image acquisition apparatus and the external object.
In addition, the image signal processor may calculate the distance between the 3D image acquisition apparatus and the external object for each of a plurality of pixels by using the depth image signal, and the calculated distance may be combined with the color image signal to produce a final 3D image.
Also, the control unit may control the optical shutter module to modulate the light having the second wavelength band into a predetermined gain waveform having the same period as the light having the second wavelength band.
BRIEF DESCRIPTION OF THE DRAWINGS
These and/or other aspects will become apparent and more readily appreciated from the following description of the exemplary embodiments, taken in conjunction with the accompanying drawings of which:
FIG. 1 is a conceptual view of an imaging optical system and a 3D image acquisition apparatus including the imaging optical system according to an exemplary embodiment;
FIG. 2 is an enlarged view illustrating a structure of an optical shutter module of FIG. 1;
FIG. 3 is a graph for explaining operation characteristics of the optical shutter module of FIG. 1;
FIG. 4 is a schematic plan view illustrating a pixel structure of an image sensor of FIG. 1;
FIGS. 5A through 5D are graphs illustrating transmittance characteristics of color filters arranged within the image sensor of FIG. 4;
FIG. 6 is a schematic conceptual view illustrating a structure of an imaging optical system according to another exemplary embodiment;
FIG. 7 is a view illustrating an example of an operation of a time-division filter and an image sensor of FIG. 6;
FIGS. 8A and 8B are graphs illustrating transmittance characteristics for explaining an operation of the time-division filter of FIG. 6;
FIGS. 9A through 9C are graphs illustrating transmittance characteristics of color filters arranged within the image sensor of FIG. 6;
FIG. 10 is a schematic conceptual view of a structure of an imaging optical system according to another exemplary embodiment;
FIG. 11 is a schematic plan view illustrating a structure of each of a plurality of sensing layers within an image sensor of FIG. 10;
FIGS. 12A through 12C are graphs illustrating an example of sensitivities according to wavelengths for each pixel within a color image sensing unit according to exemplary embodiments;
FIG. 13 is a graph illustrating an example of sensitivities according to wavelengths for each pixel within an IR image sensing unit according to exemplary embodiments;
FIG. 14 is a schematic conceptual view illustrating a structure of an imaging optical system according to another exemplary embodiment;
FIG. 15 is a schematic plan view illustrating a structure of each of a plurality of sensing layers within an image sensor of FIGS. 14; and
FIGS. 16A through 16D are graphs illustrating an example of sensitivities according to wavelengths for each of a plurality of sensing layers of an image sensor of FIG. 15.
DETAILED DESCRIPTION
Reference will now be made in detail to exemplary embodiments, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to the like elements throughout. In this regard, the present exemplary embodiments may have different forms and should not be construed as being limited to the descriptions set forth herein. Accordingly, the exemplary embodiments are merely described below, by referring to the figures, to explain aspects of the present description. Expressions such as “at least one of,” when preceding a list of elements, modify the entire list of elements and do not modify the individual elements of the list.
FIG. 1 is a conceptual view of an imaging optical system and a 3D image acquisition apparatus including the imaging optical system according to an exemplary embodiment.
Referring to FIG. 1, a 3D image acquisition apparatus 100 according to an exemplary embodiment includes a light source 101 configured to generate illumination light having a predetermined wavelength band, an imaging optical system 110 configured to generate a color image signal and a depth image signal by using visible light and the illumination light which are reflected from an external object (not shown), an image signal processing unit 105 (e.g., image signal processor) configured to produce a 3D image by using the color image signal and the depth image signal, and a control unit 102 configured to control operations of the light source 101, the imaging optical system 110, and the image signal processing unit 105. Also, the 3D image acquisition apparatus 100 may further include a memory 106 configured to store a final 3D image and a display panel 104 configured to display the final 3D image.
Also, the imaging optical system 110 according to an exemplary embodiment may include an objective lens 111 configured to focus the visible light and the illumination light which are reflected from the external object, an optical shutter module 112 configured to reflect the visible light focused by the objective lens 111 without modulating the visible light and configured to modulate the illumination light focused by the objective lens 111 to reflect the modulated illumination light, and an image sensor 113 configured to sense the visible light and the illumination light, which are reflected by the optical shutter module 112, to generate the color image signal and the depth image signal.
According to an exemplary embodiment, for example, a light emitting diode (LED) or a laser diode (LD) which emit illumination light having a near-infrared (NIR) wavelength of about 850 nm, which is invisible to the human eye for safety, may be used as the light source 101. However, this is merely an example, and thus, according to other exemplary embodiments, illumination light having an adequately different wavelength band and a different kind of light source may be used according to the design of the light source 101. Also, the light source 101 may emit illumination light having a specific waveform, for example, a sine wave, a ramp wave, a square wave, and the like according to a control signal received from the control unit 102.
FIG. 2 is an enlarged view illustrating a structure of the optical shutter module 112 of FIG. 1. Referring to FIG. 2, the optical shutter module 112 includes an optical shutter 112 b configured to modulate illumination light with a predetermined gain waveform according to a TOF method and a wavelength-selective mirror 112 a disposed on a light incident surface of the optical shutter 112 b. For example, the wavelength-selective mirror 112 a may reflect light RGB having a wavelength band corresponding to that of the visible light and transmit light IR having a wavelength band corresponding to that of the illumination light. Thus, of the light incident onto the optical shutter module 112, the visible light RGB is reflected by the wavelength-selective mirror 112 a toward the image sensor 113, and the illumination light IR is transmitted through the wavelength-selective mirror 112 a to reach the optical shutter 112 b. The wavelength-selective mirror 112 a may be, for example, coated in a thin film form on a surface of the optical shutter 112 b.
According to an exemplary embodiment, the optical shutter 112 b modulates the illumination light IR according to a modulation signal applied from the control unit 102 and reflects the modulated illumination light IR′ toward the image sensor 113. The optical shutter 112 may be, for example, a GaAs-based reflective modulator which is operable at an ultra-high speed of about several tens of MHz to about several hundreds of MHz and has a reflectance varying according to the modulation signal.
FIG. 3 is a graph for explaining operation characteristics of the optical shutter module 112 having the above-described structure. Referring to FIG. 3, since the wavelength-selective mirror 112 a of the optical shutter module 112 reflects the light RGB having the wavelength band corresponding to that of the visible light, the optical shutter module 112 may maintain a constant high reflectance at all times with respect to the visible light having a wavelength band of about 400 nm to about 700 nm. However, since the illumination light IR having the NIR light is transmitted through the wavelength-selective mirror 112 a and then is modulated by the optical shutter 112 b, the optical shutter module 112 may have a reflectance varying according to the modulation signal with respect to the illumination light IR having a wavelength band of about 845 nm to about 855 nm.
The visible light RGB reflected by the optical shutter module 112 and the illumination light IR modulated and reflected by the optical shutter module 112 may be sensed by one image sensor 113. The image sensor 113 may be implemented as a semiconductor imaging device such as a charge-coupled device (CCD) or a complementary metal oxide semiconductor (CMOS). The image sensor 113 has an array of a plurality of pixels. FIG. 4 is a schematic plan view illustrating a pixel structure of the image sensor 113 of FIG. 1. Each of the pixels of the image sensor 113 may include a plurality of sub-pixels each having different sensitivities to wavelengths. Referring to FIG. 4, the image sensor 113 may include a red sub-pixel 113R that senses red light, a green sub-pixel 113G that senses green light, a blue sub-pixel 113B that senses blue light, and an NIR sub-pixel 113IR that senses illumination light having an NIR light. Since the image sensor 113 further includes the NIR sub-pixel 113IR capable of sensing the illumination light having the NIR light in addition to the general red, green, and blue sub-pixels 113R, 113G, and 113B, the image sensor 113 may sense the visible light and the illumination light at the same time.
The image sensor 113 may be realized by respectively disposing color filters having different transmittance characteristics on light incident surfaces of the sub-pixels. For example, FIGS. 5A to 5D are graphs illustrating transmittance characteristics of color filters arranged within the image sensor 113 of FIG. 4. Referring to FIGS. 5A to 5D, a color filter transmitting only blue light having a wavelength of about 400 nm to about 500 nm may be disposed on the blue sub-pixel 113B, and a color filer transmitting only green light having a wavelength of about 500 nm to about 600 nm may be disposed on the green sub-pixel 113G. Also, a color filter transmitting only red light having a wavelength of about 600 nm to about 700 nm may be disposed on the red sub-pixel 113R, and a color filter transmitting only an NIR ray having a wavelength of about 800 nm to about 900 nm may be disposed on the NIR sub-pixel 113IR. Alternatively, a photosensitive material having different sensitivities to each wavelength may be used for each sub-pixel, instead of using the color filters.
The image sensor 113 converts the intensity of the incident light for each sub-pixel into an electrical image signal and outputs the electrical image signal for each sub-pixel to the image signal processing unit 105. The image signal processing unit 105 produces a color image using electrical signals provided from the red, green, and blue sub-pixels 113R, 113G, and 113B of the image sensor 113 and produces a depth image using an electrical signal provided from the NIR sub-pixel 113IR.
Hereinafter, an example operation of the 3D image acquisition apparatus 100 will be briefly described.
First, the light source 101 projects, for example, the NIR illumination light onto an object, the light source 101 being operated under the control of the control unit 102. For example, the light source 101 may project at least three illumination lights having the same period and different phases onto the object in a sequential order of time according to the TOF method.
Then, the NIR illumination light reflected by the object is focused by the objective lens 111. Simultaneously, the general visible light reflected from the object may also be focused by the objective lens 111. Of the light focused by the objective lens 111, the visible light may be reflected by the wavelength-selective mirror 112 a to be incident onto the image sensor 113. On the other hand, the illumination light may pass through the wavelength-selective mirror 112 a to be incident onto the optical shutter 112 b. The light incident onto the optical shutter 112 b may be modulated and reflected by the optical shutter 112 b and then be incident onto the image sensor 113. Here, the optical shutter 112 b may sequentially modulate the at least three illumination lights by using a predetermined gain waveform having the same period as each of the at least three illumination lights according to the TOF method.
The image sensor 113 converts the intensity of the visible light incident onto each of the red, green, and blue sub-pixels into an electrical color image signal and provides an electrical color image signal for each of the red, green, and blue sub-pixels to the image signal processing unit 105. Also, the image sensor 113 sequentially converts the intensities of each of the at least three modulated illumination lights incident onto the NIR sub-pixel 113IR to successively provide the produced at least three electrical depth image signals to the image signal processing unit 105. The image signal processing unit 105 produces a color image by using the electrical color image signals provided from the image sensor 113. Also, the image signal processing unit 105 calculates, for each pixel, a distance between the object and the 3D image acquisition apparatus 100 by using the at least three electrical depth image signals provided from the image sensor 113. The image signal processing unit 105 produces a depth image from the calculated results. Thus, the color image and the depth image may be combined with each other to produce a 3D image. The produced 3D image may be, for example, stored in the memory 106 or displayed on the display panel 104 according to a user selection. Since an algorithm used to calculate the distance between the object and the 3D image acquisition apparatus 100 by using the at least three electrical depth image signals may be performed according to the normal TOF method, a detailed description thereof will be omitted.
As described above, the imaging optical system 110 according to the current exemplary embodiment may obtain the color image and the depth image at the same time by using only one common objective lens 111 and only one image sensor 113. Thus, when compared to a case in which two lenses and two image sensors are provided, or one lens and two image sensors are provided, the number of components may be decreased to reduce production costs. In addition, the imaging optical system 110 and the 3D image acquisition apparatus 100 including the imaging optical system 110 may be reduced in size and weight. Also, since the color image and the depth image are photographed using the same objective lens 111 and have the same viewing angle, it is unnecessary to perform a process of matching their respective viewing angles. Also, since the color image and the depth image are obtained through the same image sensor 113, image matching between the pixels within the color and depth images may be easily performed.
Although the image sensor 113 senses the visible light and the illumination light at the same time by using a space division method according to the exemplary embodiment of FIG. 1, it may be possible to sense the visible light and the illumination light at different times by using a time division method. FIG. 6 is a schematic conceptual view illustrating a structure of an image optical system 120 according to another exemplary embodiment which implements the time division method.
Referring to FIG. 6, the imaging optical system 120 may include an objective lens 111 configured to focus visible light and illumination light which are reflected from an external object, an optical shutter module 112 configured to reflect the visible light focused by the objective lens 111 without modulating the visible light and modulate the illumination light focused by the objective lens 111 to reflect the modulated illumination light, an image sensor 123 configured to respectively sense the visible light and the illumination light, which are reflected by the optical shutter module 112, and to generate a color image signal and a depth image signal, and a time division filter 114 disposed between the optical shutter module 112 and the image sensor 123 to selectively transmit only one of the visible light and the illumination light.
According to the exemplary embodiment shown in FIG. 7, the time division filter 114 may be configured to transmit only the visible light at a time 1 so that the image sensor 123 senses only the visible light at time 1 and transmit only the illumination light at a time 2 so that the image sensor 123 senses only the illumination light at time 2. For example, the time division filter 114 may transmit visible light having a wavelength band of about 350 nm to about 700 nm at the time 1 as shown in FIG. 8A. Also, the time division filter 114 may transmit illumination light having a wavelength band in a range of about 800 nm to about 900 nm at the time 2 as shown in FIG. 8B. The time division filter 114 may be realized using a disc-shaped color wheel in which a semicircular color filter configured to transmit the visible light and a semicircular color filter configured to transmit the NIR ray are coupled to each other. For example, the color wheel may be rotated using a motor such that, at the time 1, the color filter configured to transmit the visible light is located on an optical path. Also, the color wheel may be rotated using the motor such that, at the time 2, the color filter configured to transmit the NIR ray is located on the optical path. Furthermore, according to other exemplary embodiments, instead of the color wheel, any mechanical or electrical structure which switches a location of the color filter configured to transmit the visible light and the color filter configured to transmit the NIR ray on an optical path between the optical shutter module 112 and the image sensor 123 may be implemented.
Since the visible light and the illumination light are sensed at different times, the image sensor 123 may only include three kinds of sub-pixels. For example, as shown in an upper side of FIG. 7, the image sensor 123 may include a red sub-pixel 123R that senses red light, a green sub-pixel 123G that senses green light, and a blue sub-pixel 123B that senses blue light. Also, according to the current exemplary embodiment, the sub-pixels 123R, 123G, and 123B may all be configured to sense the illumination light having the NIR light at the time 2. That is, the red sub-pixel 123R may sense the red light and the NIR ray, the green sub-pixel 123G may sense the green light and the NIR ray, and the blue sub-pixel 123B may sense the blue light and the NIR ray.
FIGS. 9A through 9C are graphs illustrating transmittance characteristics of color filters arranged on light incident surfaces of the image sensor 123 according to an exemplary embodiment. Referring to FIGS. 9A through 9C, a color filter configured to transmit blue light having a wavelength of about 400 nm to about 500 nm and an NIR ray having a wavelength of about 800 nm to about 900 nm may be disposed on the blue sub-pixel 123B. Also, a color filter configured to transmit green light having a wavelength of about 500 nm to about 600 nm and the NIR ray having the wavelength of about 800 nm to about 900 nm may be disposed on the green sub-pixel 123G, and a color filter configured to transmit red light having a wavelength of about 600 nm to about 700 nm and the NIR ray having the wavelength of about 800 nm to about 900 nm may be disposed on the red sub-pixel 123R. The sub-pixels 123R, 123G, and 123B of the image sensor 123 may sense the red, green, and blue light at a time 1, and the image sensor 123 may sense the illumination light having the NIR light in the entire region of the image sensor 123 at the time 2.
The imaging optical system 120 may be applied to the 3D image acquisition apparatus 100 of FIG. 1. Hereinafter, an example operation of the 3D image acquisition apparatus including the imaging optical system 120 of FIG. 6 will be described.
The control unit 102 controls the time division filter 114 to transmit visible light at a time 1 and sense the visible light through the image sensor 123. The image signal processing unit 105 produces a color image by using an image signal provided from the image sensor 123. The control unit 102 controls the time division filter 114 to transmit the illumination light and controls the light source 101 to emit illumination light at a time 2. Thus, the image sensor 123 senses the illumination light, and the image signal processing unit 105 produces a depth image according to the TOF method by using the image signal provided from the image sensor 123.
As described above, at least three illumination lights having different phases may be sequentially projected onto the object. The at least three illumination lights may be sensed at the time 2 all at once. Alternatively, the at least three illumination lights may be separately sensed at several different times. For example, in a case where four illumination lights are used, the visible light may be sensed at the time 1, and then, the four illumination lights may be sequentially sensed at different times (e.g., four different times 2) In this case, the light source 101 may sequentially project four illumination lights each having different phases onto the object at the time 2. Alternatively, the illumination light including the four illumination lights may be sensed over the four times 2, in the order of: visible light (time 1) -> illumination light 1 (time 2) -> visible light (time 1) -> illumination light 2 (time 2) -> visible light (time 1) -> illumination light 3 (time 2) -> visible light (time 1) -> illumination light 4 (time 2). Alternatively, the illumination light including the four illumination lights may be sensed over two times 2, in the order of: visible light (time 1) -> illumination light 1, 2 (time 2) -> visible light (time 1) -> illumination light 3, 4 (time 2) at the time 2.
According to an exemplary embodiment, when the visible light is sensed at the time 1, only three kinds of sub-pixels 123R, 123G, and 123B are used within the image sensor 123. Thus, when compared to a case in which four kinds of sub-pixels are used, the color image may be improved in resolution. Also, when the illumination light is sensed at the time 2, the illumination light may be sensed throughout the entire area of the image sensor 123. Thus, distance calculation may be increased in accuracy, and light utilization efficiency may be improved.
FIG. 10 is a schematic conceptual view of a structure of an imaging optical system 130 according to another exemplary embodiment.
Referring to FIG. 10, the imaging optical system 130 includes an objective lens 111 configured to focus visible light and illumination light which are reflected from an external object, an optical shutter module 112 configured to reflect the visible light focused by the objective lens 111 without modulating the visible light and modulate the illumination light focused by the objective lens 111 to reflect the modulated illumination light, and an image sensor 133 configured to respectively sense the visible light and the illumination light, which are reflected by the optical shutter module 112, to generate a color image signal and a depth image signal.
The exemplary embodiment of FIG. 10 is different from that of FIG. 1 in that the image sensor 133 includes two sensing layers which are vertically stacked. That is, the image sensor 133 includes a first sensing layer 133 a and a second sensing layer 133 b which are vertically stacked. The first sensing layer 133 a may be formed of a light absorption material such as silicon and may be disposed on the second sensing layer 133 b to sense the visible light. Also, the second sensing layer 133 b disposed under the first sensing layer 133 a may be formed as an illumination layer to sense the NIR light. In general, light having a relatively short wavelength may have a relatively shallow penetration depth with respect to a light absorption material such as silicon, and light having a relatively long wavelength may have a relatively deep penetration depth with respect to the light absorption material. The second sensing layer 133 b configured to sense the illumination light is disposed under the first sensing layer 133 a because the illumination light has a relatively long wavelength and is able to more deeply penetrate through the first sensing layer 133 a and reach the second sensing layer 133 b.
FIG. 11 is a schematic plan view illustrating a structure of each of a plurality of sensing layers within the image sensor 133 of FIG. 10. As shown in FIG. 11, the first sensing layer 133 a may include only three kinds of sub-pixels. For example, the first sensing layer 133 a may include a red sub-pixel 133R that senses red light, a green sub-pixel 133G that senses green light and a blue sub-pixel 133B that senses blue light. In the current exemplary embodiment, transmittance characteristics of color filters disposed on the sub-pixels 133R, 133G, and 133B may be the same as those of the color filters illustrated in FIGS. 9A to 9C. That is, referring to FIGS. 9A through 9C, a color filter configured to transmit blue light having a wavelength of about 400 nm to about 500 nm and an NIR ray having a wavelength of about 800 nm to about 900 nm may be disposed on the blue sub-pixel 133B. Also, a color filter configured to transmit green light having a wavelength of about 500 nm to about 600 nm and the NIR ray having the wavelength of about 800 nm to about 900 nm may be disposed on the green sub-pixel 133G, and a color filter configured to transmit red light having a wavelength of about 600 nm to about 700 nm and the NIR ray having the wavelength of about 800 nm to about 900 nm may be disposed on the red sub-pixel 133R.
Thus, the illumination light having the NIR ray together with the visible light are incident onto each of the sub-pixels 133R, 133G, and 133B of the first sensing layer 133 a. To remove an effect caused by the illumination light, as shown in FIGS. 12A through 12C, the sub-pixels 133R, 133G, and 133B may be formed of photosensitive materials having different sensitivities based on wavelength. For example, the blue sub-pixel 133B may be formed of a photosensitive material having high sensitivity with respect to blue light having a wavelength of about 400 nm to about 500 nm. Also, the green sub-pixel 133G may be formed of a photosensitive material having high sensitivity with respect to green light having a wavelength of about 500 nm to about 600 nm, and the red sub-pixel 133R may be formed of a photosensitive material having high sensitivity with respect to red light having a wavelength of about 600 nm to about 700 nm.
The second sensing layer 133 b senses the illumination light that is transmitted through the first sensing layer 133 a. In the current exemplary embodiment, the second sensing layer 133 b includes only a single type of sub-pixel 133IR that senses the illumination light. That is, as shown in FIG. 13, the second sensing layer 133 b may be configured so that the all pixels in the sensing layer 133 b have high sensitivities with respect to the illumination light. Here, to remove an effect caused by the remaining visible light that is not absorbed by the first sensing layer 133 a, as shown in FIG. 11, a filter 135 blocking the visible light may be disposed between the first sensing layer 133 a and the second sensing layer 133 b.
Alternatively, according to other exemplary embodiments, the sub-pixels 133R, 133G, and 133B may be formed of the same photosensitive material, instead of being formed of different photosensitive materials. For example, the light incident onto the blue sub-pixel 133B of the first sensing layer 133 a may have the same intensity as the sum of the blue light and the illumination light. On the other hand, the sub-pixel 133IR of the second sensing layer 133 a may sense only the illumination light. Thus, considering the intensity of the illumination light sensed by the sub-pixel 133IR of the second sensing layer 133 b facing the blue sub-pixel 133B of the first sensing layer 133 a, the intensity of the blue light incident onto the blue sub-pixel 133B of the first sensing layer 133 a may be calculated. Similarly, considering the intensity of the illumination light sensed by the sub-pixel 133IR of the second sensing layer 133 b, intensities of the green and red light incident onto the green and red sub-pixels 133G and 133R of the first sensing layer 133 a facing the sub-pixel 133IR of the second sensing layer 133 b may be calculated.
According to the exemplary embodiment of FIG. 10, the visible light and the illumination light may be sensed at the same time without performing a time division process, and also, advantages described in the exemplary embodiment of FIG. 6 may be obtained.
FIG. 14 is a schematic conceptual view illustrating a structure of an imaging optical system 140 according to another exemplary embodiment.
Referring to FIG. 14, the imaging optical system 140 may include an objective lens 111 configured to focus visible light and illumination light which are reflected from an external object, an optical shutter module 112 configured to reflect the visible light focused by the objective lens 111 without modulating the visible light and modulate the illumination light focused by the objective lens 111 to reflect the modulated illumination light, and an image sensor 143 configured to respectively sense the visible light and the illumination light, which are reflected by the optical shutter module 112, to generate a color image signal and a depth image signal.
The exemplary embodiment of FIG. 14 is different from the exemplary embodiment of FIG. 10 in that the image sensor 143 includes four sensing layers which are vertically stacked. That is, although the image sensor 133 includes only the first sensing layer 133 a configured to sense the visible light and the second sensing layer 133 b configured to sense the NIR ray in the exemplary embodiment of FIG. 10, the image sensor 143 may include four sensing layers 143 a to 143 d respectively configured to sense the red, green and blue light and the NIR ray in the exemplary embodiment of FIG. 14.
For example, referring to FIG. 15, the image sensor 143 includes a first sensing layer 143 a that senses the blue light, a second sensing layer 143 b that senses the green light, a third sensing layer 143 c that senses the red light, and a fourth sensing layer 143 d that senses the NIR ray, which are sequentially disposed from an upper side. As previously disclosed above, light having a relatively short wavelength may have a relatively shallow penetration depth with respect to a light absorption material such as silicon, and light having a relatively long wavelength may have a relatively deep penetration depth with respect to the light absorption material. Thus, the image sensor 143 may be configured so that the uppermost first sensing layer 143 a senses the blue light (wavelength of about 400 nm to about 500 nm), and the lowermost fourth sensing layer 143 d senses the NIR ray (wavelength of about 800 nm to about 900 nm).
Also, as shown in FIGS. 16A through 16D, the intensity of each of the sensing layers 143 a to 143 d may be selected so that each of the four sensing layers 143 a to 143 d accurately senses light of a corresponding color. For example, the first sensing layer 143 a may be formed of a photosensitive material having high sensitivity with respect to blue light having a wavelength of about 400 nm to about 500 nm, and the second sensing layer 143 b may be formed of a photosensitive material having high sensitivity with respect to green light having a wavelength of about 500 nm to about 600 nm. Also, the third sensing layer 143 c may be formed of a photosensitive material having high sensitivity with respect to red light having a wavelength of about 600 nm to about 700 nm, and the fourth sensing layer 143 d may be formed of a photosensitive material having high sensitivity with respect to light having a wavelength of about 800 nm to about 900 nm.
Alternatively, the sensing layers 143 a to 143 d may be formed of the same photosensitive material, instead of being formed of different photosensitive materials. In this case, a filter 145 a configured to block the blue light may be disposed between the first sensing layer 143 a and the second sensing layer 143 b. Also, a filter 145 b configured to block the green light may be disposed between the second sensing layer 143 b and the third sensing layer 143 c, and a filter 145 c configured to block the red light may be disposed between the third sensing layer 143 c and the fourth sensing layer 143 d. Although, according to the exemplary embodiment shown in FIG. 15, the filters 145 a, 145 b, and 145 c are added to separate the first through fourth sensing layers, according to other exemplary embodiments, the light absorption materials of each of the first through third sensing layers 143 a-143 c may also serve as optical filters by adjusting a thickness thereof. For example, the thickness of the light absorption material of the first sensing layer 143 a may be adjusted to completely absorb the blue light within the first sensing layer 143 a, the thickness of the light absorption material of the second sensing layer 143 b may be adjusted to completely absorb the green light within the second sensing layer 143 b, and the thickness of the light absorption material of the third sensing layer 143 c may be adjusted to completely absorb the red light within the third sensing layer 143 c.
As a result, the first sensing layer 143 a may sense all of the blue, green, red, and illumination light. Also, the second sensing layer 143 b may sense the green, red, and illumination light, the third sensing layer 143 c may sense the red and illumination light, and the fourth sensing layer 143 d may sense only the illumination light. Considering these relationships, the intensity of the blue light incident onto the first sensing layer 143 a, the intensity of the green light incident onto the second sensing layer 143 b, and the intensity of the red light incident onto the third sensing layer 143 c may be calculated.
According to the current exemplary embodiment, since the blue, green, red, and illumination light may be sensed through the separate sensing layers 143 a to 143 d, respectively, a high fill factor may be obtained with respect to all colors. Thus, light utilization efficiency may be improved, and high resolution color and depth images may be obtained.
The above-described 3D image acquisition apparatus 100 may be applied to various fields. For example, the 3D image acquisition apparatus 100 may be applied to a range-finding device, a gesture recognizer, a game console, and the like.
For ease of understanding, exemplary embodiments with respect to the imaging optical system and the 3D image acquisition apparatus including the imaging optical system have been described with reference to the accompanying drawings. However, it should be understood that the exemplary embodiments described therein should be considered in a descriptive sense only and not for purposes of limitation. It is to be understood that the present disclosure is not limited to the exemplary embodiments as described above. This is done because there may be variations and modifications made by those skilled in the art.

Claims (19)

What is claimed is:
1. An imaging optical system comprising:
an objective lens configured to focus light having a first wavelength band and light having a second wavelength band, wherein the light having the first wavelength band and the light having the second wavelength band are reflected from an external object;
an optical shutter module configured to reflect the light having the first wavelength band, which is focused by the objective lens, along a first reflection path without modulating the light having the first wavelength band and to modulate the light having the second wavelength band, which is focused by the objective lens, and reflect the modulated light having the second wavelength band along a second reflection path parallel to the first reflection path; and
an image sensor provided along the first reflection path and the second reflection path and configured to respectively sense the light having the first wavelength band and the modulated light having the second wavelength band, which are reflected by the optical shutter module, and to output a first image signal with respect to the light having the first wavelength band and a second image signal with respect to the modulated light having the second wavelength band.
2. The imaging optical system of claim 1, wherein the optical shutter module comprises an optical shutter configured to modulate the light having the second wavelength band with a predetermined gain waveform and a wavelength-selective mirror disposed on a light incident surface of the optical shutter to reflect the light having the first wavelength band and transmit the light having the second wavelength band,
wherein the optical shutter module maintains a constant reflectance with respect to the light having the first wavelength band and has a reflectance varying according to a modulation signal with respect to the light having the second wavelength band.
3. The imaging optical system of claim 1, wherein the light having the first wavelength band is visible light, and the light having the second wavelength band is near-infrared (NIR) light,
wherein the image sensor comprises an array of a plurality of pixels, and each of the pixels comprises a red sub-pixel configured to sense red light, a green sub-pixel configured to sense green light, a blue sub-pixel configured to sense blue light, and an NIR sub-pixel configured to sense the NIR light.
4. The imaging optical system of claim 1, further comprising a time division filter disposed between the optical shutter module and the image sensor to selectively transmit one of the light having the first wavelength band and the modulated light having the second wavelength band.
5. The imaging optical system of claim 4, wherein the time division filter is configured to transmit the light having the first wavelength band during a first time period and the modulated light having the second wavelength band during a second time period, and
wherein the time division filter is configured to switch between a first color filter configured to transmit the light having the first wavelength band on an optical path between the optical shutter module and the image sensor, and a second color filter configured to transmit the modulated light having the second wavelength band on the optical path between the optical shutter module and the image sensor.
6. The imaging optical system of claim 5, wherein the time division filter is configured to transmit the light having the first wavelength band during a first time period and the modulated light having the second wavelength band during a second time period, and
wherein the light having the first wavelength band is visible light, and the light having the second wavelength band is near-infrared (NIR) light,
wherein the image sensor comprises an array of a plurality of pixels, and each of the pixels comprises a red sub-pixel configured to sense red light and the NIR light, a green sub-pixel configured to sense green light and the NIR light, and a blue sub-pixel configured to sense blue light and the NIR light.
7. The imaging optical system of claim 5, wherein the time division filter is configured to transmit the light having the first wavelength band during a first time period and the modulated light having the second wavelength band during a second time period, and
wherein the image sensor is configured to sense the light having the first wavelength band during the first time period and the modulated light having the second wavelength band during the second time period.
8. The imaging optical system of claim 1, wherein the image sensor comprises a first sensing layer configured to sense the light having the first wavelength band and a second sensing layer disposed under the first sensing layer such that the first sensing layer is disposed between the second sensing layer and the optical shutter module, the second sensing layer being configured to sense the light having the second wavelength band.
9. The imaging optical system of claim 8, wherein the light having the first wavelength band is visible light, and the light having the second wavelength band is near-infrared (NIR) light,
wherein the first sensing layer comprises a red sub-pixel configured to sense the visible light having red light, a green sub-pixel configured to sense the visible light having green light, and a blue sub-pixel configured to sense the visible light having blue light, and
wherein the first sensing layer further comprises a first color filter disposed on the red sub-pixel to transmit the red light and the NIR light, a second color filter disposed on the green sub-pixel to transmit the green light and the NIR light, and a third color filter disposed on the blue sub-pixel to transmit the blue light and the NIR light.
10. The imaging optical system of claim 8, wherein the light having the first wavelength band is visible light, and the light having the second wavelength band is near-infrared (NIR) light,
wherein the first sensing layer comprises a red sub-pixel configured to sense the visible light having red light, a green sub-pixel configured to sense the visible light having green light, and a blue sub-pixel configured to sense the visible light having blue light, and
wherein the second sensing layer comprises an NIR sub-pixel configured to sense the NIR light.
11. The imaging optical system of claim 8, further comprising a filter disposed between the first sensing layer and the second sensing layer to block the light having the first wavelength band.
12. The imaging optical system of claim 1, wherein the image sensor comprises four sensing layers which are vertically stacked to respectively sense visible light having one of four light colors,
wherein the light having the first wavelength band is visible light, and the light having the second wavelength band is near-infrared (NIR) light, and
wherein the four sensing layers comprise a first sensing layer configured to sense visible light having a blue light, a second sensing layer configured to sense visible light having green light, a third sensing layer configured to sense visible light having red light, and a fourth sensing layer configured to sense the NIR light, which are sequentially disposed from an upper side closest to the optical shutter module to a lower side farthest from the optical shutter module.
13. The imaging optical system of claim 12, further comprising a first filter disposed between the first sensing layer and the second sensing layer to block the blue light, a second filter disposed between the second sensing layer and the third sensing layer to block the green light, and a third filter disposed between the third sensing layer and the fourth sensing layer to block the red light.
14. A 3D image acquisition apparatus comprising:
an imaging optical system comprising:
an objective lens configured to focus light having a first wavelength band and light having a second wavelength band, wherein the light having the first wavelength band and the light having the second wavelength band are reflected from an external object;
an optical shutter module configured to reflect the light having the first wavelength band, which is focused by the objective lens, along a first reflection path without modulating the light having the first wavelength band and to modulate the light having the second wavelength band, which is focused by the objective lens, and to reflect the modulated light having the second wavelength band along a second reflection path parallel to the first reflection path; and
an image sensor provided along the first reflection path and the second reflection path and configured to respectively sense the light having the first wavelength band and the modulated light having the second wavelength band, which are reflected by the optical shutter module, and to output a first image signal with respect to the light having the first wavelength band and a second image signal with respect to the modulated light having the second wavelength band;
a light source configured to emit light having the second wavelength band onto the external object;
an image signal processor configured to generate a 3D image by using the first and second image signals outputted from the image sensor; and
a control unit configured to control operations of the light source and the image signal processing unit.
15. The 3D image acquisition apparatus of claim 14, wherein the light source is configured to emit the light having the second wavelength band with a predetermined period and waveform onto the external object under control of the control unit.
16. The 3D image acquisition apparatus of claim 15, wherein the light having the first wavelength band is visible light, and the light having the second wavelength band is near-infrared (NIR) light,
wherein the first image signal is a color image signal having red, green, and blue color components, and the second image signal is a depth image signal related to a distance between the 3D image acquisition apparatus and the external object.
17. The 3D image acquisition apparatus of claim 16, wherein the image signal processor is configured to calculate the distance between the 3D image acquisition apparatus and the external object for each of a plurality of pixels by using the depth image signal, and the calculated distance is combined with the color image signal to produce a final 3D image.
18. The 3D image acquisition apparatus of claim 14, wherein the control unit controls the optical shutter module to modulate the light having the second wavelength band into a predetermined gain waveform having the same period as the light having the second wavelength band.
19. A gesture recognizer comprising the 3D image acquisition apparatus of claim 14.
US14/028,961 2012-10-10 2013-09-17 Imaging optical system and 3D image acquisition apparatus including the imaging optical system Active 2036-01-11 US9667944B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US15/589,506 US9998730B2 (en) 2012-10-10 2017-05-08 Imaging optical system and 3D image acquisition apparatus including the imaging optical system

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR10-2012-0112660 2012-10-10
KR1020120112660A KR101858577B1 (en) 2012-10-10 2012-10-10 Imaging optical system and 3D image acquisition apparatus including the imaging optical system

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US15/589,506 Continuation US9998730B2 (en) 2012-10-10 2017-05-08 Imaging optical system and 3D image acquisition apparatus including the imaging optical system

Publications (2)

Publication Number Publication Date
US20140098192A1 US20140098192A1 (en) 2014-04-10
US9667944B2 true US9667944B2 (en) 2017-05-30

Family

ID=50432375

Family Applications (2)

Application Number Title Priority Date Filing Date
US14/028,961 Active 2036-01-11 US9667944B2 (en) 2012-10-10 2013-09-17 Imaging optical system and 3D image acquisition apparatus including the imaging optical system
US15/589,506 Active US9998730B2 (en) 2012-10-10 2017-05-08 Imaging optical system and 3D image acquisition apparatus including the imaging optical system

Family Applications After (1)

Application Number Title Priority Date Filing Date
US15/589,506 Active US9998730B2 (en) 2012-10-10 2017-05-08 Imaging optical system and 3D image acquisition apparatus including the imaging optical system

Country Status (2)

Country Link
US (2) US9667944B2 (en)
KR (1) KR101858577B1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170244953A1 (en) * 2012-10-10 2017-08-24 Samsung Electronics Co., Ltd. Imaging optical system and 3d image acquisition apparatus including the imaging optical system
US20190317212A1 (en) * 2018-04-11 2019-10-17 Microsoft Technology Licensing, Llc Time of flight and picture camera

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20150010230A (en) * 2013-07-18 2015-01-28 삼성전자주식회사 Method and apparatus for generating color image and depth image of an object using singular filter
US20150245767A1 (en) * 2014-02-28 2015-09-03 Lrs Identity, Inc. Dual iris and color camera in a mobile computing device
DE102014213406A1 (en) 2014-07-10 2016-01-14 Osram Opto Semiconductors Gmbh Semiconductor laser component and camera
US9508681B2 (en) * 2014-12-22 2016-11-29 Google Inc. Stacked semiconductor chip RGBZ sensor
KR102282218B1 (en) * 2015-01-30 2021-07-26 삼성전자주식회사 Imaging Optical System for 3D Image Acquisition Apparatus, and 3D Image Acquisition Apparatus Including the Imaging Optical system
KR102372087B1 (en) * 2015-10-28 2022-03-08 삼성전자주식회사 Method and apparatus for generating depth image
CN106921856B (en) * 2015-12-25 2019-07-12 北京三星通信技术研究有限公司 Processing method, detection dividing method and the relevant apparatus and equipment of stereo-picture
KR101747603B1 (en) * 2016-05-11 2017-06-16 재단법인 다차원 스마트 아이티 융합시스템 연구단 Color night vision system and operation method thereof
KR20180016120A (en) * 2016-08-05 2018-02-14 삼성전자주식회사 3D depth sensor and Method of measuring distance using the 3D depth sensor
US10551614B2 (en) * 2017-08-14 2020-02-04 Facebook Technologies, Llc Camera assembly with programmable diffractive optical element for depth sensing
EP3474328B1 (en) * 2017-10-20 2021-09-29 Samsung Electronics Co., Ltd. Combination sensors and electronic devices
GB201804550D0 (en) 2018-03-21 2018-05-02 Trw Ltd A 3D imaging system and method of 3D imaging
KR102523762B1 (en) * 2018-06-20 2023-04-20 엘지이노텍 주식회사 Image sensor and camera module using the same
CN110737038A (en) * 2018-07-18 2020-01-31 福州高意光学有限公司 Low-angle drift multi-band-pass filter
EP3995858A4 (en) * 2019-07-05 2023-01-11 Nuvoton Technology Corporation Japan Information processing system, sensor system, information processing method, and program
KR102184210B1 (en) * 2019-12-26 2020-11-27 박배억 3d camera system

Citations (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2000102035A (en) 1998-09-21 2000-04-07 Fuji Photo Film Co Ltd Stereoscopic photograph system
US6100517A (en) 1995-06-22 2000-08-08 3Dv Systems Ltd. Three dimensional camera
US6323942B1 (en) 1999-04-30 2001-11-27 Canesta, Inc. CMOS-compatible three-dimensional image sensor IC
US7016519B1 (en) 1998-10-15 2006-03-21 Nippon Hoso Kyokai Method and device for detecting three-dimensional information
US7060957B2 (en) 2000-04-28 2006-06-13 Csem Centre Suisse D'electronique Et Microtechinique Sa Device and method for spatially resolved photodetection and demodulation of modulated electromagnetic waves
US7095487B2 (en) 2003-10-09 2006-08-22 Honda Motor Co., Ltd. Systems and methods for determining depth using shuttered light pulses
US7138619B1 (en) * 2004-09-28 2006-11-21 Rockwell Collins, Inc. Method and apparatus for coincident viewing at a plurality of wavelengths
US7224384B1 (en) 1999-09-08 2007-05-29 3Dv Systems Ltd. 3D imaging system
US20080291532A1 (en) 2007-05-26 2008-11-27 James Jianguo Xu 3-d optical microscope
US7531781B2 (en) * 2005-02-25 2009-05-12 Sony Corporation Imager with image-taking portions optimized to detect separated wavelength components
KR20090104857A (en) 2007-01-22 2009-10-06 캘리포니아 인스티튜트 오브 테크놀로지 Method and apparatus for quantitative 3-d imaging
US20100328750A1 (en) 2009-06-24 2010-12-30 Samsung Electronics Co., Ltd. High-speed optical modulator and method of modulating light by using the same
US20110074659A1 (en) 2008-12-24 2011-03-31 Samsung Electronics Co., Ltd. High speed optical shutter, method of operating the same and apparatus including the same
US20110102547A1 (en) * 2009-11-04 2011-05-05 Sul Sang-Chul Three-Dimensional Image Sensors and Methods of Manufacturing the Same
US20110134222A1 (en) 2008-08-03 2011-06-09 Microsoft International Holdings B.V. Rolling Camera System
US20110170160A1 (en) 2010-01-08 2011-07-14 Samsung Electronics Co., Ltd. Optical image shutter and method of manufacturing the same
KR20110085785A (en) 2010-01-21 2011-07-27 삼성전자주식회사 Method of extractig depth information and optical apparatus employing the method
US20110181936A1 (en) 2010-01-22 2011-07-28 Samsung Electronics Co., Ltd. Optical modulator
KR20110109031A (en) 2010-03-30 2011-10-06 김태권 Stereoscopic image camera apparatus and driving method the same
KR20120045462A (en) 2010-10-29 2012-05-09 삼성전자주식회사 Beam splitter for 3d camera and 3d image acquisition apparatus employing the same
US20120162380A1 (en) 2010-12-28 2012-06-28 Gwangju Institute Of Science And Technology Optical modulator using multiple fabry-perot resonant modes and apparatus for capturing 3d image including the optical modulator
KR20120071970A (en) 2010-12-23 2012-07-03 삼성전자주식회사 3d image acquisition apparatus and method of extractig depth information in the 3d image acquisition apparatus
US20120300038A1 (en) * 2011-05-24 2012-11-29 Inha Industry Partnership Institute Photonic crystal type light modulator and 3d image acquisition apparatus employing the same
KR20120135751A (en) 2011-06-07 2012-12-17 삼성전자주식회사 3d image acquisition apparatus employing interchangeable lens type
KR20130066287A (en) 2011-12-12 2013-06-20 삼성전자주식회사 Transmissive image modulator using multi fabry-perot resonant modes and multi absorption modes
KR20130080641A (en) 2012-01-05 2013-07-15 삼성전자주식회사 Transmissive image modulator including stacked diode structure having multi absorption modes
KR20130091194A (en) 2012-02-07 2013-08-16 삼성전자주식회사 3d image acquisition apparatus and method of extractig depth information in the 3d image acquisition apparatus
US20140014838A1 (en) * 2012-07-16 2014-01-16 Karen Denise Hendrix Optical filter and sensor system
KR20140051030A (en) 2012-10-22 2014-04-30 삼성전자주식회사 Imaging optical system for 3d image acquisition apparatus, and 3d image acquisition apparatus including the imaging optical system
US20150138620A1 (en) 2011-12-12 2015-05-21 Gwangju Institute Of Science And Technology Transmissive image modulator using multi-fabry-perot resonant mode and multi-absorption mode
US9402067B2 (en) 2012-10-22 2016-07-26 Samsung Electronics Co., Ltd. Imaging optical system for 3D image acquisition apparatus, and 3D image acquisition apparatus including the imaging optical system

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4882297B2 (en) * 2004-12-10 2012-02-22 ソニー株式会社 Physical information acquisition apparatus and semiconductor device manufacturing method
JP2008147749A (en) * 2006-12-06 2008-06-26 Umc Japan Color camera device
JP2009010627A (en) * 2007-06-27 2009-01-15 Panasonic Corp Solid-state imaging apparatus and camera using the same
KR20110140010A (en) 2010-06-24 2011-12-30 삼성전자주식회사 Image sensor using near infrared signal
KR101858577B1 (en) * 2012-10-10 2018-05-16 삼성전자주식회사 Imaging optical system and 3D image acquisition apparatus including the imaging optical system

Patent Citations (34)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6100517A (en) 1995-06-22 2000-08-08 3Dv Systems Ltd. Three dimensional camera
JP2000102035A (en) 1998-09-21 2000-04-07 Fuji Photo Film Co Ltd Stereoscopic photograph system
US7016519B1 (en) 1998-10-15 2006-03-21 Nippon Hoso Kyokai Method and device for detecting three-dimensional information
US6323942B1 (en) 1999-04-30 2001-11-27 Canesta, Inc. CMOS-compatible three-dimensional image sensor IC
US7224384B1 (en) 1999-09-08 2007-05-29 3Dv Systems Ltd. 3D imaging system
US7060957B2 (en) 2000-04-28 2006-06-13 Csem Centre Suisse D'electronique Et Microtechinique Sa Device and method for spatially resolved photodetection and demodulation of modulated electromagnetic waves
US7095487B2 (en) 2003-10-09 2006-08-22 Honda Motor Co., Ltd. Systems and methods for determining depth using shuttered light pulses
US7138619B1 (en) * 2004-09-28 2006-11-21 Rockwell Collins, Inc. Method and apparatus for coincident viewing at a plurality of wavelengths
US7531781B2 (en) * 2005-02-25 2009-05-12 Sony Corporation Imager with image-taking portions optimized to detect separated wavelength components
KR20090104857A (en) 2007-01-22 2009-10-06 캘리포니아 인스티튜트 오브 테크놀로지 Method and apparatus for quantitative 3-d imaging
US20080291532A1 (en) 2007-05-26 2008-11-27 James Jianguo Xu 3-d optical microscope
US20110134222A1 (en) 2008-08-03 2011-06-09 Microsoft International Holdings B.V. Rolling Camera System
US20110074659A1 (en) 2008-12-24 2011-03-31 Samsung Electronics Co., Ltd. High speed optical shutter, method of operating the same and apparatus including the same
US20100328750A1 (en) 2009-06-24 2010-12-30 Samsung Electronics Co., Ltd. High-speed optical modulator and method of modulating light by using the same
US20110102547A1 (en) * 2009-11-04 2011-05-05 Sul Sang-Chul Three-Dimensional Image Sensors and Methods of Manufacturing the Same
US20110170160A1 (en) 2010-01-08 2011-07-14 Samsung Electronics Co., Ltd. Optical image shutter and method of manufacturing the same
KR20110085785A (en) 2010-01-21 2011-07-27 삼성전자주식회사 Method of extractig depth information and optical apparatus employing the method
US20110181936A1 (en) 2010-01-22 2011-07-28 Samsung Electronics Co., Ltd. Optical modulator
KR20110109031A (en) 2010-03-30 2011-10-06 김태권 Stereoscopic image camera apparatus and driving method the same
KR20120045462A (en) 2010-10-29 2012-05-09 삼성전자주식회사 Beam splitter for 3d camera and 3d image acquisition apparatus employing the same
US8902411B2 (en) 2010-12-23 2014-12-02 Samsung Electronics Co., Ltd. 3-dimensional image acquisition apparatus and method of extracting depth information in the 3D image acquisition apparatus
KR20120071970A (en) 2010-12-23 2012-07-03 삼성전자주식회사 3d image acquisition apparatus and method of extractig depth information in the 3d image acquisition apparatus
US20120162380A1 (en) 2010-12-28 2012-06-28 Gwangju Institute Of Science And Technology Optical modulator using multiple fabry-perot resonant modes and apparatus for capturing 3d image including the optical modulator
US20120300038A1 (en) * 2011-05-24 2012-11-29 Inha Industry Partnership Institute Photonic crystal type light modulator and 3d image acquisition apparatus employing the same
KR20120135751A (en) 2011-06-07 2012-12-17 삼성전자주식회사 3d image acquisition apparatus employing interchangeable lens type
US20150138620A1 (en) 2011-12-12 2015-05-21 Gwangju Institute Of Science And Technology Transmissive image modulator using multi-fabry-perot resonant mode and multi-absorption mode
KR20130066287A (en) 2011-12-12 2013-06-20 삼성전자주식회사 Transmissive image modulator using multi fabry-perot resonant modes and multi absorption modes
KR20130080641A (en) 2012-01-05 2013-07-15 삼성전자주식회사 Transmissive image modulator including stacked diode structure having multi absorption modes
US9051178B2 (en) 2012-01-05 2015-06-09 Gwangju Institute Of Science And Technology Transmissive image modulator including stacked diode structure having multi absorption modes
KR20130091194A (en) 2012-02-07 2013-08-16 삼성전자주식회사 3d image acquisition apparatus and method of extractig depth information in the 3d image acquisition apparatus
US9123164B2 (en) 2012-02-07 2015-09-01 Samsung Electronics Co., Ltd. 3D image acquisition apparatus and method of extracting depth information in 3D image acquisition apparatus
US20140014838A1 (en) * 2012-07-16 2014-01-16 Karen Denise Hendrix Optical filter and sensor system
KR20140051030A (en) 2012-10-22 2014-04-30 삼성전자주식회사 Imaging optical system for 3d image acquisition apparatus, and 3d image acquisition apparatus including the imaging optical system
US9402067B2 (en) 2012-10-22 2016-07-26 Samsung Electronics Co., Ltd. Imaging optical system for 3D image acquisition apparatus, and 3D image acquisition apparatus including the imaging optical system

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Masahiro Kawakita, et al; "High-Definition real-time depth-mapping TV camera: HDTV Axi-Vision Camera"; Optics Express; vol. 12; No. 12; Jun. 14, 2004; pp. 2781-2794.
Tahito Aida, et al.; "High-speed depth-mapping Axi-Vision Camera with compact optical system"; Three Dimensional Image Capture and Applications; vol. 6805; 2008; pp. 680511-1-680511-10.

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170244953A1 (en) * 2012-10-10 2017-08-24 Samsung Electronics Co., Ltd. Imaging optical system and 3d image acquisition apparatus including the imaging optical system
US9998730B2 (en) * 2012-10-10 2018-06-12 Samsung Electronics Co., Ltd. Imaging optical system and 3D image acquisition apparatus including the imaging optical system
US20190317212A1 (en) * 2018-04-11 2019-10-17 Microsoft Technology Licensing, Llc Time of flight and picture camera
US10942274B2 (en) * 2018-04-11 2021-03-09 Microsoft Technology Licensing, Llc Time of flight and picture camera

Also Published As

Publication number Publication date
US20140098192A1 (en) 2014-04-10
US20170244953A1 (en) 2017-08-24
KR20140046335A (en) 2014-04-18
KR101858577B1 (en) 2018-05-16
US9998730B2 (en) 2018-06-12

Similar Documents

Publication Publication Date Title
US9998730B2 (en) Imaging optical system and 3D image acquisition apparatus including the imaging optical system
US10638118B2 (en) Time-of-flight camera system
US9451240B2 (en) 3-dimensional image acquisition apparatus and 3D image acquisition method for simultaneously obtaining color image and depth image
US9402067B2 (en) Imaging optical system for 3D image acquisition apparatus, and 3D image acquisition apparatus including the imaging optical system
US10412352B2 (en) Projector apparatus with distance image acquisition device and projection mapping method
US9998678B2 (en) Camera for acquiring optical properties and spatial structure properties
US10652513B2 (en) Display device, display system and three-dimension display method
US20200075652A1 (en) Pixel cell with multiple photodiodes
CN107783353A (en) For catching the apparatus and system of stereopsis
CN112823291A (en) Time-of-flight RGB-IR image sensor
TWI754809B (en) Image sensing device
US11570342B2 (en) Selective power efficient three-dimensional imaging
JP7005175B2 (en) Distance measuring device, distance measuring method and imaging device
EP3874243A2 (en) Hyper-spectral camera comprising tunable spectral illuminator
US11610339B2 (en) Imaging processing apparatus and method extracting a second RGB ToF feature points having a correlation between the first RGB and TOF feature points
JP6824833B2 (en) Distance data generation system, distance data generation method and program
JP2017133931A (en) Image creation device, and distance image and color image creation method
RU2543688C2 (en) Camera and optical system for obtaining 3d images (versions)
You et al. MOEMS-based time-of-flight camera for 3D video capturing

Legal Events

Date Code Title Description
AS Assignment

Owner name: SAMSUNG ELECTRONICS CO., LTD., KOREA, REPUBLIC OF

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PARK, YONG-HWA;YOU, JANG-WOO;REEL/FRAME:031222/0566

Effective date: 20130819

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4