US8767306B1 - Display system - Google Patents

Display system Download PDF

Info

Publication number
US8767306B1
US8767306B1 US13/240,994 US201113240994A US8767306B1 US 8767306 B1 US8767306 B1 US 8767306B1 US 201113240994 A US201113240994 A US 201113240994A US 8767306 B1 US8767306 B1 US 8767306B1
Authority
US
United States
Prior art keywords
infrared light
image
viewing location
optical system
visible light
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US13/240,994
Inventor
Xiaoyu Miao
Adrian Wong
Babak Amirparviz
Mark Spitzer
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Google LLC
Original Assignee
Google LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Google LLC filed Critical Google LLC
Priority to US13/240,994 priority Critical patent/US8767306B1/en
Assigned to GOOGLE INC. reassignment GOOGLE INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: AMIRPARVIZ, BABAK, MIAO, XIAOYU, WONG, ADRIAN, SPITZER, MARK
Application granted granted Critical
Publication of US8767306B1 publication Critical patent/US8767306B1/en
Assigned to GOOGLE LLC reassignment GOOGLE LLC CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: GOOGLE INC.
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • G02B27/0172Head mounted characterised by optical features
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/0093Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for monitoring data relating to the user, e.g. head-tracking, eye-tracking
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/09Beam shaping, e.g. changing the cross-sectional area, not otherwise provided for
    • G02B27/0938Using specific optical elements
    • G02B27/0977Reflective elements
    • G02B27/0983Reflective elements being curved
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/10Beam splitting or combining systems
    • G02B27/1006Beam splitting or combining systems for splitting or combining different wavelengths
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • G02B2027/0138Head-up displays characterised by optical features comprising image capture systems, e.g. camera
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • G02B2027/0178Eyeglass type

Definitions

  • Wearable systems can integrate various elements, such as miniaturized computers, input devices, sensors, detectors, image displays, wireless communication devices as well as image and audio processors, into a device that can be worn by a user.
  • Such devices provide a mobile and lightweight solution to communicating, computing and interacting with one's environment.
  • wearable compact optical displays that augment the wearer's experience of the real world.
  • an artificial image By placing an image display element close to the wearer's eye(s), an artificial image can be made to overlay the wearer's view of the real world.
  • image display elements are incorporated into systems also referred to as “near-eye displays”, “head-mounted displays” (HMDs) or “heads-up displays” (HUDs).
  • HMDs head-mounted displays
  • HUDs heads-up displays
  • the artificial image may fill or nearly fill the wearer's field of view.
  • an optical system in a first aspect, includes a display panel configured to generate a visible light pattern, a proximal beam splitter through which a field of view of a real-world environment and a virtual image are viewable from a viewing location, and an infrared light source configured to illuminate the viewing location with infrared light. The infrared light is reflected from the viewing location into the proximal beam splitter as collected infrared light.
  • the optical system further includes an image former optically coupled to the proximal beam splitter, wherein the image former is configured to reflect at least a portion of the visible light pattern from the display panel to form the virtual image and to transmit at least a portion of the collected infrared light.
  • the optical system further includes a camera optically coupled to the image former, wherein the camera is configured to image the viewing location using the collected infrared light transmitted through the image former.
  • a head-mounted display in a second aspect, includes a head-mounted support and an optical system attached to the head-mounted support, wherein the optical system includes a display panel configured to generate a visible light pattern, a proximal beam splitter through which a field of view of a real-world environment and a virtual image are viewable from a viewing location, and an infrared light source configured to illuminate the viewing location with infrared light such that infrared light is reflected from the viewing location into the proximal beam splitter as collected infrared light.
  • the optical system further includes an image former optically coupled to the proximal beam splitter, wherein the image former is configured to reflect at least a portion of the visible light pattern from the display panel to form the virtual image and to transmit at least a portion of the collected infrared light.
  • the optical system further includes a camera optically coupled to the image former, wherein the camera is configured to image the viewing location using the collected infrared light transmitted through the image former.
  • the head-mounted display further includes a computer, wherein the computer is configured to control the display panel and receive images of the viewing location obtained by the camera.
  • a method in a third aspect, includes generating a visible light pattern using a display panel and forming a virtual image from the visible light pattern using an image former, wherein the virtual image and a field of view of a real-world environment are viewable from a viewing location.
  • the method further includes illuminating the viewing location with infrared light, collecting infrared light reflected from the viewing location, transmitting at least a portion of the collected infrared light through the image former, and imaging the viewing location using the collected infrared light transmitted through the image former.
  • FIG. 1 is a functional block diagram of a wearable computing device, in accordance with an example embodiment.
  • FIG. 2 is a top view of an optical system, in accordance with an example embodiment.
  • FIG. 3 is a graph illustrating a variation of relative radiant power of an infrared source with respect to wavelength, in accordance with an example embodiment.
  • FIG. 4 is a graph illustrating a variation of percentage reflectance with respect to wavelength, in accordance with an example embodiment.
  • FIG. 5A is a front view of a head-mounted display, in accordance with an example embodiment.
  • FIG. 5B is a top view of the head-mounted display of FIG. 5A , in accordance with an example embodiment.
  • FIG. 5C is a side view of the head-mounted display of FIG. 5A and FIG. 5B , in accordance with an example embodiment.
  • FIG. 6 depicts side and front views of an eye as well as schematic drawings of pupil location information, in accordance with an example embodiment.
  • FIG. 7 depicts side and front views of an eye as well as schematic drawings of pupil size information, in accordance with an example embodiment.
  • FIG. 8 is a method, in accordance with an example embodiment.
  • a head-mounted display may enable its wearer to observe the wearer's real-world surroundings and also view a displayed image, such as a computer-generated image.
  • the displayed image may overlay a portion of the wearer's field of view of the real world.
  • the wearer of the HMD is going about his or her daily activities, such as walking, driving, exercising, etc., the wearer may be able to see a displayed image generated by the HMD at the same time that the wearer is looking out at his or her real-world surroundings.
  • the displayed image might include, for example, graphics, text, and/or video.
  • the content of the displayed image could relate to any number of contexts, including but not limited to the wearer's current environment, an activity in which the wearer is currently engaged, the biometric status of the wearer, and any audio, video, or textual communications that have been directed to the wearer.
  • the images displayed by the HMD may also be part of an interactive user interface.
  • the HMD could be part of a wearable computing device.
  • the images displayed by the HMD could include menus, selection boxes, navigation icons, or other user interface features that enable the wearer to invoke functions of the wearable computing device or otherwise interact with the wearable computing device.
  • the images displayed by the HMD could appear anywhere in the wearer's field of view.
  • the displayed image might occur at or near the center of the wearer's field of view, or the displayed image might be confined to the top, bottom, or a corner of the wearer's field of view.
  • the displayed image might be at the periphery of or entirely outside of the wearer's normal field of view.
  • the displayed image might be positioned such that it is not visible when the wearer looks straight ahead but is visible when the wearer looks in a specific direction, such as up, down, or to one side.
  • the displayed image might overlay only a small portion of the wearer's field of view, or the displayed image might fill most or all of the wearer's field of view.
  • the displayed image could be displayed continuously or only at certain times (e.g., only when the wearer is engaged in certain activities).
  • the HMD may utilize an optical system to present virtual images overlaid upon a real-world view to a wearer.
  • the optical system may include a light source, such as a light-emitting diode (LED), that is configured to illuminate a display panel, such as a liquid crystal-on-silicon (LCOS) display.
  • the display panel generates light patterns by spatially modulating the light from the light source, and an image former forms a virtual image from the light pattern.
  • the HMD may obtain data from the wearer in order to perform certain functions, for instance to provide context-sensitive information to the wearer.
  • the HMD may obtain information regarding the wearer and the wearer's environment and respond accordingly.
  • the HMD may use a pupil position recognition technique, wherein if the HMD recognizes that the wearer's pupil location is higher with respect to a neutral forward viewing axis, the HMD may display virtual images related to objects located above the wearer.
  • the HMD may recognize, by a similar pupil position recognition technique, that the wearer is looking downward. Accordingly the HMD may display virtual images related to objects located below the neutral forward viewing axis of the wearer.
  • the HMD may reduce the brightness or adjust other aspects of the displayed virtual images.
  • the infrared camera may image the pupil while the processor implements an image processing algorithm to find the edges or extents of the imaged pupil.
  • the image processing algorithms may include pattern recognition, Canny edge detection, thresholding, contrast detection, or differential edge detection. Those skilled in the art will understand that many other image processing techniques could be used individually or in combination with others in order to obtain pupil location and size information.
  • the processor may act to adjust various components of the displayed virtual image. For instance, if the user is looking upwards into a clear night sky, the wearable computing device may detect the upward gaze due to pupil location, and control the display to show virtual highlights around and virtual information about various stars and nebulae. Furthermore, due to a dark ambient environment, a user's pupils may be dilated. The wearable computing device may detect this and adjust the virtual image contrast and brightness accordingly.
  • FIG. 1 is a functional block diagram 100 that illustrates a wearable computing device 102 , head-mounted display (HMD) 104 and various components that comprise the system.
  • HMD 104 includes a see-through display.
  • the wearer of wearable computing device 102 may be able to look through HMD 104 and observe a portion of the real-world environment of the wearable computing device 102 , i.e., in a particular field of view provided by HMD 104 .
  • HMD 104 is operable to display images that are superimposed on the field of view, for example, to provide an “augmented reality” experience. Some of the images displayed by HMD 104 may be superimposed over particular objects in the field of view. However, HMD 104 may also display images that appear to hover within the field of view instead of being associated with particular objects in the field of view.
  • the HMD 104 may further include several components such as an infrared camera 106 , a user interface 108 , a processor 110 , optical system 112 , sensors 114 , a global positioning system (GPS) 116 , data storage 118 and a wireless communication interface 120 . These components may further work in an interconnected fashion. For instance, in an example embodiment, the infrared camera 106 may image one or both of the HMD wearer's eye pupils. The infrared camera 106 may deliver image information to the processor 110 , which may make a determination regarding the direction of HMD wearer's gaze. The wearable computing device 102 may further utilize sensors 114 and GPS 116 to gather contextual information based upon the environment and location of the HMD.
  • GPS global positioning system
  • context-specific information may be presented to the wearer in various formats such as virtual images as well as audio and vibration cues from the user interface 108 .
  • the individual components of the example embodiment will be described in more detail below.
  • HMD 104 could be configured as, for example, eyeglasses, goggles, a helmet, a hat, a visor, a headband, or in some other form that can be supported on or from the wearer's head. Further, HMD 104 may be configured to display images to both of the wearer's eyes, for example, using two see-through displays. Alternatively, HMD 104 may include only a single see-through display and may display images to only one of the wearer's eyes, either the left eye or the right eye.
  • the wearable computing device 102 may additionally include an infrared camera 106 that is configured to capture images of a point of view location associated with the HMD 104 .
  • the infrared camera 106 may be configured to image the pupil of a HMD wearer that may be located at the point of view location.
  • the images could be either video images or still images.
  • the images obtained by infrared camera 106 regarding the wearer eye pupil location may help determine where the wearer is looking within the HMD field of view.
  • the image analysis could be performed by processor 110 .
  • the imaging of the point of view location could occur continuously or at discrete times depending upon, for instance, user interactions with the user interface 108 .
  • Infrared camera 106 could be integrated into optical system 112 .
  • infrared camera 106 could additionally represent a visible light camera with sensing capabilities in the infrared wavelengths.
  • wearable computing device 102 may be controlled by a processor 110 that executes instructions stored in a non-transitory computer readable medium, such as data storage 118 .
  • processor 110 in combination with instructions stored in data storage 118 may function as a controller of wearable computing device 102 .
  • processor 110 may control HMD 104 in order to control what images are displayed by HMD 104 .
  • Processor 110 may also control wireless communication interface 120 and other components of the HMD system.
  • data storage 118 may store data that may include a set of calibrated wearer eye pupil positions and a collection of past eye pupil positions. Thus, data storage 118 may function as a database of information related to gaze direction. Such information may be used by wearable computing device 102 to anticipate where the user will look and determine what images are to be displayed to the wearer by HMD 104 .
  • Calibrated wearer eye pupil positions may include, for instance, information regarding the extents or range of the wearer's eye pupils movement (right/left and upwards/downwards) as well as wearer eye pupil positions that may relate to a neutral forward viewing axis.
  • the neutral forward viewing axis may represent the axis defined wherein the wearer is looking straight ahead and may further represent a reference axis and thus a basis for determining dynamic gaze direction. Furthermore, information may be stored in data storage 118 regarding possible control instructions that may be enacted using eye movements. For instance, two consecutive wearer eye blinks may represent a control instruction directing a second camera (not depicted) associated with the HMD 104 to capture an image.
  • Wearable computing device 102 may also include a user interface 108 for displaying information to the wearer or receiving input from the wearer.
  • User interface 108 could include, for example, the displayed virtual images, a touchpad, a keypad, buttons, a microphone, and/or other input devices.
  • Processor 110 may control the functioning of wearable computing device 102 based on input received through user interface 108 .
  • processor 110 may utilize user input from the user interface 108 to control how HMD 104 displays images or what images HMD 104 displays.
  • the wearable computing device 102 may include a wireless communication interface 120 for wirelessly communicating with the internet and/or target objects near the HMD 104 .
  • Wireless communication interface 120 could use any form of wireless communication that can support bi-directional data exchange over a packet network (such as the internet).
  • wireless communication interface 120 could use 3G cellular communication, such as CDMA, EVDO, GSM/GPRS, or 4G cellular communication, such as WiMAX or LTE.
  • wireless communication interface 120 could communicate with a wireless local area network (WLAN), for example, using WiFi.
  • WLAN wireless local area network
  • wireless communication interface 120 could be established using an infrared link, Bluetooth, or ZigBee. The wireless communications could be uni-directional or bi-directional with respect to the internet or a target object.
  • Wearable computing device 102 may further include an optical system 112 that is configured to display virtual images to a wearer.
  • Optical system 112 is described in detail below.
  • FIG. 1 shows various components of HMD 104 , i.e., wireless communication interface 120 , processor 110 , data storage 118 , infrared camera 106 , sensors 114 , GPS 116 , and user interface 108 , as being integrated into HMD 104 , one or more of these components could be mounted or associated separately from HMD 104 .
  • infrared camera 106 could be mounted on the user separate from HMD 104 .
  • wearable computing device 102 could be provided in the form of separate devices that can be worn on or carried by the wearer.
  • the separate components that make up wearable computing device 102 could be communicatively coupled together in either a wired or wireless fashion.
  • FIG. 2 illustrates a top view of an optical system 200 that is configured to display a virtual image superimposed upon a real-world scene viewable along a viewing axis 204 .
  • a distal portion 232 and a proximal portion 234 represent optically-coupled portions of the optical system 200 that may or may not be physically separated.
  • An example embodiment includes a display panel 206 that may be illuminated by a light source 208 . Light emitted from the light source 208 is incident upon the distal beam splitter 210 .
  • the light source 208 may include one or more light-emitting diodes (LEDs) and/or laser diodes.
  • the light source 208 may further include a linear polarizer that acts to pass one particular polarization to the rest of the optical system.
  • the distal beam splitter 210 is a polarizing beam splitter that reflects light depending upon the polarization of light incident upon the beam splitter.
  • s-polarized light from the light source 208 may be preferentially reflected by a distal beam-splitting interface 212 towards the display panel 206 .
  • the display panel 206 in the example embodiment is a liquid crystal-on-silicon (LCOS) display, but could also be a digital light projector (DLP) micro-mirror display, or other type of reflective display panel.
  • the display panel 206 acts to spatially-modulate the incident light to generate a light pattern.
  • the display panel 206 may be an emissive-type display such as an organic light-emitting diode (OLED) display or a transmissive liquid crystal display (LCD) with a backlight; in such cases, distal beam splitter 210 and light source 208 may be omitted.
  • OLED organic light-emitting diode
  • LCD transmissive liquid crystal display
  • the display panel 206 In the example in which the display panel 206 is a LCOS display panel, the display panel 206 generates a light pattern with a polarization perpendicular to the polarization of light initially incident upon the panel. In this example embodiment, the display panel 206 converts incident s-polarized light into a light pattern with p-polarization. The generated light pattern from the display panel 206 is directed towards the distal beam splitter 210 . The p-polarized light pattern passes through the distal beam splitter 210 and is directed along an optical axis 214 towards the proximal region of the optical system 200 . In an example embodiment, the proximal beam splitter 216 is also a polarizing beam splitter.
  • image former 218 includes a concave mirror 230 and a proximal quarter-wave plate 228 .
  • the light pattern passes through the proximal quarter-wave plate 228 and is reflected by the concave mirror 230 .
  • the reflected light pattern passes back through proximal quarter-wave plate 228 .
  • the light patterns are converted to the s-polarization and are formed into a viewable image.
  • This viewable image is incident upon the proximal beam splitter 216 and the viewable image is reflected from proximal beam splitting interface 220 towards a viewing location 222 along a viewing axis 204 .
  • a real-world scene is viewable through a viewing window 224 .
  • the viewing window 224 may include a linear polarizer in order to reduce stray light within the optical system. Light from the viewing window 224 is at least partially transmitted through the proximal beam splitter 216 .
  • both a virtual image and a real-world image are viewable to the viewing location 222 through the proximal beam splitter 216 .
  • FIG. 2 depicts the distal portion 232 of the optical system housing as to the left of the proximal portion 234 of the optical system housing when viewed from above, it is understood that other embodiments are possible to physically realize the optical system 200 , including the distal portion 232 being configured to be to the right, below and above with respect to the proximal portion 234 .
  • an image former 218 as comprising a concave mirror 230 , it is understood by those skilled in the art that the image former 218 may comprise a different optical element, such as an optical lens or a diffractive optic element.
  • the proximal beam splitter 216 , the distal beam splitter 210 , and other components of optical system 200 are made of glass. Alternatively, some or all of such optical components may be partially or entirely plastic, which can also function to reduce the weight of optical system 200 .
  • a suitable plastic material is Zeonex® E48R cyclo olefin optical grade polymer which is available from Zeon Chemicals L.P., Louisville, Ky. Another suitable plastic material is polymethyl methacrylate (PMMA).
  • An example embodiment may include an infrared light source 226 that is configured to illuminate the viewing location 222 .
  • FIG. 2 depicts the infrared light source 226 as adjacent to viewing window 224 , those skilled in the art will understand that the infrared light source 226 could be located elsewhere, such as on the side of the proximal beam splitter 216 that is adjacent to the viewing location 222 or in the distal portion 232 of the optical system 200 .
  • the infrared light source 226 may represent, for example, one or more infrared light-emitting diodes (LEDs). Infrared LEDs with a small size may be implemented, such as the Vishay Technology TSML 1000 product.
  • FIG. 3 is a graph illustrating a variation of relative radiant power of an infrared source with respect to wavelength, in accordance with an example embodiment.
  • the infrared light source 226 may include one or more LEDs located at different locations in the optical system 200 .
  • Infrared light generated from the infrared light source 226 is configured to be incident upon the viewing location 222 .
  • the wearer's eye pupil may be illuminated with the infrared light.
  • the infrared light may be reflected from the wearer's eye back along the viewing axis 204 towards the proximal beam splitter 216 .
  • a portion of the reflected infrared light may be reflected from the beam splitting interface 220 towards the image former 218 .
  • the image former 218 may include a dichroic thin film configured to selectively reflect or transmit incident light depending upon the wavelength of the incident light.
  • the dichroic thin film may be configured to pass infrared light while reflecting visible light.
  • the visible light pattern generated by the display panel 206 may be reflected by the concave mirror 230 and the visible light pattern may be formed into a viewable image. The infrared light may thus be preferably transmitted through the concave mirror 230 to infrared camera 202 .
  • Dichroic thin film coatings are available commercially from companies such as JML Optical Industries and Precision Glass & Optics (PG&O) and comprise multiple layers of dielectric and/or metal films. These dichroic coatings are also called ‘cold mirrors’.
  • FIG. 4 is a graph illustrating a variation of percentage reflectance with respect to wavelength, in accordance with an example embodiment. The graph represents example spectral reflectance characteristics for a dichroic thin film that may coat the concave mirror 230 .
  • a small aperture or apertures may be introduced into the image former 218 , which may be realized by one or more pinholes (e.g., a central pinhole) in the concave mirror 230 .
  • pinholes e.g., a central pinhole
  • most of the visible and infrared light is reflected off of and formed by the image former 218 into an image viewable by the HMD wearer.
  • Some of the visible and infrared light passes through the aperture and is incident upon the infrared camera 202 .
  • the infrared camera 202 may selectively filter and detect the infrared light from the combination of visible and infrared light to obtain information regarding the wearer's eye pupil location.
  • the infrared light source 226 may be modulated to provide a frequency reference for a lock-in amplifier or phase-locked loop in order that the infrared light signal is obtained efficiently.
  • the visible light source 208 may be modulated and infrared light detection could be performed when the visible light source 208 is off, for example.
  • FIG. 5A presents a front view of a head-mounted display (HMD) 300 in an example embodiment that includes a head-mounted support 309 .
  • FIGS. 5B and 5C present the top and side views, respectively, of the HMD in FIG. 5A .
  • the head-mounted support 309 includes lens frames 314 and 316 , a center frame support 318 , lens elements 310 and 312 , and extending side-arms 320 and 322 .
  • the center frame support 318 and side-arms 320 and 322 are configured to secure the head-mounted support 309 to the wearer's head via the wearer's nose and ears, respectively.
  • Each of the frame elements 314 , 316 , and 318 and the extending side-arms 320 and 322 may be formed of a solid structure of plastic or metal, or may be formed of a hollow structure of similar material so as to allow wiring and component interconnects to be internally routed through the head-mounted support 309 .
  • head-mounted support 309 may support external wiring.
  • Lens elements 310 and 312 are at least partially transparent so as to allow the wearer to look through them.
  • the wearer's left eye 308 may look through left lens 312 and the wearer's right eye 306 may look through right lens 310 .
  • Optical systems 302 and 304 which may be configured as shown in FIG. 2 , may be positioned in front of lenses 310 and 312 , respectively, as shown in FIGS. 5A , 5 B, and 5 C.
  • Optical systems 302 and 304 may be attached to the head-mounted support 309 using support mounts 324 and 326 , respectively.
  • optical systems 302 and 304 may be integrated partially or completely into lens elements 310 and 312 , respectively.
  • a HMD might include an optical system for only one of the wearer's eyes (either left eye 308 or right eye 306 ).
  • the HMD wearer may simultaneously observe from optical systems 302 and 304 a real-world image with an overlaid virtual image.
  • the HMD may include various elements such as a HMD computer 340 , a touchpad 342 , a microphone 344 , and a button 346 .
  • the computer 340 may use data from, among other sources, various sensors and cameras to determine the virtual image that should be displayed to the user.
  • an infrared light source or sources may illuminate the viewing position(s) 308 and 306 , i.e. the wearer's eye(s), and the reflected infrared light may be preferentially collected with an infrared camera.
  • FIG. 6 depicts side and front views of an eye as well as schematic drawings of pupil location information.
  • One way to determine gaze direction of a person is to determine the position of the person's pupil with respect to a neutral forward viewing axis.
  • To track eye pupil movements infrared light is reflected off of a person's eye. The reflected light may be collected and detected with an infrared detector. Image processing can then be conducted with a processor 110 in order to determine the extents and centroid location of the person's pupil.
  • a person may be looking directly forward.
  • the eyelid 403 is open and the pupil 404 / 410 is located centrally with respect to a reference axis 412 .
  • the position of the pupil may be determined to be at pupil location 414 .
  • the determined pupil location 414 coincides with a neutral forward viewing axis.
  • Virtual image display position and context may be adjusted due to the determined pupil location 414 .
  • a person may be looking upwards with respect to a neutral forward viewing axis.
  • imaging the person's eye with infrared light may result in a determined pupil position 428 that is above the neutral forward viewing axis.
  • Virtual images may be displayed above a person's normal field of view and contextual information regarding target objects above a person's normal field of view may be displayed.
  • a person may be looking downwards with respect to a neutral forward viewing axis.
  • the determined pupil position 442 may be determined by imaging the person's eye and may be further found to be below a neutral forward viewing axis.
  • contextual information about target objects including virtual images may be displayed for target objects below the neutral forward viewing axis of the person.
  • FIG. 7 depicts side and front views of an eye as well as schematic drawings of pupil size information.
  • One way to determine the ambient light level of a scene is to determine the diameter of a person's eye pupil who may be looking at the scene.
  • infrared light may be reflected off of a person's eye.
  • the reflected light may be collected and detected with an infrared detector.
  • Image processing can then be conducted with a processor 110 in order to determine the extents and thus the diameter of the person's pupil.
  • a person may be looking directly forward and may exhibit a relatively small diameter pupil.
  • the eyelid 504 is open and the pupil 506 / 512 is located centrally with respect to a reference axis 514 .
  • image processing which may include edge detection
  • the position of the pupil may be determined to be at pupil location 516 with a given pupil diameter. Due to a relatively small diameter pupil, the brightness or contrast of the virtual image may be adjusted assuming a bright ambient light level.
  • a person may exhibit a relatively large diameter pupil. After imaging and image processing, the size and thus the diameter of the pupil may be determined. Due to a relatively large diameter pupil, the brightness or contrast of the virtual image may be adjusted assuming a dark ambient light level.
  • an infrared eye-tracking system may include the recognition of various blink and eye-movement-based commands or control instructions.
  • image recognition algorithms could recognize a succession of blinks as a command.
  • two successive blinks with one eye within half a second may represent a command to take a picture using a second camera on the HMD.
  • an eye-tracking system may allow enhanced functionality when interacting with a user interface of the HMD or of a target object. For instance, if a HMD wearer is looking at an electronic word processing document and the wearable computing device determines that the person is looking at words near the bottom of the user interface, the wearable computing device may automatically scroll the text upwards within the user interface so the person does not need to physically scroll down the page with a mouse wheel.
  • FIG. 8 illustrates an example method 600 for an optical system to collect and image infrared light from a viewing location. It is to be understood that the steps may appear in different order and steps may be added or subtracted.
  • a visible light pattern is generated using a display panel.
  • the display panel could be a component in an optical system similar to optical systems 302 and 304 .
  • a virtual image is formed from the visible light pattern using an image former.
  • the image former could include a quarter wave plate 228 and concave mirror 230 that may act together to form the virtual image.
  • the method includes a third step 606 wherein the viewing location is illuminated with infrared light.
  • the viewing location may coincide with where a HMD wearer's eye is located while wearing the HMD.
  • the infrared light may be emitted from one or more sources, such as one or more infrared LEDs.
  • infrared light may be incident upon the viewing location from multiple locations. That is, infrared light may be incident towards the viewing location along a viewing axis 204 as well as along other axes.
  • a fourth step 608 includes collecting infrared light reflected from the viewing location. As discussed above, infrared light reflected from the wearer's eye may be passed back into the optical system through the proximal beam splitter 216 . A portion of the infrared light light may be reflected off of the proximal beam splitting interface 220 and transmitted towards the image former 218 .
  • a fifth step 610 includes transmitting at least a portion of the collected infrared light through the image former.
  • the image former 218 may comprise a concave mirror 230 with a dichroic thin film coating to selectively transmit infrared light and selectively reflect visible light.
  • the image former 218 may alternatively or additionally include an aperture through which visible and infrared light may be transmitted. In both of these situations, infrared light is transmitted through the image former 218 .
  • a sixth step 612 includes imaging the viewing location using the collected infrared light transmitted through the image former.
  • light may be detected using an infrared camera 202 sensitive to infrared light.
  • the infrared camera may convey video or still images to the processor 110 . These images may be used to form the basis of a dynamically updated database of the wearer's eye pupil and its position.

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Optics & Photonics (AREA)

Abstract

An optical system has an aperture through which virtual and real-world images are viewable along a viewing axis. The optical system may be incorporated into a head-mounted display (HMD). By illuminating a viewing location with an infrared light source, an eye pupil may be illuminated. Infrared light is reflected from the viewing location and is collected with a proximal beam splitter. An image former is configured to reflect at least a portion of the visible light pattern generated by the display panel to form the virtual image and transmit at least a portion of the collected infrared light. The transmitted infrared light may be imaged by a camera. The HMD may use images from the camera to provide, for example, context-sensitive virtual images to a wearer.

Description

BACKGROUND
Wearable systems can integrate various elements, such as miniaturized computers, input devices, sensors, detectors, image displays, wireless communication devices as well as image and audio processors, into a device that can be worn by a user. Such devices provide a mobile and lightweight solution to communicating, computing and interacting with one's environment. With the advance of technologies associated with wearable systems and miniaturized optical elements, it has become possible to consider wearable compact optical displays that augment the wearer's experience of the real world.
By placing an image display element close to the wearer's eye(s), an artificial image can be made to overlay the wearer's view of the real world. Such image display elements are incorporated into systems also referred to as “near-eye displays”, “head-mounted displays” (HMDs) or “heads-up displays” (HUDs). Depending upon the size of the display element and the distance to the wearer's eye, the artificial image may fill or nearly fill the wearer's field of view.
SUMMARY
In a first aspect, an optical system is provided. The optical system includes a display panel configured to generate a visible light pattern, a proximal beam splitter through which a field of view of a real-world environment and a virtual image are viewable from a viewing location, and an infrared light source configured to illuminate the viewing location with infrared light. The infrared light is reflected from the viewing location into the proximal beam splitter as collected infrared light. The optical system further includes an image former optically coupled to the proximal beam splitter, wherein the image former is configured to reflect at least a portion of the visible light pattern from the display panel to form the virtual image and to transmit at least a portion of the collected infrared light. The optical system further includes a camera optically coupled to the image former, wherein the camera is configured to image the viewing location using the collected infrared light transmitted through the image former.
In a second aspect, a head-mounted display is provided. The head-mounted display includes a head-mounted support and an optical system attached to the head-mounted support, wherein the optical system includes a display panel configured to generate a visible light pattern, a proximal beam splitter through which a field of view of a real-world environment and a virtual image are viewable from a viewing location, and an infrared light source configured to illuminate the viewing location with infrared light such that infrared light is reflected from the viewing location into the proximal beam splitter as collected infrared light. The optical system further includes an image former optically coupled to the proximal beam splitter, wherein the image former is configured to reflect at least a portion of the visible light pattern from the display panel to form the virtual image and to transmit at least a portion of the collected infrared light. The optical system further includes a camera optically coupled to the image former, wherein the camera is configured to image the viewing location using the collected infrared light transmitted through the image former. The head-mounted display further includes a computer, wherein the computer is configured to control the display panel and receive images of the viewing location obtained by the camera.
In a third aspect, a method is provided. The method includes generating a visible light pattern using a display panel and forming a virtual image from the visible light pattern using an image former, wherein the virtual image and a field of view of a real-world environment are viewable from a viewing location. The method further includes illuminating the viewing location with infrared light, collecting infrared light reflected from the viewing location, transmitting at least a portion of the collected infrared light through the image former, and imaging the viewing location using the collected infrared light transmitted through the image former.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a functional block diagram of a wearable computing device, in accordance with an example embodiment.
FIG. 2 is a top view of an optical system, in accordance with an example embodiment.
FIG. 3 is a graph illustrating a variation of relative radiant power of an infrared source with respect to wavelength, in accordance with an example embodiment.
FIG. 4 is a graph illustrating a variation of percentage reflectance with respect to wavelength, in accordance with an example embodiment.
FIG. 5A is a front view of a head-mounted display, in accordance with an example embodiment.
FIG. 5B is a top view of the head-mounted display of FIG. 5A, in accordance with an example embodiment.
FIG. 5C is a side view of the head-mounted display of FIG. 5A and FIG. 5B, in accordance with an example embodiment.
FIG. 6 depicts side and front views of an eye as well as schematic drawings of pupil location information, in accordance with an example embodiment.
FIG. 7 depicts side and front views of an eye as well as schematic drawings of pupil size information, in accordance with an example embodiment.
FIG. 8 is a method, in accordance with an example embodiment.
DETAILED DESCRIPTION
In the following detailed description, reference is made to the accompanying figures, which form a part thereof. In the figures, similar symbols typically identify similar components, unless context dictates otherwise. The illustrative embodiments described in the detailed description and figures are not meant to be limiting. Other embodiments may be utilized, and other changes may be made, without departing from the spirit or scope of the subject matter presented herein. It will be readily understood that the aspects of the present disclosure, as generally described herein, and illustrated in the figures, can be arranged, substituted, combined, separated, and designed in a wide variety of different configurations, all of which are contemplated herein.
1. Overview
A head-mounted display (HMD) may enable its wearer to observe the wearer's real-world surroundings and also view a displayed image, such as a computer-generated image. In some cases, the displayed image may overlay a portion of the wearer's field of view of the real world. Thus, while the wearer of the HMD is going about his or her daily activities, such as walking, driving, exercising, etc., the wearer may be able to see a displayed image generated by the HMD at the same time that the wearer is looking out at his or her real-world surroundings.
The displayed image might include, for example, graphics, text, and/or video. The content of the displayed image could relate to any number of contexts, including but not limited to the wearer's current environment, an activity in which the wearer is currently engaged, the biometric status of the wearer, and any audio, video, or textual communications that have been directed to the wearer. The images displayed by the HMD may also be part of an interactive user interface. For example, the HMD could be part of a wearable computing device. Thus, the images displayed by the HMD could include menus, selection boxes, navigation icons, or other user interface features that enable the wearer to invoke functions of the wearable computing device or otherwise interact with the wearable computing device.
The images displayed by the HMD could appear anywhere in the wearer's field of view. For example, the displayed image might occur at or near the center of the wearer's field of view, or the displayed image might be confined to the top, bottom, or a corner of the wearer's field of view. Alternatively, the displayed image might be at the periphery of or entirely outside of the wearer's normal field of view. For example, the displayed image might be positioned such that it is not visible when the wearer looks straight ahead but is visible when the wearer looks in a specific direction, such as up, down, or to one side. In addition, the displayed image might overlay only a small portion of the wearer's field of view, or the displayed image might fill most or all of the wearer's field of view. The displayed image could be displayed continuously or only at certain times (e.g., only when the wearer is engaged in certain activities).
The HMD may utilize an optical system to present virtual images overlaid upon a real-world view to a wearer. To display a virtual image to the wearer, the optical system may include a light source, such as a light-emitting diode (LED), that is configured to illuminate a display panel, such as a liquid crystal-on-silicon (LCOS) display. The display panel generates light patterns by spatially modulating the light from the light source, and an image former forms a virtual image from the light pattern.
The HMD may obtain data from the wearer in order to perform certain functions, for instance to provide context-sensitive information to the wearer. In an example embodiment, by using an infrared camera to record a wearer's pupil position and size, the HMD may obtain information regarding the wearer and the wearer's environment and respond accordingly. The HMD may use a pupil position recognition technique, wherein if the HMD recognizes that the wearer's pupil location is higher with respect to a neutral forward viewing axis, the HMD may display virtual images related to objects located above the wearer. Conversely, the HMD may recognize, by a similar pupil position recognition technique, that the wearer is looking downward. Accordingly the HMD may display virtual images related to objects located below the neutral forward viewing axis of the wearer. Further, if the HMD recognizes that the wearer's pupils are dilated, the HMD may reduce the brightness or adjust other aspects of the displayed virtual images.
In order to determine the actual position of a wearer's pupil, the infrared camera may image the pupil while the processor implements an image processing algorithm to find the edges or extents of the imaged pupil. The image processing algorithms may include pattern recognition, Canny edge detection, thresholding, contrast detection, or differential edge detection. Those skilled in the art will understand that many other image processing techniques could be used individually or in combination with others in order to obtain pupil location and size information. After image processing, the processor may act to adjust various components of the displayed virtual image. For instance, if the user is looking upwards into a clear night sky, the wearable computing device may detect the upward gaze due to pupil location, and control the display to show virtual highlights around and virtual information about various stars and nebulae. Furthermore, due to a dark ambient environment, a user's pupils may be dilated. The wearable computing device may detect this and adjust the virtual image contrast and brightness accordingly.
Certain illustrative examples of using an optical system and infrared light to view a viewing position are described below. It is to be understood, however, that other embodiments are possible and are implicitly considered within the context of the following example embodiments.
2. Optical System with Infrared Source, Camera and Image Former
FIG. 1 is a functional block diagram 100 that illustrates a wearable computing device 102, head-mounted display (HMD) 104 and various components that comprise the system. In an example embodiment, HMD 104 includes a see-through display. Thus, the wearer of wearable computing device 102 may be able to look through HMD 104 and observe a portion of the real-world environment of the wearable computing device 102, i.e., in a particular field of view provided by HMD 104. In addition, HMD 104 is operable to display images that are superimposed on the field of view, for example, to provide an “augmented reality” experience. Some of the images displayed by HMD 104 may be superimposed over particular objects in the field of view. However, HMD 104 may also display images that appear to hover within the field of view instead of being associated with particular objects in the field of view.
The HMD 104 may further include several components such as an infrared camera 106, a user interface 108, a processor 110, optical system 112, sensors 114, a global positioning system (GPS) 116, data storage 118 and a wireless communication interface 120. These components may further work in an interconnected fashion. For instance, in an example embodiment, the infrared camera 106 may image one or both of the HMD wearer's eye pupils. The infrared camera 106 may deliver image information to the processor 110, which may make a determination regarding the direction of HMD wearer's gaze. The wearable computing device 102 may further utilize sensors 114 and GPS 116 to gather contextual information based upon the environment and location of the HMD. By detecting the gaze direction of the wearer's eye(s), context-specific information may be presented to the wearer in various formats such as virtual images as well as audio and vibration cues from the user interface 108. The individual components of the example embodiment will be described in more detail below.
HMD 104 could be configured as, for example, eyeglasses, goggles, a helmet, a hat, a visor, a headband, or in some other form that can be supported on or from the wearer's head. Further, HMD 104 may be configured to display images to both of the wearer's eyes, for example, using two see-through displays. Alternatively, HMD 104 may include only a single see-through display and may display images to only one of the wearer's eyes, either the left eye or the right eye.
The wearable computing device 102 may additionally include an infrared camera 106 that is configured to capture images of a point of view location associated with the HMD 104. The infrared camera 106 may be configured to image the pupil of a HMD wearer that may be located at the point of view location. The images could be either video images or still images. The images obtained by infrared camera 106 regarding the wearer eye pupil location may help determine where the wearer is looking within the HMD field of view. The image analysis could be performed by processor 110. The imaging of the point of view location could occur continuously or at discrete times depending upon, for instance, user interactions with the user interface 108. Infrared camera 106 could be integrated into optical system 112. Furthermore, infrared camera 106 could additionally represent a visible light camera with sensing capabilities in the infrared wavelengths.
The function of wearable computing device 102 may be controlled by a processor 110 that executes instructions stored in a non-transitory computer readable medium, such as data storage 118. Thus, processor 110 in combination with instructions stored in data storage 118 may function as a controller of wearable computing device 102. As such, processor 110 may control HMD 104 in order to control what images are displayed by HMD 104. Processor 110 may also control wireless communication interface 120 and other components of the HMD system.
In addition to instructions that may be executed by processor 110, data storage 118 may store data that may include a set of calibrated wearer eye pupil positions and a collection of past eye pupil positions. Thus, data storage 118 may function as a database of information related to gaze direction. Such information may be used by wearable computing device 102 to anticipate where the user will look and determine what images are to be displayed to the wearer by HMD 104. Calibrated wearer eye pupil positions may include, for instance, information regarding the extents or range of the wearer's eye pupils movement (right/left and upwards/downwards) as well as wearer eye pupil positions that may relate to a neutral forward viewing axis. The neutral forward viewing axis may represent the axis defined wherein the wearer is looking straight ahead and may further represent a reference axis and thus a basis for determining dynamic gaze direction. Furthermore, information may be stored in data storage 118 regarding possible control instructions that may be enacted using eye movements. For instance, two consecutive wearer eye blinks may represent a control instruction directing a second camera (not depicted) associated with the HMD 104 to capture an image.
Wearable computing device 102 may also include a user interface 108 for displaying information to the wearer or receiving input from the wearer. User interface 108 could include, for example, the displayed virtual images, a touchpad, a keypad, buttons, a microphone, and/or other input devices. Processor 110 may control the functioning of wearable computing device 102 based on input received through user interface 108. For example, processor 110 may utilize user input from the user interface 108 to control how HMD 104 displays images or what images HMD 104 displays.
In one example, the wearable computing device 102 may include a wireless communication interface 120 for wirelessly communicating with the internet and/or target objects near the HMD 104. Wireless communication interface 120 could use any form of wireless communication that can support bi-directional data exchange over a packet network (such as the internet). For example, wireless communication interface 120 could use 3G cellular communication, such as CDMA, EVDO, GSM/GPRS, or 4G cellular communication, such as WiMAX or LTE. Alternatively, wireless communication interface 120 could communicate with a wireless local area network (WLAN), for example, using WiFi. Alternatively, wireless communication interface 120 could be established using an infrared link, Bluetooth, or ZigBee. The wireless communications could be uni-directional or bi-directional with respect to the internet or a target object.
Wearable computing device 102 may further include an optical system 112 that is configured to display virtual images to a wearer. Optical system 112 is described in detail below.
Although FIG. 1 shows various components of HMD 104, i.e., wireless communication interface 120, processor 110, data storage 118, infrared camera 106, sensors 114, GPS 116, and user interface 108, as being integrated into HMD 104, one or more of these components could be mounted or associated separately from HMD 104. For example, infrared camera 106 could be mounted on the user separate from HMD 104. Thus, wearable computing device 102 could be provided in the form of separate devices that can be worn on or carried by the wearer. The separate components that make up wearable computing device 102 could be communicatively coupled together in either a wired or wireless fashion.
FIG. 2 illustrates a top view of an optical system 200 that is configured to display a virtual image superimposed upon a real-world scene viewable along a viewing axis 204. For clarity, a distal portion 232 and a proximal portion 234 represent optically-coupled portions of the optical system 200 that may or may not be physically separated. An example embodiment includes a display panel 206 that may be illuminated by a light source 208. Light emitted from the light source 208 is incident upon the distal beam splitter 210. The light source 208 may include one or more light-emitting diodes (LEDs) and/or laser diodes. The light source 208 may further include a linear polarizer that acts to pass one particular polarization to the rest of the optical system.
In an example embodiment, the distal beam splitter 210 is a polarizing beam splitter that reflects light depending upon the polarization of light incident upon the beam splitter. To illustrate, s-polarized light from the light source 208 may be preferentially reflected by a distal beam-splitting interface 212 towards the display panel 206. The display panel 206 in the example embodiment is a liquid crystal-on-silicon (LCOS) display, but could also be a digital light projector (DLP) micro-mirror display, or other type of reflective display panel. The display panel 206 acts to spatially-modulate the incident light to generate a light pattern. Alternatively, the display panel 206 may be an emissive-type display such as an organic light-emitting diode (OLED) display or a transmissive liquid crystal display (LCD) with a backlight; in such cases, distal beam splitter 210 and light source 208 may be omitted.
In the example in which the display panel 206 is a LCOS display panel, the display panel 206 generates a light pattern with a polarization perpendicular to the polarization of light initially incident upon the panel. In this example embodiment, the display panel 206 converts incident s-polarized light into a light pattern with p-polarization. The generated light pattern from the display panel 206 is directed towards the distal beam splitter 210. The p-polarized light pattern passes through the distal beam splitter 210 and is directed along an optical axis 214 towards the proximal region of the optical system 200. In an example embodiment, the proximal beam splitter 216 is also a polarizing beam splitter. The light pattern is at least partially transmitted through the proximal beam splitter 216 to the image former 218. In an example embodiment, image former 218 includes a concave mirror 230 and a proximal quarter-wave plate 228. The light pattern passes through the proximal quarter-wave plate 228 and is reflected by the concave mirror 230.
The reflected light pattern passes back through proximal quarter-wave plate 228. Through the interactions with the proximal quarter-wave plate 228 and the concave mirror 230, the light patterns are converted to the s-polarization and are formed into a viewable image. This viewable image is incident upon the proximal beam splitter 216 and the viewable image is reflected from proximal beam splitting interface 220 towards a viewing location 222 along a viewing axis 204. A real-world scene is viewable through a viewing window 224. The viewing window 224 may include a linear polarizer in order to reduce stray light within the optical system. Light from the viewing window 224 is at least partially transmitted through the proximal beam splitter 216. Thus, both a virtual image and a real-world image are viewable to the viewing location 222 through the proximal beam splitter 216.
Although FIG. 2 depicts the distal portion 232 of the optical system housing as to the left of the proximal portion 234 of the optical system housing when viewed from above, it is understood that other embodiments are possible to physically realize the optical system 200, including the distal portion 232 being configured to be to the right, below and above with respect to the proximal portion 234. Further, although an example embodiment describes an image former 218 as comprising a concave mirror 230, it is understood by those skilled in the art that the image former 218 may comprise a different optical element, such as an optical lens or a diffractive optic element.
In one embodiment, the proximal beam splitter 216, the distal beam splitter 210, and other components of optical system 200 are made of glass. Alternatively, some or all of such optical components may be partially or entirely plastic, which can also function to reduce the weight of optical system 200. A suitable plastic material is Zeonex® E48R cyclo olefin optical grade polymer which is available from Zeon Chemicals L.P., Louisville, Ky. Another suitable plastic material is polymethyl methacrylate (PMMA).
An example embodiment may include an infrared light source 226 that is configured to illuminate the viewing location 222. Although FIG. 2 depicts the infrared light source 226 as adjacent to viewing window 224, those skilled in the art will understand that the infrared light source 226 could be located elsewhere, such as on the side of the proximal beam splitter 216 that is adjacent to the viewing location 222 or in the distal portion 232 of the optical system 200. The infrared light source 226 may represent, for example, one or more infrared light-emitting diodes (LEDs). Infrared LEDs with a small size may be implemented, such as the Vishay Technology TSML 1000 product. FIG. 3 is a graph illustrating a variation of relative radiant power of an infrared source with respect to wavelength, in accordance with an example embodiment.
Further, those skilled in the art will understand that, for best eye-tracking accuracy, it may be advantageous to obtain infrared images of the eye pupil using light sources that illuminate the eye from positions off-axis and/or on-axis with respect to the viewing axis 204. Therefore, the infrared light source 226 may include one or more LEDs located at different locations in the optical system 200.
Infrared light generated from the infrared light source 226 is configured to be incident upon the viewing location 222. Thus, the wearer's eye pupil may be illuminated with the infrared light. The infrared light may be reflected from the wearer's eye back along the viewing axis 204 towards the proximal beam splitter 216. A portion of the reflected infrared light may be reflected from the beam splitting interface 220 towards the image former 218.
In order to transmit infrared light to an infrared camera 202, the image former 218 may include a dichroic thin film configured to selectively reflect or transmit incident light depending upon the wavelength of the incident light. For instance, the dichroic thin film may be configured to pass infrared light while reflecting visible light. In an example embodiment, the visible light pattern generated by the display panel 206 may be reflected by the concave mirror 230 and the visible light pattern may be formed into a viewable image. The infrared light may thus be preferably transmitted through the concave mirror 230 to infrared camera 202. Dichroic thin film coatings are available commercially from companies such as JML Optical Industries and Precision Glass & Optics (PG&O) and comprise multiple layers of dielectric and/or metal films. These dichroic coatings are also called ‘cold mirrors’. FIG. 4 is a graph illustrating a variation of percentage reflectance with respect to wavelength, in accordance with an example embodiment. The graph represents example spectral reflectance characteristics for a dichroic thin film that may coat the concave mirror 230.
In an example embodiment, a small aperture or apertures may be introduced into the image former 218, which may be realized by one or more pinholes (e.g., a central pinhole) in the concave mirror 230. In this example embodiment, most of the visible and infrared light is reflected off of and formed by the image former 218 into an image viewable by the HMD wearer. Some of the visible and infrared light passes through the aperture and is incident upon the infrared camera 202. The infrared camera 202 may selectively filter and detect the infrared light from the combination of visible and infrared light to obtain information regarding the wearer's eye pupil location. Alternatively, the infrared light source 226 may be modulated to provide a frequency reference for a lock-in amplifier or phase-locked loop in order that the infrared light signal is obtained efficiently. Also, the visible light source 208 may be modulated and infrared light detection could be performed when the visible light source 208 is off, for example. Those with skill in the art will understand that there are other variations of transducing an infrared light signal mixed with a visible light signal with an infrared camera and that those variations are included implicitly in this specification.
3. Head-mounted Display with Infrared Eye-Tracking Optical System
FIG. 5A presents a front view of a head-mounted display (HMD) 300 in an example embodiment that includes a head-mounted support 309. FIGS. 5B and 5C present the top and side views, respectively, of the HMD in FIG. 5A. Although this example embodiment is provided in an eyeglasses format, it will be understood that wearable systems and HMDs may take other forms, such as hats, goggles, masks, headbands and helmets. The head-mounted support 309 includes lens frames 314 and 316, a center frame support 318, lens elements 310 and 312, and extending side- arms 320 and 322. The center frame support 318 and side- arms 320 and 322 are configured to secure the head-mounted support 309 to the wearer's head via the wearer's nose and ears, respectively. Each of the frame elements 314, 316, and 318 and the extending side- arms 320 and 322 may be formed of a solid structure of plastic or metal, or may be formed of a hollow structure of similar material so as to allow wiring and component interconnects to be internally routed through the head-mounted support 309. Alternatively or additionally, head-mounted support 309 may support external wiring. Lens elements 310 and 312 are at least partially transparent so as to allow the wearer to look through them. In particular, the wearer's left eye 308 may look through left lens 312 and the wearer's right eye 306 may look through right lens 310. Optical systems 302 and 304, which may be configured as shown in FIG. 2, may be positioned in front of lenses 310 and 312, respectively, as shown in FIGS. 5A, 5B, and 5C. Optical systems 302 and 304 may be attached to the head-mounted support 309 using support mounts 324 and 326, respectively. Furthermore, optical systems 302 and 304 may be integrated partially or completely into lens elements 310 and 312, respectively.
Although this example includes an optical system for each of the wearer's eyes, it is to be understood that a HMD might include an optical system for only one of the wearer's eyes (either left eye 308 or right eye 306). As described in FIG. 2, the HMD wearer may simultaneously observe from optical systems 302 and 304 a real-world image with an overlaid virtual image. The HMD may include various elements such as a HMD computer 340, a touchpad 342, a microphone 344, and a button 346. The computer 340 may use data from, among other sources, various sensors and cameras to determine the virtual image that should be displayed to the user. In an example embodiment, as described earlier, an infrared light source or sources may illuminate the viewing position(s) 308 and 306, i.e. the wearer's eye(s), and the reflected infrared light may be preferentially collected with an infrared camera.
Those skilled in the art would understand that other user input devices, user output devices, wireless communication devices, sensors, and cameras may be reasonably included in such a wearable computing system.
FIG. 6 depicts side and front views of an eye as well as schematic drawings of pupil location information. One way to determine gaze direction of a person is to determine the position of the person's pupil with respect to a neutral forward viewing axis. To track eye pupil movements, infrared light is reflected off of a person's eye. The reflected light may be collected and detected with an infrared detector. Image processing can then be conducted with a processor 110 in order to determine the extents and centroid location of the person's pupil. For instance, in an example embodiment 400, a person may be looking directly forward. The eyelid 403 is open and the pupil 404/410 is located centrally with respect to a reference axis 412. After image processing, which may include edge detection, the position of the pupil may be determined to be at pupil location 414. In this embodiment, the determined pupil location 414 coincides with a neutral forward viewing axis. Virtual image display position and context may be adjusted due to the determined pupil location 414.
In an example embodiment 401, a person may be looking upwards with respect to a neutral forward viewing axis. In this situation, imaging the person's eye with infrared light may result in a determined pupil position 428 that is above the neutral forward viewing axis. Virtual images may be displayed above a person's normal field of view and contextual information regarding target objects above a person's normal field of view may be displayed.
In an example embodiment 402, a person may be looking downwards with respect to a neutral forward viewing axis. The determined pupil position 442 may be determined by imaging the person's eye and may be further found to be below a neutral forward viewing axis. Thus, contextual information about target objects including virtual images may be displayed for target objects below the neutral forward viewing axis of the person.
FIG. 7 depicts side and front views of an eye as well as schematic drawings of pupil size information. One way to determine the ambient light level of a scene is to determine the diameter of a person's eye pupil who may be looking at the scene. In order to determine the diameter of a pupil, infrared light may be reflected off of a person's eye. The reflected light may be collected and detected with an infrared detector. Image processing can then be conducted with a processor 110 in order to determine the extents and thus the diameter of the person's pupil. For instance, in an example embodiment 500, a person may be looking directly forward and may exhibit a relatively small diameter pupil. The eyelid 504 is open and the pupil 506/512 is located centrally with respect to a reference axis 514. After image processing, which may include edge detection, the position of the pupil may be determined to be at pupil location 516 with a given pupil diameter. Due to a relatively small diameter pupil, the brightness or contrast of the virtual image may be adjusted assuming a bright ambient light level.
Similarly, in an example embodiment 502, a person may exhibit a relatively large diameter pupil. After imaging and image processing, the size and thus the diameter of the pupil may be determined. Due to a relatively large diameter pupil, the brightness or contrast of the virtual image may be adjusted assuming a dark ambient light level.
Further functions of an infrared eye-tracking system may include the recognition of various blink and eye-movement-based commands or control instructions. For instance, image recognition algorithms could recognize a succession of blinks as a command. In an example embodiment, two successive blinks with one eye within half a second may represent a command to take a picture using a second camera on the HMD.
Additionally, an eye-tracking system may allow enhanced functionality when interacting with a user interface of the HMD or of a target object. For instance, if a HMD wearer is looking at an electronic word processing document and the wearable computing device determines that the person is looking at words near the bottom of the user interface, the wearable computing device may automatically scroll the text upwards within the user interface so the person does not need to physically scroll down the page with a mouse wheel.
4. Method in an Optical System of Collecting and Imaging Infrared Light from a Viewing Location
FIG. 8 illustrates an example method 600 for an optical system to collect and image infrared light from a viewing location. It is to be understood that the steps may appear in different order and steps may be added or subtracted. In a first step 602, a visible light pattern is generated using a display panel. The display panel could be a component in an optical system similar to optical systems 302 and 304. In a second step 604, a virtual image is formed from the visible light pattern using an image former. The image former could include a quarter wave plate 228 and concave mirror 230 that may act together to form the virtual image. The method includes a third step 606 wherein the viewing location is illuminated with infrared light. The viewing location may coincide with where a HMD wearer's eye is located while wearing the HMD. The infrared light may be emitted from one or more sources, such as one or more infrared LEDs. Furthermore, infrared light may be incident upon the viewing location from multiple locations. That is, infrared light may be incident towards the viewing location along a viewing axis 204 as well as along other axes.
A fourth step 608 includes collecting infrared light reflected from the viewing location. As discussed above, infrared light reflected from the wearer's eye may be passed back into the optical system through the proximal beam splitter 216. A portion of the infrared light light may be reflected off of the proximal beam splitting interface 220 and transmitted towards the image former 218.
A fifth step 610 includes transmitting at least a portion of the collected infrared light through the image former. The image former 218 may comprise a concave mirror 230 with a dichroic thin film coating to selectively transmit infrared light and selectively reflect visible light. The image former 218 may alternatively or additionally include an aperture through which visible and infrared light may be transmitted. In both of these situations, infrared light is transmitted through the image former 218.
A sixth step 612 includes imaging the viewing location using the collected infrared light transmitted through the image former. In order to image the viewing location, light may be detected using an infrared camera 202 sensitive to infrared light. The infrared camera may convey video or still images to the processor 110. These images may be used to form the basis of a dynamically updated database of the wearer's eye pupil and its position.
CONCLUSION
The above detailed description describes various features and functions of the disclosed systems, devices, and methods with reference to the accompanying figures. While various aspects and embodiments have been disclosed herein, other aspects and embodiments will be apparent to those skilled in the art. The various aspects and embodiments disclosed herein are for purposes of illustration and are not intended to be limiting, with the true scope and spirit being indicated by the following claims.

Claims (23)

What is claimed is:
1. An optical system comprising:
a display panel configured to generate a visible light pattern;
an infrared light source configured to illuminate a viewing location with infrared light such that infrared light is reflected from the viewing location as reflected infrared light;
a camera, wherein the camera is configured to image the viewing location based on the reflected infrared light;
an image former optically coupled to the camera and the display panel, wherein the image former is configured to transmit the reflected infrared light through the image former to the camera, and wherein the image former is configured to reflect at least a portion of the visible light pattern to form a virtual image; and
a proximal beam splitter, wherein the proximal beam splitter is configured to transmit light from a real-world environment through the proximal beam splitter to the viewing location, transmit the visible light pattern from the display panel to the image former, reflect the virtual image from the image former to the viewing location, and reflect the reflected infrared light from the viewing location to the image former.
2. The optical system of claim 1, wherein the infrared light source comprises an infrared light-emitting diode (LED).
3. The optical system of claim 2, wherein the infrared LED is configured to illuminate the viewing location through the proximal beam splitter.
4. The optical system of claim 1, further comprising a visible light source.
5. The optical system of claim 4, wherein the display panel is configured to generate the visible light pattern by spatially modulating visible light from the visible light source.
6. The optical system of claim 5, further comprising:
a distal beam splitter optically coupled to the visible light source, display panel, and proximal beam splitter, wherein the distal beam splitter is configured to reflect visible light from the visible light source onto the display panel.
7. The optical system of claim 1, wherein the image former comprises a concave mirror.
8. The optical system of claim 7, wherein the concave mirror comprises a coating of dichroic material that reflects visible light and transmits infrared light.
9. The optical system of claim 1, wherein the image former comprises a concave mirror with a central pinhole.
10. The optical system of claim 9, wherein the camera includes a filter that blocks visible light and transmits infrared light.
11. The optical system of claim 1, wherein the camera is configured to image an eye pupil located at the viewing location.
12. A system comprising:
a head-mountable support;
an optical system attached to the head-mountable support, wherein the optical system comprises:
a display panel configured to generate a visible light pattern;
an infrared light source configured to illuminate a viewing location with infrared light such that infrared light is reflected from the viewing location as reflected infrared light;
a camera, wherein the camera is configured to image the viewing location based on the reflected infrared light;
an image former optically coupled to the camera and the display panel, wherein the image former is configured to transmit the reflected infrared light through the image former to the camera, and wherein the image former is configured to reflect at least a portion of the visible light pattern to form a virtual image; and
a proximal beam splitter, wherein the proximal beam splitter is configured to transmit light from a real-world environment through the proximal beam splitter to the viewing location, transmit the visible light pattern from the display panel to the image former, reflect the virtual image from the image former to the viewing location, and reflect the reflected infrared light from the viewing location to the image former;
a computer, wherein the computer is configured to control the display panel and receive images of the viewing location obtained by the camera.
13. The system of claim 12, wherein the viewing location corresponds to an eye of a wearer of the head-mounted display.
14. The system of claim 13, wherein the computer is configured to determine locations of the pupil of the wearer's eye from the received images.
15. The system of claim 14, wherein the computer is configured to control the display based on the determined pupil locations.
16. A method comprising:
generating a visible light pattern using a display panel;
forming, by reflection from a concave mirror, a virtual image from the visible light pattern, wherein the virtual image and a field of view of a real-world environment are viewable from a viewing location, wherein the concave mirror is configured to transmit infrared light and reflect visible light;
illuminating the viewing location with infrared light;
collecting infrared light reflected from the viewing location;
transmitting, through the concave mirror, at least a portion of the infrared light collected from the viewing location; and
imaging the viewing location using the collected infrared light transmitted through the concave mirror.
17. The method of claim 16, wherein imaging the viewing location comprises capturing at least one image of an eye pupil.
18. The method of claim 17, further comprising:
determining a location of the eye pupil from the at least one image.
19. The method of claim 18, further comprising controlling the display panel based on the determined location of the eye pupil.
20. The method of claim 17, further comprising:
determining a direction of motion of the eye pupil from the at least one image.
21. The method of claim 20, further comprising:
identifying a user instruction based on the determined direction of motion of the eye pupil.
22. The method of claim 17, further comprising:
determining a size of the eye pupil from the at least one image.
23. The method of claim 22, further comprising:
adjusting a brightness of the virtual image based on the determined size of the eye pupil.
US13/240,994 2011-09-22 2011-09-22 Display system Active 2031-12-01 US8767306B1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/240,994 US8767306B1 (en) 2011-09-22 2011-09-22 Display system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US13/240,994 US8767306B1 (en) 2011-09-22 2011-09-22 Display system

Publications (1)

Publication Number Publication Date
US8767306B1 true US8767306B1 (en) 2014-07-01

Family

ID=50982106

Family Applications (1)

Application Number Title Priority Date Filing Date
US13/240,994 Active 2031-12-01 US8767306B1 (en) 2011-09-22 2011-09-22 Display system

Country Status (1)

Country Link
US (1) US8767306B1 (en)

Cited By (71)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130176250A1 (en) * 2012-01-06 2013-07-11 Lg Electronics Inc. Mobile terminal and control method thereof
US20140267667A1 (en) * 2013-03-14 2014-09-18 Valve Corporation Outward facing camera system with identical camera and eye image picture perspective
US20150099460A1 (en) * 2013-10-04 2015-04-09 Sony Corporation Communication apparatus
US20150205104A1 (en) * 2014-01-21 2015-07-23 Osterhout Group, Inc. Eye imaging in head worn computing
US20150215581A1 (en) * 2014-01-24 2015-07-30 Avaya Inc. Enhanced communication between remote participants using augmented and virtual reality
US20160048024A1 (en) * 2014-08-13 2016-02-18 Beijing Lenovo Software Ltd. Information processing method and electronic device
CN105785570A (en) * 2014-12-26 2016-07-20 比亚迪股份有限公司 On-board head-up display system and vehicle comprising the same
US9398868B1 (en) 2012-09-11 2016-07-26 Verily Life Sciences Llc Cancellation of a baseline current signal via current subtraction within a linear relaxation oscillator-based current-to-frequency converter circuit
US9401540B2 (en) 2014-02-11 2016-07-26 Osterhout Group, Inc. Spatial location presentation in head worn computing
US9423612B2 (en) 2014-03-28 2016-08-23 Osterhout Group, Inc. Sensor dependent content position in head worn computing
US9436006B2 (en) 2014-01-21 2016-09-06 Osterhout Group, Inc. See-through computer display systems
US9494800B2 (en) 2014-01-21 2016-11-15 Osterhout Group, Inc. See-through computer display systems
WO2016191709A1 (en) * 2015-05-28 2016-12-01 Thalmic Labs Inc. Systems, devices, and methods that integrate eye tracking and scanning laser projection in wearable heads-up displays
US9523856B2 (en) 2014-01-21 2016-12-20 Osterhout Group, Inc. See-through computer display systems
US9529195B2 (en) 2014-01-21 2016-12-27 Osterhout Group, Inc. See-through computer display systems
US9547465B2 (en) 2014-02-14 2017-01-17 Osterhout Group, Inc. Object shadowing in head worn computing
US9575321B2 (en) 2014-06-09 2017-02-21 Osterhout Group, Inc. Content presentation in head worn computing
US9615742B2 (en) 2014-01-21 2017-04-11 Osterhout Group, Inc. Eye imaging in head worn computing
US9651787B2 (en) 2014-04-25 2017-05-16 Osterhout Group, Inc. Speaker assembly for headworn computer
US9651784B2 (en) 2014-01-21 2017-05-16 Osterhout Group, Inc. See-through computer display systems
US9672210B2 (en) 2014-04-25 2017-06-06 Osterhout Group, Inc. Language translation with head-worn computing
US9671613B2 (en) 2014-09-26 2017-06-06 Osterhout Group, Inc. See-through computer display systems
US9684172B2 (en) 2014-12-03 2017-06-20 Osterhout Group, Inc. Head worn computer display systems
USD792400S1 (en) 2014-12-31 2017-07-18 Osterhout Group, Inc. Computer glasses
US9715112B2 (en) 2014-01-21 2017-07-25 Osterhout Group, Inc. Suppression of stray light in head worn computing
US9720234B2 (en) 2014-01-21 2017-08-01 Osterhout Group, Inc. See-through computer display systems
USD794637S1 (en) 2015-01-05 2017-08-15 Osterhout Group, Inc. Air mouse
US9740280B2 (en) 2014-01-21 2017-08-22 Osterhout Group, Inc. Eye imaging in head worn computing
US9746686B2 (en) 2014-05-19 2017-08-29 Osterhout Group, Inc. Content position calibration in head worn computing
US9753288B2 (en) 2014-01-21 2017-09-05 Osterhout Group, Inc. See-through computer display systems
US9766463B2 (en) 2014-01-21 2017-09-19 Osterhout Group, Inc. See-through computer display systems
US9784973B2 (en) 2014-02-11 2017-10-10 Osterhout Group, Inc. Micro doppler presentations in head worn computing
US9811152B2 (en) 2014-01-21 2017-11-07 Osterhout Group, Inc. Eye imaging in head worn computing
US9810906B2 (en) 2014-06-17 2017-11-07 Osterhout Group, Inc. External user interface for head worn computing
US9829707B2 (en) 2014-08-12 2017-11-28 Osterhout Group, Inc. Measuring content brightness in head worn computing
US9836122B2 (en) 2014-01-21 2017-12-05 Osterhout Group, Inc. Eye glint imaging in see-through computer display systems
US9841599B2 (en) 2014-06-05 2017-12-12 Osterhout Group, Inc. Optical configurations for head-worn see-through displays
CN107765435A (en) * 2017-11-22 2018-03-06 深圳创维新世界科技有限公司 Head-wearing display device
US9939646B2 (en) 2014-01-24 2018-04-10 Osterhout Group, Inc. Stray light suppression for head worn computing
US9939934B2 (en) 2014-01-17 2018-04-10 Osterhout Group, Inc. External user interface for head worn computing
US9952664B2 (en) 2014-01-21 2018-04-24 Osterhout Group, Inc. Eye imaging in head worn computing
US9965681B2 (en) 2008-12-16 2018-05-08 Osterhout Group, Inc. Eye imaging in head worn computing
EP3318914A1 (en) * 2016-11-04 2018-05-09 Nokia Technologies OY Gaze tracking
US10062182B2 (en) 2015-02-17 2018-08-28 Osterhout Group, Inc. See-through computer display systems
US10111620B2 (en) 2015-02-27 2018-10-30 Microsoft Technology Licensing, Llc Enhanced motion tracking using transportable inertial sensors to determine that a frame of reference is established
US10191279B2 (en) 2014-03-17 2019-01-29 Osterhout Group, Inc. Eye imaging in head worn computing
US10254856B2 (en) 2014-01-17 2019-04-09 Osterhout Group, Inc. External user interface for head worn computing
TWI667495B (en) * 2017-03-13 2019-08-01 宏達國際電子股份有限公司 Head mounted display device, object tracking apparatus and method for tracking object thereof
US10444018B2 (en) 2015-02-27 2019-10-15 Microsoft Technology Licensing, Llc Computer-implemented method to test the sensitivity of a sensor for detecting movement of a tracking device within an established frame of reference of a moving platform
US10558050B2 (en) 2014-01-24 2020-02-11 Mentor Acquisition One, Llc Haptic systems for head-worn computers
US10649220B2 (en) 2014-06-09 2020-05-12 Mentor Acquisition One, Llc Content presentation in head worn computing
WO2020096187A1 (en) * 2018-11-08 2020-05-14 주식회사 레티널 Optical device for augmented reality
US10663740B2 (en) 2014-06-09 2020-05-26 Mentor Acquisition One, Llc Content presentation in head worn computing
US10684687B2 (en) 2014-12-03 2020-06-16 Mentor Acquisition One, Llc See-through computer display systems
CN111338085A (en) * 2020-04-21 2020-06-26 Oppo广东移动通信有限公司 Display optical system, head-mounted display apparatus, control method, and storage medium
CN111357008A (en) * 2017-11-28 2020-06-30 指纹卡有限公司 Biometric imaging system and method for controlling the same
US10783835B2 (en) * 2016-03-11 2020-09-22 Lenovo (Singapore) Pte. Ltd. Automatic control of display brightness
WO2020223228A1 (en) * 2019-04-30 2020-11-05 Intuitive Surgical Operations, Inc. Image viewing systems and methods using a black glass mirror
US10853589B2 (en) 2014-04-25 2020-12-01 Mentor Acquisition One, Llc Language translation with head-worn computing
US10932695B2 (en) 2012-09-17 2021-03-02 Verily Life Sciences Llc Sensing system
US10963103B1 (en) * 2018-12-24 2021-03-30 Facebook Technologies, Llc Display system with integrated depth detection
US11104272B2 (en) 2014-03-28 2021-08-31 Mentor Acquisition One, Llc System for assisted operator safety using an HMD
US11103122B2 (en) 2014-07-15 2021-08-31 Mentor Acquisition One, Llc Content presentation in head worn computing
CN113396356A (en) * 2019-02-22 2021-09-14 株式会社籁天那 Compact enhanced in-situ optical device
US11227294B2 (en) 2014-04-03 2022-01-18 Mentor Acquisition One, Llc Sight information collection in head worn computing
US11269182B2 (en) 2014-07-15 2022-03-08 Mentor Acquisition One, Llc Content presentation in head worn computing
US20220252875A1 (en) * 2019-05-29 2022-08-11 Facebook Technologies, Llc Dual purkinje imaging with ellipsoidal lensing structure
US11487110B2 (en) 2014-01-21 2022-11-01 Mentor Acquisition One, Llc Eye imaging in head worn computing
US11669163B2 (en) 2014-01-21 2023-06-06 Mentor Acquisition One, Llc Eye glint imaging in see-through computer display systems
US11737666B2 (en) 2014-01-21 2023-08-29 Mentor Acquisition One, Llc Eye imaging in head worn computing
US11892644B2 (en) 2014-01-21 2024-02-06 Mentor Acquisition One, Llc See-through computer display systems

Citations (57)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5526184A (en) 1992-12-10 1996-06-11 Olympus Optical Co., Ltd. Head-mounted display apparatus for observing an outside world image and/or a display image
US5715337A (en) 1996-09-19 1998-02-03 The Mirco Optical Corporation Compact display system
US5886822A (en) 1996-10-08 1999-03-23 The Microoptical Corporation Image combining system for eyeglasses and face masks
US5943171A (en) 1998-06-03 1999-08-24 International Business Machines Corporation Head mounted displays utilizing reflection light valves
US5949583A (en) 1992-02-07 1999-09-07 I-O Display Systems Llc Head-mounted display with image generator, fold mirror and mirror for transmission to the eye position of the user
US6023372A (en) 1997-10-30 2000-02-08 The Microoptical Corporation Light weight, compact remountable electronic display device for eyeglasses or other head-borne eyewear frames
US6091546A (en) 1997-10-30 2000-07-18 The Microoptical Corporation Eyeglass interface system
US6191892B1 (en) * 1996-04-02 2001-02-20 Canon Kabushiki Kaisha Image display apparatus
US6201629B1 (en) 1997-08-27 2001-03-13 Microoptical Corporation Torsional micro-mechanical mirror system
US6204974B1 (en) 1996-10-08 2001-03-20 The Microoptical Corporation Compact image display system for eyeglasses or other head-borne frames
US6353503B1 (en) 1999-06-21 2002-03-05 The Micropitical Corporation Eyeglass display lens system employing off-axis optical design
US20030090439A1 (en) 2001-09-07 2003-05-15 Spitzer Mark B. Light weight, compact, remountable face-supported electronic display
US6618099B1 (en) 1999-06-21 2003-09-09 The Microoptical Corporation Display device with eyepiece assembly and display on opto-mechanical support
US6693749B2 (en) 2001-01-31 2004-02-17 Raytheon Company Low-observability, wide-field-of-view, situation awareness viewing device
US6701038B2 (en) 2001-03-05 2004-03-02 The Microoptical Corporation Micro-electromechanical optical switch assembly for optical data networks
US6724354B1 (en) 1999-06-21 2004-04-20 The Microoptical Corporation Illumination systems for eyeglass and facemask display systems
US6847336B1 (en) 1996-10-02 2005-01-25 Jerome H. Lemelson Selectively controllable heads-up display system
US6862006B2 (en) * 2002-05-17 2005-03-01 Seiko Epson Corporation Image processing apparatus and image processing method, and image processing program and recording medium of the same
US6879443B2 (en) 2003-04-25 2005-04-12 The Microoptical Corporation Binocular viewing system
US6916096B2 (en) * 2000-09-23 2005-07-12 Heinrich A. Eberl System and method for recording the retinal reflex image
US20060192307A1 (en) 2005-02-25 2006-08-31 Eugene Giller Method for producing high quality optical parts by casting
US20060192306A1 (en) 2005-02-25 2006-08-31 The Microoptical Corporation Manufacturing methods for embedded optical system
US20060238877A1 (en) * 2003-05-12 2006-10-26 Elbit Systems Ltd. Advanced Technology Center Method and system for improving audiovisual communication
US7158096B1 (en) 1999-06-21 2007-01-02 The Microoptical Corporation Compact, head-mountable display device with suspended eyepiece assembly
US7192136B2 (en) 2003-04-15 2007-03-20 Howell Thomas A Tethered electrical components for eyeglasses
US7242527B2 (en) 2005-03-22 2007-07-10 The Microoptical Corporation Optical system using total internal reflection images
US7255437B2 (en) 2003-10-09 2007-08-14 Howell Thomas A Eyeglasses with activity monitoring
US7380936B2 (en) 2003-10-09 2008-06-03 Ipventure, Inc. Eyeglasses with a clock or other electrical component
US20080219025A1 (en) 2007-03-07 2008-09-11 Spitzer Mark B Bi-directional backlight assembly
US7438410B1 (en) 2003-10-09 2008-10-21 Ip Venture, Inc. Tethered electrical components for eyeglasses
US7457040B2 (en) 2002-03-21 2008-11-25 Lumus Ltd. Light guide optical device
US7500747B2 (en) 2003-10-09 2009-03-10 Ipventure, Inc. Eyeglasses with electrical components
US7500746B1 (en) 2004-04-15 2009-03-10 Ip Venture, Inc. Eyewear with radiation detection system
US7522344B1 (en) 2005-12-14 2009-04-21 University Of Central Florida Research Foundation, Inc. Projection-based head-mounted display with eye-tracking capabilities
US20090122414A1 (en) 2005-02-10 2009-05-14 Lumus Ltd. Substrate-Guided Optical Device Utilzing Thin Transparent Layer
US7542210B2 (en) 2006-06-29 2009-06-02 Chirieleison Sr Anthony Eye tracking head mounted display
US7543934B2 (en) 2006-09-20 2009-06-09 Ipventures, Inc. Eyeglasses with activity monitoring and acoustic dampening
US7581833B2 (en) 2003-10-09 2009-09-01 Ipventure, Inc. Eyewear supporting after-market electrical components
US20090231687A1 (en) * 2008-03-11 2009-09-17 Kakuya Yamamoto Display apparatus, display method, goggle-type head-mounted display, and vehicle
US7663805B2 (en) 2007-10-09 2010-02-16 Myvu Corporation Eyewear display and media device interconnection system
US20100046070A1 (en) 2008-08-21 2010-02-25 Sony Corporation Head-mounted display
US7672055B2 (en) 2003-09-10 2010-03-02 Lumus Ltd. Substrate-guided optical devices
US7677723B2 (en) 2003-10-09 2010-03-16 Ipventure, Inc. Eyeglasses with a heart rate monitor
US20100103078A1 (en) 2008-10-23 2010-04-29 Sony Corporation Head-mounted display apparatus
US20100149073A1 (en) 2008-11-02 2010-06-17 David Chaum Near to Eye Display System and Appliance
US7751122B2 (en) 2005-02-10 2010-07-06 Lumus Ltd. Substrate-guided optical device particularly for vision enhanced optical systems
US7760898B2 (en) 2003-10-09 2010-07-20 Ip Venture, Inc. Eyeglasses with hearing enhanced and other audio signal-generating capabilities
US7792552B2 (en) 2003-04-15 2010-09-07 Ipventure, Inc. Eyeglasses for wireless communications
US7806525B2 (en) 2003-10-09 2010-10-05 Ipventure, Inc. Eyeglasses having a camera
US20100278480A1 (en) 2009-04-21 2010-11-04 Vasylyev Sergiy V Light collection and illumination systems employing planar waveguide
US20110043644A1 (en) 2008-04-02 2011-02-24 Esight Corp. Apparatus and Method for a Dynamic "Region of Interest" in a Display System
US7900068B2 (en) 2006-09-14 2011-03-01 Hon Hai Precision Industry Co., Ltd. Mobile multi-media interface and power pack for portable entertainment devices
US20110077548A1 (en) 2004-04-01 2011-03-31 Torch William C Biosensors, communicators, and controllers monitoring eye movement and methods for using them
US7922321B2 (en) 2003-10-09 2011-04-12 Ipventure, Inc. Eyewear supporting after-market electrical components
US7936518B2 (en) * 2006-11-27 2011-05-03 Nippon Seiki Co., Ltd. Head-up display apparatus
US20110227820A1 (en) 2010-02-28 2011-09-22 Osterhout Group, Inc. Lock virtual keyboard position in an augmented reality eyepiece
US8109629B2 (en) 2003-10-09 2012-02-07 Ipventure, Inc. Eyewear supporting electrical components and apparatus therefor

Patent Citations (71)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5949583A (en) 1992-02-07 1999-09-07 I-O Display Systems Llc Head-mounted display with image generator, fold mirror and mirror for transmission to the eye position of the user
US5526184A (en) 1992-12-10 1996-06-11 Olympus Optical Co., Ltd. Head-mounted display apparatus for observing an outside world image and/or a display image
US6191892B1 (en) * 1996-04-02 2001-02-20 Canon Kabushiki Kaisha Image display apparatus
US5715337A (en) 1996-09-19 1998-02-03 The Mirco Optical Corporation Compact display system
US6847336B1 (en) 1996-10-02 2005-01-25 Jerome H. Lemelson Selectively controllable heads-up display system
US6356392B1 (en) 1996-10-08 2002-03-12 The Microoptical Corporation Compact image display system for eyeglasses or other head-borne frames
US5886822A (en) 1996-10-08 1999-03-23 The Microoptical Corporation Image combining system for eyeglasses and face masks
US6204974B1 (en) 1996-10-08 2001-03-20 The Microoptical Corporation Compact image display system for eyeglasses or other head-borne frames
US6384982B1 (en) 1996-10-08 2002-05-07 The Microoptical Corporation Compact image display system for eyeglasses or other head-borne frames
US6538799B2 (en) 1997-08-27 2003-03-25 The Microoptical Corporation Magnetically actuated torsional micro-mechanical mirror system
US6201629B1 (en) 1997-08-27 2001-03-13 Microoptical Corporation Torsional micro-mechanical mirror system
US6353492B2 (en) 1997-08-27 2002-03-05 The Microoptical Corporation Method of fabrication of a torsional micro-mechanical mirror system
US6023372A (en) 1997-10-30 2000-02-08 The Microoptical Corporation Light weight, compact remountable electronic display device for eyeglasses or other head-borne eyewear frames
US6349001B1 (en) 1997-10-30 2002-02-19 The Microoptical Corporation Eyeglass interface system
US6091546A (en) 1997-10-30 2000-07-18 The Microoptical Corporation Eyeglass interface system
US5943171A (en) 1998-06-03 1999-08-24 International Business Machines Corporation Head mounted displays utilizing reflection light valves
US6353503B1 (en) 1999-06-21 2002-03-05 The Micropitical Corporation Eyeglass display lens system employing off-axis optical design
US7158096B1 (en) 1999-06-21 2007-01-02 The Microoptical Corporation Compact, head-mountable display device with suspended eyepiece assembly
US6618099B1 (en) 1999-06-21 2003-09-09 The Microoptical Corporation Display device with eyepiece assembly and display on opto-mechanical support
US7843403B2 (en) 1999-06-21 2010-11-30 Myvu Corporation Compact, head-mountable display device with suspended eyepiece assembly
US6724354B1 (en) 1999-06-21 2004-04-20 The Microoptical Corporation Illumination systems for eyeglass and facemask display systems
US6916096B2 (en) * 2000-09-23 2005-07-12 Heinrich A. Eberl System and method for recording the retinal reflex image
US6693749B2 (en) 2001-01-31 2004-02-17 Raytheon Company Low-observability, wide-field-of-view, situation awareness viewing device
US6701038B2 (en) 2001-03-05 2004-03-02 The Microoptical Corporation Micro-electromechanical optical switch assembly for optical data networks
US20030090439A1 (en) 2001-09-07 2003-05-15 Spitzer Mark B. Light weight, compact, remountable face-supported electronic display
US8004765B2 (en) 2002-03-21 2011-08-23 Lumus Ltd. Light guide optical device
US7457040B2 (en) 2002-03-21 2008-11-25 Lumus Ltd. Light guide optical device
US6862006B2 (en) * 2002-05-17 2005-03-01 Seiko Epson Corporation Image processing apparatus and image processing method, and image processing program and recording medium of the same
US7192136B2 (en) 2003-04-15 2007-03-20 Howell Thomas A Tethered electrical components for eyeglasses
US7792552B2 (en) 2003-04-15 2010-09-07 Ipventure, Inc. Eyeglasses for wireless communications
US6879443B2 (en) 2003-04-25 2005-04-12 The Microoptical Corporation Binocular viewing system
US20050174651A1 (en) 2003-04-25 2005-08-11 The Microoptical Corporation Binocular viewing system
US20060238877A1 (en) * 2003-05-12 2006-10-26 Elbit Systems Ltd. Advanced Technology Center Method and system for improving audiovisual communication
US7672055B2 (en) 2003-09-10 2010-03-02 Lumus Ltd. Substrate-guided optical devices
US7401918B2 (en) 2003-10-09 2008-07-22 Ipventure, Inc. Eyeglasses with activity monitoring
US7771046B2 (en) 2003-10-09 2010-08-10 I p Venture, Inc. Eyewear with monitoring capability
US7438410B1 (en) 2003-10-09 2008-10-21 Ip Venture, Inc. Tethered electrical components for eyeglasses
US7380936B2 (en) 2003-10-09 2008-06-03 Ipventure, Inc. Eyeglasses with a clock or other electrical component
US7481531B2 (en) 2003-10-09 2009-01-27 Ipventure, Inc. Eyeglasses with user monitoring
US7500747B2 (en) 2003-10-09 2009-03-10 Ipventure, Inc. Eyeglasses with electrical components
US7922321B2 (en) 2003-10-09 2011-04-12 Ipventure, Inc. Eyewear supporting after-market electrical components
US7255437B2 (en) 2003-10-09 2007-08-14 Howell Thomas A Eyeglasses with activity monitoring
US7760898B2 (en) 2003-10-09 2010-07-20 Ip Venture, Inc. Eyeglasses with hearing enhanced and other audio signal-generating capabilities
US7806525B2 (en) 2003-10-09 2010-10-05 Ipventure, Inc. Eyeglasses having a camera
US8109629B2 (en) 2003-10-09 2012-02-07 Ipventure, Inc. Eyewear supporting electrical components and apparatus therefor
US7581833B2 (en) 2003-10-09 2009-09-01 Ipventure, Inc. Eyewear supporting after-market electrical components
US7677723B2 (en) 2003-10-09 2010-03-16 Ipventure, Inc. Eyeglasses with a heart rate monitor
US7621634B2 (en) 2003-10-09 2009-11-24 Ipventure, Inc. Tethered electrical components for eyeglasses
US20110077548A1 (en) 2004-04-01 2011-03-31 Torch William C Biosensors, communicators, and controllers monitoring eye movement and methods for using them
US7500746B1 (en) 2004-04-15 2009-03-10 Ip Venture, Inc. Eyewear with radiation detection system
US7724443B2 (en) 2005-02-10 2010-05-25 Lumus Ltd. Substrate-guided optical device utilizing thin transparent layer
US7751122B2 (en) 2005-02-10 2010-07-06 Lumus Ltd. Substrate-guided optical device particularly for vision enhanced optical systems
US20090122414A1 (en) 2005-02-10 2009-05-14 Lumus Ltd. Substrate-Guided Optical Device Utilzing Thin Transparent Layer
US20060192307A1 (en) 2005-02-25 2006-08-31 Eugene Giller Method for producing high quality optical parts by casting
US20060192306A1 (en) 2005-02-25 2006-08-31 The Microoptical Corporation Manufacturing methods for embedded optical system
US7242527B2 (en) 2005-03-22 2007-07-10 The Microoptical Corporation Optical system using total internal reflection images
US7522344B1 (en) 2005-12-14 2009-04-21 University Of Central Florida Research Foundation, Inc. Projection-based head-mounted display with eye-tracking capabilities
US7542210B2 (en) 2006-06-29 2009-06-02 Chirieleison Sr Anthony Eye tracking head mounted display
US7900068B2 (en) 2006-09-14 2011-03-01 Hon Hai Precision Industry Co., Ltd. Mobile multi-media interface and power pack for portable entertainment devices
US7543934B2 (en) 2006-09-20 2009-06-09 Ipventures, Inc. Eyeglasses with activity monitoring and acoustic dampening
US7936518B2 (en) * 2006-11-27 2011-05-03 Nippon Seiki Co., Ltd. Head-up display apparatus
US20080219025A1 (en) 2007-03-07 2008-09-11 Spitzer Mark B Bi-directional backlight assembly
US7663805B2 (en) 2007-10-09 2010-02-16 Myvu Corporation Eyewear display and media device interconnection system
US20090231687A1 (en) * 2008-03-11 2009-09-17 Kakuya Yamamoto Display apparatus, display method, goggle-type head-mounted display, and vehicle
US20110043644A1 (en) 2008-04-02 2011-02-24 Esight Corp. Apparatus and Method for a Dynamic "Region of Interest" in a Display System
US20100046070A1 (en) 2008-08-21 2010-02-25 Sony Corporation Head-mounted display
US20100103078A1 (en) 2008-10-23 2010-04-29 Sony Corporation Head-mounted display apparatus
US20100149073A1 (en) 2008-11-02 2010-06-17 David Chaum Near to Eye Display System and Appliance
US20100278480A1 (en) 2009-04-21 2010-11-04 Vasylyev Sergiy V Light collection and illumination systems employing planar waveguide
WO2011060525A1 (en) 2009-11-19 2011-05-26 Esight Corporation Image magnification on a head mounted display
US20110227820A1 (en) 2010-02-28 2011-09-22 Osterhout Group, Inc. Lock virtual keyboard position in an augmented reality eyepiece

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
Cakmakci, Ozan et al., "Head-Worn Displays: A Review," Journal of Display Technology, vol. 2, pp. 199-216, 2006.
Levola, Tapani, "Diffractive Optics for Virtual Reality Displays", Academic Dissertation, Joensuu 2005, University of Joensuu, Department of Physics, Vaisala Laboratory, 26 pages.
Mukawa, Hiroshi et al., "Distinguished Paper: A Full Color Eyewear Display using Holographic Planar Waveguides", SID Symposium Digest of Technical Papers-May 2008-vol. 39, Issue 1, pp. 89-92.
Mukawa, Hiroshi et al., "Distinguished Paper: A Full Color Eyewear Display using Holographic Planar Waveguides", SID Symposium Digest of Technical Papers—May 2008—vol. 39, Issue 1, pp. 89-92.
Selker, Ted et al., "Eye-R, a Glasses-Mounted Eye Motion Detection Interface," CHI 2001, pp. 179-180, Mar. 31-Apr. 5, 2001.

Cited By (163)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9965681B2 (en) 2008-12-16 2018-05-08 Osterhout Group, Inc. Eye imaging in head worn computing
US20130176250A1 (en) * 2012-01-06 2013-07-11 Lg Electronics Inc. Mobile terminal and control method thereof
US9398868B1 (en) 2012-09-11 2016-07-26 Verily Life Sciences Llc Cancellation of a baseline current signal via current subtraction within a linear relaxation oscillator-based current-to-frequency converter circuit
US9737248B1 (en) 2012-09-11 2017-08-22 Verily Life Sciences Llc Cancellation of a baseline current signal via current subtraction within a linear relaxation oscillator-based current-to-frequency converter circuit
US10729363B1 (en) 2012-09-11 2020-08-04 Verily Life Sciences Llc Cancellation of a baseline current signal via current subtraction within a linear relaxation oscillator-based current-to-frequency converter circuit
US10932695B2 (en) 2012-09-17 2021-03-02 Verily Life Sciences Llc Sensing system
US20140267667A1 (en) * 2013-03-14 2014-09-18 Valve Corporation Outward facing camera system with identical camera and eye image picture perspective
US20150099460A1 (en) * 2013-10-04 2015-04-09 Sony Corporation Communication apparatus
US9537529B2 (en) * 2013-10-04 2017-01-03 Sony Corporation Communication apparatus
US11782529B2 (en) 2014-01-17 2023-10-10 Mentor Acquisition One, Llc External user interface for head worn computing
US11169623B2 (en) 2014-01-17 2021-11-09 Mentor Acquisition One, Llc External user interface for head worn computing
US10254856B2 (en) 2014-01-17 2019-04-09 Osterhout Group, Inc. External user interface for head worn computing
US11231817B2 (en) 2014-01-17 2022-01-25 Mentor Acquisition One, Llc External user interface for head worn computing
US9939934B2 (en) 2014-01-17 2018-04-10 Osterhout Group, Inc. External user interface for head worn computing
US11507208B2 (en) 2014-01-17 2022-11-22 Mentor Acquisition One, Llc External user interface for head worn computing
US11126003B2 (en) 2014-01-21 2021-09-21 Mentor Acquisition One, Llc See-through computer display systems
US9811152B2 (en) 2014-01-21 2017-11-07 Osterhout Group, Inc. Eye imaging in head worn computing
US9529192B2 (en) * 2014-01-21 2016-12-27 Osterhout Group, Inc. Eye imaging in head worn computing
US9529199B2 (en) 2014-01-21 2016-12-27 Osterhout Group, Inc. See-through computer display systems
US9529195B2 (en) 2014-01-21 2016-12-27 Osterhout Group, Inc. See-through computer display systems
US9523856B2 (en) 2014-01-21 2016-12-20 Osterhout Group, Inc. See-through computer display systems
US10579140B2 (en) 2014-01-21 2020-03-03 Mentor Acquisition One, Llc Eye glint imaging in see-through computer display systems
US11947126B2 (en) 2014-01-21 2024-04-02 Mentor Acquisition One, Llc See-through computer display systems
US9594246B2 (en) 2014-01-21 2017-03-14 Osterhout Group, Inc. See-through computer display systems
US9615742B2 (en) 2014-01-21 2017-04-11 Osterhout Group, Inc. Eye imaging in head worn computing
US11892644B2 (en) 2014-01-21 2024-02-06 Mentor Acquisition One, Llc See-through computer display systems
US9651788B2 (en) 2014-01-21 2017-05-16 Osterhout Group, Inc. See-through computer display systems
US9651784B2 (en) 2014-01-21 2017-05-16 Osterhout Group, Inc. See-through computer display systems
US9651783B2 (en) 2014-01-21 2017-05-16 Osterhout Group, Inc. See-through computer display systems
US9651789B2 (en) 2014-01-21 2017-05-16 Osterhout Group, Inc. See-Through computer display systems
US9658457B2 (en) 2014-01-21 2017-05-23 Osterhout Group, Inc. See-through computer display systems
US9658458B2 (en) 2014-01-21 2017-05-23 Osterhout Group, Inc. See-through computer display systems
US10866420B2 (en) 2014-01-21 2020-12-15 Mentor Acquisition One, Llc See-through computer display systems
US11796805B2 (en) 2014-01-21 2023-10-24 Mentor Acquisition One, Llc Eye imaging in head worn computing
US11796799B2 (en) 2014-01-21 2023-10-24 Mentor Acquisition One, Llc See-through computer display systems
US9684171B2 (en) 2014-01-21 2017-06-20 Osterhout Group, Inc. See-through computer display systems
US9684165B2 (en) 2014-01-21 2017-06-20 Osterhout Group, Inc. Eye imaging in head worn computing
US20150205104A1 (en) * 2014-01-21 2015-07-23 Osterhout Group, Inc. Eye imaging in head worn computing
US11737666B2 (en) 2014-01-21 2023-08-29 Mentor Acquisition One, Llc Eye imaging in head worn computing
US9715112B2 (en) 2014-01-21 2017-07-25 Osterhout Group, Inc. Suppression of stray light in head worn computing
US9720234B2 (en) 2014-01-21 2017-08-01 Osterhout Group, Inc. See-through computer display systems
US9720235B2 (en) 2014-01-21 2017-08-01 Osterhout Group, Inc. See-through computer display systems
US11669163B2 (en) 2014-01-21 2023-06-06 Mentor Acquisition One, Llc Eye glint imaging in see-through computer display systems
US9720227B2 (en) 2014-01-21 2017-08-01 Osterhout Group, Inc. See-through computer display systems
US11650416B2 (en) 2014-01-21 2023-05-16 Mentor Acquisition One, Llc See-through computer display systems
US9740012B2 (en) 2014-01-21 2017-08-22 Osterhout Group, Inc. See-through computer display systems
US9740280B2 (en) 2014-01-21 2017-08-22 Osterhout Group, Inc. Eye imaging in head worn computing
US11619820B2 (en) 2014-01-21 2023-04-04 Mentor Acquisition One, Llc See-through computer display systems
US11622426B2 (en) 2014-01-21 2023-04-04 Mentor Acquisition One, Llc See-through computer display systems
US9746676B2 (en) 2014-01-21 2017-08-29 Osterhout Group, Inc. See-through computer display systems
US9753288B2 (en) 2014-01-21 2017-09-05 Osterhout Group, Inc. See-through computer display systems
US9766463B2 (en) 2014-01-21 2017-09-19 Osterhout Group, Inc. See-through computer display systems
US9772492B2 (en) * 2014-01-21 2017-09-26 Osterhout Group, Inc. Eye imaging in head worn computing
US9933622B2 (en) 2014-01-21 2018-04-03 Osterhout Group, Inc. See-through computer display systems
US9811159B2 (en) 2014-01-21 2017-11-07 Osterhout Group, Inc. Eye imaging in head worn computing
US10698223B2 (en) 2014-01-21 2020-06-30 Mentor Acquisition One, Llc See-through computer display systems
US9811153B2 (en) 2014-01-21 2017-11-07 Osterhout Group, Inc. Eye imaging in head worn computing
US10890760B2 (en) 2014-01-21 2021-01-12 Mentor Acquisition One, Llc See-through computer display systems
US20150205105A1 (en) * 2014-01-21 2015-07-23 Osterhout Group, Inc. Eye imaging in head worn computing
US9829703B2 (en) 2014-01-21 2017-11-28 Osterhout Group, Inc. Eye imaging in head worn computing
US9836122B2 (en) 2014-01-21 2017-12-05 Osterhout Group, Inc. Eye glint imaging in see-through computer display systems
US11487110B2 (en) 2014-01-21 2022-11-01 Mentor Acquisition One, Llc Eye imaging in head worn computing
US11002961B2 (en) 2014-01-21 2021-05-11 Mentor Acquisition One, Llc See-through computer display systems
US10321821B2 (en) 2014-01-21 2019-06-18 Mentor Acquisition One, Llc Eye imaging in head worn computing
US9885868B2 (en) * 2014-01-21 2018-02-06 Osterhout Group, Inc. Eye imaging in head worn computing
US11353957B2 (en) 2014-01-21 2022-06-07 Mentor Acquisition One, Llc Eye glint imaging in see-through computer display systems
US20150205103A1 (en) * 2014-01-21 2015-07-23 Osterhout Group, Inc. Eye imaging in head worn computing
US11054902B2 (en) 2014-01-21 2021-07-06 Mentor Acquisition One, Llc Eye glint imaging in see-through computer display systems
US9927612B2 (en) 2014-01-21 2018-03-27 Osterhout Group, Inc. See-through computer display systems
US11099380B2 (en) 2014-01-21 2021-08-24 Mentor Acquisition One, Llc Eye imaging in head worn computing
US11103132B2 (en) 2014-01-21 2021-08-31 Mentor Acquisition One, Llc Eye imaging in head worn computing
US9494800B2 (en) 2014-01-21 2016-11-15 Osterhout Group, Inc. See-through computer display systems
US9952664B2 (en) 2014-01-21 2018-04-24 Osterhout Group, Inc. Eye imaging in head worn computing
US10001644B2 (en) 2014-01-21 2018-06-19 Osterhout Group, Inc. See-through computer display systems
US9958674B2 (en) 2014-01-21 2018-05-01 Osterhout Group, Inc. Eye imaging in head worn computing
US9436006B2 (en) 2014-01-21 2016-09-06 Osterhout Group, Inc. See-through computer display systems
US10558050B2 (en) 2014-01-24 2020-02-11 Mentor Acquisition One, Llc Haptic systems for head-worn computers
US9959676B2 (en) 2014-01-24 2018-05-01 Avaya Inc. Presentation of enhanced communication between remote participants using augmented and virtual reality
US10013805B2 (en) 2014-01-24 2018-07-03 Avaya Inc. Control of enhanced communication between remote participants using augmented and virtual reality
US11822090B2 (en) 2014-01-24 2023-11-21 Mentor Acquisition One, Llc Haptic systems for head-worn computers
US9939646B2 (en) 2014-01-24 2018-04-10 Osterhout Group, Inc. Stray light suppression for head worn computing
US9524588B2 (en) * 2014-01-24 2016-12-20 Avaya Inc. Enhanced communication between remote participants using augmented and virtual reality
US20150215581A1 (en) * 2014-01-24 2015-07-30 Avaya Inc. Enhanced communication between remote participants using augmented and virtual reality
US9784973B2 (en) 2014-02-11 2017-10-10 Osterhout Group, Inc. Micro doppler presentations in head worn computing
US9843093B2 (en) 2014-02-11 2017-12-12 Osterhout Group, Inc. Spatial location presentation in head worn computing
US9841602B2 (en) 2014-02-11 2017-12-12 Osterhout Group, Inc. Location indicating avatar in head worn computing
US9401540B2 (en) 2014-02-11 2016-07-26 Osterhout Group, Inc. Spatial location presentation in head worn computing
US9928019B2 (en) 2014-02-14 2018-03-27 Osterhout Group, Inc. Object shadowing in head worn computing
US9547465B2 (en) 2014-02-14 2017-01-17 Osterhout Group, Inc. Object shadowing in head worn computing
US10191279B2 (en) 2014-03-17 2019-01-29 Osterhout Group, Inc. Eye imaging in head worn computing
US9423612B2 (en) 2014-03-28 2016-08-23 Osterhout Group, Inc. Sensor dependent content position in head worn computing
US11104272B2 (en) 2014-03-28 2021-08-31 Mentor Acquisition One, Llc System for assisted operator safety using an HMD
US11227294B2 (en) 2014-04-03 2022-01-18 Mentor Acquisition One, Llc Sight information collection in head worn computing
US11880041B2 (en) 2014-04-25 2024-01-23 Mentor Acquisition One, Llc Speaker assembly for headworn computer
US11474360B2 (en) 2014-04-25 2022-10-18 Mentor Acquisition One, Llc Speaker assembly for headworn computer
US11727223B2 (en) 2014-04-25 2023-08-15 Mentor Acquisition One, Llc Language translation with head-worn computing
US9651787B2 (en) 2014-04-25 2017-05-16 Osterhout Group, Inc. Speaker assembly for headworn computer
US10853589B2 (en) 2014-04-25 2020-12-01 Mentor Acquisition One, Llc Language translation with head-worn computing
US9672210B2 (en) 2014-04-25 2017-06-06 Osterhout Group, Inc. Language translation with head-worn computing
US10634922B2 (en) 2014-04-25 2020-04-28 Mentor Acquisition One, Llc Speaker assembly for headworn computer
US9746686B2 (en) 2014-05-19 2017-08-29 Osterhout Group, Inc. Content position calibration in head worn computing
US9841599B2 (en) 2014-06-05 2017-12-12 Osterhout Group, Inc. Optical configurations for head-worn see-through displays
US10877270B2 (en) 2014-06-05 2020-12-29 Mentor Acquisition One, Llc Optical configurations for head-worn see-through displays
US11402639B2 (en) 2014-06-05 2022-08-02 Mentor Acquisition One, Llc Optical configurations for head-worn see-through displays
US11960089B2 (en) 2014-06-05 2024-04-16 Mentor Acquisition One, Llc Optical configurations for head-worn see-through displays
US11663794B2 (en) 2014-06-09 2023-05-30 Mentor Acquisition One, Llc Content presentation in head worn computing
US9575321B2 (en) 2014-06-09 2017-02-21 Osterhout Group, Inc. Content presentation in head worn computing
US9720241B2 (en) 2014-06-09 2017-08-01 Osterhout Group, Inc. Content presentation in head worn computing
US10649220B2 (en) 2014-06-09 2020-05-12 Mentor Acquisition One, Llc Content presentation in head worn computing
US11790617B2 (en) 2014-06-09 2023-10-17 Mentor Acquisition One, Llc Content presentation in head worn computing
US10976559B2 (en) 2014-06-09 2021-04-13 Mentor Acquisition One, Llc Content presentation in head worn computing
US11327323B2 (en) 2014-06-09 2022-05-10 Mentor Acquisition One, Llc Content presentation in head worn computing
US11022810B2 (en) 2014-06-09 2021-06-01 Mentor Acquisition One, Llc Content presentation in head worn computing
US10663740B2 (en) 2014-06-09 2020-05-26 Mentor Acquisition One, Llc Content presentation in head worn computing
US10139635B2 (en) 2014-06-09 2018-11-27 Osterhout Group, Inc. Content presentation in head worn computing
US11887265B2 (en) 2014-06-09 2024-01-30 Mentor Acquisition One, Llc Content presentation in head worn computing
US11360318B2 (en) 2014-06-09 2022-06-14 Mentor Acquisition One, Llc Content presentation in head worn computing
US11294180B2 (en) 2014-06-17 2022-04-05 Mentor Acquisition One, Llc External user interface for head worn computing
US10698212B2 (en) 2014-06-17 2020-06-30 Mentor Acquisition One, Llc External user interface for head worn computing
US9810906B2 (en) 2014-06-17 2017-11-07 Osterhout Group, Inc. External user interface for head worn computing
US11789267B2 (en) 2014-06-17 2023-10-17 Mentor Acquisition One, Llc External user interface for head worn computing
US11054645B2 (en) 2014-06-17 2021-07-06 Mentor Acquisition One, Llc External user interface for head worn computing
US11269182B2 (en) 2014-07-15 2022-03-08 Mentor Acquisition One, Llc Content presentation in head worn computing
US11786105B2 (en) 2014-07-15 2023-10-17 Mentor Acquisition One, Llc Content presentation in head worn computing
US11103122B2 (en) 2014-07-15 2021-08-31 Mentor Acquisition One, Llc Content presentation in head worn computing
US11630315B2 (en) 2014-08-12 2023-04-18 Mentor Acquisition One, Llc Measuring content brightness in head worn computing
US10908422B2 (en) 2014-08-12 2021-02-02 Mentor Acquisition One, Llc Measuring content brightness in head worn computing
US11360314B2 (en) 2014-08-12 2022-06-14 Mentor Acquisition One, Llc Measuring content brightness in head worn computing
US9829707B2 (en) 2014-08-12 2017-11-28 Osterhout Group, Inc. Measuring content brightness in head worn computing
US9696551B2 (en) * 2014-08-13 2017-07-04 Beijing Lenovo Software Ltd. Information processing method and electronic device
US20160048024A1 (en) * 2014-08-13 2016-02-18 Beijing Lenovo Software Ltd. Information processing method and electronic device
US9671613B2 (en) 2014-09-26 2017-06-06 Osterhout Group, Inc. See-through computer display systems
US10684687B2 (en) 2014-12-03 2020-06-16 Mentor Acquisition One, Llc See-through computer display systems
US11262846B2 (en) 2014-12-03 2022-03-01 Mentor Acquisition One, Llc See-through computer display systems
US9684172B2 (en) 2014-12-03 2017-06-20 Osterhout Group, Inc. Head worn computer display systems
US11809628B2 (en) 2014-12-03 2023-11-07 Mentor Acquisition One, Llc See-through computer display systems
CN105785570B (en) * 2014-12-26 2019-06-25 比亚迪股份有限公司 A kind of vehicle-mounted head-up-display system and the vehicle comprising the system
CN105785570A (en) * 2014-12-26 2016-07-20 比亚迪股份有限公司 On-board head-up display system and vehicle comprising the same
USD792400S1 (en) 2014-12-31 2017-07-18 Osterhout Group, Inc. Computer glasses
USD794637S1 (en) 2015-01-05 2017-08-15 Osterhout Group, Inc. Air mouse
US10062182B2 (en) 2015-02-17 2018-08-28 Osterhout Group, Inc. See-through computer display systems
US10444018B2 (en) 2015-02-27 2019-10-15 Microsoft Technology Licensing, Llc Computer-implemented method to test the sensitivity of a sensor for detecting movement of a tracking device within an established frame of reference of a moving platform
US10111620B2 (en) 2015-02-27 2018-10-30 Microsoft Technology Licensing, Llc Enhanced motion tracking using transportable inertial sensors to determine that a frame of reference is established
US10139633B2 (en) * 2015-05-28 2018-11-27 Thalmic Labs Inc. Eyebox expansion and exit pupil replication in wearable heads-up display having integrated eye tracking and laser projection
WO2016191709A1 (en) * 2015-05-28 2016-12-01 Thalmic Labs Inc. Systems, devices, and methods that integrate eye tracking and scanning laser projection in wearable heads-up displays
CN107710048A (en) * 2015-05-28 2018-02-16 赛尔米克实验室公司 The system, apparatus and method of eye tracks and scanning laser projection are integrated in wearable head-up display
US10783835B2 (en) * 2016-03-11 2020-09-22 Lenovo (Singapore) Pte. Ltd. Automatic control of display brightness
EP3318914A1 (en) * 2016-11-04 2018-05-09 Nokia Technologies OY Gaze tracking
TWI667495B (en) * 2017-03-13 2019-08-01 宏達國際電子股份有限公司 Head mounted display device, object tracking apparatus and method for tracking object thereof
CN107765435A (en) * 2017-11-22 2018-03-06 深圳创维新世界科技有限公司 Head-wearing display device
CN111357008B (en) * 2017-11-28 2023-09-12 指纹卡安娜卡敦知识产权有限公司 Biometric imaging system and method for controlling the same
CN111357008A (en) * 2017-11-28 2020-06-30 指纹卡有限公司 Biometric imaging system and method for controlling the same
US11391954B2 (en) 2018-11-08 2022-07-19 Letinar Co., Ltd. Optical device for augmented reality
WO2020096187A1 (en) * 2018-11-08 2020-05-14 주식회사 레티널 Optical device for augmented reality
US11188175B1 (en) 2018-12-24 2021-11-30 Facebook Technologies, Llc Display system with integrated depth detection
US10963103B1 (en) * 2018-12-24 2021-03-30 Facebook Technologies, Llc Display system with integrated depth detection
CN113396356A (en) * 2019-02-22 2021-09-14 株式会社籁天那 Compact enhanced in-situ optical device
CN113396356B (en) * 2019-02-22 2023-09-15 株式会社籁天那 Compact enhanced on-site practical optical device
WO2020223228A1 (en) * 2019-04-30 2020-11-05 Intuitive Surgical Operations, Inc. Image viewing systems and methods using a black glass mirror
US20220252875A1 (en) * 2019-05-29 2022-08-11 Facebook Technologies, Llc Dual purkinje imaging with ellipsoidal lensing structure
US11953679B2 (en) * 2019-05-29 2024-04-09 Meta Platforms Technologies, Llc Dual Purkinje imaging with ellipsoidal lensing structure
CN111338085B (en) * 2020-04-21 2023-10-03 Oppo广东移动通信有限公司 Display optical system, head-mounted display device, control method, and storage medium
CN111338085A (en) * 2020-04-21 2020-06-26 Oppo广东移动通信有限公司 Display optical system, head-mounted display apparatus, control method, and storage medium

Similar Documents

Publication Publication Date Title
US8767306B1 (en) Display system
US8970452B2 (en) Imaging method
US8611015B2 (en) User interface
US10055642B2 (en) Staredown to produce changes in information density and type
US8982471B1 (en) HMD image source as dual-purpose projector/near-eye display
US9213185B1 (en) Display scaling based on movement of a head-mounted display
US10031579B2 (en) Automatic calibration for reflective lens
US8955973B2 (en) Method and system for input detection using structured light projection
US20130088413A1 (en) Method to Autofocus on Near-Eye Display
US8971570B1 (en) Dual LED usage for glint detection
US20150153572A1 (en) Adjustment of Location of Superimposed Image
US10488661B2 (en) Systems, devices, and methods that integrate eye tracking and scanning laser projection in wearable heads-up displays
US8866702B1 (en) Use of optical display system as a visual indicator for a wearable computing device
US9007301B1 (en) User interface
US20130241805A1 (en) Using Convergence Angle to Select Among Different UI Elements
US9261959B1 (en) Input detection
US8823740B1 (en) Display system
US20220382064A1 (en) Metalens for use in an eye-tracking system of a mixed-reality display device
US20170090557A1 (en) Systems and Devices for Implementing a Side-Mounted Optical Sensor
US20170261750A1 (en) Co-Aligned Retinal Imaging And Display System
WO2016101861A1 (en) Head-worn display device
Peddie et al. Technology issues
US20230087172A1 (en) Helmet projector system for virtual display
KR20240030881A (en) Method for outputting a virtual content and an electronic device supporting the same

Legal Events

Date Code Title Description
AS Assignment

Owner name: GOOGLE INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MIAO, XIAOYU;WONG, ADRIAN;AMIRPARVIZ, BABAK;AND OTHERS;SIGNING DATES FROM 20110906 TO 20110921;REEL/FRAME:026951/0845

STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: GOOGLE LLC, CALIFORNIA

Free format text: CHANGE OF NAME;ASSIGNOR:GOOGLE INC.;REEL/FRAME:044277/0001

Effective date: 20170929

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551)

Year of fee payment: 4

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8