US20130027772A1 - Variable-depth stereoscopic display - Google Patents
Variable-depth stereoscopic display Download PDFInfo
- Publication number
- US20130027772A1 US20130027772A1 US13/192,321 US201113192321A US2013027772A1 US 20130027772 A1 US20130027772 A1 US 20130027772A1 US 201113192321 A US201113192321 A US 201113192321A US 2013027772 A1 US2013027772 A1 US 2013027772A1
- Authority
- US
- United States
- Prior art keywords
- light
- viewer
- distance
- lens structure
- display
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/0075—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for altering, e.g. increasing, the depth of field or depth of focus
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B30/00—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images
- G02B30/20—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes
- G02B30/26—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes of the autostereoscopic type
- G02B30/27—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes of the autostereoscopic type involving lenticular arrays
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B30/00—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images
- G02B30/20—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes
- G02B30/26—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes of the autostereoscopic type
- G02B30/27—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes of the autostereoscopic type involving lenticular arrays
- G02B30/28—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes of the autostereoscopic type involving lenticular arrays involving active lenticular arrays
-
- G—PHYSICS
- G02—OPTICS
- G02F—OPTICAL DEVICES OR ARRANGEMENTS FOR THE CONTROL OF LIGHT BY MODIFICATION OF THE OPTICAL PROPERTIES OF THE MEDIA OF THE ELEMENTS INVOLVED THEREIN; NON-LINEAR OPTICS; FREQUENCY-CHANGING OF LIGHT; OPTICAL LOGIC ELEMENTS; OPTICAL ANALOGUE/DIGITAL CONVERTERS
- G02F1/00—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics
- G02F1/01—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the intensity, phase, polarisation or colour
- G02F1/13—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the intensity, phase, polarisation or colour based on liquid crystals, e.g. single liquid crystal display cells
- G02F1/133—Constructional arrangements; Operation of liquid crystal cells; Circuit arrangements
- G02F1/1333—Constructional arrangements; Manufacturing methods
- G02F1/1335—Structural association of cells with optical devices, e.g. polarisers or reflectors
- G02F1/1336—Illuminating devices
- G02F1/133615—Edge-illuminating devices, i.e. illuminating from the side
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G3/00—Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes
- G09G3/20—Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes for presentation of an assembly of a number of characters, e.g. a page, by composing the assembly by combination of individual elements arranged in a matrix no fixed position being assigned to or needed to be assigned to the individual characters or partial characters
- G09G3/34—Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes for presentation of an assembly of a number of characters, e.g. a page, by composing the assembly by combination of individual elements arranged in a matrix no fixed position being assigned to or needed to be assigned to the individual characters or partial characters by control of light from an independent source
- G09G3/3406—Control of illumination source
Definitions
- 3D content providers are increasingly offering content in three-dimensional (3D) format to enhance a viewing experience.
- This 3D content is often shown on specialized 3D displays that split the 3D content into different images targeted to each eye of a viewer.
- Different imagery can be displayed to each eye of a viewer in a variety of ways including the use of passive eyewear, active eyewear, dual head-mounted displays, or dividing the image directionally without the use of eyewear (auto-stereoscopy).
- auto-stereoscopy does not require the use of specialized glasses or multiple displays, a distance at which the divided images are effective to produce 3D imagery is generally limited.
- stereoscopic displays are unable to display 3D imagery to a viewer that moves out of a focal plane of the display or to multiple viewers at varying distances from the display.
- This document describes various apparatuses and techniques for implementing a variable-depth stereoscopic display. These apparatuses and techniques may enable a stereoscopic display to provide three-dimensional (3D) content to a viewer over different distances from the stereoscopic display or to multiple viewers at varying distances from the stereoscopic display.
- a first distance at which a viewer is disposed relative to a stereoscopic display can be received.
- a second distance by which to change a front focal distance of a lens structure of the stereoscopic display is determined based on the first distance.
- the front focal distance of the lens structure is then caused to change by the second distance effective to display a stereoscopic image at the first distance.
- a front focal distance of the stereoscopic display can be varied to provide 3D content to a viewer at different distances from the stereoscopic display.
- FIG. 1 illustrates an example environment in which these techniques may be implemented.
- FIG. 2 illustrates an example display of the display device of FIG. 1 in more detail.
- FIG. 3 illustrates another example display of the display device of FIG. 1 in more detail.
- FIG. 4 illustrates an example of scanning light with components of the display of FIG. 2 .
- FIG. 5 illustrates another example of scanning light with components of the display of FIG. 2 .
- FIG. 6 illustrates a detailed example of the light injection system of FIG. 2 .
- FIG. 7 illustrates an example light guide configuration of the display of FIG. 2 and ray traces associated with a light source.
- FIG. 8 illustrates a plan view of example ray traces associated with the light guide configuration and the light source of FIG. 7 .
- FIG. 9 illustrates an example light guide configuration of the display of FIG. 2 and ray traces associated with another light source.
- FIG. 10 illustrates a plan view of example ray traces associated with the light guide configuration and the other light source of FIG. 9 .
- FIG. 11 illustrates a plan view of example ray traces associated with the liquid crystal lens of FIG. 3 .
- FIG. 12 illustrates a plan view of example of ray traces of re-focused light associated with the liquid crystal lens of FIG. 3 .
- FIG. 13 is a flow diagram depicting an example method for causing a front focal distance of a stereoscopic display to change.
- FIG. 14 illustrates an example viewing environment in which various embodiments of these apparatuses and techniques may be employed.
- FIG. 15 is a flow diagram depicting an example method for selecting a light source to change a focal plane of a stereoscopic display.
- FIG. 16 illustrates the example viewing environment of FIG. 14 in which a viewer has moved out of a focal plane of a stereoscopic display.
- FIG. 17 is a flow diagram depicting an example method for selecting a light source based on a distance to a viewer.
- FIG. 18 illustrates an example device in which techniques for enabling a variable-depth stereoscopic display can be implemented.
- This document describes various apparatuses and techniques for implementing a variable-depth stereoscopic display.
- Various embodiments of these techniques change a front focal distance of a stereoscopic display, which allows the stereoscopic display to provide 3D content to a viewer over different distances from the stereoscopic display or to multiple viewers at varying distances from the stereoscopic display.
- 3D content e.g., a movie or video-game
- both viewers are positioned at approximately the same distance from the television device, such as sitting at a couch roughly parallel with the television device, and are able to view the 3D content.
- one of the viewers would like to move to a position closer to or farther from to the television device (e.g., a chair located closer to the television device).
- the described apparatuses and techniques enable the 3D content to be provided to both viewers at different distances from the television device or to a viewer that moves closer to or farther from the television device.
- FIG. 1 is an illustration of an example environment 100 in which a variable-depth stereoscopic display can be implemented.
- Environment 100 includes display device 102 which is illustrated, by way of example and not limitation, as one of a smart phone 104 , laptop computer 106 , television device 108 , desktop computer 110 , or tablet computer 112 .
- display device 102 can provide three-dimensional (3D) content to viewers without the use of special 3D eyewear.
- 3D content may comprise images (e.g., stereoscopic imagery) and/or video effective to cause a viewer to be able to perceive depth within the content when displayed.
- the 3D content is manipulable by a viewer or user, such as when displayed in connection with an application or user interface of a display device capable of accepting user input (e.g., a touch-screen enabled smart phone 104 or television device 108 operably associated with gaming controllers or optical-spatial sensors).
- a display device capable of accepting user input (e.g., a touch-screen enabled smart phone 104 or television device 108 operably associated with gaming controllers or optical-spatial sensors).
- Display device 102 includes processor(s) 114 and computer-readable media 116 , which includes memory media 118 and storage media 120 .
- Applications and/or an operating system (not shown) embodied as computer-readable instructions on computer-readable memory 116 can be executed by processor(s) 114 to provide some or all of the functionalities described herein.
- Computer-readable media also includes stereoscopic-depth controller (controller) 122 . How controller 122 is implemented and used varies, and is described as part of the methods discussed below.
- Display device 102 also includes stereoscopic display 124 , sensor 126 , input/output (I/O) ports 128 , and network interface(s) 130 .
- Stereoscopic display 124 is capable of generating stereoscopic 3D content that can be viewed without the use of special eyewear.
- Stereoscopic display 124 may be separate or integral with display device 102 ; integral examples include smart phone 104 , laptop 106 , and tablet 112 ; separate examples include television device 108 and, in some instances, desktop computer 110 (e.g., when embodied as a separate tower and monitor (shown)).
- Sensor 126 collects viewer positional data useful to determine a position of a viewer relative stereoscopic display 124 .
- the viewer positional data can be useful determine a distance of a viewer from stereoscopic display 124 (relative Z position), a distance of a viewer from a horizontal axis of stereoscopic display 124 (relative Y position), or a distance of a viewer from a vertical axis of stereoscopic display 124 (relative X position).
- an approximate position of a viewer's eyes can be determined with the viewer positional data provided by sensor 126 .
- Sensor 126 may be separate or integral with display device 102 ; integral examples include sensor 126 - 1 of television device 108 and sensor 126 - 2 of tablet computer 112 ; separate examples include stand-alone sensors, such as sensors operably coupled with display device 102 , a set-top box, or a gaming device.
- Sensor 126 can collect viewer position data by way of various sensing technologies, either working alone or in conjunction with one another. Sensing technologies may include, by way of example and not limitation, optical, radio-frequency, acoustic (active or passive), micro-electro-mechanical systems (MEMS), ultrasonic, infrared, pressure sensitive, and the like. In some embodiments, sensor 126 may receive additional data or work in conjunction with a remote control device or gaming controller associated with one or more viewers to generate the viewer positional data.
- Sensing technologies may include, by way of example and not limitation, optical, radio-frequency, acoustic (active or passive), micro-electro-mechanical systems (MEMS), ultrasonic, infrared, pressure sensitive, and the like.
- MEMS micro-electro-mechanical systems
- ultrasonic ultrasonic, infrared, pressure sensitive, and the like.
- sensor 126 may receive additional data or work in conjunction with a remote control device or gaming controller associated with one or more viewers to generate the viewer position
- I/O ports 128 of display device 102 enable interaction with other devices, media, or users.
- I/O ports 128 can include a variety of ports, such as by way of example and not limitation, high-definition multimedia (HDMI), digital video interface (DVI), display port, fiber-optic or light-based, audio ports (e.g., analog, optical, or digital), USB ports, serial advanced technology attachment (SATA) ports, peripheral component interconnect (PCI) express based ports or card slots, serial ports, parallel ports, or other legacy ports.
- HDMI high-definition multimedia
- DVI digital video interface
- display port fiber-optic or light-based
- audio ports e.g., analog, optical, or digital
- USB ports e.g., serial advanced technology attachment (SATA) ports
- PCI peripheral component interconnect express based ports or card slots
- serial ports parallel ports, or other legacy ports.
- 3D content is received by display device 102 via one or more I/O ports 128 from another device or source (e.g.
- Display device 102 may also include network interface(s) 130 for communicating data over wired, wireless, or optical networks. Data communicated over such networks may include 3D content that can be displayed or interacted with via stereoscopic display 124 .
- network interface 130 may communicate data over a local-area-network (LAN), a wireless local-area-network (WLAN), a personal-area-network (PAN), a wide-area-network (WAN), an intranet, the Internet, a peer-to-peer network, point-to-point network, a mesh network, and the like.
- LAN local-area-network
- WLAN wireless local-area-network
- PAN personal-area-network
- WAN wide-area-network
- intranet the Internet
- peer-to-peer network point-to-point network
- mesh network and the like.
- FIG. 2 illustrates a detailed example of stereoscopic display 124 of FIG. 1 .
- Stereoscopic display 124 includes lens structure 202 , light injection system 204 , light re-director 206 , and spatial light modulator 208 .
- Stereoscopic display 124 may be configured as a non-projection based flat panel display having a depth or thickness similar to that of a liquid crystal display (LCD) panel and the like.
- Lens structure 202 emits light from a surface when light is received from light injection system 204 .
- the light emitted from lens structure may be collimated light.
- lens structure 202 is an optical wedge having a thin end 210 to receive light, a thick end 212 effective to reflect the light (e.g., via an end reflector or reflective cladding), and a viewing surface 214 at which the light is emitted as collimated light.
- an optical wedge may comprise an optical lens or light guide that permits light input at an edge of the optical wedge (e.g., thin end 210 ) to fan out within the optical wedge via total internal reflection before reaching the critical angle for internal reflection and exiting via another surface of the optical wedge (e.g., viewing surface 214 ). The light may exit the optical wedge at a glancing angle relative to viewing surface 214 .
- the light emitted by lens structure 202 can be scanned by varying light generated by light injection system 204 or an injection location thereof. Generally, scanning the light enables the display of 3D content that is viewable without the use of special eyewear. The scanned light enables display of different stereoscopic imagery to each eye of a respective viewer.
- Light injection system 204 may include any suitable type or number of light sources, such as by way of example and not limitation, cold-cathode fluorescent lamps (CCFL), light emitting diodes (LEDs), light engines, lasers (chemical or solid-state), and the like. How the light of light injection source 204 is generated and controlled varies, and is described in conjunction with subsequent figures and methods.
- CCFL cold-cathode fluorescent lamps
- LEDs light emitting diodes
- lasers chemical or solid-state
- stereoscopic display 124 includes light re-director 206 located adjacent to viewing surface 214 to diffuse collimated light emitted by lens structure 202 .
- This collimated light may exit lens structure 202 at a glancing angle with respect to viewing surface 214 .
- light re-director 206 can re-direct the emitted light of lens structure 202 towards viewer 216 and may provide a diffusing function in one dimension.
- Light re-director 206 can be configured as any suitable structure, such as a film of prisms or a light-guide panel having a prismatic textured surface.
- light re-director 206 can be configured to have optical power capable of focusing the collimated light emitted from viewing surface 214 of lens structure 202 towards the eyes of viewer 216 .
- Spatial light modulator 208 modulates the light with visual information to form imagery displayed by the light converging on the eyes of viewer 216 .
- the visual information is parallax information directed to different eyes of viewer 216 in order to provide the 3D content.
- spatial light modulator 208 can modulate light directed towards a viewer's left eye with a frame of stereoscopic imagery, and then modulate light directed to a viewer's right eye with another frame of stereoscopic imagery.
- 3D content can be provided to a viewer.
- Spatial light modulator 208 can be any suitable structure such as a liquid crystal display (LCD) cell composed of an array of pixels or sub-pixels capable of light modulation and/or manipulation.
- LCD liquid crystal display
- controller 122 is operably coupled to light injection system 204 and sensor 126 .
- controller 122 is operably coupled with spatial light modulator 208 or a modulation-controller associated therewith.
- Controller 122 receives viewer position information, such as a distance to a viewer, collected by sensor 126 and, as described in the following sections and associated methods, can control light injection source 204 effective to display 3D imagery via stereoscopic display 124 over various distances.
- FIG. 3 illustrates another detailed example of stereoscopic display 124 of FIG. 1 .
- Stereoscopic display 124 includes lens structure 202 , light injection system 204 , liquid crystal lens (LQ lens) 302 , and spatial light modulator 208 .
- Lens structure 202 emits light (collimated or otherwise) from a surface when light is received from light injection system 204 .
- lens structure 202 is an optical wedge having a thin end 210 to receive light, a thick end 212 effective to reflect the light, and a viewing surface 214 at which the light is emitted as collimated light.
- Display 124 can be configured as a non-projection based flat panel display having a depth or thickness similar to that of a liquid crystal display (LCD) panel and the like.
- LCD liquid crystal display
- the light emitted by lens structure 202 can be scanned by varying light generated by light injection system 204 or an injection location thereof. Generally, scanning the light enables the display of 3D content that is viewable without the use of special eyewear. The scanned light enables display of different stereoscopic imagery to each eye of a respective viewer.
- Light injection system 204 may include any suitable type or number of light sources, such as by way of example and not limitation, cold-cathode fluorescent lamps (CCFL), light emitting diodes (LEDs), light engines, lasers (chemical or solid-state), and the like. How the light of light injection source 204 is generated and controlled varies, and is described in conjunction with subsequent figures and methods.
- CCFL cold-cathode fluorescent lamps
- LEDs light emitting diodes
- lasers chemical or solid-state
- Stereoscopic display 124 may also include light re-director 206 located adjacent to viewing surface 214 to diffuse collimated light emitted by lens structure 202 . This collimated light may exit lens structure 202 at a glancing angle with respect to viewing surface 214 .
- light re-director 206 can re-direct the emitted light of lens structure 202 towards viewer 216 and may provide a diffusing function in one dimension.
- Light re-director 206 can be configured as any suitable structure, such as a film of prisms or a light-guide panel having a prismatic textured surface.
- light re-director 206 can be configured to have optical power capable of focusing the collimated light emitted from viewing surface 214 of lens structure 202 towards the eyes of viewer 216 .
- stereoscopic display 124 includes LQ lens 302 located adjacent to light re-director 206 capable of varying a depth at which stereoscopic imagery is provided.
- LQ lens 302 located adjacent to light re-director 206 capable of varying a depth at which stereoscopic imagery is provided.
- a focal depth (or focal length) of lens structure 202 can be varied by focusing or re-focusing LQ lens 302 .
- Affecting or manipulating liquid crystal molecules of LQ lens 302 may focus or re-focus light emitted by lens structure 202 .
- LQ lens 302 may comprise two transparent substrates, such as glass or other transparent material, between which these liquid crystal molecules are disposed.
- LQ lens 302 also includes electrodes 304 and 306 for applying excitation voltage to LQ lens 302 .
- these electrodes may be composed of a transparent material such as indium tin oxide (ITO) so as not to compromise clarity or transparency of LQ lens 302 .
- ITO indium tin oxide
- LQ lens 302 may include any suitable number of electrodes proximate the perimeter of the lens to apply excitation voltage between any two peripheral points of LQ lens 302 .
- one or both of the transparent substrates can be coated, either in part or entirely, with transparent conductive coating, such as ITO and the like.
- Applying excitation voltage to LQ lens 302 can re-align the liquid crystal molecules effective to re-focus LQ lens 302 at various distances.
- other layers or substrates located proximate the liquid crystal molecules of LQ lens 302 may affect a uniformity of a distortion or alignment of the liquid crystal molecules when under the influence of an electrical excitation field. In some cases, affect enables LQ lens 302 to focus light emitted from lens structure 202 at various distances.
- varying a focal distance of lens structure 202 enables implementation of a variable depth stereoscopic display.
- LQ lens 302 may be used in place of (e.g., instead of light guides) or in conjunction with any other elements described herein in order to implement a stereoscopic display.
- Spatial light modulator 208 modulates the light exiting LQ lens 302 with visual information to form imagery displayed by the light converging on the eyes of viewer 216 .
- the visual information is parallax information directed to different eyes of viewer 216 in order to provide the 3D content.
- spatial light modulator 208 can modulate light directed towards a viewer's left eye with a frame of stereoscopic imagery, and then modulate light directed to a viewer's right eye with another frame of stereoscopic imagery.
- 3D content can be provided to a viewer.
- Spatial light modulator 208 can be any suitable structure such as a liquid crystal display (LCD) cell composed of an array of pixels or sub-pixels capable of light modulation and/or manipulation.
- LCD liquid crystal display
- controller 122 is operably coupled to light injection system 204 , LQ lens 302 , and sensor 126 .
- Controller 122 receives viewer position information, such as a distance to a viewer, collected by sensor 126 and, as described in the following sections and associated methods, can control light injection source 204 or LQ lens 302 effective to display 3D imagery via stereoscopic display 124 over various distances.
- FIGS. 4 and 5 illustrate an example of scanning light by varying a location at which light is injected into lens structure 202 at 400 and 500 respectively.
- scanning light is performed by directing the light horizontally across a viewing plane or area as shown in FIGS. 4 and 5 .
- 3D imagery can be displayed to a viewer with scanned light that directs different images to each of the viewer's eyes.
- 3D video content split into imagery intended for a viewer's right eye and a viewer's left eye, when interleaved and displayed at sufficient frequency, enables the viewer to experience the 3D video without special eyewear.
- the light can be directed to the left ( 300 ) by shifting a location of light injection to the right, and vice versa ( 400 ) as illustrated by ray traces within lens structure 202 in each respective figure. Additionally, in each figure, a visible portion of a single pixel of light, shown respectively at 402 and 502 , is illustrated for clarity. Illustrating perspective, lines tracing a path of the light from the corners of lens structure 202 to the points of light 402 and 502 are shown respectively with respect to center line 404 and focal plane 406 .
- a location at which light is injected can be varied in any suitable way.
- a light injection system 204 may comprise LEDs that are individually selectable to provide light at different locations along the lens structure effective to scan light emitted by the lens structure. In this particular example, selecting one or more LEDs proximate one end of the lens structure (right-hand side) directs light in an opposite direction (left-hand side), as illustrated by 400 .
- light injection system 204 may include a laser and an acousto-optic modulator or liquid crystal hologram for controlling directionality of a beam produced by the laser.
- FIG. 6 illustrates a detailed example of light injection system 204 of FIG. 2 in more detail.
- light injection system 204 includes light sources 602 , 604 , and 606 located at different distances from lens structure 202 . Each light source located at a different distance from a light receiving surface of lens structure 202 has a different effective back focal length. Although shown as three light sources, light injection system 204 may include any suitable number or configuration of light sources.
- Light sources 602 , 604 , and 606 include individually selectable light generating elements (elements) 608 , such as LEDs arranged in a roughly linear configuration along a length of the light source (e.g., LED bars or strips). Elements 608 can be selected such that light sources 602 , 604 , or 606 provide light at different locations of a light receiving surface of lens structure 202 to scan light as described above.
- elements 608 such as LEDs arranged in a roughly linear configuration along a length of the light source (e.g., LED bars or strips).
- Elements 608 can be selected such that light sources 602 , 604 , or 606 provide light at different locations of a light receiving surface of lens structure 202 to scan light as described above.
- Light sources 602 , 604 , and 606 are associated with light guides 610 , 612 , and 614 respectively.
- light guides 610 , 612 , and 614 transmit light from a respective light source to a light receiving surface of lens structure 202 , such as then end 210 of an optical wedge.
- light guides 610 , 612 , and 614 are useful to fold down back focal lengths of lens structure 202 .
- any suitable optical structure may be associated with light sources 602 , 604 , or 606 to transmit light thereof, such as light pipes, prisms, films, and the like.
- Light guides of different lengths can provide different optical path lengths for light transmitted through the light guide (or similar structure).
- one or more of the light sources may not be associated with a light guide and/or located adjacent to lens structure 202 (not shown).
- light guides 610 , 612 , and 614 can be configured to prevent light sources 602 , 604 , and 606 from occluding one another and/or to reduce light pollution among the light sources. Although shown as having plate-like geometry, light guides 610 , 612 , and 614 may be configured in any suitable geometry. Additionally, light guides 610 , 612 , and 614 may comprise any suitable optically transmissive material, such as plastic, acrylic, glass, and the like. In some cases, a geometry or material composition of a light guide effects an optical path length associated with the light guide.
- FIGS. 7 and 8 illustrate an example light pattern associated with a light guide of FIG. 6 .
- a light guide associated with a respective light source will transmit light from that respective light source.
- FIG. 7 which illustrates a side view of light injection system 204
- light 702 is generated by light source 602 and transmitted by light guide 610 .
- light 702 is intended to direct light towards a viewer's right eye, capable of delivering a frame of stereoscopic imagery to the viewer's right eye.
- FIG. 8 illustrates a plan view of example light patterns associated with the light guide of FIG. 7 .
- light 802 is intended to direct light towards a viewer's left eye, capable delivering a frame of stereoscopic imagery to the viewer's left eye.
- Light 802 is also generated by light source 602 ; however, light 802 is generated at a different location by a set of elements 608 that is different from a set that generated light 702 .
- elements 608 located proximate an end of light source 602 generate light 802
- elements 608 proximate the other end of light source 602 generate light 702 .
- light 702 and light 802 are not generated concurrently, but sequentially to direct light in different directions.
- Lens structure 202 receives light 702 and 802 and emits light 804 and 806 respectively.
- Spatial light modulator 208 modulates light 804 and 806 with stereoscopic information intended for a viewer's right eye and left eye respectively.
- the stereoscopic imagery can be displayed to a viewer at focal plane 808 where the light is appropriately directed to each eye of the viewer.
- Focal plane 808 can be any suitable distance from lens structure 202 and may vary depending on a display device 102 in which stereoscopic display 124 is implemented.
- focal plane 808 may be approximately 500-700 mm for laptop device 106 or approximately 3-5 m for television device 108 .
- a depth of the focal plane at which the stereoscopic imagery can be viewed may vary, ranging for example from 50 mm to 300 mm.
- FIGS. 9 and 10 illustrate another example light pattern associated with another light guide of FIG. 6 .
- light 902 is generated by light source 606 and transmitted by light guide 614 .
- light 902 is intended to direct light towards a viewer's right eye, capable delivering a frame of stereoscopic imagery to the viewer's right eye.
- light guide 614 is longer than light guide 610 by distance 904 , thus selecting light source 606 changes a back focal length associated with light 902 .
- FIG. 10 illustrates a plan view of example light patterns associated with the light guide of FIG. 9 .
- light 1002 is intended to direct light towards a viewer's left eye, capable delivering a frame of stereoscopic imagery to the viewer's left eye.
- Light 1002 is also generated by light source 606 ; however, light 1002 is generated at a different location by a set of elements 608 , which is different from a set that generated light 902 .
- elements 608 located proximate an end of light source 606 generate light 1002
- elements 608 proximate the other end of light source 606 generate light 902 .
- over-lapping in FIG.
- light guide 614 is a different length than the other light guides and configured such that the other light guides do not occlude light generated by light source 606 .
- Lens structure 202 receives light 902 and 1002 and emits light 1006 and 1008 , respectively.
- Spatial light modulator 208 modulates light 1006 and 1008 with stereoscopic information intended for a viewer's right eye and left eye respectively.
- the stereoscopic imagery is displayed to a viewer at focal plane 1010 , where the light is appropriately directed to each eye of the viewer.
- focal plane 1010 moves towards lens structure 202 when another light source is selected to inject light into lens structure 202 .
- a front focal distance of lens structure 202 at which the focal planes are located, can be varied by any suitable distance when different light sources are selected.
- a light source has a back focal length that is different from those associated with other light sources. Selecting another light source having a different back focal length can change a distance at which the focal plane is located from a stereoscopic display 124 .
- the distance varies based on the type of device having the stereoscopic display. For example, a distance between focal plane 1010 and focal plane 808 can be approximately 100-150 mm for laptop device 106 or approximately 0.5-1.0 m for a television device.
- the light sources may also be configured such that the focal planes of a stereoscopic display overlap, allowing stereoscopic imagery to be delivered to a moving viewer and/or to eliminate gaps between focal planes.
- FIGS. 11 and 12 illustrate a plan view of example ray traces associated with the liquid crystal lens of FIG. 3 .
- LQ lens 302 is capable of changing a front focal distance of lens structure 202 by re-focusing light emitted by lens structure 202 .
- light 1102 and 1104 are generated by a light source (not show) and transmitted to lens structure by light guide 610 .
- light 1102 and 1104 are not generated concurrently, but sequentially to direct light in different directions. Assume here that light 1102 and 1104 are intended to direct light towards a viewer's right and left eyes respectively.
- Lens structure 202 receives light 1102 and 1104 and emits light 1106 and 1108 respectively. As the light passes through LQ lens 302 , it is focused to converge on the viewers eyes at focal plane 1110 . Additionally or alternately, spatial light modulator 208 modulates light 1106 and 1108 with stereoscopic information intended for a viewer's right eye and left eye respectively.
- Focal plane 1110 can be any suitable distance from lens structure 202 and may vary depending on a display device 102 in which stereoscopic display 124 is implemented, examples of which are discussed above.
- FIG. 12 illustrates a plan view of an example of re-focused ray traces associated with the liquid crystal lens of FIG. 3 .
- light 1102 and 1104 are the same light shown in FIG. 11 , although other light may be generated, such as when a viewer's relative Y position has changed.
- lens structure 202 receives light 1102 and 1104 from light guide 610 and emits light 1202 and 1204 respectively.
- light 1202 and 1204 are intended to direct light towards a viewer's right and left eyes respectively.
- light 1204 and 1204 converge on a viewer's eyes at a focal plane.
- LQ lens 302 re-focuses and changes a focal distance of lens structure 202 as shown by focal plane 1206 .
- focal plane 1206 moves towards lens structure 202 when LQ lens 302 is re-focused.
- a front focal distance of lens structure 202 can be increased or decreased as needed by refocusing LQ lens 302 to implement a variable depth stereoscopic display.
- FIG. 13 is flow diagram depicting an example method 1300 for causing a front focal distance of a stereoscopic display to change.
- Block 1302 receives a first distance at which a viewer is disposed relative a stereoscopic display. In some cases, block 1302 receives a distance to a viewer that has moved towards or away from the stereoscopic display. A distance to the viewer may be received in the form of viewer positional data or viewer-tracking data.
- the viewer positional data can be collected by any suitable sensing technology, such as optical or imaging sensors, examples of which are described above.
- Viewing environment 1400 includes television device 108 and viewers 1402 and 1404 located at viewing plane 1406 .
- viewing plane 1406 is located at distance 1408 from television device 108 .
- gaming device 1410 is operably coupled with television device 108 and sensor 126 (shown as an external sensor) that collects positional data of viewers 1402 and 1404 .
- controller 122 (not shown) of television device 108 receives a first distance 1408 from sensor 126 via gaming device 1410 .
- controller receives viewer positional data from sensor 126 .
- Controller 122 analyzes the viewer positional data to determine distance 1408 to viewer 1402 .
- Block 1304 determines a second distance by which to change a front focal distance of the stereoscopic display based on the first distance. In some cases, determining the second distance is based on comparing the first distance with a current front focal distance of the stereoscopic display or a lens structure thereof. For example, the second distance may be the difference between the first distance to the viewer and the current front focal distance of the stereoscopic display. In the context of the present example, assume that the current front focal distance of television device 108 is approximately half of distance 1408 to viewer 1402 . Here, controller 122 determines a second distance (half of distance 1408 ) by which to change a front focal distance of television device 108 such that a front focal distance of lens structure 202 (not shown) of television device 108 is approximately distance 1408 .
- Block 1306 causes the front focal distance of the stereoscopic display to change by the second distance.
- Changing the front focal distance of the stereoscopic display by the second distance can be effective to display a stereoscopic image at the first distance.
- the front focal distance is caused to change by the difference between the first distance to the viewer and the current front focal distance of the stereoscopic display.
- the front focal distance of the stereoscopic display, or a lens structure thereof may be caused to change in any suitable way.
- a liquid crystal lens may re-focus light emitted by a lens structure of the stereoscopic display.
- LQ lens 302 may focus or refocus light emitted by lens structure 202 effective to change a front focal distance of stereoscopic display 124 .
- selecting or enabling a light source of a light injection system may cause a front focal distance of the stereoscopic display to change.
- the selected light source that injects the light may be one of multiple light sources that are individually selectable to generate or emit light.
- a back focal length associated with each of the multiple light sources is different. By selecting a light source having a particular back focal length to inject light, a front focal distance of a stereoscopic display can be changed such that a stereoscopic image is visible to viewers located in a focal plane proximate the first distance.
- a light source may be selected to cause the front focal distance to change based on the first distance and a lens structure's front focal distance associated with the light source.
- the first distance to the viewer may correspond approximately to a front focal distance associated with a particular light source.
- controller 122 enables a light source (e.g., light source 604 ) based on the first distance to display stereoscopic imagery at distance 1408 enabling both viewers 1402 and 1404 to view 3D content.
- both viewers 1402 and 1404 are able to view the stereoscopic imagery as they are within the same focal plane of television device 108 .
- controller 122 enables the same light source based on a distances to either viewer 1402 or viewer 1404 that are about the same distance from television device 108 .
- method 1300 may repeat blocks 1302 , 1304 , and 1306 to display a stereoscopic image to a viewer that moves out of a focal plane or to multiple viewers in multiple different focal planes of a stereoscopic display.
- FIG. 15 is a flow diagram depicting an example method for selecting a light source to change a focal plane of a stereoscopic display.
- Block 1502 detects movement of a viewer of a stereoscopic display.
- the stereoscopic display may be displaying 3D content to one or more viewers located within multiple focal planes.
- movement of a viewer is detected by analyzing data collected by an optical or imaging sensor, examples of which are described above. The movement may be detected as a viewer moves towards or away from the stereoscopic display.
- controller 122 detects movement of viewer 1402 as she moves away from television device 108 by analyzing viewer positional data received from sensor 126 .
- Block 1504 determines whether the viewer is within a focal plane of the stereoscopic display.
- a current focal plane or a front focal distance of the stereoscopic display may correspond with a currently selected light source having a particular back focal length.
- a viewer may be determined to be within a focal plane when a distance to the viewer corresponds approximately to that of a front focal distance of the stereoscopic display.
- Distances to one or more viewers may be received or determined by analyzing viewer positional data. In some cases, a viewer may move, yet remain within a focal plane of the stereoscopic display. If block 1504 determines that the viewer is within a current focal plane of the stereoscopic display, method 1500 returns to block 1502 where further movement may be detected. If block 1504 determines that the viewer is not within the current focal plane of the stereoscopic display, method 1500 proceeds to block 1506 .
- a viewer may be determined to be within two overlapping focal planes. In such a case, a further determination is made as to whether the viewer is closer to a previously occupied focal plane or closer to another focal plane towards which the viewer is moving. In the context of the present example, controller 122 determines that viewer 1402 is not within focal plane 1406 and is moving towards another focal plane of television device 108 .
- Block 1506 selects another light source to change a focal length of the stereoscopic display.
- Changing the focal length of the stereoscopic display varies a distance at which the focal plane of the stereoscopic display is located.
- changing a front focal length of the stereoscopic display includes selecting another light source having a back focal length that is different from a previously selected light source.
- method 1500 may return to block 1502 to detect subsequent movement of the viewer or movement of another viewer.
- controller 122 selects another light source (e.g., light source 602 ) operably associated with a lens structure of television device 108 .
- selecting the other light source moves or creates an additional focal plane 1602 at distance 1604 from the original focal plane 1406 .
- television device 108 is able to display stereoscopic imagery to viewer 1402 now located within focal plane 1602 .
- television device 108 can display stereoscopic imagery to one or more viewers as they move out of or between focal planes.
- FIG. 17 is a flow diagram depicting an example method for selecting a light source based on a distance to a viewer.
- Block 1702 determines a distance to a viewer of a stereoscopic display. The distance to the viewer may be determined by analyzing viewer positional data. Viewer positional data associated with the viewer can be collected using any suitable sensing technology, such as optical or imaging sensors, examples of which are described above. In some cases, determining a distance to a viewer may be responsive to detecting movement of the viewer. Additionally or alternately, block 1702 may determine distances to multiple viewers of the stereoscopic display.
- Sensor 126 collects viewer positional data associated with viewer 1402 and transmits this data to controller 122 .
- Controller 122 determines a distance of viewer 1402 from television device 108 by analyzing the viewer positional data received from sensor 126 .
- Block 1704 selects a light source based on the distance from the stereoscopic display to the viewer. Selecting the light source based on the distance enables stereoscopic imagery to be displayed to the viewer.
- the light source may be selected from multiple light sources individually selectable to emit or generate light.
- the selected light source may have an associated back focal length that corresponds to a front focal distance of a lens structure. In some cases, the front focal distance of the lens structure is approximately the distance to the viewer.
- a lens structure collimates the light generated by the selected light source.
- the light is then modulated by a spatial light modulator to display stereoscopic imagery to the viewer.
- controller 122 selects a light source (e.g., light source 602 ) of television device 108 effective to display stereoscopic imagery to viewer 1402 at focal plane 1602 .
- method 1700 may repeat blocks 1702 and 1704 to display stereoscopic imagery to another viewer.
- another light source can be selected based on another distance to the other viewer.
- the other light source may have an associated back focal length that is different than a currently selected light source and corresponds with a different front focal distance of the lens structure. In such cases, this different front focal distance of the lens structure may be approximately the distance to the other viewer.
- controller 122 selects another light source (e.g., light source 604 ) effective to display stereoscopic imagery to viewer 1404 at focal plane 1406 .
- another light source e.g., light source 604
- viewers 1402 and 1404 are able to concurrently view 3D content displayed by television device 108 .
- aspects of these methods may be implemented in hardware (e.g., fixed logic circuitry), firmware, a System-on-Chip (SoC), software, manual processing, or any combination thereof.
- a software implementation represents program code that performs specified tasks when executed by a computer processor, such as software, applications, routines, programs, objects, components, data structures, procedures, modules, functions, and the like.
- the program code can be stored in one or more computer-readable memory devices, both local and/or remote to a computer processor.
- the methods may also be practiced in a distributed computing environment by multiple computing devices.
- FIG. 18 illustrates various components of example device 1800 that can be implemented as any type of client, server, and/or display device as described with reference to the previous FIGS. 1-14 to implement techniques enabling a variable-depth stereoscopic display.
- device 1800 can be implemented as one or a combination of a wired and/or wireless device, as a form of flat panel display, television, television client device (e.g., television set-top box, digital video recorder (DVR), etc.), consumer device, computer device, server device, portable computer device, user device, communication device, video processing and/or rendering device, appliance device, gaming device, electronic device, and/or as another type of device.
- Device 1800 may also be associated with a viewer (e.g., a person or user) and/or an entity that operates the device such that a device describes logical devices that include users, software, firmware, and/or a combination of devices.
- Device 1800 includes communication devices 1802 that enable wired and/or wireless communication of device data 1804 (e.g., received data, data that is being received, data scheduled for broadcast, data packets of the data, etc.).
- the device data 1804 or other device content can include configuration settings of the device, media content stored on the device, and/or information associated with a user of the device.
- Media content stored on device 1800 can include any type of audio, video, and/or image data.
- Device 1800 includes one or more data inputs 1806 via which any type of data, media content, and/or inputs can be received, such as user-selectable inputs, messages, music, television media content, recorded video content, and any other type of audio, video, and/or image data received from any content and/or data source.
- Device 1800 also includes communication interfaces 1808 , which can be implemented as any one or more of a serial and/or parallel interface, a wireless interface, any type of network interface, a modem, and as any other type of communication interface.
- the communication interfaces 1808 provide a connection and/or communication links between device 1800 and a communication network by which other electronic, computing, and communication devices communicate data with device 1800 .
- Device 1800 includes one or more processors 1810 (e.g., any of microprocessors, controllers, and the like), which process various computer-executable instructions to control the operation of device 1800 and to enable techniques for implementing a variable-depth stereoscopic display.
- processors 1810 e.g., any of microprocessors, controllers, and the like
- device 1800 can be implemented with any one or combination of hardware, firmware, or fixed logic circuitry that is implemented in connection with processing and control circuits which are generally identified at 1812 .
- device 1800 can include a system bus or data transfer system that couples the various components within the device.
- a system bus can include any one or combination of different bus structures, such as a memory bus or memory controller, a peripheral bus, a universal serial bus, and/or a processor or local bus that utilizes any of a variety of bus architectures.
- Device 1800 also includes computer-readable storage media 1814 , such as one or more memory devices that enable persistent and/or non-transitory data storage (i.e., in contrast to mere signal transmission), examples of which include random access memory (RAM), non-volatile memory (e.g., any one or more of a read-only memory (ROM), non-volatile RAM (NVRAM), flash memory, EPROM, EEPROM, etc.), and a disk storage device.
- RAM random access memory
- non-volatile memory e.g., any one or more of a read-only memory (ROM), non-volatile RAM (NVRAM), flash memory, EPROM, EEPROM, etc.
- a disk storage device may be implemented as any type of magnetic or optical storage device, such as a hard disk drive, a recordable and/or rewriteable compact disc (CD), any type of a digital versatile disc (DVD), and the like.
- Device 1800 can also include a mass storage media device 1816 .
- Computer-readable storage media 1814 provides data storage mechanisms to store the device data 1804 , as well as various device applications 1818 and any other types of information and/or data related to operational aspects of device 1800 .
- an operating system 1820 can be maintained as a computer application with the computer-readable storage media 1814 and executed on processors 1810 .
- the device applications 1818 may include a device manager, such as any form of a control application, software application, signal-processing and control module, code that is native to a particular device, a hardware abstraction layer for a particular device, and so on.
- the device applications 1818 also include any system components or modules to implement techniques using or enabling stereoscopic display.
- the device applications 1818 can include controller 122 and sensor module 1822 for analyzing and/or processing sensor information received from sensor 126 .
Landscapes
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Optics & Photonics (AREA)
- Engineering & Computer Science (AREA)
- Nonlinear Science (AREA)
- Computer Hardware Design (AREA)
- Theoretical Computer Science (AREA)
- Mathematical Physics (AREA)
- Chemical & Material Sciences (AREA)
- Crystallography & Structural Chemistry (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
This document describes various techniques for implementing a variable-depth stereoscopic display. A first distance at which a viewer is disposed relative to a stereoscopic display is received. Once received, a second distance by which to change a front focal distance of a lens structure of the stereoscopic display is determined based on the first distance. The front focal distance of the lens structure is then caused to change by the second distance effective to display a stereoscopic image at the first distance.
Description
- Content providers are increasingly offering content in three-dimensional (3D) format to enhance a viewing experience. This 3D content is often shown on specialized 3D displays that split the 3D content into different images targeted to each eye of a viewer. Different imagery can be displayed to each eye of a viewer in a variety of ways including the use of passive eyewear, active eyewear, dual head-mounted displays, or dividing the image directionally without the use of eyewear (auto-stereoscopy). Although auto-stereoscopy does not require the use of specialized glasses or multiple displays, a distance at which the divided images are effective to produce 3D imagery is generally limited. As such, stereoscopic displays are unable to display 3D imagery to a viewer that moves out of a focal plane of the display or to multiple viewers at varying distances from the display.
- This document describes various apparatuses and techniques for implementing a variable-depth stereoscopic display. These apparatuses and techniques may enable a stereoscopic display to provide three-dimensional (3D) content to a viewer over different distances from the stereoscopic display or to multiple viewers at varying distances from the stereoscopic display. A first distance at which a viewer is disposed relative to a stereoscopic display can be received. Once received, a second distance by which to change a front focal distance of a lens structure of the stereoscopic display is determined based on the first distance. The front focal distance of the lens structure is then caused to change by the second distance effective to display a stereoscopic image at the first distance. By so doing, a front focal distance of the stereoscopic display can be varied to provide 3D content to a viewer at different distances from the stereoscopic display.
- This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
- The detailed description is described with reference to the accompanying figures. In the figures, the left-most digit of a reference number identifies the figure in which the reference number first appears. The use of the same reference number in different instances in the description and the figures may indicate similar or identical items.
-
FIG. 1 illustrates an example environment in which these techniques may be implemented. -
FIG. 2 illustrates an example display of the display device ofFIG. 1 in more detail. -
FIG. 3 illustrates another example display of the display device ofFIG. 1 in more detail. -
FIG. 4 illustrates an example of scanning light with components of the display ofFIG. 2 . -
FIG. 5 illustrates another example of scanning light with components of the display ofFIG. 2 . -
FIG. 6 illustrates a detailed example of the light injection system ofFIG. 2 . -
FIG. 7 illustrates an example light guide configuration of the display ofFIG. 2 and ray traces associated with a light source. -
FIG. 8 illustrates a plan view of example ray traces associated with the light guide configuration and the light source ofFIG. 7 . -
FIG. 9 illustrates an example light guide configuration of the display ofFIG. 2 and ray traces associated with another light source. -
FIG. 10 illustrates a plan view of example ray traces associated with the light guide configuration and the other light source ofFIG. 9 . -
FIG. 11 illustrates a plan view of example ray traces associated with the liquid crystal lens ofFIG. 3 . -
FIG. 12 illustrates a plan view of example of ray traces of re-focused light associated with the liquid crystal lens ofFIG. 3 . -
FIG. 13 is a flow diagram depicting an example method for causing a front focal distance of a stereoscopic display to change. -
FIG. 14 illustrates an example viewing environment in which various embodiments of these apparatuses and techniques may be employed. -
FIG. 15 is a flow diagram depicting an example method for selecting a light source to change a focal plane of a stereoscopic display. -
FIG. 16 illustrates the example viewing environment ofFIG. 14 in which a viewer has moved out of a focal plane of a stereoscopic display. -
FIG. 17 is a flow diagram depicting an example method for selecting a light source based on a distance to a viewer. -
FIG. 18 illustrates an example device in which techniques for enabling a variable-depth stereoscopic display can be implemented. - Overview
- This document describes various apparatuses and techniques for implementing a variable-depth stereoscopic display. Various embodiments of these techniques change a front focal distance of a stereoscopic display, which allows the stereoscopic display to provide 3D content to a viewer over different distances from the stereoscopic display or to multiple viewers at varying distances from the stereoscopic display.
- Consider a case where two viewers are watching 3D content (e.g., a movie or video-game) on a stereoscopic television device. Assume that both viewers are positioned at approximately the same distance from the television device, such as sitting at a couch roughly parallel with the television device, and are able to view the 3D content. Assume now, that one of the viewers would like to move to a position closer to or farther from to the television device (e.g., a chair located closer to the television device). The described apparatuses and techniques enable the 3D content to be provided to both viewers at different distances from the television device or to a viewer that moves closer to or farther from the television device.
- Example Environment
-
FIG. 1 is an illustration of anexample environment 100 in which a variable-depth stereoscopic display can be implemented.Environment 100 includesdisplay device 102 which is illustrated, by way of example and not limitation, as one of asmart phone 104,laptop computer 106,television device 108,desktop computer 110, ortablet computer 112. Generally,display device 102 can provide three-dimensional (3D) content to viewers without the use of special 3D eyewear. 3D content may comprise images (e.g., stereoscopic imagery) and/or video effective to cause a viewer to be able to perceive depth within the content when displayed. In some embodiments, the 3D content is manipulable by a viewer or user, such as when displayed in connection with an application or user interface of a display device capable of accepting user input (e.g., a touch-screen enabledsmart phone 104 ortelevision device 108 operably associated with gaming controllers or optical-spatial sensors). -
Display device 102 includes processor(s) 114 and computer-readable media 116, which includesmemory media 118 andstorage media 120. Applications and/or an operating system (not shown) embodied as computer-readable instructions on computer-readable memory 116 can be executed by processor(s) 114 to provide some or all of the functionalities described herein. Computer-readable media also includes stereoscopic-depth controller (controller) 122. Howcontroller 122 is implemented and used varies, and is described as part of the methods discussed below. -
Display device 102 also includesstereoscopic display 124,sensor 126, input/output (I/O)ports 128, and network interface(s) 130.Stereoscopic display 124 is capable of generating stereoscopic 3D content that can be viewed without the use of special eyewear.Stereoscopic display 124 may be separate or integral withdisplay device 102; integral examples includesmart phone 104,laptop 106, andtablet 112; separate examples includetelevision device 108 and, in some instances, desktop computer 110 (e.g., when embodied as a separate tower and monitor (shown)). -
Sensor 126 collects viewer positional data useful to determine a position of a viewer relativestereoscopic display 124. The viewer positional data can be useful determine a distance of a viewer from stereoscopic display 124 (relative Z position), a distance of a viewer from a horizontal axis of stereoscopic display 124 (relative Y position), or a distance of a viewer from a vertical axis of stereoscopic display 124 (relative X position). In at least some embodiments, an approximate position of a viewer's eyes can be determined with the viewer positional data provided bysensor 126.Sensor 126 may be separate or integral withdisplay device 102; integral examples include sensor 126-1 oftelevision device 108 and sensor 126-2 oftablet computer 112; separate examples include stand-alone sensors, such as sensors operably coupled withdisplay device 102, a set-top box, or a gaming device. -
Sensor 126 can collect viewer position data by way of various sensing technologies, either working alone or in conjunction with one another. Sensing technologies may include, by way of example and not limitation, optical, radio-frequency, acoustic (active or passive), micro-electro-mechanical systems (MEMS), ultrasonic, infrared, pressure sensitive, and the like. In some embodiments,sensor 126 may receive additional data or work in conjunction with a remote control device or gaming controller associated with one or more viewers to generate the viewer positional data. - I/
O ports 128 ofdisplay device 102 enable interaction with other devices, media, or users. I/O ports 128 can include a variety of ports, such as by way of example and not limitation, high-definition multimedia (HDMI), digital video interface (DVI), display port, fiber-optic or light-based, audio ports (e.g., analog, optical, or digital), USB ports, serial advanced technology attachment (SATA) ports, peripheral component interconnect (PCI) express based ports or card slots, serial ports, parallel ports, or other legacy ports. In at least some instances, 3D content is received bydisplay device 102 via one or more I/O ports 128 from another device or source (e.g., a set-top box or content receiving device). -
Display device 102 may also include network interface(s) 130 for communicating data over wired, wireless, or optical networks. Data communicated over such networks may include 3D content that can be displayed or interacted with viastereoscopic display 124. By way of example and not limitation,network interface 130 may communicate data over a local-area-network (LAN), a wireless local-area-network (WLAN), a personal-area-network (PAN), a wide-area-network (WAN), an intranet, the Internet, a peer-to-peer network, point-to-point network, a mesh network, and the like. -
FIG. 2 illustrates a detailed example ofstereoscopic display 124 ofFIG. 1 .Stereoscopic display 124 includeslens structure 202,light injection system 204,light re-director 206, and spatiallight modulator 208.Stereoscopic display 124 may be configured as a non-projection based flat panel display having a depth or thickness similar to that of a liquid crystal display (LCD) panel and the like.Lens structure 202 emits light from a surface when light is received fromlight injection system 204. The light emitted from lens structure may be collimated light. In some case,lens structure 202 is an optical wedge having a thin end 210 to receive light, a thick end 212 effective to reflect the light (e.g., via an end reflector or reflective cladding), and aviewing surface 214 at which the light is emitted as collimated light. - In some implementations, an optical wedge may comprise an optical lens or light guide that permits light input at an edge of the optical wedge (e.g., thin end 210) to fan out within the optical wedge via total internal reflection before reaching the critical angle for internal reflection and exiting via another surface of the optical wedge (e.g., viewing surface 214). The light may exit the optical wedge at a glancing angle relative to
viewing surface 214. - The light emitted by
lens structure 202 can be scanned by varying light generated bylight injection system 204 or an injection location thereof. Generally, scanning the light enables the display of 3D content that is viewable without the use of special eyewear. The scanned light enables display of different stereoscopic imagery to each eye of a respective viewer. -
Light injection system 204 may include any suitable type or number of light sources, such as by way of example and not limitation, cold-cathode fluorescent lamps (CCFL), light emitting diodes (LEDs), light engines, lasers (chemical or solid-state), and the like. How the light oflight injection source 204 is generated and controlled varies, and is described in conjunction with subsequent figures and methods. - In some cases,
stereoscopic display 124 includes light re-director 206 located adjacent toviewing surface 214 to diffuse collimated light emitted bylens structure 202. This collimated light may exitlens structure 202 at a glancing angle with respect toviewing surface 214. Here, light re-director 206 can re-direct the emitted light oflens structure 202 towardsviewer 216 and may provide a diffusing function in one dimension.Light re-director 206 can be configured as any suitable structure, such as a film of prisms or a light-guide panel having a prismatic textured surface. By varying angles of prisms or prismatic features over a surface of light re-director 206, light re-director 206 can be configured to have optical power capable of focusing the collimated light emitted from viewingsurface 214 oflens structure 202 towards the eyes ofviewer 216. - Spatial
light modulator 208 modulates the light with visual information to form imagery displayed by the light converging on the eyes ofviewer 216. In some cases, the visual information is parallax information directed to different eyes ofviewer 216 in order to provide the 3D content. For instance, spatiallight modulator 208 can modulate light directed towards a viewer's left eye with a frame of stereoscopic imagery, and then modulate light directed to a viewer's right eye with another frame of stereoscopic imagery. Thus, by synchronizing scanning and modulation of light (collimated or otherwise), 3D content can be provided to a viewer. Spatiallight modulator 208 can be any suitable structure such as a liquid crystal display (LCD) cell composed of an array of pixels or sub-pixels capable of light modulation and/or manipulation. - In this particular example,
controller 122 is operably coupled tolight injection system 204 andsensor 126. In some cases,controller 122 is operably coupled with spatiallight modulator 208 or a modulation-controller associated therewith.Controller 122 receives viewer position information, such as a distance to a viewer, collected bysensor 126 and, as described in the following sections and associated methods, can controllight injection source 204 effective to display 3D imagery viastereoscopic display 124 over various distances. -
FIG. 3 illustrates another detailed example ofstereoscopic display 124 ofFIG. 1 .Stereoscopic display 124 includeslens structure 202,light injection system 204, liquid crystal lens (LQ lens) 302, and spatiallight modulator 208.Lens structure 202 emits light (collimated or otherwise) from a surface when light is received fromlight injection system 204. In some cases,lens structure 202 is an optical wedge having a thin end 210 to receive light, a thick end 212 effective to reflect the light, and aviewing surface 214 at which the light is emitted as collimated light.Display 124 can be configured as a non-projection based flat panel display having a depth or thickness similar to that of a liquid crystal display (LCD) panel and the like. - The light emitted by
lens structure 202 can be scanned by varying light generated bylight injection system 204 or an injection location thereof. Generally, scanning the light enables the display of 3D content that is viewable without the use of special eyewear. The scanned light enables display of different stereoscopic imagery to each eye of a respective viewer. -
Light injection system 204 may include any suitable type or number of light sources, such as by way of example and not limitation, cold-cathode fluorescent lamps (CCFL), light emitting diodes (LEDs), light engines, lasers (chemical or solid-state), and the like. How the light oflight injection source 204 is generated and controlled varies, and is described in conjunction with subsequent figures and methods. -
Stereoscopic display 124 may also include light re-director 206 located adjacent toviewing surface 214 to diffuse collimated light emitted bylens structure 202. This collimated light may exitlens structure 202 at a glancing angle with respect toviewing surface 214. Here, light re-director 206 can re-direct the emitted light oflens structure 202 towardsviewer 216 and may provide a diffusing function in one dimension.Light re-director 206 can be configured as any suitable structure, such as a film of prisms or a light-guide panel having a prismatic textured surface. By varying angles of prisms or prismatic features over a surface of light re-director 206, light re-director 206 can be configured to have optical power capable of focusing the collimated light emitted from viewingsurface 214 oflens structure 202 towards the eyes ofviewer 216. - In this particular example,
stereoscopic display 124 includesLQ lens 302 located adjacent to light re-director 206 capable of varying a depth at which stereoscopic imagery is provided. As light re-directed by light re-director 206 passes throughLQ lens 302, a focal depth (or focal length) oflens structure 202 can be varied by focusing or re-focusingLQ lens 302. Affecting or manipulating liquid crystal molecules ofLQ lens 302 may focus or re-focus light emitted bylens structure 202.LQ lens 302 may comprise two transparent substrates, such as glass or other transparent material, between which these liquid crystal molecules are disposed.LQ lens 302 also includeselectrodes LQ lens 302. In some cases, these electrodes may be composed of a transparent material such as indium tin oxide (ITO) so as not to compromise clarity or transparency ofLQ lens 302. Although illustrated as having twoelectrodes LQ lens 302 may include any suitable number of electrodes proximate the perimeter of the lens to apply excitation voltage between any two peripheral points ofLQ lens 302. Alternately or additionally, one or both of the transparent substrates can be coated, either in part or entirely, with transparent conductive coating, such as ITO and the like. - Applying excitation voltage to
LQ lens 302 can re-align the liquid crystal molecules effective to re-focusLQ lens 302 at various distances. Although not shown, other layers or substrates located proximate the liquid crystal molecules ofLQ lens 302 may affect a uniformity of a distortion or alignment of the liquid crystal molecules when under the influence of an electrical excitation field. In some cases, affect enablesLQ lens 302 to focus light emitted fromlens structure 202 at various distances. As is described in conjunction with subsequent figures and methods, varying a focal distance oflens structure 202 enables implementation of a variable depth stereoscopic display.LQ lens 302 may be used in place of (e.g., instead of light guides) or in conjunction with any other elements described herein in order to implement a stereoscopic display. - Spatial
light modulator 208 modulates the light exitingLQ lens 302 with visual information to form imagery displayed by the light converging on the eyes ofviewer 216. In some cases, the visual information is parallax information directed to different eyes ofviewer 216 in order to provide the 3D content. For instance, spatiallight modulator 208 can modulate light directed towards a viewer's left eye with a frame of stereoscopic imagery, and then modulate light directed to a viewer's right eye with another frame of stereoscopic imagery. Thus, by synchronizing scanning and modulation of light (collimated or otherwise), 3D content can be provided to a viewer. Spatiallight modulator 208 can be any suitable structure such as a liquid crystal display (LCD) cell composed of an array of pixels or sub-pixels capable of light modulation and/or manipulation. - In this particular example,
controller 122 is operably coupled tolight injection system 204,LQ lens 302, andsensor 126.Controller 122 receives viewer position information, such as a distance to a viewer, collected bysensor 126 and, as described in the following sections and associated methods, can controllight injection source 204 orLQ lens 302 effective to display 3D imagery viastereoscopic display 124 over various distances. -
FIGS. 4 and 5 illustrate an example of scanning light by varying a location at which light is injected intolens structure 202 at 400 and 500 respectively. Generally, scanning light is performed by directing the light horizontally across a viewing plane or area as shown inFIGS. 4 and 5 . 3D imagery can be displayed to a viewer with scanned light that directs different images to each of the viewer's eyes. For example, 3D video content split into imagery intended for a viewer's right eye and a viewer's left eye, when interleaved and displayed at sufficient frequency, enables the viewer to experience the 3D video without special eyewear. - Specifically, the light can be directed to the left (300) by shifting a location of light injection to the right, and vice versa (400) as illustrated by ray traces within
lens structure 202 in each respective figure. Additionally, in each figure, a visible portion of a single pixel of light, shown respectively at 402 and 502, is illustrated for clarity. Illustrating perspective, lines tracing a path of the light from the corners oflens structure 202 to the points oflight center line 404 andfocal plane 406. - A location at which light is injected can be varied in any suitable way. For example, a
light injection system 204 may comprise LEDs that are individually selectable to provide light at different locations along the lens structure effective to scan light emitted by the lens structure. In this particular example, selecting one or more LEDs proximate one end of the lens structure (right-hand side) directs light in an opposite direction (left-hand side), as illustrated by 400. In other cases,light injection system 204 may include a laser and an acousto-optic modulator or liquid crystal hologram for controlling directionality of a beam produced by the laser. -
FIG. 6 illustrates a detailed example oflight injection system 204 ofFIG. 2 in more detail. In this particular example,light injection system 204 includeslight sources lens structure 202. Each light source located at a different distance from a light receiving surface oflens structure 202 has a different effective back focal length. Although shown as three light sources,light injection system 204 may include any suitable number or configuration of light sources. -
Light sources Elements 608 can be selected such thatlight sources lens structure 202 to scan light as described above. -
Light sources light guides lens structure 202, such as then end 210 of an optical wedge. In some cases, light guides 610, 612, and 614 are useful to fold down back focal lengths oflens structure 202. Although shown as light guides, any suitable optical structure may be associated withlight sources - Additionally, light guides 610, 612, and 614 can be configured to prevent
light sources -
FIGS. 7 and 8 illustrate an example light pattern associated with a light guide ofFIG. 6 . Generally, a light guide associated with a respective light source will transmit light from that respective light source. As shown inFIG. 7 , which illustrates a side view oflight injection system 204, light 702 is generated bylight source 602 and transmitted bylight guide 610. Here, assume that light 702 is intended to direct light towards a viewer's right eye, capable of delivering a frame of stereoscopic imagery to the viewer's right eye. -
FIG. 8 illustrates a plan view of example light patterns associated with the light guide ofFIG. 7 . Here, assume that light 802 is intended to direct light towards a viewer's left eye, capable delivering a frame of stereoscopic imagery to the viewer's left eye.Light 802 is also generated bylight source 602; however, light 802 is generated at a different location by a set ofelements 608 that is different from a set that generated light 702. For example,elements 608 located proximate an end oflight source 602 generate light 802, whileelements 608 proximate the other end oflight source 602 generate light 702. Thus, although shown as over-lapping inFIG. 8 , light 702 and light 802 are not generated concurrently, but sequentially to direct light in different directions. -
Lens structure 202 receives light 702 and 802 and emits light 804 and 806 respectively. Spatiallight modulator 208 modulates light 804 and 806 with stereoscopic information intended for a viewer's right eye and left eye respectively. The stereoscopic imagery can be displayed to a viewer atfocal plane 808 where the light is appropriately directed to each eye of the viewer.Focal plane 808 can be any suitable distance fromlens structure 202 and may vary depending on adisplay device 102 in whichstereoscopic display 124 is implemented. For example,focal plane 808 may be approximately 500-700 mm forlaptop device 106 or approximately 3-5 m fortelevision device 108. A depth of the focal plane at which the stereoscopic imagery can be viewed may vary, ranging for example from 50 mm to 300 mm. -
FIGS. 9 and 10 illustrate another example light pattern associated with another light guide ofFIG. 6 . As shown inFIG. 9 which illustrates a side view oflight injection system 204, light 902 is generated bylight source 606 and transmitted bylight guide 614. Here, assume that light 902 is intended to direct light towards a viewer's right eye, capable delivering a frame of stereoscopic imagery to the viewer's right eye. Note here, thatlight guide 614 is longer thanlight guide 610 bydistance 904, thus selectinglight source 606 changes a back focal length associated withlight 902. -
FIG. 10 illustrates a plan view of example light patterns associated with the light guide ofFIG. 9 . Here, assume that light 1002 is intended to direct light towards a viewer's left eye, capable delivering a frame of stereoscopic imagery to the viewer's left eye.Light 1002 is also generated bylight source 606; however, light 1002 is generated at a different location by a set ofelements 608, which is different from a set that generated light 902. For example,elements 608 located proximate an end oflight source 606 generate light 1002, whileelements 608 proximate the other end oflight source 606 generate light 902. Thus, although shown as over-lapping inFIG. 10 , light 902 and light 1002 are not generated concurrently, but sequentially to direct light in different directions. Also note thatlight guide 614 is a different length than the other light guides and configured such that the other light guides do not occlude light generated bylight source 606. -
Lens structure 202 receives light 902 and 1002 and emits light 1006 and 1008, respectively. Spatiallight modulator 208 modulates light 1006 and 1008 with stereoscopic information intended for a viewer's right eye and left eye respectively. In this example, the stereoscopic imagery is displayed to a viewer atfocal plane 1010, where the light is appropriately directed to each eye of the viewer. As shown bydistance 1012,focal plane 1010 moves towardslens structure 202 when another light source is selected to inject light intolens structure 202. - A front focal distance of
lens structure 202, at which the focal planes are located, can be varied by any suitable distance when different light sources are selected. In some cases, a light source has a back focal length that is different from those associated with other light sources. Selecting another light source having a different back focal length can change a distance at which the focal plane is located from astereoscopic display 124. In some cases, the distance varies based on the type of device having the stereoscopic display. For example, a distance betweenfocal plane 1010 andfocal plane 808 can be approximately 100-150 mm forlaptop device 106 or approximately 0.5-1.0 m for a television device. These distances are given merely as examples, and should not be construed as limitations, as other distances are contemplated to suit the viewing needs associated with different devices or environments. Additionally, the light sources may also be configured such that the focal planes of a stereoscopic display overlap, allowing stereoscopic imagery to be delivered to a moving viewer and/or to eliminate gaps between focal planes. -
FIGS. 11 and 12 illustrate a plan view of example ray traces associated with the liquid crystal lens ofFIG. 3 . Generally,LQ lens 302 is capable of changing a front focal distance oflens structure 202 by re-focusing light emitted bylens structure 202. As shown inFIG. 11 , light 1102 and 1104 are generated by a light source (not show) and transmitted to lens structure bylight guide 610. Although shown as over-lapping inFIG. 11 , light 1102 and 1104 are not generated concurrently, but sequentially to direct light in different directions. Assume here that light 1102 and 1104 are intended to direct light towards a viewer's right and left eyes respectively. -
Lens structure 202 receives light 1102 and 1104 and emits light 1106 and 1108 respectively. As the light passes throughLQ lens 302, it is focused to converge on the viewers eyes atfocal plane 1110. Additionally or alternately, spatiallight modulator 208 modulates light 1106 and 1108 with stereoscopic information intended for a viewer's right eye and left eye respectively.Focal plane 1110 can be any suitable distance fromlens structure 202 and may vary depending on adisplay device 102 in whichstereoscopic display 124 is implemented, examples of which are discussed above. -
FIG. 12 illustrates a plan view of an example of re-focused ray traces associated with the liquid crystal lens ofFIG. 3 . Here, assume that light 1102 and 1104 are the same light shown inFIG. 11 , although other light may be generated, such as when a viewer's relative Y position has changed. As inFIG. 11 ,lens structure 202 receives light 1102 and 1104 fromlight guide 610 and emits light 1202 and 1204 respectively. Assume here that light 1202 and 1204 are intended to direct light towards a viewer's right and left eyes respectively. As shown inFIG. 11 , light 1204 and 1204 converge on a viewer's eyes at a focal plane. Here, however,LQ lens 302 re-focuses and changes a focal distance oflens structure 202 as shown byfocal plane 1206. As shown bydistance 1208,focal plane 1206 moves towardslens structure 202 whenLQ lens 302 is re-focused. Thus, a front focal distance oflens structure 202 can be increased or decreased as needed by refocusingLQ lens 302 to implement a variable depth stereoscopic display. - Example Methods
-
FIG. 13 is flow diagram depicting anexample method 1300 for causing a front focal distance of a stereoscopic display to change.Block 1302 receives a first distance at which a viewer is disposed relative a stereoscopic display. In some cases,block 1302 receives a distance to a viewer that has moved towards or away from the stereoscopic display. A distance to the viewer may be received in the form of viewer positional data or viewer-tracking data. The viewer positional data can be collected by any suitable sensing technology, such as optical or imaging sensors, examples of which are described above. - Consider
method 1300 in the context of anexample viewing environment 1400 illustrated byFIG. 14 .Viewing environment 1400 includestelevision device 108 andviewers viewing plane 1406. In this particular example,viewing plane 1406 is located atdistance 1408 fromtelevision device 108. - In this example,
gaming device 1410 is operably coupled withtelevision device 108 and sensor 126 (shown as an external sensor) that collects positional data ofviewers television device 108 receives afirst distance 1408 fromsensor 126 viagaming device 1410. Here, assume that controller receives viewer positional data fromsensor 126.Controller 122 then analyzes the viewer positional data to determinedistance 1408 toviewer 1402. -
Block 1304 determines a second distance by which to change a front focal distance of the stereoscopic display based on the first distance. In some cases, determining the second distance is based on comparing the first distance with a current front focal distance of the stereoscopic display or a lens structure thereof. For example, the second distance may be the difference between the first distance to the viewer and the current front focal distance of the stereoscopic display. In the context of the present example, assume that the current front focal distance oftelevision device 108 is approximately half ofdistance 1408 toviewer 1402. Here,controller 122 determines a second distance (half of distance 1408) by which to change a front focal distance oftelevision device 108 such that a front focal distance of lens structure 202 (not shown) oftelevision device 108 is approximatelydistance 1408. -
Block 1306 causes the front focal distance of the stereoscopic display to change by the second distance. Changing the front focal distance of the stereoscopic display by the second distance can be effective to display a stereoscopic image at the first distance. In some case, the front focal distance is caused to change by the difference between the first distance to the viewer and the current front focal distance of the stereoscopic display. The front focal distance of the stereoscopic display, or a lens structure thereof, may be caused to change in any suitable way. In some cases, a liquid crystal lens may re-focus light emitted by a lens structure of the stereoscopic display. For example,LQ lens 302 may focus or refocus light emitted bylens structure 202 effective to change a front focal distance ofstereoscopic display 124. - In other cases, selecting or enabling a light source of a light injection system may cause a front focal distance of the stereoscopic display to change. The selected light source that injects the light may be one of multiple light sources that are individually selectable to generate or emit light. In some cases, a back focal length associated with each of the multiple light sources is different. By selecting a light source having a particular back focal length to inject light, a front focal distance of a stereoscopic display can be changed such that a stereoscopic image is visible to viewers located in a focal plane proximate the first distance.
- Alternately or additionally, a light source may be selected to cause the front focal distance to change based on the first distance and a lens structure's front focal distance associated with the light source. For example, the first distance to the viewer may correspond approximately to a front focal distance associated with a particular light source. Concluding the present example,
controller 122 enables a light source (e.g., light source 604) based on the first distance to display stereoscopic imagery atdistance 1408 enabling bothviewers viewers television device 108. In the context of the present example,controller 122 enables the same light source based on a distances to eitherviewer 1402 orviewer 1404 that are about the same distance fromtelevision device 108. - Optionally,
method 1300 may repeatblocks -
FIG. 15 is a flow diagram depicting an example method for selecting a light source to change a focal plane of a stereoscopic display.Block 1502 detects movement of a viewer of a stereoscopic display. The stereoscopic display may be displaying 3D content to one or more viewers located within multiple focal planes. In some cases, movement of a viewer is detected by analyzing data collected by an optical or imaging sensor, examples of which are described above. The movement may be detected as a viewer moves towards or away from the stereoscopic display. - Consider
FIG. 16 , which illustratesexample viewing environment 1400 in whichviewer 1402 has moved away fromtelevision device 108. In the context of this example,controller 122 detects movement ofviewer 1402 as she moves away fromtelevision device 108 by analyzing viewer positional data received fromsensor 126. -
Block 1504 determines whether the viewer is within a focal plane of the stereoscopic display. A current focal plane or a front focal distance of the stereoscopic display may correspond with a currently selected light source having a particular back focal length. A viewer may be determined to be within a focal plane when a distance to the viewer corresponds approximately to that of a front focal distance of the stereoscopic display. Distances to one or more viewers may be received or determined by analyzing viewer positional data. In some cases, a viewer may move, yet remain within a focal plane of the stereoscopic display. Ifblock 1504 determines that the viewer is within a current focal plane of the stereoscopic display,method 1500 returns to block 1502 where further movement may be detected. Ifblock 1504 determines that the viewer is not within the current focal plane of the stereoscopic display,method 1500 proceeds to block 1506. - In some cases, a viewer may be determined to be within two overlapping focal planes. In such a case, a further determination is made as to whether the viewer is closer to a previously occupied focal plane or closer to another focal plane towards which the viewer is moving. In the context of the present example,
controller 122 determines thatviewer 1402 is not withinfocal plane 1406 and is moving towards another focal plane oftelevision device 108. -
Block 1506 selects another light source to change a focal length of the stereoscopic display. Changing the focal length of the stereoscopic display varies a distance at which the focal plane of the stereoscopic display is located. In some cases, changing a front focal length of the stereoscopic display includes selecting another light source having a back focal length that is different from a previously selected light source. Optionally,method 1500 may return to block 1502 to detect subsequent movement of the viewer or movement of another viewer. - Concluding the present example,
controller 122 selects another light source (e.g., light source 602) operably associated with a lens structure oftelevision device 108. In the context of the present example, selecting the other light source moves or creates an additionalfocal plane 1602 atdistance 1604 from the originalfocal plane 1406. As shown inFIG. 16 ,television device 108 is able to display stereoscopic imagery toviewer 1402 now located withinfocal plane 1602. By repeating operations ofmethod 1500 and other methods described herein,television device 108 can display stereoscopic imagery to one or more viewers as they move out of or between focal planes. -
FIG. 17 is a flow diagram depicting an example method for selecting a light source based on a distance to a viewer.Block 1702 determines a distance to a viewer of a stereoscopic display. The distance to the viewer may be determined by analyzing viewer positional data. Viewer positional data associated with the viewer can be collected using any suitable sensing technology, such as optical or imaging sensors, examples of which are described above. In some cases, determining a distance to a viewer may be responsive to detecting movement of the viewer. Additionally or alternately, block 1702 may determine distances to multiple viewers of the stereoscopic display. - In the context of the above-mentioned example, assume that
viewers focal planes Sensor 126 collects viewer positional data associated withviewer 1402 and transmits this data tocontroller 122.Controller 122 then determines a distance ofviewer 1402 fromtelevision device 108 by analyzing the viewer positional data received fromsensor 126. -
Block 1704 selects a light source based on the distance from the stereoscopic display to the viewer. Selecting the light source based on the distance enables stereoscopic imagery to be displayed to the viewer. The light source may be selected from multiple light sources individually selectable to emit or generate light. The selected light source may have an associated back focal length that corresponds to a front focal distance of a lens structure. In some cases, the front focal distance of the lens structure is approximately the distance to the viewer. - A lens structure collimates the light generated by the selected light source. The light is then modulated by a spatial light modulator to display stereoscopic imagery to the viewer. Continuing the present example,
controller 122 selects a light source (e.g., light source 602) oftelevision device 108 effective to display stereoscopic imagery toviewer 1402 atfocal plane 1602. - Additionally or alternately,
method 1700 may repeatblocks - Light generated by the other light source is then and modulated to display stereoscopic imagery to the other viewer. Concluding the present example,
controller 122 selects another light source (e.g., light source 604) effective to display stereoscopic imagery toviewer 1404 atfocal plane 1406. Within their respective focal planes,viewers television device 108. - The preceding discussion describes methods in which the techniques may enable the display of variable-depth stereoscopic imagery. These methods are shown as sets of blocks that specify operations performed but are not necessarily limited to the order shown for performing the operations by the respective blocks.
- Aspects of these methods may be implemented in hardware (e.g., fixed logic circuitry), firmware, a System-on-Chip (SoC), software, manual processing, or any combination thereof. A software implementation represents program code that performs specified tasks when executed by a computer processor, such as software, applications, routines, programs, objects, components, data structures, procedures, modules, functions, and the like. The program code can be stored in one or more computer-readable memory devices, both local and/or remote to a computer processor. The methods may also be practiced in a distributed computing environment by multiple computing devices.
- Example Device
-
FIG. 18 illustrates various components ofexample device 1800 that can be implemented as any type of client, server, and/or display device as described with reference to the previousFIGS. 1-14 to implement techniques enabling a variable-depth stereoscopic display. In embodiments,device 1800 can be implemented as one or a combination of a wired and/or wireless device, as a form of flat panel display, television, television client device (e.g., television set-top box, digital video recorder (DVR), etc.), consumer device, computer device, server device, portable computer device, user device, communication device, video processing and/or rendering device, appliance device, gaming device, electronic device, and/or as another type of device.Device 1800 may also be associated with a viewer (e.g., a person or user) and/or an entity that operates the device such that a device describes logical devices that include users, software, firmware, and/or a combination of devices. -
Device 1800 includescommunication devices 1802 that enable wired and/or wireless communication of device data 1804 (e.g., received data, data that is being received, data scheduled for broadcast, data packets of the data, etc.). Thedevice data 1804 or other device content can include configuration settings of the device, media content stored on the device, and/or information associated with a user of the device. Media content stored ondevice 1800 can include any type of audio, video, and/or image data.Device 1800 includes one ormore data inputs 1806 via which any type of data, media content, and/or inputs can be received, such as user-selectable inputs, messages, music, television media content, recorded video content, and any other type of audio, video, and/or image data received from any content and/or data source. -
Device 1800 also includescommunication interfaces 1808, which can be implemented as any one or more of a serial and/or parallel interface, a wireless interface, any type of network interface, a modem, and as any other type of communication interface. The communication interfaces 1808 provide a connection and/or communication links betweendevice 1800 and a communication network by which other electronic, computing, and communication devices communicate data withdevice 1800. -
Device 1800 includes one or more processors 1810 (e.g., any of microprocessors, controllers, and the like), which process various computer-executable instructions to control the operation ofdevice 1800 and to enable techniques for implementing a variable-depth stereoscopic display. Alternatively or in addition,device 1800 can be implemented with any one or combination of hardware, firmware, or fixed logic circuitry that is implemented in connection with processing and control circuits which are generally identified at 1812. Although not shown,device 1800 can include a system bus or data transfer system that couples the various components within the device. A system bus can include any one or combination of different bus structures, such as a memory bus or memory controller, a peripheral bus, a universal serial bus, and/or a processor or local bus that utilizes any of a variety of bus architectures. -
Device 1800 also includes computer-readable storage media 1814, such as one or more memory devices that enable persistent and/or non-transitory data storage (i.e., in contrast to mere signal transmission), examples of which include random access memory (RAM), non-volatile memory (e.g., any one or more of a read-only memory (ROM), non-volatile RAM (NVRAM), flash memory, EPROM, EEPROM, etc.), and a disk storage device. A disk storage device may be implemented as any type of magnetic or optical storage device, such as a hard disk drive, a recordable and/or rewriteable compact disc (CD), any type of a digital versatile disc (DVD), and the like.Device 1800 can also include a massstorage media device 1816. - Computer-
readable storage media 1814 provides data storage mechanisms to store thedevice data 1804, as well asvarious device applications 1818 and any other types of information and/or data related to operational aspects ofdevice 1800. For example, anoperating system 1820 can be maintained as a computer application with the computer-readable storage media 1814 and executed onprocessors 1810. Thedevice applications 1818 may include a device manager, such as any form of a control application, software application, signal-processing and control module, code that is native to a particular device, a hardware abstraction layer for a particular device, and so on. - The
device applications 1818 also include any system components or modules to implement techniques using or enabling stereoscopic display. In this example, thedevice applications 1818 can includecontroller 122 andsensor module 1822 for analyzing and/or processing sensor information received fromsensor 126. - This document describes various apparatuses and techniques for implementing a variable-depth stereoscopic display. By so doing, a front focal distance of a stereoscopic display can be varied to display stereoscopic imagery to a viewer at different distances. Although the invention has been described in language specific to structural features and/or methodological acts, it is to be understood that the invention defined in the appended claims is not necessarily limited to the specific features or acts described. Rather, the specific features and acts are disclosed as example forms of implementing the claimed invention.
Claims (20)
1. A method comprising:
receiving a first distance at which a viewer is disposed relative to a stereoscopic display;
determining, based on the first distance, a second distance by which to change a front focal distance of a lens structure of the stereoscopic display; and
causing the focal distance of the lens structure to change by the second distance effective to display a stereoscopic image at the first distance.
2. The method as recited in claim 1 , wherein the stereoscopic display includes a liquid crystal lens configured to re-focus light emitted by the lens structure effective to change the front focal distance of the lens structure by the second distance.
3. The method as recited in claim 1 further comprising determining, based on the first distance, which of a first or a second light source of the stereoscopic display to use and wherein the causing causes the determined light source to inject light into the lens structure of the stereoscopic display.
4. The method as recited in claim 3 , wherein the lens structure an optical wedge configured to receive light from the determined light source and to emit light to display the stereoscopic image at the first distance.
5. The method as recited in claim 3 , wherein the first light source is operably coupled with a light guide, the light guide configured such that a back focal length associated with the first light source is different than a back focal length associated with the second light source.
6. The method as recited in claim 5 , wherein the respective back focal lengths associated with the first or the second light sources are effective to cause a front focal distance associated with the lens structure to be different depending on which of the first or the second light sources is caused to inject the light into the lens structure.
7. The method as recited in claim 3 , wherein the first or the second light source includes an array of light emitting diodes (LEDs) that are individually selectable to inject the light into the lens structure.
8. An optical system comprising:
two or more light sources individually selectable to emit light, each of the two or more light sources associated with a respective light guide having an optical path length that is different from an optical path length of a light guide associated with another of the two or more light sources;
a lens structure configured to receive light from one of the light guides associated with one of the two or more light sources and emit light useful to display stereoscopic imagery to a viewer;
a controller configured to:
determine whether the viewer is within a front focal plane of the stereoscopic display; and
if the viewer is not in the front focal plane of the stereoscopic display, select another of the two or more light sources to change the front focal plane of the stereoscopic display.
9. The optical system as recited in claim 8 , wherein the lens structure is an optical wedge.
10. The optical system as recited in claim 8 , wherein the light guide associated with the other of the two or more light sources is configured to not occlude light emitted by the light source.
11. The optical system as recited in claim 8 , wherein the controller is further configured to receive viewer positional data from a sensor based on which the controller is configured to determine whether the viewer is within the front focal plane of the stereoscopic display.
12. The optical system as recited in claim 8 , wherein the controller is further configured to detect movement of the viewer in response to which the controller is configured to determine whether the viewer is within the front focal plane of the stereoscopic display.
13. The optical system as recited in claim 8 , wherein the optical system is configured as a non-projective flat panel display.
14. The optical system as recited in claim 8 , further comprising a spatial light modulator configured to modulate the light with parallax information associated with the stereoscopic imagery.
15. A system comprising:
a first light source and a second light source individually selectable to emit light;
a first light guide and a second light guide, the first and the second light guides configured to receive the light emitted from the first or the second light source respectively, the first light guide having an optical path length that is different from an optical path length of the second light guide;
a lens structure configured to receive the light from the first or the second light guides and emit light useful to display stereoscopic imagery to a viewer; and
a controller configured to select, based on an distance from the lens structure to the viewer, one of the first or the second light sources effective to set a front focal distance of the lens structure such that the front focal distance corresponds to the distance to the viewer.
16. The system as recited in claim 15 , wherein the lens structure is an optical wedge.
17. The system as recited in claim 15 , further comprising a light re-director configured to diffuse the light emitted by the lens structure.
18. The system as recited in claim 15 , wherein selecting the first light source is effective to cause the lens structure to have a first front focal distance and selecting the second light source is effective to cause the lens structure to have a second front focal distance.
19. The system as recited in claim 15 , wherein the first and the second front focal distances of the lens structure sufficiently overlap to provide the stereoscopic imagery continuously over the first and second front focal distances.
20. The system as recited in claim 15 , further comprising a spatial light modulator configured to modulate the light emitted by the lens structure.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/192,321 US20130027772A1 (en) | 2011-07-27 | 2011-07-27 | Variable-depth stereoscopic display |
US15/047,859 US10082669B2 (en) | 2011-07-27 | 2016-02-19 | Variable-depth stereoscopic display |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/192,321 US20130027772A1 (en) | 2011-07-27 | 2011-07-27 | Variable-depth stereoscopic display |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/047,859 Continuation US10082669B2 (en) | 2011-07-27 | 2016-02-19 | Variable-depth stereoscopic display |
Publications (1)
Publication Number | Publication Date |
---|---|
US20130027772A1 true US20130027772A1 (en) | 2013-01-31 |
Family
ID=47597012
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/192,321 Abandoned US20130027772A1 (en) | 2011-07-27 | 2011-07-27 | Variable-depth stereoscopic display |
US15/047,859 Active US10082669B2 (en) | 2011-07-27 | 2016-02-19 | Variable-depth stereoscopic display |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/047,859 Active US10082669B2 (en) | 2011-07-27 | 2016-02-19 | Variable-depth stereoscopic display |
Country Status (1)
Country | Link |
---|---|
US (2) | US20130027772A1 (en) |
Cited By (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130127839A1 (en) * | 2011-11-17 | 2013-05-23 | Lg Electronics Inc. | Stereoscopic display device and mobile device having the same |
US20130242067A1 (en) * | 2012-03-19 | 2013-09-19 | Dongman Jeong | Three-dimensional image processing apparatus and method for adjusting location of sweet spot for displaying multi-view image |
US20130249895A1 (en) * | 2012-03-23 | 2013-09-26 | Microsoft Corporation | Light guide display and field of view |
US20150039061A1 (en) * | 2013-07-30 | 2015-02-05 | Samsung Display Co., Ltd. | Display apparatus providing light therapy |
US9297996B2 (en) | 2012-02-15 | 2016-03-29 | Microsoft Technology Licensing, Llc | Laser illumination scanning |
US9298012B2 (en) | 2012-01-04 | 2016-03-29 | Microsoft Technology Licensing, Llc | Eyebox adjustment for interpupillary distance |
US9368546B2 (en) | 2012-02-15 | 2016-06-14 | Microsoft Technology Licensing, Llc | Imaging structure with embedded light sources |
US9558590B2 (en) | 2012-03-28 | 2017-01-31 | Microsoft Technology Licensing, Llc | Augmented reality light guide display |
US9578318B2 (en) | 2012-03-14 | 2017-02-21 | Microsoft Technology Licensing, Llc | Imaging structure emitter calibration |
US9581820B2 (en) | 2012-06-04 | 2017-02-28 | Microsoft Technology Licensing, Llc | Multiple waveguide imaging structure |
US9717981B2 (en) | 2012-04-05 | 2017-08-01 | Microsoft Technology Licensing, Llc | Augmented reality and physical games |
US9726887B2 (en) | 2012-02-15 | 2017-08-08 | Microsoft Technology Licensing, Llc | Imaging structure color conversion |
US9779643B2 (en) | 2012-02-15 | 2017-10-03 | Microsoft Technology Licensing, Llc | Imaging structure emitter configurations |
US10018844B2 (en) | 2015-02-09 | 2018-07-10 | Microsoft Technology Licensing, Llc | Wearable image display system |
US10082669B2 (en) | 2011-07-27 | 2018-09-25 | Microsoft Technology Licensing, Llc | Variable-depth stereoscopic display |
US10191515B2 (en) | 2012-03-28 | 2019-01-29 | Microsoft Technology Licensing, Llc | Mobile device light guide display |
US10192358B2 (en) | 2012-12-20 | 2019-01-29 | Microsoft Technology Licensing, Llc | Auto-stereoscopic augmented reality display |
CN109725430A (en) * | 2019-03-06 | 2019-05-07 | 成都工业学院 | A kind of 3 d display device of empty poly- mixing imaging |
US10317677B2 (en) | 2015-02-09 | 2019-06-11 | Microsoft Technology Licensing, Llc | Display system |
US10502876B2 (en) | 2012-05-22 | 2019-12-10 | Microsoft Technology Licensing, Llc | Waveguide optics focus elements |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2018175649A1 (en) | 2017-03-21 | 2018-09-27 | Magic Leap, Inc. | Methods, devices, and systems for illuminating spatial light modulators |
KR102665359B1 (en) | 2017-03-21 | 2024-05-10 | 매직 립, 인코포레이티드 | Low-profile beam splitter |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7369192B2 (en) * | 2004-05-25 | 2008-05-06 | General Electric Company | Methods of making and using optical film, light-diffusing film |
US20080284801A1 (en) * | 2007-05-18 | 2008-11-20 | 3M Innovative Properties Company | Stereoscopic 3d liquid crystal display apparatus with black data insertion |
US7660047B1 (en) * | 2008-09-03 | 2010-02-09 | Microsoft Corporation | Flat panel lens |
US8035771B2 (en) * | 2006-10-06 | 2011-10-11 | 3M Innovative Properties Company | Stereoscopic 3D liquid crystal display apparatus with structured light guide surface |
US20110273906A1 (en) * | 2010-04-16 | 2011-11-10 | Anthony John Nichol | Front illumination device comprising a film-based lightguide |
Family Cites Families (53)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4735495A (en) | 1986-12-12 | 1988-04-05 | General Electric Co. | Light source for liquid crystal display panels utilizing internally reflecting light pipes and integrating sphere |
US5319455A (en) | 1990-09-28 | 1994-06-07 | Ictv Inc. | System for distributing customized commercials to television viewers |
US5806955A (en) | 1992-04-16 | 1998-09-15 | Tir Technologies, Inc. | TIR lens for waveguide injection |
US6463585B1 (en) | 1992-12-09 | 2002-10-08 | Discovery Communications, Inc. | Targeted advertisement using television delivery systems |
EP0793815B1 (en) | 1995-09-22 | 2004-06-09 | Koninklijke Philips Electronics N.V. | Flat-panel picture display device |
US6072551A (en) | 1996-02-14 | 2000-06-06 | Physical Optics Corporation | Backlight apparatus for illuminating a display with controlled light output characteristics |
US5838403A (en) | 1996-02-14 | 1998-11-17 | Physical Optics Corporation | Liquid crystal display system with internally reflecting waveguide for backlighting and non-Lambertian diffusing |
JP3874224B2 (en) | 1997-09-12 | 2007-01-31 | インターナショナル・ビジネス・マシーンズ・コーポレーション | Light guide unit and liquid crystal display device for increasing polarization component |
US6710920B1 (en) | 1998-03-27 | 2004-03-23 | Sanyo Electric Co., Ltd | Stereoscopic display |
US7146627B1 (en) | 1998-06-12 | 2006-12-05 | Metabyte Networks, Inc. | Method and apparatus for delivery of targeted video programming |
TWI228610B (en) | 1998-07-31 | 2005-03-01 | Toshiba Corp | Flat panel display unit |
JP2000106021A (en) | 1998-09-30 | 2000-04-11 | Casio Comput Co Ltd | Surface light source device |
US7260823B2 (en) | 2001-01-11 | 2007-08-21 | Prime Research Alliance E., Inc. | Profiling and identification of television viewers |
US6129444A (en) | 1998-12-10 | 2000-10-10 | L-3 Communications Corporation | Display backlight with white balance compensation |
US6256447B1 (en) | 1998-12-31 | 2001-07-03 | Physical Optics Corporation | Backlight for correcting diagonal line distortion |
KR20010039013A (en) | 1999-10-28 | 2001-05-15 | 최태현 | A light guide plate in a backlight unit for LCD |
GB0024112D0 (en) | 2000-10-03 | 2000-11-15 | Cambridge 3D Display Ltd | Flat panel display |
US6648485B1 (en) | 2000-11-13 | 2003-11-18 | International Business Machines Corporation | Highly collimating tapered light guide for uniform illumination of flat panel displays |
JP2002162912A (en) | 2000-11-24 | 2002-06-07 | Minolta Co Ltd | Sheet lighting system and liquid crystal display |
US6752498B2 (en) | 2001-05-14 | 2004-06-22 | Eastman Kodak Company | Adaptive autostereoscopic display system |
US7631327B2 (en) | 2001-08-08 | 2009-12-08 | Accenture Global Services Gmbh | Enhanced custom content television |
JP4242772B2 (en) | 2001-11-02 | 2009-03-25 | ハネウェル・インターナショナル・インコーポレーテッド | Hollow wedge shaped light guide for rear lighting |
KR100835005B1 (en) | 2001-12-24 | 2008-06-04 | 엘지디스플레이 주식회사 | BackLight Unit |
US7153017B2 (en) | 2002-01-31 | 2006-12-26 | Mitsubishi Rayon Co., Ltd. | Light deflection element and light source apparatus using the same |
WO2003083530A1 (en) | 2002-03-28 | 2003-10-09 | Koninklijke Philips Electronics N.V. | Compact lighting system and display device |
AU2003289149A1 (en) | 2002-12-18 | 2004-07-09 | Sharp Kabushiki Kaisha | Light guide plate, illuminating device using same, area light source and display |
CN1768536B (en) | 2003-03-31 | 2010-06-23 | 皇家飞利浦电子股份有限公司 | Autostereoscopic display |
US7503684B2 (en) | 2003-06-12 | 2009-03-17 | Omron Corporation | Planar light source and image display |
US6834961B1 (en) | 2003-09-12 | 2004-12-28 | Eastman Kodak Company | Autostereoscopic optical apparatus |
IL165376A0 (en) | 2003-12-02 | 2006-01-15 | Electro Optics Ind Ltd | Vehicle display system |
WO2005059874A1 (en) | 2003-12-16 | 2005-06-30 | U-Marketing Intellectual Properties Pte. Ltd | A display system |
DE10359403B4 (en) * | 2003-12-18 | 2005-12-15 | Seereal Technologies Gmbh | Autostereoscopic multi-user display |
US8425102B2 (en) * | 2004-04-30 | 2013-04-23 | Modilis Holdings Llc | Ultrathin lighting element |
GB0410551D0 (en) | 2004-05-12 | 2004-06-16 | Ller Christian M | 3d autostereoscopic display |
JP2005331565A (en) | 2004-05-18 | 2005-12-02 | Seiko Epson Corp | Illumination device, liquid crystal display device, and electronic device |
WO2006026743A1 (en) | 2004-08-31 | 2006-03-09 | Fusion Optix, Inc. | Enhanced light diffusing sheet |
US20060083004A1 (en) | 2004-10-15 | 2006-04-20 | Eastman Kodak Company | Flat-panel area illumination system |
WO2006055872A2 (en) | 2004-11-17 | 2006-05-26 | Fusion Optix, Inc. | Enhanced light fixture |
US7773849B2 (en) * | 2004-12-14 | 2010-08-10 | Oms Displays Ltd. | Device and method for optical resizing and backlighting |
GB0502453D0 (en) | 2005-02-05 | 2005-03-16 | Cambridge Flat Projection | Flat panel lens |
JP4600317B2 (en) * | 2005-03-31 | 2010-12-15 | カシオ計算機株式会社 | Illumination device that emits at least two illumination lights having directivity and display device using the same |
JP4556749B2 (en) | 2005-04-08 | 2010-10-06 | 凸版印刷株式会社 | Light guide plate and display device |
US7528374B2 (en) | 2006-03-03 | 2009-05-05 | Vidar Systems Corporation | Sensing apparatus having optical assembly that collimates emitted light for detection |
US20070274099A1 (en) | 2006-05-25 | 2007-11-29 | Clio Technologies, Inc. | Light expanding system for producing a planar light beam from point light sources |
KR100843287B1 (en) | 2006-07-24 | 2008-07-03 | 주식회사 영실업 | Light controlling structure and the surface lightning device using the same |
KR20080055051A (en) | 2006-12-14 | 2008-06-19 | 엘지전자 주식회사 | Back light unit for display panel |
US7957082B2 (en) | 2007-10-03 | 2011-06-07 | Skc Haas Display Films Co., Ltd. | Turning film having multiple slopes |
DE102009003069A1 (en) | 2009-05-13 | 2010-11-25 | Seereal Technologies S.A. | 3D display with controllable visibility tracker |
US20110044582A1 (en) | 2009-08-21 | 2011-02-24 | Microsoft Corporation | Efficient collimation of light with optical wedge |
US8354806B2 (en) | 2009-08-21 | 2013-01-15 | Microsoft Corporation | Scanning collimation of light via flat panel lamp |
US20110242298A1 (en) | 2009-08-21 | 2011-10-06 | Microsoft Corporation | Private video presentation |
US20110075257A1 (en) * | 2009-09-14 | 2011-03-31 | The Arizona Board Of Regents On Behalf Of The University Of Arizona | 3-Dimensional electro-optical see-through displays |
US20130027772A1 (en) | 2011-07-27 | 2013-01-31 | Microsoft Corporation | Variable-depth stereoscopic display |
-
2011
- 2011-07-27 US US13/192,321 patent/US20130027772A1/en not_active Abandoned
-
2016
- 2016-02-19 US US15/047,859 patent/US10082669B2/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7369192B2 (en) * | 2004-05-25 | 2008-05-06 | General Electric Company | Methods of making and using optical film, light-diffusing film |
US8035771B2 (en) * | 2006-10-06 | 2011-10-11 | 3M Innovative Properties Company | Stereoscopic 3D liquid crystal display apparatus with structured light guide surface |
US8059217B2 (en) * | 2006-10-06 | 2011-11-15 | 3M Innovative Properties Company | Stereoscopic 3D liquid crystal display apparatus with slatted light guide |
US20080284801A1 (en) * | 2007-05-18 | 2008-11-20 | 3M Innovative Properties Company | Stereoscopic 3d liquid crystal display apparatus with black data insertion |
US7660047B1 (en) * | 2008-09-03 | 2010-02-09 | Microsoft Corporation | Flat panel lens |
US20110273906A1 (en) * | 2010-04-16 | 2011-11-10 | Anthony John Nichol | Front illumination device comprising a film-based lightguide |
Cited By (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10082669B2 (en) | 2011-07-27 | 2018-09-25 | Microsoft Technology Licensing, Llc | Variable-depth stereoscopic display |
US20130127839A1 (en) * | 2011-11-17 | 2013-05-23 | Lg Electronics Inc. | Stereoscopic display device and mobile device having the same |
US9324251B2 (en) * | 2011-11-17 | 2016-04-26 | Lg Electronics Inc. | Stereoscopic display device and mobile device having the same |
US9298012B2 (en) | 2012-01-04 | 2016-03-29 | Microsoft Technology Licensing, Llc | Eyebox adjustment for interpupillary distance |
US9779643B2 (en) | 2012-02-15 | 2017-10-03 | Microsoft Technology Licensing, Llc | Imaging structure emitter configurations |
US9297996B2 (en) | 2012-02-15 | 2016-03-29 | Microsoft Technology Licensing, Llc | Laser illumination scanning |
US9726887B2 (en) | 2012-02-15 | 2017-08-08 | Microsoft Technology Licensing, Llc | Imaging structure color conversion |
US9368546B2 (en) | 2012-02-15 | 2016-06-14 | Microsoft Technology Licensing, Llc | Imaging structure with embedded light sources |
US9684174B2 (en) | 2012-02-15 | 2017-06-20 | Microsoft Technology Licensing, Llc | Imaging structure with embedded light sources |
US9578318B2 (en) | 2012-03-14 | 2017-02-21 | Microsoft Technology Licensing, Llc | Imaging structure emitter calibration |
US9807381B2 (en) | 2012-03-14 | 2017-10-31 | Microsoft Technology Licensing, Llc | Imaging structure emitter calibration |
US9854229B2 (en) * | 2012-03-19 | 2017-12-26 | Lg Electronics Inc. | Three-dimensional image processing apparatus and method for adjusting location of sweet spot for displaying multi-view image |
US20130242067A1 (en) * | 2012-03-19 | 2013-09-19 | Dongman Jeong | Three-dimensional image processing apparatus and method for adjusting location of sweet spot for displaying multi-view image |
US11068049B2 (en) * | 2012-03-23 | 2021-07-20 | Microsoft Technology Licensing, Llc | Light guide display and field of view |
US20130249895A1 (en) * | 2012-03-23 | 2013-09-26 | Microsoft Corporation | Light guide display and field of view |
US10388073B2 (en) | 2012-03-28 | 2019-08-20 | Microsoft Technology Licensing, Llc | Augmented reality light guide display |
US9558590B2 (en) | 2012-03-28 | 2017-01-31 | Microsoft Technology Licensing, Llc | Augmented reality light guide display |
US10191515B2 (en) | 2012-03-28 | 2019-01-29 | Microsoft Technology Licensing, Llc | Mobile device light guide display |
US9717981B2 (en) | 2012-04-05 | 2017-08-01 | Microsoft Technology Licensing, Llc | Augmented reality and physical games |
US10478717B2 (en) | 2012-04-05 | 2019-11-19 | Microsoft Technology Licensing, Llc | Augmented reality and physical games |
US10502876B2 (en) | 2012-05-22 | 2019-12-10 | Microsoft Technology Licensing, Llc | Waveguide optics focus elements |
US9581820B2 (en) | 2012-06-04 | 2017-02-28 | Microsoft Technology Licensing, Llc | Multiple waveguide imaging structure |
US10192358B2 (en) | 2012-12-20 | 2019-01-29 | Microsoft Technology Licensing, Llc | Auto-stereoscopic augmented reality display |
US20150039061A1 (en) * | 2013-07-30 | 2015-02-05 | Samsung Display Co., Ltd. | Display apparatus providing light therapy |
US11130001B2 (en) * | 2013-07-30 | 2021-09-28 | Samsung Display Co., Ltd. | Display apparatus including a lens |
US10317677B2 (en) | 2015-02-09 | 2019-06-11 | Microsoft Technology Licensing, Llc | Display system |
US10018844B2 (en) | 2015-02-09 | 2018-07-10 | Microsoft Technology Licensing, Llc | Wearable image display system |
CN109725430A (en) * | 2019-03-06 | 2019-05-07 | 成都工业学院 | A kind of 3 d display device of empty poly- mixing imaging |
Also Published As
Publication number | Publication date |
---|---|
US20160161739A1 (en) | 2016-06-09 |
US10082669B2 (en) | 2018-09-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10082669B2 (en) | Variable-depth stereoscopic display | |
US20220365361A1 (en) | Directional backlight | |
US10025089B2 (en) | Backlight for viewing three-dimensional images from a display from variable viewing angles | |
US9191661B2 (en) | Virtual image display device | |
US8988474B2 (en) | Wide field-of-view virtual image projector | |
JP4813366B2 (en) | 2D / 3D image display | |
US9535537B2 (en) | Hover detection in an interactive display device | |
JP5829798B2 (en) | Video display apparatus and method | |
US20130201094A1 (en) | Virtual image device | |
US11683472B2 (en) | Superstereoscopic display with enhanced off-angle separation | |
US20100079584A1 (en) | 2D/3D switchable autostereoscopic display apparatus and method | |
US20130222353A1 (en) | Prism illumination-optic | |
Travis et al. | Wedge optics in flat panel displays | |
US10750159B2 (en) | 3D image display apparatus and image display method thereof | |
Xia et al. | Time-multiplexed multi-view three-dimensional display with projector array and steering screen | |
KR102142594B1 (en) | Phase control backlight | |
KR102070811B1 (en) | Display apparatus and touch panel | |
US10295835B2 (en) | Stereoscopic display device comprising at least two active scattering panels arranged in different planes each having a scattering function and a transmission function and stereoscopic display method | |
JP5942150B2 (en) | Image display device | |
Zhuang et al. | Directional view method for a time-sequential autostereoscopic display with full resolution | |
CN111308698A (en) | Directional display screen, induction type three-dimensional display device and display method thereof | |
Travis et al. | Flat projection for 3-D | |
KR20130106725A (en) | Integral image system and driving method of the same | |
KR102293367B1 (en) | Display device for displying 2-dimension and 3-dimension images | |
JP2013045073A (en) | Display device, electronic equipment and illumination device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: MICROSOFT CORPORATION, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:LARGE, TIMOTHY ANDREW;REEL/FRAME:026662/0125 Effective date: 20110724 |
|
AS | Assignment |
Owner name: MICROSOFT TECHNOLOGY LICENSING, LLC, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICROSOFT CORPORATION;REEL/FRAME:034544/0001 Effective date: 20141014 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |