US20060238442A1 - Direct ocular virtual 3D workspace - Google Patents
Direct ocular virtual 3D workspace Download PDFInfo
- Publication number
- US20060238442A1 US20060238442A1 US11/187,771 US18777105A US2006238442A1 US 20060238442 A1 US20060238442 A1 US 20060238442A1 US 18777105 A US18777105 A US 18777105A US 2006238442 A1 US2006238442 A1 US 2006238442A1
- Authority
- US
- United States
- Prior art keywords
- workspace
- image space
- user
- workspace user
- laser
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/0093—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for monitoring data relating to the user, e.g. head-tracking, eye-tracking
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B27/0172—Head mounted characterised by optical features
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B27/0172—Head mounted characterised by optical features
- G02B2027/0174—Head mounted characterised by optical features holographic
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0179—Display position adjusting means not related to the information to be displayed
- G02B2027/0187—Display position adjusting means not related to the information to be displayed slaved to motion of at least a part of the body of the user, e.g. head, eye
Definitions
- This invention relates generally to displays, and more particularly to a laser-based, three-dimensional (3D) display system that generates a holographic-like image in space for simultaneous observation and interaction by a plurality of display system users.
- 3D three-dimensional
- Modern laser-based 3D display systems are capable of providing such tools and resources for use by a single system user, such as that described by John R. Lewis in his article In the Eye of the Beholder , IEEE Spectrum, pp. 24-28, May 2004.
- the present invention is directed to a direct ocular virtual 3D workspace (laser-based 3D display system) that utilizes traditional binocular methods of image composition to produce a 3D workspace in the observers' psyche.
- This direct ocular virtual 3D workspace is implemented such that in moving environments, such as on a ship or vehicle, the immersion of the individual into the display environment will not result in vertigo and/or motion sickness.
- This is resolved by utilizing a system that accomplishes eye-tracking, and produces the display image within the eye by means of laser projection. Two images can thus be projected into an individual's eyes to produce a 3D holographic-like image space. The individual will still maintain partial vision, resulting in a holographic-like display in space that the observer can look through.
- the images provided to multiple collaborators/users are geometrically corrected for each collaborator, allowing multiple individuals to see the same 3D image space from their particular viewing location without overwriting other visual data cues. This allows an individual collaborator to point within the image space such that other collaborators could see the pointing finger within their own 3D image space; and all parties would see the same item identified by the individual collaborator.
- a feature of the direct ocular virtual 3D workspace is the generation of a rich three-dimensional information display.
- Another feature of the direct ocular virtual 3D workspace is the provision of user interaction with systems and other users in a 3D environment.
- Yet another feature of the direct ocular virtual 3D workspace is enhanced situational awareness.
- Still another feature of the direct ocular virtual 3D workspace is improved safety/reduced collateral damage.
- Still another feature of the direct ocular virtual 3D workspace is more effective planning and execution.
- Still another feature of the direct ocular virtual 3D workspace is improved target detection, tracking, classification, identification and targeting.
- FIG. 1 depicts a pedestal assembly having an integral laser projector to implement a direct ocular virtual 3D workspace
- FIG. 2 depicts the pedestal assembly shown in FIG. 1 generating a holographic-like image in space
- FIG. 3 illustrates a side view of a two user direct ocular virtual 3D workspace layout including one pair of laser projectors such as that shown in FIG. 1 according to one embodiment of the present invention
- FIGS. 4A, 4B and 4 C depict different views of a workspace having stereoscopic displays for sensing and tracking immersion of an external object into the workspace.
- a pedestal assembly 10 is configured with a laser projector 20 to implement a direct ocular virtual 3D workspace.
- Pedestal assembly 10 utilizes traditional binocular methods of image composition to produce a 3D workspace in the observer's psyche.
- the immersion of the individual into the display environment can cause vertigo and motion sickness, as stated herein before.
- Pedestal assembly 10 resolves this undesirable characteristic by use of eye-tracking, described in further detail herein below with reference to FIG. 3 , to produce the display image within the observer's eye by means of laser projection. Two images can thus be projected into an individual's eye to produce a 3D image space. The individual will still maintain partial vision, resulting in a holographic-like display in space that the observer can look through.
- FIG. 2 depicts the pedestal assembly 10 shown in FIG. 1 generating a holographic-like image 12 in space. Due to eye-tracking, as discussed above, an individual can move to a different point around the pedestal assembly 10 to view the same 3D image space from a different viewing location without overwriting other visual data cues.
- Current techniques for achieving the aforesaid feature employ an immersive environment such as the CAVE system technology developed by Iowa State University, or an individual display, wherein pointing is achieved via a pointing device such as a mouse, trackball, etc.
- the CAVE system technology is undesirable since movement of an individual within the CAVE system environment causes disruption of the resultant holographic-like image in space.
- FIG. 3 a side view of a two user direct ocular virtual 3D workspace system layout 100 can be seen to include one pair of laser projector assemblies 20 such as the one shown in FIG. 1 according to one embodiment of the present invention.
- the two laser projector assemblies 20 are configured with optical transceivers and tracking software to continuously track eye movement and location of the observers 102 , 104 .
- a rotational gimbal apparatus for example, can be employed to provide the requisite movement of the laser projector assemblies 20 .
- Eye movement and location may, for example, be accomplished using an optical transceiver in association with each laser projector 20 to generate positional signals.
- a second optical transceiver 108 , 110 can then be employed via each workspace user to receive and retransmit signals back to the respective optical transmitter.
- Appropriate algorithmic tracking and calibration software including handshaking software between the laser projector transceiver and workspace user transceiver 108 , 110 , then functions to establish a calibrated wireless link between the laser projector 20 transceivers and the respective workspace user transceiver 108 , 110 , thus allowing the laser projector 20 to track the respective workspace user eye movement and location.
- the image 106 provided to multiple observers (collaborators) 102 , 104 is geometrically corrected for each collaborator, allowing multiple individuals to see the same 3D image space from their particular viewing location without overwriting other visual data cues.
- the algorithmic geometric correction can be implemented, for example, via software that may be an integral part of the workspace user's transceiver, implemented via micro-electro-mechanical system (MEMS) technology.
- MEMS micro-electro-mechanical system
- the requisite eye-tracking can be accomplished using any number of known tracking technologies such as, but not limited to, for example, optical mirrors, gimbal, gyro, phased-array, MEMS technology, and the like familiar to those skilled in the particular arts employed to implement eye-tracking.
- the selected eye-tracking technology is then configured, using appropriate algorithmic software, to geometrically correct the image provided to each collaborator 102 , 104 , allowing multiple individuals to see the same 3D image space from their particular viewing location without overwriting other visual data cues, as stated herein before.
- An individual collaborator can then point a finger within the image space, and other collaborators could see the pointing finger within their own 3D image space such that all parties would see the same item identified.
- the two user system layout 100 employs more than one projector assembly 20 configured for simultaneous use by multiple observers (collaborators) 102 , 104 , the individual collaborators 102 , 104 can communicate with one another by means of body language, tone and content not available using known laser-based 3D display system technology.
- the foregoing techniques also allow remote collaborators to be included in the environment, allowing true remote collaboration between individual sites. Further, allowing additional visual cues not currently present in an immersive environment will lessen the likelihood of vertigo and motion sickness that is inherently associated with known immersive environments.
- a direct ocular virtual 3D workspace 100 employs multiple projectors 20 within a table-like pedestal assembly 10 .
- the projectors 20 utilize traditional technology to provide binocular visual data for each user utilizing eye position scanning and projection to the location of the user's iris. Since the workspace 100 employs multiple output devices (projectors) 20 , each device is dedicated to a specific user location to ensure specific user data is geometrically correct with respect to other users of the workspace 100 . This technique allows collaboration within the 3-dimensional viewing space, with semi-opaque images provided to allow direct interface between users of the system.
- Direct interfacing may employ any number of techniques such as, but not limited to, using stereoscopic displays in association with the pedestal such that when a collaborator's finger is encased in a white glove which is then inserted into the holographic-like image space and between the stereoscopic displays, the appropriate algorithmic software will synchronize the finger location data with the holographic-like image data to allow direct interfacing between the users of the workspace 100 .
- FIG. 4 illustrates different views showing one embodiment of a workspace 200 having stereoscopic displays 202 , 204 .
- the present invention presents a significant advancement in the art of workstation design. Further, this invention has been described in considerable detail in order to provide those skilled in the ocular arts with the information needed to apply the novel principles and to construct and use such specialized components as are required.
Landscapes
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Optics & Photonics (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
Description
- This application claims the benefit, under 35 U.S.C. §119(e)(1), of U.S. Provisional Application No. 60/590,898, entitled Direct Ocular Virtual 3D Workspace, filed Jul. 23, 2004 by Brian L. Uhlhorn, Robert J. Monson and Steven H. Ernst.
- 1. Field of the Invention
- This invention relates generally to displays, and more particularly to a laser-based, three-dimensional (3D) display system that generates a holographic-like image in space for simultaneous observation and interaction by a plurality of display system users.
- 2. Description of the Prior Art
- Military systems are becoming more complex with a larger array of information being supplied at a faster rate to fewer decision makers (war fighters) at all levels of the decision chain. The war fighter, for example, needs tools and resources that provide the sufficient and requisite leverage to respond to this challenge. Such tools and resources are presently limited to use of 1) an immersive environment such as the CAVE system designed and developed by Iowa State University, or 2) individual displays for each participant, wherein collaboration is achieved via a non-intuitive pointing device within the display system, e.g. mouse, track, etc.
- Modern laser-based 3D display systems are capable of providing such tools and resources for use by a single system user, such as that described by John R. Lewis in his article In the Eye of the Beholder, IEEE Spectrum, pp. 24-28, May 2004. A need exists however, for a display system that allows simultaneous immersion of multiple users into a holographic-like display environment without disruption of the holographic-like image produced by the laser-based 3D display system, or without causing individual user vertigo or motion sickness.
- In view of the foregoing background, it would be extremely beneficial and advantageous to provide a next generation interactive display system (direct ocular virtual 3D workspace) that allows simultaneous observation and interaction via immersion of multiple users into the display environment without overwriting other visual data cues.
- The present invention is directed to a direct ocular virtual 3D workspace (laser-based 3D display system) that utilizes traditional binocular methods of image composition to produce a 3D workspace in the observers' psyche. This direct ocular virtual 3D workspace is implemented such that in moving environments, such as on a ship or vehicle, the immersion of the individual into the display environment will not result in vertigo and/or motion sickness. This is resolved by utilizing a system that accomplishes eye-tracking, and produces the display image within the eye by means of laser projection. Two images can thus be projected into an individual's eyes to produce a 3D holographic-like image space. The individual will still maintain partial vision, resulting in a holographic-like display in space that the observer can look through. The images provided to multiple collaborators/users are geometrically corrected for each collaborator, allowing multiple individuals to see the same 3D image space from their particular viewing location without overwriting other visual data cues. This allows an individual collaborator to point within the image space such that other collaborators could see the pointing finger within their own 3D image space; and all parties would see the same item identified by the individual collaborator.
- A feature of the direct ocular virtual 3D workspace is the generation of a rich three-dimensional information display.
- Another feature of the direct ocular virtual 3D workspace is the provision of user interaction with systems and other users in a 3D environment.
- Yet another feature of the direct ocular virtual 3D workspace is enhanced situational awareness.
- Still another feature of the direct ocular virtual 3D workspace is improved safety/reduced collateral damage.
- Still another feature of the direct ocular virtual 3D workspace is more effective planning and execution.
- Still another feature of the direct ocular virtual 3D workspace is improved target detection, tracking, classification, identification and targeting.
- Other aspects, features and advantages of the present invention will be readily appreciated as the invention becomes better understood by reference to the following detailed description when considered in connection with the accompanying drawing figures wherein:
-
FIG. 1 depicts a pedestal assembly having an integral laser projector to implement a direct ocular virtual 3D workspace; -
FIG. 2 depicts the pedestal assembly shown inFIG. 1 generating a holographic-like image in space; -
FIG. 3 illustrates a side view of a two user direct ocular virtual 3D workspace layout including one pair of laser projectors such as that shown inFIG. 1 according to one embodiment of the present invention; and -
FIGS. 4A, 4B and 4C depict different views of a workspace having stereoscopic displays for sensing and tracking immersion of an external object into the workspace. - While the above-identified drawing figures set forth particular embodiments, other embodiments of the present invention are also contemplated, as noted in the discussion. In all cases, this disclosure presents illustrated embodiments of the present invention by way of representation and not limitation. Numerous other modifications and embodiments can be devised by those skilled in the art which fall within the scope and spirit of the principles of this invention.
- Looking now at
FIG. 1 , apedestal assembly 10 is configured with alaser projector 20 to implement a direct ocular virtual 3D workspace.Pedestal assembly 10 utilizes traditional binocular methods of image composition to produce a 3D workspace in the observer's psyche. In moving environments, such as on a ship or vehicle, the immersion of the individual into the display environment can cause vertigo and motion sickness, as stated herein before.Pedestal assembly 10 resolves this undesirable characteristic by use of eye-tracking, described in further detail herein below with reference toFIG. 3 , to produce the display image within the observer's eye by means of laser projection. Two images can thus be projected into an individual's eye to produce a 3D image space. The individual will still maintain partial vision, resulting in a holographic-like display in space that the observer can look through. -
FIG. 2 depicts thepedestal assembly 10 shown inFIG. 1 generating a holographic-like image 12 in space. Due to eye-tracking, as discussed above, an individual can move to a different point around thepedestal assembly 10 to view the same 3D image space from a different viewing location without overwriting other visual data cues. Current techniques for achieving the aforesaid feature employ an immersive environment such as the CAVE system technology developed by Iowa State University, or an individual display, wherein pointing is achieved via a pointing device such as a mouse, trackball, etc. The CAVE system technology is undesirable since movement of an individual within the CAVE system environment causes disruption of the resultant holographic-like image in space. - Moving now to
FIG. 3 , a side view of a two user direct ocular virtual 3D workspace system layout 100 can be seen to include one pair oflaser projector assemblies 20 such as the one shown inFIG. 1 according to one embodiment of the present invention. The two laser projector assemblies 20 are configured with optical transceivers and tracking software to continuously track eye movement and location of the 102, 104.observers - Many techniques are suitable to implement the requisite vertical and horizontal rotation of the laser projector assemblies 20 to accommodate the eye tracking. A rotational gimbal apparatus, for example, can be employed to provide the requisite movement of the laser projector assemblies 20.
- Several known techniques can be employed to accomplish the eye tracking. Eye movement and location may, for example, be accomplished using an optical transceiver in association with each
laser projector 20 to generate positional signals. A secondoptical transceiver 108, 110, can then be employed via each workspace user to receive and retransmit signals back to the respective optical transmitter. Appropriate algorithmic tracking and calibration software, including handshaking software between the laser projector transceiver andworkspace user transceiver 108, 110, then functions to establish a calibrated wireless link between thelaser projector 20 transceivers and the respectiveworkspace user transceiver 108, 110, thus allowing thelaser projector 20 to track the respective workspace user eye movement and location. - Traditional binocular methods of image composition are employed to produce a 3D workspace in each observer's 102, 104 psyche by generating the display within each observer's 102, 104 eye by means of laser projection. Two images, as stated herein before, are projected into each observer's eye to produce a 3D image space. Each
102, 104 will still maintain partial vision, resulting in a common holographic-like display in space that theobserver 102, 104 can look through.observer - The
image 106 provided to multiple observers (collaborators) 102, 104 is geometrically corrected for each collaborator, allowing multiple individuals to see the same 3D image space from their particular viewing location without overwriting other visual data cues. The algorithmic geometric correction can be implemented, for example, via software that may be an integral part of the workspace user's transceiver, implemented via micro-electro-mechanical system (MEMS) technology. The requisite eye-tracking can be accomplished using any number of known tracking technologies such as, but not limited to, for example, optical mirrors, gimbal, gyro, phased-array, MEMS technology, and the like familiar to those skilled in the particular arts employed to implement eye-tracking. The selected eye-tracking technology is then configured, using appropriate algorithmic software, to geometrically correct the image provided to each 102, 104, allowing multiple individuals to see the same 3D image space from their particular viewing location without overwriting other visual data cues, as stated herein before. An individual collaborator can then point a finger within the image space, and other collaborators could see the pointing finger within their own 3D image space such that all parties would see the same item identified.collaborator - Because the two user system layout 100 employs more than one
projector assembly 20 configured for simultaneous use by multiple observers (collaborators) 102, 104, the 102, 104 can communicate with one another by means of body language, tone and content not available using known laser-based 3D display system technology. The foregoing techniques also allow remote collaborators to be included in the environment, allowing true remote collaboration between individual sites. Further, allowing additional visual cues not currently present in an immersive environment will lessen the likelihood of vertigo and motion sickness that is inherently associated with known immersive environments.individual collaborators - In summary explanation, a direct ocular virtual 3D workspace 100 employs
multiple projectors 20 within a table-like pedestal assembly 10. Theprojectors 20 utilize traditional technology to provide binocular visual data for each user utilizing eye position scanning and projection to the location of the user's iris. Since the workspace 100 employs multiple output devices (projectors) 20, each device is dedicated to a specific user location to ensure specific user data is geometrically correct with respect to other users of the workspace 100. This technique allows collaboration within the 3-dimensional viewing space, with semi-opaque images provided to allow direct interface between users of the system. - Direct interfacing may employ any number of techniques such as, but not limited to, using stereoscopic displays in association with the pedestal such that when a collaborator's finger is encased in a white glove which is then inserted into the holographic-like image space and between the stereoscopic displays, the appropriate algorithmic software will synchronize the finger location data with the holographic-like image data to allow direct interfacing between the users of the workspace 100.
FIG. 4 illustrates different views showing one embodiment of aworkspace 200 having 202, 204.stereoscopic displays - In view of the above, it can be seen the present invention presents a significant advancement in the art of workstation design. Further, this invention has been described in considerable detail in order to provide those skilled in the ocular arts with the information needed to apply the novel principles and to construct and use such specialized components as are required.
- It should be apparent that the present invention represents a significant departure from the prior art in construction and operation. However, while particular embodiments of the present invention have been described herein in detail, it is to be understood that various alterations, modifications and substitutions can be made therein without departing in any way from the spirit and scope of the present invention, as defined in the claims which follow.
- The complete disclosure of all patents, patent documents, and publications are incorporated herein by reference as if individually incorporated.
Claims (18)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US11/187,771 US7538746B2 (en) | 2004-07-23 | 2005-07-22 | Direct ocular virtual 3D workspace |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US59089804P | 2004-07-23 | 2004-07-23 | |
| US11/187,771 US7538746B2 (en) | 2004-07-23 | 2005-07-22 | Direct ocular virtual 3D workspace |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20060238442A1 true US20060238442A1 (en) | 2006-10-26 |
| US7538746B2 US7538746B2 (en) | 2009-05-26 |
Family
ID=37186332
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US11/187,771 Expired - Fee Related US7538746B2 (en) | 2004-07-23 | 2005-07-22 | Direct ocular virtual 3D workspace |
Country Status (1)
| Country | Link |
|---|---|
| US (1) | US7538746B2 (en) |
Cited By (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20110161998A1 (en) * | 2009-12-31 | 2011-06-30 | Motorola, Inc. | Systems and Methods Providing Content on a Display Based Upon Facial Recognition of a Viewer |
| US20130038599A1 (en) * | 2011-08-11 | 2013-02-14 | Aaron I. Krakowski | System and method for motion sickness minimization using integration of attended and unattended datastreams |
| US9118911B2 (en) | 2013-02-07 | 2015-08-25 | Delphi Technologies, Inc. | Variable disparity three-dimensional (3D) display system and method of operating the same |
| US9265458B2 (en) | 2012-12-04 | 2016-02-23 | Sync-Think, Inc. | Application of smooth pursuit cognitive testing paradigms to clinical drug development |
| US9380976B2 (en) | 2013-03-11 | 2016-07-05 | Sync-Think, Inc. | Optical neuroinformatics |
| US20180063205A1 (en) * | 2016-08-30 | 2018-03-01 | Augre Mixed Reality Technologies, Llc | Mixed reality collaboration |
| US11631224B2 (en) | 2016-11-21 | 2023-04-18 | Hewlett-Packard Development Company, L.P. | 3D immersive visualization of a radial array |
Families Citing this family (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20090109215A1 (en) * | 2007-10-31 | 2009-04-30 | Fein Gene S | Method and apparatus for user interface communication with an image manipulator |
| US8477098B2 (en) | 2007-10-31 | 2013-07-02 | Gene S. Fein | Method and apparatus for user interface of input devices |
| US8885882B1 (en) | 2011-07-14 | 2014-11-11 | The Research Foundation For The State University Of New York | Real time eye tracking for human computer interaction |
Citations (8)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US5661667A (en) * | 1994-03-14 | 1997-08-26 | Virtek Vision Corp. | 3D imaging using a laser projector |
| US5781229A (en) * | 1997-02-18 | 1998-07-14 | Mcdonnell Douglas Corporation | Multi-viewer three dimensional (3-D) virtual display system and operating method therefor |
| US5855425A (en) * | 1996-07-19 | 1999-01-05 | Sanyo Electric Co., Ltd. | Stereoscopic display |
| US6278418B1 (en) * | 1995-12-29 | 2001-08-21 | Kabushiki Kaisha Sega Enterprises | Three-dimensional imaging system, game device, method for same and recording medium |
| US20020135738A1 (en) * | 2001-01-22 | 2002-09-26 | Eastman Kodak Company | Image display system with body position compensation |
| US20030043146A1 (en) * | 2000-06-30 | 2003-03-06 | Yoshifumi Kitamura | Multiple sharing type display device |
| US6937210B1 (en) * | 2002-11-06 | 2005-08-30 | The United States Of America As Represented By The Secretary Of Commerce | Projecting images on a sphere |
| US20070127121A1 (en) * | 2003-10-21 | 2007-06-07 | Bart Maximus | Method and device for performing stereoscopic image display based on color selective filters |
-
2005
- 2005-07-22 US US11/187,771 patent/US7538746B2/en not_active Expired - Fee Related
Patent Citations (8)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US5661667A (en) * | 1994-03-14 | 1997-08-26 | Virtek Vision Corp. | 3D imaging using a laser projector |
| US6278418B1 (en) * | 1995-12-29 | 2001-08-21 | Kabushiki Kaisha Sega Enterprises | Three-dimensional imaging system, game device, method for same and recording medium |
| US5855425A (en) * | 1996-07-19 | 1999-01-05 | Sanyo Electric Co., Ltd. | Stereoscopic display |
| US5781229A (en) * | 1997-02-18 | 1998-07-14 | Mcdonnell Douglas Corporation | Multi-viewer three dimensional (3-D) virtual display system and operating method therefor |
| US20030043146A1 (en) * | 2000-06-30 | 2003-03-06 | Yoshifumi Kitamura | Multiple sharing type display device |
| US20020135738A1 (en) * | 2001-01-22 | 2002-09-26 | Eastman Kodak Company | Image display system with body position compensation |
| US6937210B1 (en) * | 2002-11-06 | 2005-08-30 | The United States Of America As Represented By The Secretary Of Commerce | Projecting images on a sphere |
| US20070127121A1 (en) * | 2003-10-21 | 2007-06-07 | Bart Maximus | Method and device for performing stereoscopic image display based on color selective filters |
Cited By (9)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20110161998A1 (en) * | 2009-12-31 | 2011-06-30 | Motorola, Inc. | Systems and Methods Providing Content on a Display Based Upon Facial Recognition of a Viewer |
| US8589968B2 (en) | 2009-12-31 | 2013-11-19 | Motorola Mobility Llc | Systems and methods providing content on a display based upon facial recognition of a viewer |
| US20130038599A1 (en) * | 2011-08-11 | 2013-02-14 | Aaron I. Krakowski | System and method for motion sickness minimization using integration of attended and unattended datastreams |
| US9123143B2 (en) * | 2011-08-11 | 2015-09-01 | Aaron I. Krakowski | System and method for motion sickness minimization using integration of attended and unattended datastreams |
| US9265458B2 (en) | 2012-12-04 | 2016-02-23 | Sync-Think, Inc. | Application of smooth pursuit cognitive testing paradigms to clinical drug development |
| US9118911B2 (en) | 2013-02-07 | 2015-08-25 | Delphi Technologies, Inc. | Variable disparity three-dimensional (3D) display system and method of operating the same |
| US9380976B2 (en) | 2013-03-11 | 2016-07-05 | Sync-Think, Inc. | Optical neuroinformatics |
| US20180063205A1 (en) * | 2016-08-30 | 2018-03-01 | Augre Mixed Reality Technologies, Llc | Mixed reality collaboration |
| US11631224B2 (en) | 2016-11-21 | 2023-04-18 | Hewlett-Packard Development Company, L.P. | 3D immersive visualization of a radial array |
Also Published As
| Publication number | Publication date |
|---|---|
| US7538746B2 (en) | 2009-05-26 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US20220328021A1 (en) | Eye mounted displays and eye tracking systems, with toroidal battery | |
| Youngblut et al. | Review of Virtual Environment Interface Technology. | |
| US11156843B2 (en) | End-to-end artificial reality calibration testing | |
| Azuma | A survey of augmented reality | |
| US6614407B2 (en) | Virtual display system and method | |
| US8786675B2 (en) | Systems using eye mounted displays | |
| US7804507B2 (en) | Face-mounted display apparatus for mixed reality environment | |
| US9122321B2 (en) | Collaboration environment using see through displays | |
| US20200142480A1 (en) | Immersive displays | |
| WO2020140078A1 (en) | Systems and methods for virtual and augmented reality | |
| EP3729166B1 (en) | Head-mountable apparatus and methods | |
| US11695913B1 (en) | Mixed reality system | |
| CN116778120A (en) | Augmented reality display system | |
| CN102540464A (en) | Head-mounted display device which provides surround video | |
| US7538746B2 (en) | Direct ocular virtual 3D workspace | |
| CN118001116B (en) | Head-mounted display device for vision training and vision training method | |
| JP2018007180A (en) | Image display device, image display method and image display program | |
| US20260037059A1 (en) | Systems and methods for virtual and augmented reality | |
| CN118631982A (en) | Line-of-sight (LoS) communication capabilities for near-eye display devices |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: LOCKHEED MARTIN CORPORATION, MARYLAND Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:UHLHORN, BRIAN L.;MONSON, ROBERT J.;ERNST, STEVEN H.;REEL/FRAME:017428/0644 Effective date: 20060320 |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
| FPAY | Fee payment |
Year of fee payment: 4 |
|
| FPAY | Fee payment |
Year of fee payment: 8 |
|
| FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
| FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20210526 |