Cloud native EDA tools & pre-optimized hardware platforms
Augmented reality (AR) uses technology to combine a simulated environment with a real environment. AR relies on optics to create a simulated environment that annotates or enhances the real environment so that the user can experience them as one environment. The hardware for augmented reality typically includes a computer capable of real-time simulation that synchronizes and maps the simulated to the real environment using a visual output display.
Currently, the AR optics of most interest are see-through head-mounted displays (HMDs), which are also known as see-through near-eye displays or head-up displays (HUDs). Like virtual reality, wearable devices (e.g., haptic gloves) that sense and respond to motions of the user; devices for audio feedback; and trackers for body, head, and eye may be used to interact with the simulation. However, in augmented reality, the user also interacts with objects in the real world.
For augmented reality to work, an optical system must project an image on a transparent display in front of your eyes to overlay the current environment. This may be in the form of a head-mounted display, handheld display (such as a digital tablet), or a mounted display (such as a windshield).
Typically, an optical system for augmented reality includes light sources (display), receivers (eyes), and optical elements (lenses).
Figure 1: Schematic diagram of Prescription AR:
(a) The side view and the beam path of the AR image of the proposed system. The prescription lens works both for vision-correction and for wave-guide of the AR image. Light rays from a micro display refracted by a beam shaping lens enter to the prescription lens through an in-coupling prism and create a magnified virtual image located a distance from the lens.
(b) The detailed diagram for geometric parameters in the Prescription AR.
(c) The 3D diagram of optical components.
Reprinted with permission. ? The Optical Society.
The effects of aberrations to image quality in HMDs are similar to that in other optical systems. Aberrations such as axial chromatic aberration, spherical aberration, coma, astigmatism, and field curvature introduce blur. Aberrations such as distortion, coma, and lateral chromatic aberration induce warping. Aberration control is important in the design of AR HMD optics. An example all-reflective freeform design is shown in Figure 2, where (a) displays the system design, (b) shows the test object, (c) provides the image simulation results of an initial system, and (d) demonstrates the image after optimization. In this example, we can appreciate the power of optimization in reducing aberrations and distortions.
As It is extremely challenging to match both the FOV and the resolution of human eyes, tradeoffs are often made for specific tasks among FOV, weight (volume, number of optical element), resolution, pupil size (eye box), eye clearance, and the size of the micro-displays. Some of the tradeoffs can be addressed with improved technology. A few potential solutions are listed as below: the tradeoffs between FOV and resolution can be addressed with high-resolution insets, partial binocular overlap, spatial tiling, temporal division, and diffraction-order tiling. The tradeoffs between FOV and pupil size can be addressed by duplicating the exit pupil into an array and employing eye-tracking devices.
Figure 2: Simulating an all-reflective freeform design
Also see the VR page for HMDs.
Synopsys provides a complete set of tools to study AR/VR devices. There are multiple software needs for the design of AR optical systems. The optical engineer needs software to create and optimize the imaging system, analyze straylight in the optical path, and designing diffractive optical elements. The mechanical engineer needs a CAD package to draw the system layout and accomplish thermal and structural analysis. The AR system may also require electrical engineer to track the eye motion and send the signal to the optical system.
The workflow:
Once the gratings have been built, the Bidirectional Scattering Distribution Function (BSDF) information and layout files can be exported directly to LightTools to define a surface property. All diffractive properties are included in the RSoft BSDF files, which contain information about how a surface (thin film, patterns, etc.) scatters light.
A Head¡¯s Up Display (HUD) augments a driver¡¯s field of view with an image from a display. Software is needed to model the rays traveling through the windshield, and also to evaluate the quality of the projected image.
CODE V has powerful features for applications in the HUD design space when tackling a wide range of opto-mechanical systems design challenges. Engineers can use this optical design software for CAD visualization and import for ray tracing. CODE V can accommodate new freeform surfaces with flexibility for design degrees of freedom for compact designs. when modeling windshields as the combiner. New freeform surface types allow for enhanced aberration control as display resolution increases at the viewer¡¯s eye (higher density display pixels) combined with more compact form factors.
After design completion, it is important to check the final system performance not only against nominal criteria but also for actual system as-built performance. For this, LightTools is the logical next step for viewer simulation. A reverse ray trace from spectral color objects representing a display image in LightTools shows the projected HUD image for a viewer (onto a model scene). A LightTools simulation can also help uncover any unforeseen issues with stray images or reflections in the system. Also, engineers can use LightTools CAD import and measurement tools to determine:
Eyebox to windshield distance
Approximate angle of incidence on windshield
Windshield to dashboard distance
LightTools simulation of real world optical system performance is an excellent strength for Synopsys Optical 91³Ô¹ÏÍø product users in their engineering and design work.
The key difference from virtual reality is that the former simulates the entire visual environment, while the optics for augmented reality captures the real environment and maps the simulated environment to the real environment, presenting them together using a visual display.
In augmented reality, the display is often see-through to combine the simulated environment with the real environment. In virtual reality (VR), the display only needs to output the simulated environment.
There are a few differences between AR and VR optics.
References:
[1] Jui-Yi Wu and Jonghyun Kim, "," Opt. Express 28, 6225-6241 (2020).