<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://vrarwiki.com/index.php?action=history&amp;feed=atom&amp;title=Stereoscopic_rendering</id>
	<title>Stereoscopic rendering - Revision history</title>
	<link rel="self" type="application/atom+xml" href="https://vrarwiki.com/index.php?action=history&amp;feed=atom&amp;title=Stereoscopic_rendering"/>
	<link rel="alternate" type="text/html" href="https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;action=history"/>
	<updated>2026-04-14T04:20:05Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.43.0</generator>
	<entry>
		<id>https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;diff=36673&amp;oldid=prev</id>
		<title>Xinreality at 00:19, 28 October 2025</title>
		<link rel="alternate" type="text/html" href="https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;diff=36673&amp;oldid=prev"/>
		<updated>2025-10-28T00:19:41Z</updated>

		<summary type="html">&lt;p&gt;&lt;/p&gt;
&lt;table style=&quot;background-color: #fff; color: #202122;&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;Revision as of 00:19, 28 October 2025&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l21&quot;&gt;Line 21:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 21:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[File:stereoscopic rendering2.jpg|300px|right]]&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[File:stereoscopic rendering2.jpg|300px|right]]&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&#039;&#039;&#039;Stereoscopic rendering&#039;&#039;&#039; is the foundational [[computer graphics]] technique that creates the perception of three-dimensional depth in [[virtual reality]] (VR) and [[augmented reality]] (AR) systems by generating two slightly different images from distinct viewpoints corresponding to the left and right eyes.&amp;lt;ref name=&quot;arm2021&quot;&amp;gt;ARM Software. &quot;Introduction to Stereo Rendering - VR SDK for Android.&quot; ARM Developer Documentation, 2021. https://arm-software.github.io/vr-sdk-for-android/IntroductionToStereoRendering.html&amp;lt;/ref&amp;gt; This technique exploits [[binocular disparity]]&lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;—the &lt;/del&gt;horizontal displacement between corresponding points in the two &lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;images—enabling &lt;/del&gt;the [[visual cortex]] to reconstruct depth information through [[stereopsis]], the same process human eyes use to perceive the real world.&amp;lt;ref name=&quot;numberanalytics2024&quot;&amp;gt;Number Analytics. &quot;Stereoscopy in VR: A Comprehensive Guide.&quot; 2024. https://www.numberanalytics.com/blog/ultimate-guide-stereoscopy-vr-ar-development&amp;lt;/ref&amp;gt; By delivering two offset images (one per eye) that the brain combines into a single scene, stereoscopic rendering produces an illusion of depth that mimics natural [[binocular vision]].&amp;lt;ref name=&quot;drawandcode&quot;&amp;gt;Draw &amp;amp; Code. &quot;What Is Stereoscopic VR Technology.&quot; January 23, 2024. https://drawandcode.com/learning-zone/what-is-stereoscopic-vr-technology/&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&#039;&#039;&#039;Stereoscopic rendering&#039;&#039;&#039; is the foundational [[computer graphics]] technique that creates the perception of three-dimensional depth in [[virtual reality]] (VR) and [[augmented reality]] (AR) systems by generating two slightly different images from distinct viewpoints corresponding to the left and right eyes.&amp;lt;ref name=&quot;arm2021&quot;&amp;gt;ARM Software. &quot;Introduction to Stereo Rendering - VR SDK for Android.&quot; ARM Developer Documentation, 2021. https://arm-software.github.io/vr-sdk-for-android/IntroductionToStereoRendering.html&amp;lt;/ref&amp;gt; This technique exploits [[binocular disparity]]&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;, the &lt;/ins&gt;horizontal displacement between corresponding points in the two &lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;images, enabling &lt;/ins&gt;the [[visual cortex]] to reconstruct depth information through [[stereopsis]], the same process human eyes use to perceive the real world.&amp;lt;ref name=&quot;numberanalytics2024&quot;&amp;gt;Number Analytics. &quot;Stereoscopy in VR: A Comprehensive Guide.&quot; 2024. https://www.numberanalytics.com/blog/ultimate-guide-stereoscopy-vr-ar-development&amp;lt;/ref&amp;gt; By delivering two offset images (one per eye) that the brain combines into a single scene, stereoscopic rendering produces an illusion of depth that mimics natural [[binocular vision]].&amp;lt;ref name=&quot;drawandcode&quot;&amp;gt;Draw &amp;amp; Code. &quot;What Is Stereoscopic VR Technology.&quot; January 23, 2024. https://drawandcode.com/learning-zone/what-is-stereoscopic-vr-technology/&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The approach doubles computational requirements compared to traditional rendering but delivers the immersive depth perception that defines modern VR experiences, powering a $15.9 billion industry serving 171 million users worldwide as of 2024.&amp;lt;ref name=&amp;quot;mordor2024&amp;quot;&amp;gt;Mordor Intelligence. &amp;quot;Virtual Reality (VR) Market Size, Report, Share &amp;amp; Growth Trends 2025-2030.&amp;quot; 2024. https://www.mordorintelligence.com/industry-reports/virtual-reality-market&amp;lt;/ref&amp;gt; Unlike monoscopic imagery (showing the same image to both eyes), stereoscopic rendering presents each eye with a slightly different perspective, closely matching how humans view the real world and thereby greatly enhancing the sense of presence and realism in VR/AR.&amp;lt;ref name=&amp;quot;borisfx2024&amp;quot;&amp;gt;Boris FX. &amp;quot;Monoscopic vs Stereoscopic 360 VR: Key Differences.&amp;quot; 2024. https://borisfx.com/blog/monoscopic-vs-stereoscopic-360-vr-key-differences/&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The approach doubles computational requirements compared to traditional rendering but delivers the immersive depth perception that defines modern VR experiences, powering a $15.9 billion industry serving 171 million users worldwide as of 2024.&amp;lt;ref name=&amp;quot;mordor2024&amp;quot;&amp;gt;Mordor Intelligence. &amp;quot;Virtual Reality (VR) Market Size, Report, Share &amp;amp; Growth Trends 2025-2030.&amp;quot; 2024. https://www.mordorintelligence.com/industry-reports/virtual-reality-market&amp;lt;/ref&amp;gt; Unlike monoscopic imagery (showing the same image to both eyes), stereoscopic rendering presents each eye with a slightly different perspective, closely matching how humans view the real world and thereby greatly enhancing the sense of presence and realism in VR/AR.&amp;lt;ref name=&amp;quot;borisfx2024&amp;quot;&amp;gt;Boris FX. &amp;quot;Monoscopic vs Stereoscopic 360 VR: Key Differences.&amp;quot; 2024. https://borisfx.com/blog/monoscopic-vs-stereoscopic-360-vr-key-differences/&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l93&quot;&gt;Line 93:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 93:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;=== Computer Graphics Era ===&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;=== Computer Graphics Era ===&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Computer-generated stereoscopy began with [[Ivan Sutherland]]&#039;s 1968 head-mounted display at [[Harvard University]], nicknamed the &quot;[[Sword of Damocles (virtual reality)|Sword of Damocles]]&quot; due to its unwieldy overhead suspension system. This wireframe graphics prototype established the technical &lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;template—&lt;/del&gt;[[head tracking]], stereoscopic displays, and real-time &lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;rendering—that &lt;/del&gt;would define VR development for decades.&amp;lt;ref name=&quot;nextgen2024&quot;&amp;gt;Nextgeninvent. &quot;Virtual Reality&#039;s Evolution From Science Fiction to Mainstream Technology.&quot; 2024. https://nextgeninvent.com/blogs/the-evolution-of-virtual-reality/&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Computer-generated stereoscopy began with [[Ivan Sutherland]]&#039;s 1968 head-mounted display at [[Harvard University]], nicknamed the &quot;[[Sword of Damocles (virtual reality)|Sword of Damocles]]&quot; due to its unwieldy overhead suspension system. This wireframe graphics prototype established the technical &lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;template, &lt;/ins&gt;[[head tracking]], stereoscopic displays, and real-time &lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;rendering, that &lt;/ins&gt;would define VR development for decades.&amp;lt;ref name=&quot;nextgen2024&quot;&amp;gt;Nextgeninvent. &quot;Virtual Reality&#039;s Evolution From Science Fiction to Mainstream Technology.&quot; 2024. https://nextgeninvent.com/blogs/the-evolution-of-virtual-reality/&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The gaming industry drove early consumer adoption with [[Sega]]&amp;#039;s SubRoc-3D in 1982, the world&amp;#039;s first commercial stereoscopic video game featuring an active shutter 3D system jointly developed with [[Matsushita Electric Industrial Co.|Matsushita]].&amp;lt;ref name=&amp;quot;siggraph2024&amp;quot;&amp;gt;ACM SIGGRAPH. &amp;quot;Remember Stereo 3D on the PC? Have You Ever Wondered What Happened to It?&amp;quot; 2024. https://blog.siggraph.org/2024/10/stereo-3d-pc-history-decline.html/&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The gaming industry drove early consumer adoption with [[Sega]]&amp;#039;s SubRoc-3D in 1982, the world&amp;#039;s first commercial stereoscopic video game featuring an active shutter 3D system jointly developed with [[Matsushita Electric Industrial Co.|Matsushita]].&amp;lt;ref name=&amp;quot;siggraph2024&amp;quot;&amp;gt;ACM SIGGRAPH. &amp;quot;Remember Stereo 3D on the PC? Have You Ever Wondered What Happened to It?&amp;quot; 2024. https://blog.siggraph.org/2024/10/stereo-3d-pc-history-decline.html/&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l99&quot;&gt;Line 99:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 99:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;=== Modern VR Revolution ===&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;=== Modern VR Revolution ===&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The modern VR revolution began with [[Palmer Luckey]]&#039;s 2012 [[Oculus Rift]] [[Kickstarter]] campaign, which raised $2.5 million. [[Facebook]]&#039;s $2 billion acquisition of [[Oculus VR|Oculus]] in 2014 validated the market potential. The watershed 2016 launches of the [[Oculus Rift#Consumer version|Oculus Rift CV1]] and [[HTC Vive]]&lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;—offering &lt;/del&gt;2160×1200 combined resolution at 90Hz with [[room-scale tracking]]&lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;—established &lt;/del&gt;the technical baseline for modern VR.&amp;lt;ref name=&quot;cavendish2024&quot;&amp;gt;Cavendishprofessionals. &quot;The Evolution of VR and AR in Gaming: A Historical Perspective.&quot; 2024. https://www.cavendishprofessionals.com/the-evolution-of-vr-and-ar-in-gaming-a-historical-perspective/&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The modern VR revolution began with [[Palmer Luckey]]&#039;s 2012 [[Oculus Rift]] [[Kickstarter]] campaign, which raised $2.5 million. [[Facebook]]&#039;s $2 billion acquisition of [[Oculus VR|Oculus]] in 2014 validated the market potential. The watershed 2016 launches of the [[Oculus Rift#Consumer version|Oculus Rift CV1]] and [[HTC Vive]]&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;, offering &lt;/ins&gt;2160×1200 combined resolution at 90Hz with [[room-scale tracking]]&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;, established &lt;/ins&gt;the technical baseline for modern VR.&amp;lt;ref name=&quot;cavendish2024&quot;&amp;gt;Cavendishprofessionals. &quot;The Evolution of VR and AR in Gaming: A Historical Perspective.&quot; 2024. https://www.cavendishprofessionals.com/the-evolution-of-vr-and-ar-in-gaming-a-historical-perspective/&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Mathematical Foundations ==&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Mathematical Foundations ==&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l137&quot;&gt;Line 137:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 137:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;=== Multi-Pass Rendering ===&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;=== Multi-Pass Rendering ===&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Traditional multi-pass rendering takes the straightforward approach of rendering the complete scene twice sequentially, once per eye. Each eye uses separate camera parameters, performing independent [[draw call]]s, culling operations, and shader executions. While conceptually simple and compatible with all rendering pipelines, this approach imposes nearly 2× computational &lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;cost—doubling &lt;/del&gt;[[CPU]] overhead from draw call submission, duplicating geometry processing on the [[GPU]], and requiring full iteration through all rendering stages twice.&amp;lt;ref name=&quot;unity2024&quot;&amp;gt;Unity. &quot;How to maximize AR and VR performance with advanced stereo rendering.&quot; Unity Blog, 2024. https://blog.unity.com/technology/how-to-maximize-ar-and-vr-performance-with-advanced-stereo-rendering&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Traditional multi-pass rendering takes the straightforward approach of rendering the complete scene twice sequentially, once per eye. Each eye uses separate camera parameters, performing independent [[draw call]]s, culling operations, and shader executions. While conceptually simple and compatible with all rendering pipelines, this approach imposes nearly 2× computational &lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;cost, doubling &lt;/ins&gt;[[CPU]] overhead from draw call submission, duplicating geometry processing on the [[GPU]], and requiring full iteration through all rendering stages twice.&amp;lt;ref name=&quot;unity2024&quot;&amp;gt;Unity. &quot;How to maximize AR and VR performance with advanced stereo rendering.&quot; Unity Blog, 2024. https://blog.unity.com/technology/how-to-maximize-ar-and-vr-performance-with-advanced-stereo-rendering&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;=== Single-Pass Stereo Rendering ===&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;=== Single-Pass Stereo Rendering ===&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l155&quot;&gt;Line 155:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 155:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[NVIDIA]]&amp;#039;s [[Pascal (microarchitecture)|Pascal]] architecture introduced &amp;#039;&amp;#039;&amp;#039;Simultaneous Multi-Projection (SMP)&amp;#039;&amp;#039;&amp;#039; enabling true Single Pass Stereo where geometry processes once and projects to both eyes simultaneously using hardware acceleration.&amp;lt;ref name=&amp;quot;anandtech2016&amp;quot;&amp;gt;AnandTech. &amp;quot;Simultaneous Multi-Projection: Reusing Geometry on the Cheap.&amp;quot; 2016. https://www.anandtech.com/show/10325/the-nvidia-geforce-gtx-1080-and-1070-founders-edition-review/11&amp;lt;/ref&amp;gt; The [[Turing (microarchitecture)|Turing]] architecture expanded this to Multi-View Rendering supporting up to 4 projection views in a single pass.&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[NVIDIA]]&amp;#039;s [[Pascal (microarchitecture)|Pascal]] architecture introduced &amp;#039;&amp;#039;&amp;#039;Simultaneous Multi-Projection (SMP)&amp;#039;&amp;#039;&amp;#039; enabling true Single Pass Stereo where geometry processes once and projects to both eyes simultaneously using hardware acceleration.&amp;lt;ref name=&amp;quot;anandtech2016&amp;quot;&amp;gt;AnandTech. &amp;quot;Simultaneous Multi-Projection: Reusing Geometry on the Cheap.&amp;quot; 2016. https://www.anandtech.com/show/10325/the-nvidia-geforce-gtx-1080-and-1070-founders-edition-review/11&amp;lt;/ref&amp;gt; The [[Turing (microarchitecture)|Turing]] architecture expanded this to Multi-View Rendering supporting up to 4 projection views in a single pass.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&#039;&#039;&#039;Lens Matched Shading&#039;&#039;&#039; divides each eye&#039;s view into 4 quadrants with adjusted projections approximating the barrel-distorted output shape after lens correction, reducing rendered pixels from 2.1 megapixels to 1.4 megapixels per &lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;eye—a &lt;/del&gt;50% increase in available pixel shading throughput.&amp;lt;ref name=&quot;roadtovr2016&quot;&amp;gt;Road to VR. &quot;NVIDIA Explains Pascal&#039;s &#039;Lens Matched Shading&#039; for VR.&quot; 2016. https://www.roadtovr.com/nvidia-explains-pascal-simultaneous-multi-projection-lens-matched-shading-for-vr/&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&#039;&#039;&#039;Lens Matched Shading&#039;&#039;&#039; divides each eye&#039;s view into 4 quadrants with adjusted projections approximating the barrel-distorted output shape after lens correction, reducing rendered pixels from 2.1 megapixels to 1.4 megapixels per &lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;eye, a &lt;/ins&gt;50% increase in available pixel shading throughput.&amp;lt;ref name=&quot;roadtovr2016&quot;&amp;gt;Road to VR. &quot;NVIDIA Explains Pascal&#039;s &#039;Lens Matched Shading&#039; for VR.&quot; 2016. https://www.roadtovr.com/nvidia-explains-pascal-simultaneous-multi-projection-lens-matched-shading-for-vr/&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;=== Advanced Optimization Techniques ===&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;=== Advanced Optimization Techniques ===&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l200&quot;&gt;Line 200:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 200:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;|}&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;|}&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Modern VR rendering demands GPU capabilities significantly beyond traditional gaming.&amp;lt;ref name=&quot;computercity2024&quot;&amp;gt;ComputerCity. &quot;VR PC Hardware Requirements: Minimum and Recommended Specs.&quot; 2024. https://computercity.com/hardware/vr/vr-pc-hardware-requirements&amp;lt;/ref&amp;gt; To prevent [[simulation sickness]], VR applications must maintain consistently high frame &lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;rates—typically &lt;/del&gt;90 frames per second or &lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;higher—and &lt;/del&gt;motion-to-photon latency under 20 milliseconds.&amp;lt;ref name=&quot;daqri2024&quot;&amp;gt;DAQRI. &quot;Motion to Photon Latency in Mobile AR and VR.&quot; Medium, 2024. https://medium.com/@DAQRI/motion-to-photon-latency-in-mobile-ar-and-vr-99f82c480926&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Modern VR rendering demands GPU capabilities significantly beyond traditional gaming.&amp;lt;ref name=&quot;computercity2024&quot;&amp;gt;ComputerCity. &quot;VR PC Hardware Requirements: Minimum and Recommended Specs.&quot; 2024. https://computercity.com/hardware/vr/vr-pc-hardware-requirements&amp;lt;/ref&amp;gt; To prevent [[simulation sickness]], VR applications must maintain consistently high frame &lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;rates, typically &lt;/ins&gt;90 frames per second or &lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;higher, and &lt;/ins&gt;motion-to-photon latency under 20 milliseconds.&amp;lt;ref name=&quot;daqri2024&quot;&amp;gt;DAQRI. &quot;Motion to Photon Latency in Mobile AR and VR.&quot; Medium, 2024. https://medium.com/@DAQRI/motion-to-photon-latency-in-mobile-ar-and-vr-99f82c480926&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Display Technologies ==&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Display Technologies ==&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l220&quot;&gt;Line 220:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 220:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;=== Gaming ===&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;=== Gaming ===&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Gaming dominates current VR usage with 48.3% of market revenue and 70% of VR users playing games regularly.&amp;lt;ref name=&quot;marketgrowth2024&quot;&amp;gt;Marketgrowthreports. &quot;Virtual and Augmented Reality Industry Market Size, Trends 2033.&quot; 2024. https://www.marketgrowthreports.com/market-reports/virtual-and-augmented-reality-market-100490&amp;lt;/ref&amp;gt; The depth cues from stereoscopic rendering prove essential for gameplay mechanics requiring accurate spatial &lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;judgment—from &lt;/del&gt;grabbing objects in [[Beat Saber]] to navigating complex environments in [[Half-Life: Alyx]].&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Gaming dominates current VR usage with 48.3% of market revenue and 70% of VR users playing games regularly.&amp;lt;ref name=&quot;marketgrowth2024&quot;&amp;gt;Marketgrowthreports. &quot;Virtual and Augmented Reality Industry Market Size, Trends 2033.&quot; 2024. https://www.marketgrowthreports.com/market-reports/virtual-and-augmented-reality-market-100490&amp;lt;/ref&amp;gt; The depth cues from stereoscopic rendering prove essential for gameplay mechanics requiring accurate spatial &lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;judgment, from &lt;/ins&gt;grabbing objects in [[Beat Saber]] to navigating complex environments in [[Half-Life: Alyx]].&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;=== Healthcare ===&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;=== Healthcare ===&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l310&quot;&gt;Line 310:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 310:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Stereoscopic rendering remains indispensable for VR and AR experiences requiring depth perception. The technique&amp;#039;s evolution from mechanical stereoscopes to real-time GPU-accelerated rendering reflects advancing hardware capabilities and algorithmic innovations. Modern implementations reduce computational overhead by 30-70% compared to naive approaches, making immersive VR accessible on $300 standalone headsets rather than requiring $2000 gaming PCs.&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Stereoscopic rendering remains indispensable for VR and AR experiences requiring depth perception. The technique&amp;#039;s evolution from mechanical stereoscopes to real-time GPU-accelerated rendering reflects advancing hardware capabilities and algorithmic innovations. Modern implementations reduce computational overhead by 30-70% compared to naive approaches, making immersive VR accessible on $300 standalone headsets rather than requiring $2000 gaming PCs.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The fundamental vergence-accommodation conflict represents a limitation of current display technology rather than stereoscopic rendering &lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;itself—one &lt;/del&gt;being actively solved through light field displays, holographic waveguides, and varifocal systems. The industry&#039;s convergence on OpenXR as a unified standard, combined with mature optimization techniques integrated into Unity and Unreal Engine, enables developers to target diverse platforms efficiently. The 171 million VR users in 2024 represent early adoption, with enterprise applications demonstrating stereoscopic rendering&#039;s value extends far beyond entertainment into training, healthcare, and industrial visualization.&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The fundamental vergence-accommodation conflict represents a limitation of current display technology rather than stereoscopic rendering &lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;itself, one &lt;/ins&gt;being actively solved through light field displays, holographic waveguides, and varifocal systems. The industry&#039;s convergence on OpenXR as a unified standard, combined with mature optimization techniques integrated into Unity and Unreal Engine, enables developers to target diverse platforms efficiently. The 171 million VR users in 2024 represent early adoption, with enterprise applications demonstrating stereoscopic rendering&#039;s value extends far beyond entertainment into training, healthcare, and industrial visualization.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== See Also ==&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== See Also ==&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Xinreality</name></author>
	</entry>
	<entry>
		<id>https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;diff=36631&amp;oldid=prev</id>
		<title>Xinreality: /* Display Technologies */</title>
		<link rel="alternate" type="text/html" href="https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;diff=36631&amp;oldid=prev"/>
		<updated>2025-10-26T03:29:21Z</updated>

		<summary type="html">&lt;p&gt;&lt;span class=&quot;autocomment&quot;&gt;Display Technologies&lt;/span&gt;&lt;/p&gt;
&lt;table style=&quot;background-color: #fff; color: #202122;&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;Revision as of 03:29, 26 October 2025&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l208&quot;&gt;Line 208:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 208:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* &amp;#039;&amp;#039;&amp;#039;[[Head-Mounted Display]]s (HMDs)&amp;#039;&amp;#039;&amp;#039;: Modern VR and AR headsets achieve perfect image separation using either two separate micro-displays (one for each eye) or a single display partitioned by optics. This direct-view approach completely isolates the left and right eye views, eliminating [[crosstalk]].&amp;lt;ref name=&amp;quot;drawandcode&amp;quot;/&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* &amp;#039;&amp;#039;&amp;#039;[[Head-Mounted Display]]s (HMDs)&amp;#039;&amp;#039;&amp;#039;: Modern VR and AR headsets achieve perfect image separation using either two separate micro-displays (one for each eye) or a single display partitioned by optics. This direct-view approach completely isolates the left and right eye views, eliminating [[crosstalk]].&amp;lt;ref name=&amp;quot;drawandcode&amp;quot;/&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* &#039;&#039;&#039;Color Filtering ([[Anaglyph 3D|Anaglyph]])&#039;&#039;&#039;: Uses glasses with filters of different colors, typically red and cyan. Very inexpensive but suffers from severe color distortion and ghosting.&amp;lt;ref name=&quot;basic_principles&quot;/&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* &#039;&#039;&#039;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;[[&lt;/ins&gt;Color Filtering&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;]] &lt;/ins&gt;([[Anaglyph 3D|Anaglyph]])&#039;&#039;&#039;: Uses glasses with filters of different colors, typically red and cyan. Very inexpensive but suffers from severe color distortion and ghosting.&amp;lt;ref name=&quot;basic_principles&quot;/&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* &amp;#039;&amp;#039;&amp;#039;[[Polarized 3D system|Polarization]]&amp;#039;&amp;#039;&amp;#039;: Uses glasses with differently polarized lenses. Linear polarization orients filters at 90 degrees; circular polarization uses opposite clockwise/counter-clockwise polarization. Commonly used in 3D cinemas.&amp;lt;ref name=&amp;quot;palušová2023&amp;quot;&amp;gt;Palušová, P. &amp;quot;Stereoscopy in Extended Reality: Utilizing Natural Binocular Disparity.&amp;quot; 2023. https://www.petrapalusova.com/stereoscopy&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* &amp;#039;&amp;#039;&amp;#039;[[Polarized 3D system|Polarization]]&amp;#039;&amp;#039;&amp;#039;: Uses glasses with differently polarized lenses. Linear polarization orients filters at 90 degrees; circular polarization uses opposite clockwise/counter-clockwise polarization. Commonly used in 3D cinemas.&amp;lt;ref name=&amp;quot;palušová2023&amp;quot;&amp;gt;Palušová, P. &amp;quot;Stereoscopy in Extended Reality: Utilizing Natural Binocular Disparity.&amp;quot; 2023. https://www.petrapalusova.com/stereoscopy&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* &#039;&#039;&#039;Time Multiplexing (Active Shutter)&#039;&#039;&#039;: Display alternates between left and right images at high speed (120+ Hz). Viewer wears LCD shutter glasses synchronized to the display. Delivers full resolution to each eye.&amp;lt;ref name=&quot;basic_principles&quot;/&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* &#039;&#039;&#039;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;[[&lt;/ins&gt;Time Multiplexing&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;]] &lt;/ins&gt;(Active Shutter)&#039;&#039;&#039;: Display alternates between left and right images at high speed (120+ Hz). Viewer wears LCD shutter glasses synchronized to the display. Delivers full resolution to each eye.&amp;lt;ref name=&quot;basic_principles&quot;/&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* &amp;#039;&amp;#039;&amp;#039;[[Autostereoscopy]] (Glasses-Free 3D)&amp;#039;&amp;#039;&amp;#039;: Uses optical elements like [[parallax barrier]]s or [[lenticular lens]]es to direct different pixels to each eye. Limited by narrow optimal viewing angle.&amp;lt;ref name=&amp;quot;palušová2023&amp;quot;/&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* &amp;#039;&amp;#039;&amp;#039;[[Autostereoscopy]] (Glasses-Free 3D)&amp;#039;&amp;#039;&amp;#039;: Uses optical elements like [[parallax barrier]]s or [[lenticular lens]]es to direct different pixels to each eye. Limited by narrow optimal viewing angle.&amp;lt;ref name=&amp;quot;palušová2023&amp;quot;/&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Xinreality</name></author>
	</entry>
	<entry>
		<id>https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;diff=36630&amp;oldid=prev</id>
		<title>Xinreality: /* Single-Pass Stereo Rendering */</title>
		<link rel="alternate" type="text/html" href="https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;diff=36630&amp;oldid=prev"/>
		<updated>2025-10-26T03:29:00Z</updated>

		<summary type="html">&lt;p&gt;&lt;span class=&quot;autocomment&quot;&gt;Single-Pass Stereo Rendering&lt;/span&gt;&lt;/p&gt;
&lt;table style=&quot;background-color: #fff; color: #202122;&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;Revision as of 03:29, 26 October 2025&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l142&quot;&gt;Line 142:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 142:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Single-pass stereo rendering optimizes by traversing the scene graph once while rendering to both eye buffers.&amp;lt;ref name=&amp;quot;nvidia2018&amp;quot;&amp;gt;NVIDIA Developer. &amp;quot;Turing Multi-View Rendering in VRWorks.&amp;quot; NVIDIA Technical Blog, 2018. https://developer.nvidia.com/blog/turing-multi-view-rendering-vrworks/&amp;lt;/ref&amp;gt; Single-pass instanced approach uses GPU instancing with instance count of 2, where the [[vertex shader]] outputs positions for both views simultaneously. Example shader code:&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Single-pass stereo rendering optimizes by traversing the scene graph once while rendering to both eye buffers.&amp;lt;ref name=&amp;quot;nvidia2018&amp;quot;&amp;gt;NVIDIA Developer. &amp;quot;Turing Multi-View Rendering in VRWorks.&amp;quot; NVIDIA Technical Blog, 2018. https://developer.nvidia.com/blog/turing-multi-view-rendering-vrworks/&amp;lt;/ref&amp;gt; Single-pass instanced approach uses GPU instancing with instance count of 2, where the [[vertex shader]] outputs positions for both views simultaneously. Example shader code:&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;&lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;code&lt;/del&gt;&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;pre&lt;/ins&gt;&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;uniform EyeUniforms {&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;uniform EyeUniforms {&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;     mat4 mMatrix[2];&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;     mat4 mMatrix[2];&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;};&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;};&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;vec4 pos = mMatrix[gl_InvocationID] * vertex;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;vec4 pos = mMatrix[gl_InvocationID] * vertex;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;/&lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;code&lt;/del&gt;&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;/&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;pre&lt;/ins&gt;&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;This technique halves draw call count compared to multi-pass, reducing CPU bottlenecks in complex scenes.&amp;lt;ref name=&amp;quot;iquilez&amp;quot;&amp;gt;Quilez, Inigo. &amp;quot;Stereo rendering.&amp;quot; 2024. https://iquilezles.org/articles/stereo/&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;This technique halves draw call count compared to multi-pass, reducing CPU bottlenecks in complex scenes.&amp;lt;ref name=&amp;quot;iquilez&amp;quot;&amp;gt;Quilez, Inigo. &amp;quot;Stereo rendering.&amp;quot; 2024. https://iquilezles.org/articles/stereo/&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Xinreality</name></author>
	</entry>
	<entry>
		<id>https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;diff=36629&amp;oldid=prev</id>
		<title>Xinreality: /* Single-Pass Stereo Rendering */</title>
		<link rel="alternate" type="text/html" href="https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;diff=36629&amp;oldid=prev"/>
		<updated>2025-10-26T03:28:37Z</updated>

		<summary type="html">&lt;p&gt;&lt;span class=&quot;autocomment&quot;&gt;Single-Pass Stereo Rendering&lt;/span&gt;&lt;/p&gt;
&lt;table style=&quot;background-color: #fff; color: #202122;&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;Revision as of 03:28, 26 October 2025&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l142&quot;&gt;Line 142:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 142:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Single-pass stereo rendering optimizes by traversing the scene graph once while rendering to both eye buffers.&amp;lt;ref name=&amp;quot;nvidia2018&amp;quot;&amp;gt;NVIDIA Developer. &amp;quot;Turing Multi-View Rendering in VRWorks.&amp;quot; NVIDIA Technical Blog, 2018. https://developer.nvidia.com/blog/turing-multi-view-rendering-vrworks/&amp;lt;/ref&amp;gt; Single-pass instanced approach uses GPU instancing with instance count of 2, where the [[vertex shader]] outputs positions for both views simultaneously. Example shader code:&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Single-pass stereo rendering optimizes by traversing the scene graph once while rendering to both eye buffers.&amp;lt;ref name=&amp;quot;nvidia2018&amp;quot;&amp;gt;NVIDIA Developer. &amp;quot;Turing Multi-View Rendering in VRWorks.&amp;quot; NVIDIA Technical Blog, 2018. https://developer.nvidia.com/blog/turing-multi-view-rendering-vrworks/&amp;lt;/ref&amp;gt; Single-pass instanced approach uses GPU instancing with instance count of 2, where the [[vertex shader]] outputs positions for both views simultaneously. Example shader code:&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;```glsl&lt;/del&gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&amp;lt;code&amp;gt;&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;uniform EyeUniforms {&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;uniform EyeUniforms {&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;     mat4 mMatrix[2];&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;     mat4 mMatrix[2];&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;};&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;};&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;vec4 pos = mMatrix[gl_InvocationID] * vertex;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;vec4 pos = mMatrix[gl_InvocationID] * vertex;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;```&lt;/del&gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&amp;lt;/code&amp;gt;&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;This technique halves draw call count compared to multi-pass, reducing CPU bottlenecks in complex scenes.&amp;lt;ref name=&amp;quot;iquilez&amp;quot;&amp;gt;Quilez, Inigo. &amp;quot;Stereo rendering.&amp;quot; 2024. https://iquilezles.org/articles/stereo/&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;This technique halves draw call count compared to multi-pass, reducing CPU bottlenecks in complex scenes.&amp;lt;ref name=&amp;quot;iquilez&amp;quot;&amp;gt;Quilez, Inigo. &amp;quot;Stereo rendering.&amp;quot; 2024. https://iquilezles.org/articles/stereo/&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Xinreality</name></author>
	</entry>
	<entry>
		<id>https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;diff=36628&amp;oldid=prev</id>
		<title>Xinreality: /* Historical Evolution */</title>
		<link rel="alternate" type="text/html" href="https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;diff=36628&amp;oldid=prev"/>
		<updated>2025-10-26T03:28:19Z</updated>

		<summary type="html">&lt;p&gt;&lt;span class=&quot;autocomment&quot;&gt;Historical Evolution&lt;/span&gt;&lt;/p&gt;
&lt;table style=&quot;background-color: #fff; color: #202122;&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;Revision as of 03:28, 26 October 2025&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l78&quot;&gt;Line 78:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 78:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;| 1991 || Virtuality VR arcades || Real-time stereoscopic multiplayer VR&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;| 1991 || Virtuality VR arcades || Real-time stereoscopic multiplayer VR&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;|-&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;|-&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;| 1995 || Nintendo Virtual Boy || Portable stereoscopic gaming console&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;| 1995 || &lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;[[&lt;/ins&gt;Nintendo Virtual Boy&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;]] &lt;/ins&gt;|| Portable stereoscopic gaming console&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;|-&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;|-&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;| 2010 || Oculus Rift prototype || Modern stereoscopic HMD revival&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;| 2010 || &lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;[[&lt;/ins&gt;Oculus Rift&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;]] &lt;/ins&gt;prototype || Modern stereoscopic HMD revival&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;|-&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;|-&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;| 2016 || HTC Vive/Oculus Rift CV1 release || Consumer room-scale stereoscopic VR&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;| 2016 || &lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;[[&lt;/ins&gt;HTC Vive&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;]]&lt;/ins&gt;/&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;[[&lt;/ins&gt;Oculus Rift CV1&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;]] &lt;/ins&gt;release || Consumer room-scale stereoscopic VR&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;|-&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;|-&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;| 2023 || Apple Vision Pro || High-resolution stereoscopic mixed reality (70 pixels per degree)&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;| 2023 || &lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;[[&lt;/ins&gt;Apple Vision Pro&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;]] &lt;/ins&gt;|| High-resolution stereoscopic mixed reality (70 pixels per degree)&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;|}&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;|}&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Xinreality</name></author>
	</entry>
	<entry>
		<id>https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;diff=36627&amp;oldid=prev</id>
		<title>Xinreality at 03:27, 26 October 2025</title>
		<link rel="alternate" type="text/html" href="https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;diff=36627&amp;oldid=prev"/>
		<updated>2025-10-26T03:27:07Z</updated>

		<summary type="html">&lt;p&gt;&lt;/p&gt;
&lt;table style=&quot;background-color: #fff; color: #202122;&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;Revision as of 03:27, 26 October 2025&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l17&quot;&gt;Line 17:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 17:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;| users = 171 million (2024)&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;| users = 171 million (2024)&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;}}&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;}}&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-side-deleted&quot;&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;[[File:stereoscopic rendering1.jpg|300px|right]]&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-side-deleted&quot;&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-side-deleted&quot;&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;[[File:stereoscopic rendering2.jpg|300px|right]]&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;#039;&amp;#039;&amp;#039;Stereoscopic rendering&amp;#039;&amp;#039;&amp;#039; is the foundational [[computer graphics]] technique that creates the perception of three-dimensional depth in [[virtual reality]] (VR) and [[augmented reality]] (AR) systems by generating two slightly different images from distinct viewpoints corresponding to the left and right eyes.&amp;lt;ref name=&amp;quot;arm2021&amp;quot;&amp;gt;ARM Software. &amp;quot;Introduction to Stereo Rendering - VR SDK for Android.&amp;quot; ARM Developer Documentation, 2021. https://arm-software.github.io/vr-sdk-for-android/IntroductionToStereoRendering.html&amp;lt;/ref&amp;gt; This technique exploits [[binocular disparity]]—the horizontal displacement between corresponding points in the two images—enabling the [[visual cortex]] to reconstruct depth information through [[stereopsis]], the same process human eyes use to perceive the real world.&amp;lt;ref name=&amp;quot;numberanalytics2024&amp;quot;&amp;gt;Number Analytics. &amp;quot;Stereoscopy in VR: A Comprehensive Guide.&amp;quot; 2024. https://www.numberanalytics.com/blog/ultimate-guide-stereoscopy-vr-ar-development&amp;lt;/ref&amp;gt; By delivering two offset images (one per eye) that the brain combines into a single scene, stereoscopic rendering produces an illusion of depth that mimics natural [[binocular vision]].&amp;lt;ref name=&amp;quot;drawandcode&amp;quot;&amp;gt;Draw &amp;amp; Code. &amp;quot;What Is Stereoscopic VR Technology.&amp;quot; January 23, 2024. https://drawandcode.com/learning-zone/what-is-stereoscopic-vr-technology/&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;#039;&amp;#039;&amp;#039;Stereoscopic rendering&amp;#039;&amp;#039;&amp;#039; is the foundational [[computer graphics]] technique that creates the perception of three-dimensional depth in [[virtual reality]] (VR) and [[augmented reality]] (AR) systems by generating two slightly different images from distinct viewpoints corresponding to the left and right eyes.&amp;lt;ref name=&amp;quot;arm2021&amp;quot;&amp;gt;ARM Software. &amp;quot;Introduction to Stereo Rendering - VR SDK for Android.&amp;quot; ARM Developer Documentation, 2021. https://arm-software.github.io/vr-sdk-for-android/IntroductionToStereoRendering.html&amp;lt;/ref&amp;gt; This technique exploits [[binocular disparity]]—the horizontal displacement between corresponding points in the two images—enabling the [[visual cortex]] to reconstruct depth information through [[stereopsis]], the same process human eyes use to perceive the real world.&amp;lt;ref name=&amp;quot;numberanalytics2024&amp;quot;&amp;gt;Number Analytics. &amp;quot;Stereoscopy in VR: A Comprehensive Guide.&amp;quot; 2024. https://www.numberanalytics.com/blog/ultimate-guide-stereoscopy-vr-ar-development&amp;lt;/ref&amp;gt; By delivering two offset images (one per eye) that the brain combines into a single scene, stereoscopic rendering produces an illusion of depth that mimics natural [[binocular vision]].&amp;lt;ref name=&amp;quot;drawandcode&amp;quot;&amp;gt;Draw &amp;amp; Code. &amp;quot;What Is Stereoscopic VR Technology.&amp;quot; January 23, 2024. https://drawandcode.com/learning-zone/what-is-stereoscopic-vr-technology/&amp;lt;/ref&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Xinreality</name></author>
	</entry>
	<entry>
		<id>https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;diff=36624&amp;oldid=prev</id>
		<title>Xinreality at 03:23, 26 October 2025</title>
		<link rel="alternate" type="text/html" href="https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;diff=36624&amp;oldid=prev"/>
		<updated>2025-10-26T03:23:44Z</updated>

		<summary type="html">&lt;p&gt;&lt;/p&gt;
&lt;table style=&quot;background-color: #fff; color: #202122;&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;Revision as of 03:23, 26 October 2025&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l369&quot;&gt;Line 369:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 369:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;/references&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;/references&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[Category:&lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;Virtual reality]]&lt;/del&gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[Category:&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;Terms&lt;/ins&gt;]]&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;[[Category:Augmented reality&lt;/del&gt;]]&lt;/div&gt;&lt;/td&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-side-added&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[Category:Computer graphics]]&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[Category:Computer graphics]]&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[Category:3D rendering]]&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[Category:3D rendering]]&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[Category:Display technology]]&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[Category:Display technology]]&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[Category:Human–computer interaction]]&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[Category:Human–computer interaction]]&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Xinreality</name></author>
	</entry>
	<entry>
		<id>https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;diff=36623&amp;oldid=prev</id>
		<title>Xinreality at 03:23, 26 October 2025</title>
		<link rel="alternate" type="text/html" href="https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;diff=36623&amp;oldid=prev"/>
		<updated>2025-10-26T03:23:20Z</updated>

		<summary type="html">&lt;p&gt;&lt;/p&gt;
&lt;a href=&quot;https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;amp;diff=36623&amp;amp;oldid=36622&quot;&gt;Show changes&lt;/a&gt;</summary>
		<author><name>Xinreality</name></author>
	</entry>
	<entry>
		<id>https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;diff=36622&amp;oldid=prev</id>
		<title>Xinreality at 03:22, 26 October 2025</title>
		<link rel="alternate" type="text/html" href="https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;diff=36622&amp;oldid=prev"/>
		<updated>2025-10-26T03:22:59Z</updated>

		<summary type="html">&lt;p&gt;&lt;/p&gt;
&lt;a href=&quot;https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;amp;diff=36622&amp;amp;oldid=36621&quot;&gt;Show changes&lt;/a&gt;</summary>
		<author><name>Xinreality</name></author>
	</entry>
	<entry>
		<id>https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;diff=36621&amp;oldid=prev</id>
		<title>Xinreality: Created page with &quot;{{Infobox technology | name = Stereoscopic Rendering | image =  | caption =  | type = Computer graphics technique | inventor =  | inception = 1838 (concept), 1968 (computer graphics) | manufacturer =  | available =  | current_supplier =  | last_production =  | introduced =  | discontinued =  | cost =  | applications = Virtual reality, Augmented reality, 3D gaming, Medical visualization | industry = VR/AR ($15.9 billion as of 2024) | users = 171 millio...&quot;</title>
		<link rel="alternate" type="text/html" href="https://vrarwiki.com/index.php?title=Stereoscopic_rendering&amp;diff=36621&amp;oldid=prev"/>
		<updated>2025-10-26T03:22:11Z</updated>

		<summary type="html">&lt;p&gt;Created page with &amp;quot;{{Infobox technology | name = Stereoscopic Rendering | image =  | caption =  | type = &lt;a href=&quot;/index.php?title=Computer_graphics&amp;amp;action=edit&amp;amp;redlink=1&quot; class=&quot;new&quot; title=&quot;Computer graphics (page does not exist)&quot;&gt;Computer graphics&lt;/a&gt; technique | inventor =  | inception = 1838 (concept), 1968 (computer graphics) | manufacturer =  | available =  | current_supplier =  | last_production =  | introduced =  | discontinued =  | cost =  | applications = &lt;a href=&quot;/wiki/Virtual_reality&quot; class=&quot;mw-redirect&quot; title=&quot;Virtual reality&quot;&gt;Virtual reality&lt;/a&gt;, &lt;a href=&quot;/wiki/Augmented_reality&quot; title=&quot;Augmented reality&quot;&gt;Augmented reality&lt;/a&gt;, &lt;a href=&quot;/index.php?title=3D_gaming&amp;amp;action=edit&amp;amp;redlink=1&quot; class=&quot;new&quot; title=&quot;3D gaming (page does not exist)&quot;&gt;3D gaming&lt;/a&gt;, &lt;a href=&quot;/index.php?title=Medical_visualization&amp;amp;action=edit&amp;amp;redlink=1&quot; class=&quot;new&quot; title=&quot;Medical visualization (page does not exist)&quot;&gt;Medical visualization&lt;/a&gt; | industry = VR/AR ($15.9 billion as of 2024) | users = 171 millio...&amp;quot;&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;{{Infobox technology&lt;br /&gt;
| name = Stereoscopic Rendering&lt;br /&gt;
| image = &lt;br /&gt;
| caption = &lt;br /&gt;
| type = [[Computer graphics]] technique&lt;br /&gt;
| inventor = &lt;br /&gt;
| inception = 1838 (concept), 1968 (computer graphics)&lt;br /&gt;
| manufacturer = &lt;br /&gt;
| available = &lt;br /&gt;
| current_supplier = &lt;br /&gt;
| last_production = &lt;br /&gt;
| introduced = &lt;br /&gt;
| discontinued = &lt;br /&gt;
| cost = &lt;br /&gt;
| applications = [[Virtual reality]], [[Augmented reality]], [[3D gaming]], [[Medical visualization]]&lt;br /&gt;
| industry = VR/AR ($15.9 billion as of 2024)&lt;br /&gt;
| users = 171 million (2024)&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Stereoscopic rendering&amp;#039;&amp;#039;&amp;#039; is the foundational [[computer graphics]] technique that creates the perception of three-dimensional depth in [[virtual reality]] (VR) and [[augmented reality]] (AR) systems by generating two slightly different images from distinct viewpoints corresponding to the left and right eyes.&amp;lt;ref name=&amp;quot;arm2021&amp;quot;&amp;gt;ARM Software. &amp;quot;Introduction to Stereo Rendering - VR SDK for Android.&amp;quot; ARM Developer Documentation, 2021. https://arm-software.github.io/vr-sdk-for-android/IntroductionToStereoRendering.html&amp;lt;/ref&amp;gt; This technique exploits [[binocular disparity]]—the horizontal displacement between corresponding points in the two images—enabling the [[visual cortex]] to reconstruct depth information through [[stereopsis]], the same process human eyes use to perceive the real world.&amp;lt;ref name=&amp;quot;numberanalytics2024&amp;quot;&amp;gt;Number Analytics. &amp;quot;Stereoscopy in VR: A Comprehensive Guide.&amp;quot; 2024. https://www.numberanalytics.com/blog/ultimate-guide-stereoscopy-vr-ar-development&amp;lt;/ref&amp;gt; By delivering two offset images (one per eye) that the brain combines into a single scene, stereoscopic rendering produces an illusion of depth that mimics natural [[binocular vision]].&amp;lt;ref name=&amp;quot;drawandcode&amp;quot;&amp;gt;Draw &amp;amp; Code. &amp;quot;What Is Stereoscopic VR Technology.&amp;quot; January 23, 2024. https://drawandcode.com/learning-zone/what-is-stereoscopic-vr-technology/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The approach doubles computational requirements compared to traditional rendering but delivers the immersive depth perception that defines modern VR experiences, powering a $15.9 billion industry serving 171 million users worldwide as of 2024.&amp;lt;ref name=&amp;quot;mordor2024&amp;quot;&amp;gt;Mordor Intelligence. &amp;quot;Virtual Reality (VR) Market Size, Report, Share &amp;amp; Growth Trends 2025-2030.&amp;quot; 2024. https://www.mordorintelligence.com/industry-reports/virtual-reality-market&amp;lt;/ref&amp;gt; Unlike monoscopic imagery (showing the same image to both eyes), stereoscopic rendering presents each eye with a slightly different perspective, closely matching how humans view the real world and thereby greatly enhancing the sense of presence and realism in VR/AR.&amp;lt;ref name=&amp;quot;borisfx2024&amp;quot;&amp;gt;Boris FX. &amp;quot;Monoscopic vs Stereoscopic 360 VR: Key Differences.&amp;quot; 2024. https://borisfx.com/blog/monoscopic-vs-stereoscopic-360-vr-key-differences/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Fundamental Principles ==&lt;br /&gt;
&lt;br /&gt;
=== How It Works ===&lt;br /&gt;
&lt;br /&gt;
In stereoscopic rendering, a scene is captured or rendered from two viewpoints separated by roughly the distance between the human eyes (the [[interpupillary distance]] or IPD), typically calibrated to the average human IPD of 64mm (ranging from 54-74mm in adults).&amp;lt;ref name=&amp;quot;afifi2020&amp;quot;&amp;gt;Afifi, Mahmoud. &amp;quot;Basics of stereoscopic imaging in virtual and augmented reality systems.&amp;quot; Medium, 2020. https://medium.com/@mahmoudnafifi/basics-of-stereoscopic-imaging-6f69a7916cfd&amp;lt;/ref&amp;gt; Each viewpoint (often called the left-eye and right-eye camera) generates a 2D image of the scene. When these images are presented to the corresponding eyes of the user, the slight horizontal disparity between them is interpreted by the visual system as depth information through [[stereopsis]].&amp;lt;ref name=&amp;quot;wikipedia_depth&amp;quot;&amp;gt;Wikipedia. &amp;quot;Depth perception.&amp;quot; https://en.wikipedia.org/wiki/Depth_perception&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Modern VR headsets have two displays or a split display, one for each eye, and each display shows an image rendered from the appropriate perspective. Similarly, AR headsets (such as see-through [[head-mounted display]]s) project stereoscopic digital overlays so that virtual objects appear integrated into the real world with correct depth. The result is that the user perceives a unified 3D scene with depth, providing critical spatial awareness for interaction.&amp;lt;ref name=&amp;quot;drawandcode&amp;quot;/&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Binocular Disparity and Parallax ===&lt;br /&gt;
&lt;br /&gt;
The primary depth cue exploited by stereoscopic rendering is [[binocular disparity]]. Because the two virtual cameras are separated horizontally, objects in the 3D scene are projected onto different locations in the left and right images. This difference in projection is called [[parallax]].&amp;lt;ref name=&amp;quot;basic_principles&amp;quot;&amp;gt;Newcastle University. &amp;quot;Basic Principles of Stereoscopic 3D.&amp;quot; 2024. https://www.ncl.ac.uk/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Positive Parallax (Uncrossed Disparity)&amp;#039;&amp;#039;&amp;#039;: Occurs when an object appears behind the display screen. The object&amp;#039;s image is shifted to the left in the left eye&amp;#039;s view and to the right in the right eye&amp;#039;s view.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Negative Parallax (Crossed Disparity)&amp;#039;&amp;#039;&amp;#039;: Occurs when an object appears in front of the display screen, &amp;quot;popping out&amp;quot; toward the viewer. The object&amp;#039;s image is shifted to the right in the left eye&amp;#039;s view and to the left in the right eye&amp;#039;s view.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Zero Parallax&amp;#039;&amp;#039;&amp;#039;: Occurs when an object appears exactly at the depth of the display screen. The object&amp;#039;s image is in the same position in both the left and right eye views. This plane is also known as the convergence plane or stereo window.&lt;br /&gt;
&lt;br /&gt;
The magnitude of parallax for an object is inversely proportional to its distance from the cameras. The mathematical relationship between disparity and depth follows: &amp;#039;&amp;#039;&amp;#039;Z = b×f/d&amp;#039;&amp;#039;&amp;#039;, where Z is depth, b is baseline (interocular distance), f is focal length, and d is disparity.&amp;lt;ref name=&amp;quot;scratchapixel2024&amp;quot;&amp;gt;Scratchapixel. &amp;quot;The Perspective and Orthographic Projection Matrix.&amp;quot; 2024. https://www.scratchapixel.com/lessons/3d-basic-rendering/perspective-and-orthographic-projection-matrix/building-basic-perspective-projection-matrix.html&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Convergence and Accommodation ===&lt;br /&gt;
&lt;br /&gt;
In the real world, two functions of the eyes are perfectly synchronized:&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;[[Vergence]]&amp;#039;&amp;#039;&amp;#039;: The simultaneous movement of both eyes in opposite directions to obtain or maintain single binocular vision. When looking at a nearby object, the eyes rotate inward (converge).&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;[[Accommodation (eye)|Accommodation]]&amp;#039;&amp;#039;&amp;#039;: The process by which the eye&amp;#039;s lens changes shape to focus light on the retina.&lt;br /&gt;
&lt;br /&gt;
Stereoscopic displays create a conflict between these two systems, known as the &amp;#039;&amp;#039;&amp;#039;[[Vergence-Accommodation Conflict]]&amp;#039;&amp;#039;&amp;#039; (VAC).&amp;lt;ref name=&amp;quot;wikipedia_vac&amp;quot;&amp;gt;Wikipedia. &amp;quot;Vergence-accommodation conflict.&amp;quot; https://en.wikipedia.org/wiki/Vergence-accommodation_conflict&amp;lt;/ref&amp;gt; The viewer&amp;#039;s eyes must always accommodate (focus) on the fixed physical distance of the display screen, while their vergence system is directed to objects that appear at various depths within the virtual scene. This mismatch is a primary cause of visual fatigue and discomfort in VR and AR.&amp;lt;ref name=&amp;quot;packet39_2017&amp;quot;&amp;gt;Packet39. &amp;quot;The Accommodation-Vergence conflict and how it affects your kids (and yourself).&amp;quot; 2017. https://packet39.com/blog/2017/12/25/the-accommodation-vergence-conflict-and-how-it-affects-your-kids-and-yourself/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Historical Evolution ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+ Timeline of Key Milestones in Stereoscopic Rendering&lt;br /&gt;
|-&lt;br /&gt;
! Year !! Event !! Relevance&lt;br /&gt;
|-&lt;br /&gt;
| 1838 || Charles Wheatstone invents [[stereoscope]] || Foundational demonstration of binocular depth perception&lt;br /&gt;
|-&lt;br /&gt;
| 1849 || David Brewster&amp;#039;s lenticular stereoscope || First portable commercial stereoscopic device&lt;br /&gt;
|-&lt;br /&gt;
| 1939 || [[View-Master]] patented || Popular consumer stereoscopic viewer&lt;br /&gt;
|-&lt;br /&gt;
| 1956 || Morton Heilig&amp;#039;s Sensorama || Multi-sensory stereoscopic experience machine&lt;br /&gt;
|-&lt;br /&gt;
| 1960 || Heilig&amp;#039;s Telesphere Mask || First stereoscopic head-mounted display patent&lt;br /&gt;
|-&lt;br /&gt;
| 1968 || Ivan Sutherland&amp;#039;s Sword of Damocles || First computer-generated stereoscopic VR display&lt;br /&gt;
|-&lt;br /&gt;
| 1972 || General Electric flight simulator || 180-degree stereoscopic views for training&lt;br /&gt;
|-&lt;br /&gt;
| 1980 || StereoGraphics stereo glasses || Electronic stereoscopic viewing for PCs&lt;br /&gt;
|-&lt;br /&gt;
| 1982 || Sega SubRoc-3D || First commercial stereoscopic video game&lt;br /&gt;
|-&lt;br /&gt;
| 1991 || Virtuality VR arcades || Real-time stereoscopic multiplayer VR&lt;br /&gt;
|-&lt;br /&gt;
| 1995 || Nintendo Virtual Boy || Portable stereoscopic gaming console&lt;br /&gt;
|-&lt;br /&gt;
| 2010 || Oculus Rift prototype || Modern stereoscopic HMD revival&lt;br /&gt;
|-&lt;br /&gt;
| 2016 || HTC Vive/Oculus Rift CV1 release || Consumer room-scale stereoscopic VR&lt;br /&gt;
|-&lt;br /&gt;
| 2023 || Apple Vision Pro || High-resolution stereoscopic mixed reality (70 pixels per degree)&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=== Early Mechanical Stereoscopy ===&lt;br /&gt;
&lt;br /&gt;
The conceptual foundation emerged in 1838 when Sir [[Charles Wheatstone]] invented the first stereoscope using mirrors to present two offset images, formally describing binocular vision in a paper to the [[Royal Society]], earning him the Royal Medal in 1840.&amp;lt;ref name=&amp;quot;googlearts2024&amp;quot;&amp;gt;Google Arts &amp;amp; Culture. &amp;quot;Stereoscopy: the birth of 3D technology.&amp;quot; 2024. https://artsandculture.google.com/story/stereoscopy-the-birth-of-3d-technology-the-royal-society/pwWRTNS-hqDN5g&amp;lt;/ref&amp;gt; Sir [[David Brewster]] improved the design in 1849 with a lens-based portable stereoscope that became the first commercially successful stereoscopic device after the [[Great Exhibition]] of 1851.&amp;lt;ref name=&amp;quot;googlearts2024&amp;quot;/&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Computer Graphics Era ===&lt;br /&gt;
&lt;br /&gt;
Computer-generated stereoscopy began with [[Ivan Sutherland]]&amp;#039;s 1968 head-mounted display at [[Harvard University]], nicknamed the &amp;quot;[[Sword of Damocles (virtual reality)|Sword of Damocles]]&amp;quot; due to its unwieldy overhead suspension system. This wireframe graphics prototype established the technical template—[[head tracking]], stereoscopic displays, and real-time rendering—that would define VR development for decades.&amp;lt;ref name=&amp;quot;nextgen2024&amp;quot;&amp;gt;Nextgeninvent. &amp;quot;Virtual Reality&amp;#039;s Evolution From Science Fiction to Mainstream Technology.&amp;quot; 2024. https://nextgeninvent.com/blogs/the-evolution-of-virtual-reality/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The gaming industry drove early consumer adoption with [[Sega]]&amp;#039;s SubRoc-3D in 1982, the world&amp;#039;s first commercial stereoscopic video game featuring an active shutter 3D system jointly developed with [[Matsushita Electric Industrial Co.|Matsushita]].&amp;lt;ref name=&amp;quot;siggraph2024&amp;quot;&amp;gt;ACM SIGGRAPH. &amp;quot;Remember Stereo 3D on the PC? Have You Ever Wondered What Happened to It?&amp;quot; 2024. https://blog.siggraph.org/2024/10/stereo-3d-pc-history-decline.html/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Modern VR Revolution ===&lt;br /&gt;
&lt;br /&gt;
The modern VR revolution began with [[Palmer Luckey]]&amp;#039;s 2012 [[Oculus Rift]] [[Kickstarter]] campaign, which raised $2.5 million. [[Facebook]]&amp;#039;s $2 billion acquisition of [[Oculus VR|Oculus]] in 2014 validated the market potential. The watershed 2016 launches of the [[Oculus Rift#Consumer version|Oculus Rift CV1]] and [[HTC Vive]]—offering 2160×1200 combined resolution at 90Hz with [[room-scale tracking]]—established the technical baseline for modern VR.&amp;lt;ref name=&amp;quot;cavendish2024&amp;quot;&amp;gt;Cavendishprofessionals. &amp;quot;The Evolution of VR and AR in Gaming: A Historical Perspective.&amp;quot; 2024. https://www.cavendishprofessionals.com/the-evolution-of-vr-and-ar-in-gaming-a-historical-perspective/&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Mathematical Foundations ==&lt;br /&gt;
&lt;br /&gt;
=== Perspective Projection ===&lt;br /&gt;
&lt;br /&gt;
The core mathematics begins with [[perspective projection]], where a 3D point projects onto a 2D image plane based on its depth. For a single viewpoint with center of projection at (0, 0, -d), the projected coordinates follow:&lt;br /&gt;
* x&amp;lt;sub&amp;gt;p&amp;lt;/sub&amp;gt; = (x×d)/(d+z)&lt;br /&gt;
* y&amp;lt;sub&amp;gt;p&amp;lt;/sub&amp;gt; = (y×d)/(d+z)&lt;br /&gt;
&lt;br /&gt;
Stereoscopic rendering extends this with off-axis projection:&amp;lt;ref name=&amp;quot;songho2024&amp;quot;&amp;gt;Song Ho Ahn. &amp;quot;OpenGL Projection Matrix.&amp;quot; 2024. https://www.songho.ca/opengl/gl_projectionmatrix.html&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Left eye at (-e/2, 0, -d): x&amp;lt;sub&amp;gt;left&amp;lt;/sub&amp;gt; = (x×d - z×e/2)/(d+z)&lt;br /&gt;
* Right eye at (e/2, 0, -d): x&amp;lt;sub&amp;gt;right&amp;lt;/sub&amp;gt; = (x×d + z×e/2)/(d+z)&lt;br /&gt;
&lt;br /&gt;
Where e represents eye separation. The horizontal displacement varies with depth z, creating the disparity cues that enable stereopsis.&lt;br /&gt;
&lt;br /&gt;
=== Asymmetric Frustum and Off-Axis Projection ===&lt;br /&gt;
&lt;br /&gt;
The correct method for stereoscopic rendering is known as &amp;#039;&amp;#039;&amp;#039;off-axis projection&amp;#039;&amp;#039;&amp;#039;. It involves keeping the camera view vectors parallel and instead shearing the [[viewing frustum]] for each eye horizontally.&amp;lt;ref name=&amp;quot;bourke&amp;quot;&amp;gt;Bourke, P. &amp;quot;Stereoscopic Rendering.&amp;quot; 2024. http://paulbourke.net/stereographics/stereorender/&amp;lt;/ref&amp;gt; This avoids the vertical parallax issues that would be introduced by &amp;quot;toeing-in&amp;quot; the cameras.&lt;br /&gt;
&lt;br /&gt;
The horizontal shift (s) for the frustum boundaries on the near plane is given by:&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;s = (e/2) × (n/d)&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
* e is the eye separation (interaxial distance)&lt;br /&gt;
* n is the distance to the near clipping plane&lt;br /&gt;
* d is the convergence distance (the distance to the zero parallax plane)&lt;br /&gt;
&lt;br /&gt;
This shift value modifies the left and right parameters of the frustum definition:&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;For the left eye&amp;#039;&amp;#039;&amp;#039;: left = left_original + s, right = right_original + s&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;For the right eye&amp;#039;&amp;#039;&amp;#039;: left = left_original - s, right = right_original - s&lt;br /&gt;
&lt;br /&gt;
Graphics APIs like OpenGL provide functions such as `glFrustum(left, right, bottom, top, near, far)` that allow for the explicit definition of an asymmetric frustum.&amp;lt;ref name=&amp;quot;utah2024&amp;quot;&amp;gt;University of Utah. &amp;quot;Projection and View Frustums.&amp;quot; Computer Graphics Course Material, 2024. https://my.eng.utah.edu/~cs6360/Lectures/frustum.pdf&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Rendering Techniques ==&lt;br /&gt;
&lt;br /&gt;
=== Multi-Pass Rendering ===&lt;br /&gt;
&lt;br /&gt;
Traditional multi-pass rendering takes the straightforward approach of rendering the complete scene twice sequentially, once per eye. Each eye uses separate camera parameters, performing independent [[draw call]]s, culling operations, and shader executions. While conceptually simple and compatible with all rendering pipelines, this approach imposes nearly 2× computational cost—doubling [[CPU]] overhead from draw call submission, duplicating geometry processing on the [[GPU]], and requiring full iteration through all rendering stages twice.&amp;lt;ref name=&amp;quot;unity2024&amp;quot;&amp;gt;Unity. &amp;quot;How to maximize AR and VR performance with advanced stereo rendering.&amp;quot; Unity Blog, 2024. https://blog.unity.com/technology/how-to-maximize-ar-and-vr-performance-with-advanced-stereo-rendering&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Single-Pass Stereo Rendering ===&lt;br /&gt;
&lt;br /&gt;
Single-pass stereo rendering optimizes by traversing the scene graph once while rendering to both eye buffers.&amp;lt;ref name=&amp;quot;nvidia2018&amp;quot;&amp;gt;NVIDIA Developer. &amp;quot;Turing Multi-View Rendering in VRWorks.&amp;quot; NVIDIA Technical Blog, 2018. https://developer.nvidia.com/blog/turing-multi-view-rendering-vrworks/&amp;lt;/ref&amp;gt; Single-pass instanced approach uses GPU instancing with instance count of 2, where the [[vertex shader]] outputs positions for both views simultaneously. Example shader code:&lt;br /&gt;
&amp;lt;code&amp;gt;&lt;br /&gt;
uniform EyeUniforms {&lt;br /&gt;
mat4 mMatrix[2];&lt;br /&gt;
};&lt;br /&gt;
vec4 pos = mMatrix[gl_InvocationID] * vertex;&lt;br /&gt;
&amp;lt;/code&amp;gt;&lt;/div&gt;</summary>
		<author><name>Xinreality</name></author>
	</entry>
</feed>