1. What Is Spatial Computing?
Spatial computing refers to human-machine interaction in which the digital world is merged with or overlaid onto three-dimensional physical space. Unlike traditional screens — flat, fixed, 2D rectangles — spatial computers understand and interact with the geometry of the real world.
The term was first popularised by Simon Greenwold in his 2003 MIT thesis, but it was Apple's introduction of the Marketing term in 2023 at WWDC that brought it into mainstream technology discourse. Spatial computing encompasses:
- Augmented Reality (AR) — Digital overlays on the real world (e.g., iPhone AR apps, HoloLens).
- Mixed Reality (MR) — Digital objects that interact with real-world surfaces and objects.
- Virtual Reality (VR) — Fully immersive digital environments.
- Passthrough XR — Camera-based real-world view with digital content overlaid (Apple Vision Pro's primary mode).
2. Apple Vision Pro Overview
Apple Vision Pro launched on February 2, 2024 in the United States at a starting price of $3,499. It is Apple's first entirely new product category since the Apple Watch (2015) and represents over seven years of development.
Apple describes it not as a VR or AR headset but as a spatial computer — a device that runs a full operating system (visionOS), supports multiple floating windows, accepts keyboard and trackpad input, and integrates with the rest of the Apple ecosystem via Continuity.
Key differentiators
- Eye + hand + voice input — No controllers needed. You look to focus, pinch to click, and speak to dictate.
- EyeSight — An outward-facing display shows the wearer's eyes to people nearby, reducing social isolation.
- M2 + R1 dual-chip — M2 handles general compute; the dedicated R1 chip processes sensor data with <12ms latency for a natural passthrough view.
- micro-OLED displays — 23 million pixels across two displays — more than a 4K TV per eye.
- Optic ID — Iris-based biometric authentication.
3. Hardware Specifications
- Chipset: Apple M2 (main) + Apple R1 (sensor processing)
- RAM: 16 GB unified memory (M2 die)
- Storage: 256 GB, 512 GB, or 1 TB SSD
- Display: Two micro-OLED panels, ~3660 × 3200 px per eye, 100 Hz
- FOV: ~100° horizontal, ~90° vertical (approximate)
- Cameras & Sensors: 12 cameras, 5 sensors, 6 microphones, TrueDepth camera
- Eye tracking: High-resolution IR eye tracking (used for input and personalized calibration)
- Audio: Spatial Audio, Personalized Spatial Audio via ear canal mapping
- Battery: External battery pack — 2 hours continuous use (video); 2.5 hours general use
- Weight: ~600–650 g (Solo Knit Band configuration)
- Connectivity: Wi-Fi 6, Bluetooth 5.3
- Price: $3,499 (256 GB) to ~$3,899 (1 TB, with accessories)
4. visionOS Architecture
visionOS is Apple's spatial operating system, derived from iPadOS and macOS but rebuilt from the ground up for three-dimensional interactions.
Key layers
- Shared Space — The default context. Multiple apps coexist as floating windows in the user's real environment. Apps render in a shared scene and must respect each other's boundaries.
- Full Space — An app requests exclusive control of the environment, showing only its own content. Used for immersive games and VR experiences.
- Ornaments — UI elements that float outside a window's bounds, used for contextual tools and controls.
- Volumes — 3D app windows that display objects with depth, visible from multiple angles.
- RealityKit compositor — The rendering engine that composites virtual content onto the passthrough camera feed in real time with sub-frame latency.
5. Input System: Eyes, Hands & Voice
Apple Vision Pro has no physical controllers. All interaction happens through three input modalities:
Eye tracking
IR cameras map the user's gaze at <2 ms latency. Where you look is your cursor. UI elements subtly highlight as your eyes rest on them — a design principle Apple calls hover state.
Hand gestures
- Tap (pinch) — Touch thumb to index finger to confirm a selection.
- Scroll — Flick fingers upward/downward while looking at a scrollable area.
- Zoom/rotate — Use two hand pinch and spread/twist motions.
- Drag — Pinch and hold, then move the hand.
- Hands can be at your lap — cameras track them without needing to raise arms.
Voice (Siri)
Siri handles dictation, app launching, search, and system commands. Many tasks are faster by voice ("Hey Siri, open Safari", "Hey Siri, take a spatial photo").
External input
- Bluetooth keyboard and trackpad for productivity.
- Mac Virtual Display — Mirror your Mac screen as a massive floating display inside Vision Pro via Continuity.
- iPhone and iPad can be used as input devices via Continuity.
6. Display & Passthrough Technology
Vision Pro does not have optical see-through like HoloLens. Instead, it uses video passthrough: cameras capture the real world and display it on the inner micro-OLED screens at <12 ms latency (thanks to the R1 chip), making it imperceptible to most users.
- Passthrough quality: Near-retina resolution, wide dynamic range, accurate colour (key for professional use).
- Digital Crown: Rotating the crown on the device frame shifts between fully immersive VR and full passthrough, continuously and in real time.
- Environments: Built-in virtual environments (Mt Hood, Haleakalā, etc.) replace the passthrough with an immersive backdrop at adjustable opacity.
7. Spatial Audio
Vision Pro includes Personalized Spatial Audio: it uses the TrueDepth camera to scan the shape of the user's ears, then generates a Head-Related Transfer Function (HRTF) specific to that person. The result is audio that sounds like it originates from a precise point in three-dimensional space around you.
Audio does not require headphones — the built-in speaker array projects audio toward the user's ears with enough precision for private listening in quiet environments. Bluetooth AirPods are supported for higher fidelity and noise isolation.
8. Using Apple Vision Pro Day-to-Day
After two years of retail availability, a clearer picture has emerged of how people actually use Vision Pro:
Common patterns
- Multi-window workspace — Floating Safari, Mail, Slack, and Notes windows arranged around a physical desk. Some users report a 20–30% productivity boost from the infinite screen real estate.
- Movie & TV watching — Apple TV+ spatial videos, 3D films, and the 180° immersive format have strong user satisfaction scores.
- Video calls — FaceTime with Persona (Apple's avatar of the user, generated from face scans) enabled natural eye contact in spatial calls.
- Reading & research — Floating a massive virtual PDF over a real desk is genuinely useful for document-heavy work.
Ergonomics
The weight (~600 g) remains the primary complaint. Apple recommends using the Dual Loop Band for extended use. Most users report comfortable use up to 2–3 hours before needing a break.
9. Key Spatial Apps & Experiences
Apple first-party
- Apple TV+ Immersive Video — 180° 8K spatial video with binaural audio. Titles include Prehistoric Planet Immersive and Submerged.
- Spatial Photos & Videos — Shot with iPhone 15 Pro+ or Vision Pro; viewable in stunning depth.
- Mindfulness — Guided meditation in virtual environments.
- Freeform — Collaborative whiteboard in 3D space.
Third-party highlights
- Microsoft 365 — Word, Excel, PowerPoint, Teams all available as visionOS apps.
- DJI Fly — View drone footage and telemetry spatially.
- PTC Vuforia — Industrial AR maintenance guidance.
- Blackbox AR — Surgical planning overlaid on patient anatomy.
- Sky Guide — Point at the night sky and see star maps spatially overlaid in passthrough.
10. Productivity & Work Use Cases
- Virtual screen expansion — Use Mac Virtual Display to give yourself a 100-inch screen in any room.
- Reference material while working — Float documentation, designs, or spreadsheets alongside your primary Mac window.
- Focused deep work — Use Environments to replace a noisy office with a quiet mountain retreat.
- Spatial presentations — Present 3D models, architectural plans, and medical imaging to clients in a shared spatial session.
- Code review — Float multiple terminal windows and code editor panes around you.
11. Entertainment & Immersive Media
Vision Pro's entertainment credentials are its strongest consumer selling point:
- Cinema mode — Any video app fills your field of view with a virtual 30-foot cinema screen.
- Spatial video — The "you are there" feeling of spatial video shot on iPhone 15 Pro or Vision Pro is the most common reason users cite for purchase satisfaction.
- Gaming — Apple Arcade games run spatially. Developer interest has grown significantly, with titles like Synth Riders and Rec Room offering full spatial experiences.
- Live events — MLS Season Pass (Apple TV+) added a spatial viewing mode with floating stats and multi-angle feeds.
12. Enterprise Adoption
Priced at enterprise levels, Vision Pro has found its strongest growth in B2B markets:
Industries leading adoption
- Healthcare — Surgical planning, anatomy visualization, telemedicine consultations with 3D imaging. Stryker and Medtronic ship Vision Pro-compatible surgical guidance apps.
- Aerospace & manufacturing — Boeing uses Vision Pro for aircraft assembly guidance, reducing wiring errors by 25% in pilot programs.
- Architecture & construction — Autodesk's spatial tools allow architects to walk through full-scale 3D building models before construction begins.
- Field service — Technicians overlay repair instructions on real equipment. Companies report 15–40% reduction in service time.
- Training & simulation — Military, police, and medical training use fully immersive simulated environments.
Apple's dedicated Apple for Business and AppleCare for Enterprise programs support volume deployment of Vision Pro with MDM integration (Jamf, Mosyle).
13. Developer Tools: RealityKit, ARKit & SwiftUI
Developing for visionOS uses Apple's existing frameworks, extended for spatial interaction:
RealityKit
Apple's 3D rendering framework. Handles entity-component-system (ECS) scene graphs, physics simulation, animation, lighting, and shader materials optimised for visionOS's compositor.
ARKit
Scene understanding — detects planes, objects, meshes, hands, and the room layout. On Vision Pro, ARKit provides the spatial understanding that lets apps anchor content to real surfaces.
SwiftUI
2D windows in visionOS are built with SwiftUI exactly as on iOS/iPadOS. SwiftUI gains visionOS-specific modifiers like .glassBackgroundEffect(), .hoverEffect(), and 3D layout primitives.
Reality Composer Pro
A macOS tool for building and previewing 3D scenes, adding behaviours, and fine-tuning RealityKit entities before embedding them in Xcode projects.
xrOS Simulator
Run and test visionOS apps on a Mac without physical hardware. Supports gaze simulation and window placement testing.
14. Building Your First Spatial App
// SwiftUI Volume app — a 3D rotating globe
import SwiftUI
import RealityKit
struct GlobeView: View {
var body: some View {
RealityView { content in
// Load a USDZ model
if let globe = try? await ModelEntity(
named: "Earth",
in: Bundle.main
) {
globe.scale = [0.5, 0.5, 0.5]
content.add(globe)
}
}
.frame(depth: 200) // visionOS depth modifier
}
}
@main
struct GlobeApp: App {
var body: some Scene {
WindowGroup {
GlobeView()
}
.windowStyle(.volumetric) // Enable 3D Volume
.defaultSize(width: 0.4, height: 0.4, depth: 0.4,
in: .meters)
}
}
Build in Xcode 16+, select the visionOS destination, and run on the Simulator or device.
15. Spatial Design Guidelines
Apple's Human Interface Guidelines for visionOS define key principles:
- Comfort & ergonomics — Place content within a comfortable field of view (not too close, not requiring neck strain). Default placement is ~2 metres away.
- Familiar 2D + new spatial — Start with 2D window paradigms users know; introduce 3D volumes and immersive spaces purposefully.
- Glass morphism — The default visionOS material is a frosted glass that adapts to the environment behind it. Use it to feel native.
- Visual feedback for gaze — Hover states must be visible but subtle. Users need confirmation that the system has registered where they are looking.
- Persistence — Windows remember their position in the real world. Apps should save and restore window positions.
16. Apple Vision Pro 2: What to Expect
Apple has not officially announced a Vision Pro successor, but supply chain reports and analyst forecasts point to a Vision Pro 2 in late 2026 or 2027:
- M4 chip — Significantly faster performance for complex spatial scenes and on-device AI processing.
- Lighter design — Apple is engineering a lighter form factor (~100–150 g lighter) via new materials and a redesigned battery.
- Lower price point — Rumoured $2,499 to expand the addressable market.
- Wi-Fi 7 support for higher-bandwidth wireless streaming.
- Improved Persona quality — More realistic avatar generation for FaceTime.
Separately, a lower-cost Apple Vision device (without the Pro designation) targeting the $1,500–$2,000 range has been reported by Bloomberg's Mark Gurman as a longer-term project.
17. Vision Pro vs Meta Quest 3 & 3S
| Feature | Apple Vision Pro | Meta Quest 3 | Meta Quest 3S |
|---|---|---|---|
| Price | $3,499+ | $499 | $299 |
| Chipset | M2 + R1 | Snapdragon XR2 Gen 2 | Snapdragon XR2 Gen 2 |
| Display | micro-OLED (23M px total) | LCD pancake lens | LCD Fresnel lens |
| Input | Eyes + hands + voice | Touch controllers + hand tracking | Touch controllers + hand tracking |
| Battery | ~2 h (external pack) | ~2.5 h (built-in) | ~2.5 h (built-in) |
| OS | visionOS | Meta Horizon OS | Meta Horizon OS |
| Enterprise focus | Strong | Growing | Limited |
| Gaming library | Limited | Very strong | Very strong |
18. Privacy & Security
- Optic ID — Iris data is encrypted and never leaves the device. Neither Apple nor apps can access raw iris data.
- Eye tracking privacy — Apps receive only the result of gaze-based interactions (which UI element was tapped), not the raw gaze direction. This prevents apps from building gaze heatmaps without consent.
- Camera data — Passthrough and depth camera streams are processed entirely in a restricted hardware enclave and are not accessible to apps without explicit user permission (the same App Privacy model as iPhone cameras).
- Persona data — The face mesh used to generate your Persona is stored and processed on-device.
19. Current Limitations & Criticisms
- Price — $3,499 remains the primary barrier. Consumer adoption is limited compared to Meta Quest.
- Weight — ~600 g causes neck strain in long sessions. The external battery pack adds awkwardness.
- Battery life — 2 hours is insufficient for a full workday without the tethered battery.
- App library — While growing, the truly spatial (not just iPad port) app library is limited compared to Meta Quest's gaming library.
- Social experience — The EyeSight display is unconvincing close up, and wearing Vision Pro in public remains socially awkward.
- Requires Apple ecosystem — Best features (Mac Virtual Display, Continuity, Spatial Photos) only work with other Apple devices.
20. The Future of Spatial Computing
Spatial computing in 2026 is where the smartphone was in 2009 — present, promising, and not yet mainstream. The trajectory points toward:
- Lighter form factors — As optics and batteries advance, headsets will approach the size and weight of sunglasses within 5–7 years.
- AI integration — On-device LLMs will enable real-time contextual awareness — Vision Pro could identify objects in your room, translate signs, or recognize people with your permission.
- Persistent shared spaces — Multiple users co-inhabiting a spatial computing environment, collaborating on 3D objects in real time.
- Mainstream price points — Below $1,000 within 3 years as component costs fall and competition intensifies.
- Healthcare as a driving market — Surgical navigation, physical therapy, and diagnostic imaging are already generating compelling ROI.
21. FAQ
- Can I use Apple Vision Pro with glasses?
- Yes, using custom ZEISS Optical Inserts — magnetic prescription lens attachments. Standard Apple Vision Pro prescription inserts ($149 for readers, $299 for custom prescription) clip into the device and match your corrective lenses.
- Does Apple Vision Pro work as a standalone device?
- Yes. All compute is on-device (M2 + R1). It does not require an iPhone, iPad, or Mac to function, though those devices unlock additional features via Continuity.
- Can developers outside the US build visionOS apps?
- Yes. You can develop and publish visionOS apps from any country using Xcode and an Apple Developer account. The device itself is sold in 30+ countries as of 2026.
- Is visionOS based on iPadOS or macOS?
- visionOS is its own operating system derived from both iPadOS and macOS. iPadOS apps run unmodified in a compatibility mode, but native visionOS apps use platform-specific frameworks.
- Can I use Apple Vision Pro for gaming?
- Yes, but the library is smaller than Meta Quest. Apple Arcade provides dozens of spatial games, and the App Store has growing third-party support. For hardcore gaming, Meta Quest 3 still has a better library.
22. Glossary
- visionOS
- Apple's spatial operating system that powers Apple Vision Pro.
- RealityKit
- Apple's 3D rendering and simulation framework for building spatial experiences.
- Passthrough
- Video feed from external cameras displayed on inner screens to create a see-through effect.
- EyeSight
- The outward-facing display on the front of Apple Vision Pro that shows the wearer's eyes to nearby people.
- Persona
- A digital avatar generated from a face scan used to represent the user in FaceTime and spatial calls.
- Shared Space
- The default visionOS context where multiple apps coexist floating in the real environment.
- HRTF
- Head-Related Transfer Function — a mathematical model of how the ears perceive the direction of sound, used to create spatial audio.
- micro-OLED
- A display technology combining the high contrast and black levels of OLED with the sub-micron pixel sizes achievable on silicon substrates, enabling extremely high pixel density in small displays.
23. References & Further Reading
- Apple Vision Pro — Official Page
- Apple Developer: visionOS
- RealityKit Documentation
- HIG: Designing for visionOS
- Hello World — visionOS Sample App
- Bloomberg: Apple Vision Pro Review
24. Conclusion
Apple Vision Pro is the first serious entry in what Apple is betting will be the next major computing platform. In 2026, it remains a premium professional and enterprise device — genuinely useful for specific high-value workflows — rather than a mainstream consumer product.
The critical questions for the next 3 years are weight reduction, battery life, price accessibility, and whether a truly killer spatial-native app emerges that makes mass-market consumers want one. The technology and platform are ready; the use case and form factor still need to catch up.
If you work in healthcare, architecture, engineering, design, or media production — Apple Vision Pro may already be worth the investment. For everyone else, keep watching: the device that changes everything may still be one generation away.