visionOS
visionOS is a spatial operating system developed by Apple Inc. for the Apple Vision Pro headset, marking the company's entry into mixed reality computing.[1] Announced on June 5, 2023, as the world's first operating system designed specifically for spatial experiences, visionOS blends digital content with the physical world to enable immersive interactions.[1] It powers the Apple Vision Pro, a wearable device that launched in the United States on February 2, 2024, and has since expanded to additional markets.[2] Built on the foundations of macOS, iOS, and iPadOS, visionOS allows developers to create applications using familiar Apple frameworks and tools, supporting deployment across the company's ecosystem.[3] Users navigate the interface through natural inputs including eye tracking for selection and scrolling, hand gestures for manipulation at up to 90Hz refresh rates, and voice commands via Siri.[4] Key features include spatial widgets that anchor in the user's environment, 3D Safari browsing for immersive web experiences, and the ability to transform photos into spatial memories shareable via SharePlay.[5] The system integrates seamlessly with other Apple devices, such as unlocking an iPhone with a glance and incorporating Apple Intelligence for enhanced communication, writing, and productivity tools starting with visionOS 2.4 in March 2025.[6] As of visionOS 2.6, released in June 2025, the platform introduces advanced spatial experiences like anchored folders in the Home View, improved accessibility options including Live Recognition for object identification, and support for the upgraded Apple Vision Pro with the M5 chip announced in October 2025.[7][8] These updates emphasize visionOS's role in redefining computing by providing an infinite canvas for apps, games, and productivity, fostering an ecosystem with over 600 native applications at launch and ongoing expansions.[9]History and development
Announcement and initial development
Apple's development of visionOS stemmed from its broader augmented reality (AR) and virtual reality (VR) initiatives, which began in earnest around 2015 under the Technology Development Group led by Mike Rockwell.[10] The operating system evolved from prototypes like the N301 mixed-reality headset and N421 lightweight AR glasses, concepts tied to early explorations of "Apple Glass" as a future AR platform.[10] These efforts ramped up significantly by 2017, with the project initially codenamed xrOS internally before its rebranding.[11] By 2020, approximately 1,000 engineers were dedicated to AR and VR projects, focusing on hardware-software integration to create immersive experiences.[10] visionOS drew foundational influences from macOS and iOS, incorporating shared frameworks such as SwiftUI and RealityKit to enable seamless app development across Apple's ecosystem.[12] The operating system was publicly announced on June 5, 2023, during Apple's Worldwide Developers Conference (WWDC 2023) keynote, positioned as the software foundation for the Apple Vision Pro spatial computer.[1] This reveal highlighted visionOS's role in blending digital content with the physical world through spatial computing. Development emphasized co-design between software and the Vision Pro's hardware, powered by an M2 chip running visionOS, paired with dual 4K micro-OLED displays delivering over 23 million pixels and eye-tracking via four infrared cameras for precise gaze-based interactions.[13][1] At the event, Apple showcased visionOS through live demos of spatial applications, including immersive 3D FaceTime calls with participant avatars and a Photos app rendering volumetric memories, alongside dynamic environments such as a serene Yosemite landscape or a cosmic starry night to enhance user immersion.[1][14]Developer previews and betas
At the 2023 Worldwide Developers Conference (WWDC), Apple introduced updates to Xcode 15 beta, including the visionOS software development kit (SDK) to enable developers to build spatial computing applications for Apple Vision Pro.[15] This release featured extensions to RealityKit for rendering 3D content and immersive experiences, as well as enhancements to SwiftUI for creating adaptive user interfaces that respond to spatial environments.[15] These tools allowed developers to prototype and preview apps using familiar Apple frameworks without immediate access to physical hardware.[16] The first developer beta of visionOS 1.0 was made available on June 21, 2023, bundled with Xcode 15 beta 2, marking the start of pre-launch testing for enrolled Apple Developer Program members.[15] Subsequent iterative betas followed throughout 2023, including beta 2 on July 25, beta 3 in August, and up to beta 7 in December, with each update focusing on improving system stability, refining gesture recognition, and resolving performance issues in simulated environments.[17] Unlike other Apple operating systems, no public beta was released for visionOS 1.0 in 2023 due to the limited availability of Vision Pro hardware for testing.[18] Developers faced key challenges in testing spatial features, which were addressed through the visionOS simulator integrated into Xcode, enabling simulation of 3D spatial environments and various room layouts on Apple Silicon Macs.[12] The simulator also handled emulation of passthrough camera feeds, allowing developers to preview mixed reality overlays without real-time hardware access.[19] Specific APIs, such as extensions to ARKit tailored for visionOS, provided tools for anchoring virtual content to the physical world and integrating spatial tracking. In late 2023, Apple transitioned to broader enterprise testing via the AppleSeed for IT program, inviting select organizations to evaluate beta software in controlled work environments ahead of the February 2024 launch.[20] This shift supported validation of visionOS in professional settings, including compatibility with managed devices and security protocols.[21]Launch and post-launch updates
visionOS was officially launched on February 2, 2024, alongside the availability of the Apple Vision Pro headset in all U.S. Apple Store locations and online, marking Apple's entry into spatial computing for consumers.[2] The operating system powered the device's immersive experiences from day one, with core features like spatial multitasking and eye-and-hand tracking integrated seamlessly into the hardware.[22] Initial adoption was strong, with pre-orders surpassing 200,000 units within the first 10 days of availability on January 19, 2024, indicating robust early demand despite the device's premium pricing.[23] Software updates for visionOS follow an over-the-air (OTA) delivery model akin to iOS and other Apple platforms, allowing users to install patches and enhancements directly on the Vision Pro without needing external hardware. Post-launch stability improvements began promptly, with visionOS 1.0.3 released on February 12, 2024, providing essential bug fixes and introducing a passcode reset option to enhance user security and usability. While specific eye-tracking calibration challenges were reported by early users, subsequent updates like visionOS 1.1 in March 2024 addressed related inaccuracies, such as drift after Persona capture, improving tracking reliability for daily interactions. These patches were crucial for refining the headset's core input methods in real-world scenarios. International expansion accelerated in mid-2024, with Vision Pro and visionOS rolling out to additional markets starting June 28, including China mainland, Hong Kong, Japan, and Singapore, supported by localized language and region-specific updates to broaden accessibility.[24] Further rollouts continued to other regions throughout the year, aligning software enhancements with global hardware availability. visionOS releases are closely synchronized with Vision Pro firmware updates, ensuring that sensor calibrations for components like cameras and inertial measurement units remain optimized across software versions for consistent performance. This integrated approach minimizes compatibility issues and supports ongoing refinements to spatial tracking and environmental passthrough features. In 2025, visionOS continued to evolve with the release of version 2.4 in March, introducing Apple Intelligence features for enhanced communication, writing, and productivity.[6] This was followed by visionOS 2.6 in June, which added anchored folders in the Home View, improved accessibility including Live Recognition, and support for advanced spatial experiences.[7] October saw the announcement of an upgraded Apple Vision Pro model featuring the M5 chip and a dual-knit band, further advancing the platform's hardware-software integration.[8]Technical architecture
Core framework and hardware integration
visionOS is built upon a variant of Apple's established operating system foundation, sharing the Darwin core with iOS and macOS, which utilizes the XNU hybrid kernel to manage system resources and hardware abstraction.[25] This architecture enables visionOS to leverage familiar development tools and APIs while extending them for spatial computing, ensuring efficient multitasking and resource allocation in a mixed-reality environment. The system supports Metal 3, Apple's graphics and compute API, which facilitates high-performance rendering of 3D content and visual effects directly on the device's GPU. At the hardware level, visionOS is tightly integrated with the Apple Vision Pro's components, including dual micro-OLED displays that collectively deliver over 23 million pixels for immersive visuals exceeding 4K resolution per eye.[13] The operating system processes input from 12 outward-facing cameras, each with 12-megapixel resolution, to enable high-fidelity color passthrough that blends the real world with digital overlays in real time.[13] This integration initially relied on the Apple M2 chip, featuring an 8-core CPU, 10-core GPU, and 16-core Neural Engine, paired with 16 GB of unified memory to handle complex spatial computations without compromising responsiveness. In October 2025, Apple released an upgraded Apple Vision Pro with the M5 chip, featuring a 10-core CPU (4 performance cores and 6 efficiency cores), 10-core GPU with hardware-accelerated ray tracing, 16-core Neural Engine, and 16 GB unified memory, enhancing performance for AI-driven spatial experiences and efficiency in a wearable form factor.[8][13] The spatial rendering engine in visionOS combines RealityKit for authoring and rendering 3D content with AVFoundation for handling mixed-reality video streams, allowing developers to create seamless immersive experiences.[26] RealityKit manages entity-based 3D scenes, physics simulations, and animations, while AVFoundation supports immersive video playback in formats like MV-HEVC, enabling spatial audio and video integration.[26] Power management is optimized through the chip's efficiency cores and the external battery pack, providing up to 2 hours of untethered use or extended sessions when connected, balancing performance with thermal constraints in a wearable form factor, with the M5 upgrade offering improved battery efficiency.[13] System-level APIs further enhance hardware synergy, with SceneKit providing tools for importing, manipulating, and rendering 3D assets in spatial contexts, though RealityKit is increasingly recommended for new visionOS development.[27] Core Motion fuses data from the device's sensors, including head and eye tracking via infrared cameras and LEDs, to deliver precise motion estimates for intuitive interactions and scene stabilization.[28] These APIs ensure low-latency tracking, with the Neural Engine accelerating fusion algorithms to maintain fluid rendering even during rapid movements.[29]Security and privacy features
visionOS incorporates robust security and privacy features tailored to its spatial computing environment, emphasizing on-device processing and user consent to protect sensitive data such as eye and hand movements. These mechanisms build on Apple's established iOS security model while addressing unique challenges like real-time environmental awareness and biometric interactions.[30][31] Eye and hand tracking data is processed entirely on-device to prevent leakage, with the Secure Enclave ensuring isolation of biometric information. Gaze data remains private until a user explicitly selects content via a gesture like a finger tap, and it is not shared with apps or Apple otherwise. Hand setup data for gesture recognition is also stored locally and only shared with apps upon explicit permission, with no automatic upload to the cloud without user consent. This on-device approach minimizes privacy risks in spatial interactions.[30][32][33] App sandboxing in visionOS extends iOS protections by isolating applications to prevent interference with the system or other apps, using unique home directories and runtime checks like Address Space Layout Randomization (ASLR). For spatial elements, apps running in Shared Space cannot access surroundings data or head orientation, while those in Full Space require user permission to access environmental data limited to a 5-meter radius, enforcing spatial boundaries that restrict apps from intruding on the user's physical environment. Entitlements further limit access to system resources, ensuring apps cannot escalate privileges.[34][31][30] Optic ID provides biometric authentication through iris scanning, using infrared cameras and LEDs to create encrypted mathematical representations of the iris stored exclusively in the Secure Enclave. This data never leaves the device or backs up to iCloud, and apps receive only a success or failure signal without accessing raw biometrics. Integrated with passkeys for passwordless sign-ins, Optic ID secures device unlocking, purchases, and app access, leveraging Apple silicon's hardware isolation for protection.[30][32][35] Passthrough security relies on on-device mapping of surroundings via cameras and LiDAR, with the resulting scene mesh encrypted using the device passcode to safeguard real-time AR overlays. Apps must obtain explicit permission to access camera or microphone feeds for passthrough features, and visual indicators (green for camera/mic use, orange for mic alone) alert users to active access, preventing unauthorized environmental scanning. This framework protects against unintended data exposure in mixed-reality scenarios.[30][33][36] Update mechanisms in visionOS use digitally signed firmware and software to verify integrity and prevent tampering during installation. Apple issues regular security updates addressing vulnerabilities, such as the 2024 GAZEploit exploit in eye-tracking that enabled keystroke inference (CVE-2024-40865), which was patched in visionOS 1.3 by suspending the Persona view when the virtual keyboard is active. Automatic update options ensure devices receive these protections promptly, with incident responses including rapid CVE disclosures and mitigations. The Secure Enclave's chip-level isolation further bolsters overall hardware security by segregating sensitive operations.[37][38]Compatibility with other Apple platforms
visionOS integrates seamlessly with other Apple platforms, enabling users to leverage apps, data, and features across devices such as iPhone, iPad, and Mac. This interoperability is facilitated through compatibility layers, shared services, and Continuity features, allowing for a unified ecosystem without requiring users to switch devices frequently.[39] iPadOS and iOS apps can run on visionOS via a dedicated compatibility layer that simulates an iPad-like environment, adapting the app to the spatial interface by rendering it in a stereoscopic window. Developers build these compatible versions by linking against the iOS SDK in Xcode 15 or later and enabling the "Designed for iPad" runtime destination, making the apps automatically available in the visionOS App Store upon updating the Apple Developer Program agreement. However, certain features like Core Motion sensors, location services, HealthKit access, and camera functionalities are unavailable in this mode, and apps must use adaptive layouts to handle the spatial display. For macOS productivity, visionOS supports Mac Virtual Display, which virtualizes a 4K external display for any compatible Mac running macOS Sonoma 14 or later, allowing users to interact with full macOS apps in a spatial window alongside native visionOS content. This feature streams the Mac's screen wirelessly, supporting multiple virtual displays for enhanced multitasking, though it requires both devices to be signed into the same iCloud account.[40][41] Continuity features extend this integration by enabling fluid transitions between devices. Handoff allows starting tasks like FaceTime calls on Apple Vision Pro and continuing them on iPhone, iPad, or Mac, provided the devices meet minimum software requirements such as iOS 16 or macOS Ventura 13. Universal Clipboard supports copying text, images, or files on one device and pasting them on another, including into spatial views on visionOS, with iOS 10 or macOS Sierra 10.12 as prerequisites. Additional capabilities include Instant Hotspot for tethering to an iPhone's cellular data (iOS 8.1 or later) and text message forwarding from iPhone to visionOS (iOS 8.1 or later). Mirror My View uses AirPlay to project the visionOS interface onto a Mac screen for demonstrations, requiring macOS Monterey 12 or later.[39] Data synchronization across platforms is handled primarily through iCloud, which ensures seamless sharing of photos, notes, files, and even spatial annotations created in visionOS apps. For instance, passwords and passkeys sync via iCloud Keychain, keeping credentials updated across iPhone, iPad, Mac, and Vision Pro when enabled in settings. AirDrop facilitates quick transfers of 3D assets, photos, videos, and documents between visionOS and nearby Apple devices supporting iOS 7, iPadOS 13, or macOS 10.10, operating over Bluetooth and Wi-Fi for proximity-based sharing.[42][43][44] Cross-device APIs further enhance interoperability, with HealthKit providing shared access to fitness and health data stored on iPhone, Apple Watch, or iPad, allowing visionOS apps to read and write metrics like workouts or state of mind in a spatial context while maintaining privacy controls. CarPlay extensions enable navigation apps on visionOS to integrate in-VR driving directions, extending iPhone-based CarPlay functionality into immersive environments without invoking full CarPlay code in compatible apps.[45][46][47] Despite these integrations, limitations exist: there is no direct access to the iOS App Store on visionOS; instead, compatible iPadOS apps appear in the visionOS store, but developers must opt-in to create native visionOS versions for full spatial optimization. Background execution modes for location or Bluetooth are restricted in compatible apps, and core kernel similarities with other platforms do not extend to unrestricted feature parity.[40][48]User interface and experience
Spatial computing paradigms
Spatial computing in visionOS represents a paradigm that seamlessly integrates digital content into the physical environment, enabling users to interact with virtual elements as if they were part of the real world. This is achieved through advanced techniques such as spatial anchors, which fix virtual objects to specific locations in the user's physical space for persistent placement, and occlusion, where real-world objects realistically obscure virtual ones to maintain depth perception and immersion.[49][50] These mechanisms, powered by RealityKit and ARKit frameworks, allow developers to create experiences that blend augmented reality (AR) overlays with the user's surroundings without disrupting spatial awareness.[51] Central to this paradigm are volumetric environments, where windows and volumes serve as flexible containers for content that can be resized, rotated, and pinned to real-world surfaces or locations. Unlike traditional flat interfaces, these volumetric windows enable viewing 3D content from multiple angles and support dynamic scaling to fit user needs, fostering an infinite workspace free from bezels or screen edges that constrain traditional displays.[52][53][54] Users can arrange multiple windows in a boundless 3D canvas, positioning them at varying depths and orientations to enhance productivity and immersion.[54] As of visionOS 2.6 (June 2025), spatial widgets integrate into the user's environment, anchoring in place with customizable frame thickness, color, and depth for persistent access to information like clocks or weather. The Home View supports anchored folders for organizing apps, and the Control Center has been redesigned for quicker management of settings such as Guest User and Focus modes.[7] visionOS supports a spectrum of reality modes to cater to different levels of immersion: full virtual reality (VR) through Immersive Spaces that envelop the user in entirely digital environments; mixed reality (MR) via passthrough in Shared or Full Spaces, where high-fidelity camera feeds overlay AR elements with depth mapping for realistic integration; and lighter AR overlays that augment the physical world without full occlusion.[55][56] Depth mapping ensures virtual objects respect real-world geometry, such as planes and surfaces detected by the device's sensors.[50] Enhancing these visual paradigms is spatial audio, which creates immersive 3D soundscapes by simulating directional and distance-based audio cues using head-related transfer functions (HRTF) for binaural rendering. The system employs a six-microphone array with directional beamforming to capture and process environmental sounds in real time, allowing audio to adapt dynamically to the user's head movements and room acoustics via ray tracing.[13][57] This integration ensures that sound sources appear anchored in the spatial environment, complementing visual elements for a cohesive experience.[58] To deliver smooth interactions, visionOS targets a 90 Hz refresh rate as the baseline for low-latency rendering, with support for higher rates like 96 Hz, 100 Hz, and 120 Hz depending on content.[13] Eye-tracking capabilities enable foveated rendering, which prioritizes high-resolution detail in the user's gaze direction while optimizing peripheral areas, reducing computational load and enhancing performance without perceptible lag.[59]Gestures and input methods
visionOS primarily facilitates user interaction through a combination of eye tracking, hand gestures, and voice commands, enabling hands-free and intuitive navigation in spatial environments without the need for physical controllers. Users target elements by gazing at them and confirm actions via subtle hand movements, creating a seamless blend of natural inputs that prioritize comfort and precision. This input paradigm supports both indirect manipulation, where users interact with virtual objects from a distance, and direct touch interactions in closer proximity, as outlined in Apple's human interface guidelines.[60] Eye tracking serves as the foundational selection mechanism in visionOS, allowing users to direct their gaze toward app icons, buttons, or content to highlight and prepare them for activation. Upon initial setup, users calibrate eye tracking by following on-screen instructions in the Settings app under Eyes & Hands, which involves looking at specific points to map gaze accurately; this process can be redone at any time for optimal performance. Selection occurs through gaze combined with a hand gesture rather than dwell-time alone, ensuring deliberate interactions while maintaining responsiveness in dynamic spatial scenes.[61] As of visionOS 2.6, the Look to Scroll feature enables customizable eye-based scrolling for navigation.[7] Hand gestures provide the primary means for confirming selections and manipulating content, tracked using the device's outward-facing cameras and ARKit framework without requiring handheld controllers. Core gestures include pinching the thumb and index finger together to select or tap an item after gazing at it, pinching and dragging to reposition windows or objects, and flicking the wrist after a pinch to scroll through lists or dismiss panels. Additional gestures support zooming via two-handed pinch and drag, rotation through circular pinches, and contextual menus by pinching and holding, all designed for ergonomic use in mixed reality. These inputs leverage machine learning models within ARKit for robust 3D hand pose estimation, achieving fluid tracking up to 90 Hz in immersive applications for enhanced responsiveness.[62][60][63] Starting with visionOS 2.6, visionOS supports additional input methods including the PlayStation VR2 Sense controllers, which provide 6DoF motion tracking, finger touch detection, and haptic feedback, and the Logitech Muse accessory for precise pointing in collaboration scenarios.[7] Voice integration enhances hands-free control through Siri, which responds to natural language commands tailored for spatial computing, such as opening apps, sending messages, or adjusting environments— for instance, users can say "Siri, open the Notes app" or "Siri, remind me to leave in 20 minutes" to perform tasks without manual input. Siri activation occurs by saying "Siri" followed by the request, with support for setting up features like eye tracking via voice prompts. This system integrates seamlessly with gesture-based navigation, allowing verbal overrides for complex or repetitive actions in shared or multitasking scenarios.[64][65] For text input and precise cursor control, visionOS supports virtual keyboards that users can type on directly using pinch gestures to tap keys, alongside compatibility with Bluetooth peripherals like the Magic Keyboard, Magic Trackpad, or third-party mice. Users connect these devices via Bluetooth settings, enabling traditional input methods for productivity tasks while the headset provides adaptive visibility of keyboards in immersive spaces through features like Keyboard Awareness. In accessibility contexts, fallbacks such as Switch Control allow navigation via head movements or switches when standard eye and hand inputs are unavailable, promoting inclusive use. Input accuracy benefits from ARKit's hand tracking, which delivers sub-centimeter precision in joint detection through on-device machine learning, though environmental factors like lighting can influence performance.[66][67][68]Accessibility and customization
visionOS incorporates robust accessibility features tailored to spatial computing, enabling users with visual, motor, and other impairments to navigate and interact with 3D environments effectively. VoiceOver, the built-in screen reader, delivers audible descriptions of elements in the user's view, including battery status, incoming calls, open applications, and spatial layouts such as app windows and surroundings.[69] When users open apps or shift perspectives in 3D space, VoiceOver plays directional audio cues to indicate changes in the environment, facilitating orientation without visual reliance.[69] This gesture-based system supports single-hand interactions like pinching and swiping to select and explore items, with adjustable speaking rates and pitches for personalized audio feedback.[69] Introduced in visionOS 2.6, Live Recognition uses the device's cameras and on-device machine learning to provide real-time descriptions of surroundings, identify and locate objects (e.g., keys or doors), read printed text, and detect faces for VoiceOver users.[70] For users with low vision, visionOS offers dynamic zoom and magnification tools that scale spatial elements in real time. The Zoom feature provides full-screen magnification or a resizable window lens, adjustable via the Digital Crown or controller, with options to include apps, passthrough video of surroundings, or both, and lock depth to hand interactions for stable focus in 3D.[71] Maximum zoom levels and border colors can be customized in settings, allowing seamless enlargement of virtual content or real-world views captured by the device's cameras.[71] Complementing this, color filters enable adjustments for visual impairments by altering hues, intensity, and contrast across the interface and passthrough video; options include grayscale, inverted colors, or specific tints to reduce strain or enhance readability.[72] Features like Reduce White Point and Reduce Transparency further tone down bright elements and solidify overlays, promoting clearer perception in mixed-reality contexts.[72] Customization in visionOS extends to personalizing interfaces and environments for comfort and efficiency. Users can select light or dark modes for compatible apps, adjusting text size, boldness, and window scaling to suit preferences, with temporary two-handed resizing for dynamic layouts.[73] Widgets, which float in the user's personal spatial volume, allow customization of frame thickness, color, and depth to blend with surroundings, enabling persistent placement in defined areas like desks or walls for quick access to information.[74] Environment themes adapt to these modes, supporting immersive VR spaces with toggled lighting to match user-selected appearances.[4] Assistive technologies in visionOS prioritize inclusive input methods for motor challenges. Switch Control facilitates navigation through item or point scanning, where users trigger selections via external switches or built-in options like head tracking, which moves the pointer based on subtle head movements for precise control in spatial interfaces.[75] This extends to brain-computer interfaces for severe mobility limitations, integrating with the system's gesture adaptations.[70] For tactile feedback, braille display integration works seamlessly with VoiceOver over Bluetooth, supporting international tables, contracted/uncontracted output, and Nemeth Code for math; users can input commands, edit text with auto-conversion, and access note-taking tools directly on the display.[76] In visionOS 2.6, users can save eye and hand calibration data, vision prescriptions, and accessibility settings to an iPhone running iOS 18.6 or later for easy transfer to shared Vision Pro devices, and unlock the iPhone using Face ID while wearing the headset.[7] To accommodate shared use, visionOS supports multiple user profiles through Guest User mode, where temporary setups capture eye and hand data for sessions without overwriting the owner's configuration.[77] Guests create their own Personas—digital avatars—for interactions, but access is restricted from the owner's Optic ID, Apple Pay, and sensitive data, with customizable app permissions to maintain privacy.[77] Owners can monitor sessions via mirroring and end them remotely, ensuring controlled shared experiences in collaborative or family settings.[77] As of visionOS 2.6, Personas feature volumetric rendering for more realistic depth, full side profiles, and over 1,000 variations of glasses and accessories. The system also supports shared spatial experiences, allowing multiple Vision Pro users to collaborate in the same immersive environment via FaceTime, whether in-room or remote.[7]Key features
Immersive environments and multitasking
visionOS provides users with a library of pre-built immersive environments designed to enhance focus and ambiance during use. These include scenic worlds such as Yosemite National Park, Mount Hood, and the Moon surface, which can be selected to create a calming backdrop while interacting with apps.[78] Additionally, the Memories environment leverages spatial photos and videos from the user's library to generate personalized, photo-based scenes that relive captured moments in three dimensions.[1] For customization, developers and users can import bespoke environments created with Reality Composer Pro, Apple's tool for building interactive 3D content that integrates seamlessly into the system.[79] Multitasking in visionOS revolves around an infinite spatial canvas that supports unlimited floating windows, allowing users to position and scale apps freely beyond physical room constraints.[3] Windows exhibit spatial snapping, automatically aligning to nearby surfaces or other elements for organized layouts, which extends the underlying spatial computing paradigms by enabling intuitive placement in mixed reality.[50] For focused experiences, Immersive Spaces provide a theater-like mode where a single app expands to fill the user's field of view, minimizing distractions while maintaining environmental context.[55] Window management emphasizes natural hand gestures for control. Users resize windows by pinching the corner handles and dragging, with the system dynamically scaling content to ensure legibility regardless of distance or size.[80] Arrays of windows can be arranged across the infinite canvas, supporting complex workflows with multiple apps visible simultaneously; while direct linking for synced scrolling across windows is not natively implemented, users can manually align related content for efficient reference.[81] This approach fosters productivity by treating the surrounding space as an extensible desktop. To heighten immersion across setups, visionOS incorporates shared spatial audio, where sound from multiple apps emanates from their respective window positions, creating a layered auditory environment that enhances spatial awareness.[55] Occlusion handling ensures realistic overlaps, with virtual windows and objects respecting depth to appear behind closer elements, including passthrough real-world items via materials like OcclusionMaterial.[82] Performance remains consistent even with demanding multitasking, as the system targets a 90 frames per second refresh rate and employs dynamic scaling to adapt resolution based on content proximity and thermal conditions, supporting fluid operation across more than 10 open windows without perceptible lag.[83]Personas and collaboration tools
Personas in visionOS serve as digital 3D avatars that represent users during remote interactions, generated on-device through a brief facial scan using the Apple Vision Pro headset.[84] These avatars capture a user's facial expressions and hand movements in real time via the device's sensors, enabling natural and expressive communication in virtual spaces.[85] Users can customize their Personas, including adjustments to lighting and backgrounds, to enhance personalization during calls or shared sessions.[85] Spatial Personas extend this functionality by rendering avatars as life-sized, volumetric figures within the user's physical environment, facilitating immersive eye contact and spatial awareness during interactions.[86] Introduced in visionOS 1.1 and refined in subsequent updates, Spatial Personas appear opt-in during compatible sessions, allowing participants to move freely and engage as if co-located.[87] In visionOS 2.6, enhancements to hair, eyelashes, complexion, and side-profile views further improve realism and expressivity.[7] The visionOS 2.6.1 update in November 2025 includes minor refinements to these features along with bug fixes.[63] FaceTime on visionOS integrates spatial computing for life-sized video calls, where participants using Apple Vision Pro can view each other as Spatial Personas, simulating direct eye contact through gaze correction algorithms.[86] This setup supports SharePlay, enabling synchronized activities such as co-watching movies, playing games, or browsing media together in a shared virtual space.[86] For non-Vision Pro users, traditional 2D video feeds are displayed alongside spatial elements, ensuring compatibility while prioritizing immersive experiences for headset wearers.[88] Collaboration tools in visionOS emphasize shared spatial canvases for creative and professional workflows. The Freeform app functions as a 3D whiteboard, allowing multiple users to draw, add notes, and manipulate infinite boards in real time via SharePlay or nearby connections.[89] Participants can import 3D objects, photos, or files, fostering collaborative brainstorming where annotations appear volumetrically in the shared environment.[90] In the Photos app, SharePlay sessions enable joint viewing of spatial photos and videos, with users able to navigate and discuss content together, enhancing remote storytelling and review processes.[91] Remote access features provide secure ways to share the device and experiences. Guest User mode allows controlled access, where the owner selects permitted apps and limits session duration, with guests completing a temporary eye and hand setup upon donning the headset.[92] In visionOS 2.4 and later, integration with nearby iPhones or iPads streamlines guest authentication, returning the device to the owner's configuration automatically after use.[93] For enterprise scenarios, visionOS supports virtual meetings through apps like Zoom, which leverage spatial audio, Personas, and diagramming tools to create immersive sessions with 3D annotations and shared diagrams.[94] Enterprise APIs further enable customized spatial experiences, such as enhanced sensor access for professional diagramming in collaborative environments.[95] Privacy measures in social features prioritize user control and data minimization. Persona creation and processing occur entirely on-device, preventing transmission of raw biometric data.[30] Sharing Spatial Personas requires explicit opt-in during calls, with users able to toggle between 2D and 3D representations.[96] During FaceTime or SharePlay sessions, surroundings captured via passthrough cameras are not shared externally; instead, the system blurs or isolates the user's environment to protect bystanders and maintain focus on the avatar.[30] Settings allow granular control over features like microphone, camera, and Persona access for individual apps, ensuring opt-in consent for all social interactions.[97]Health and productivity integrations
visionOS integrates with Apple's HealthKit framework starting from version 2, enabling developers to build spatial health and fitness applications that leverage the device's expansive canvas for immersive experiences, such as guided meditations or wellness visualizations. This extension allows seamless synchronization of health data across Apple ecosystems, including metrics like heart rate captured by an Apple Watch during virtual reality sessions, which can be accessed and displayed in visionOS apps for real-time monitoring without leaving the spatial environment.[46][98][45] While visionOS does not natively implement eye wellness reminders based on headset usage, its built-in eye-tracking system—powered by high-precision infrared cameras—supports accessibility features that promote comfortable viewing, and developers can incorporate HealthKit data to create custom alerts for prolonged sessions in third-party wellness apps. The Mindfulness app, native to visionOS, further supports behavioral health by offering spatial breathing exercises and reflections to build emotional resilience, integrating with HealthKit for mood logging and progress tracking.[99][100] For productivity, Safari on visionOS introduces spatial browsing, which transforms standard web pages into interactive 3D environments, allowing users to explore articles, videos, and content with depth and immersion for enhanced focus during research or reading tasks. The Notes app facilitates spatial sketches through gesture-based handwriting and drawing in the user's physical space, with Apple Intelligence features like Image Wand converting rough sketches into polished images or diagrams directly within the app.[101][7][102] Focus modes in visionOS adapt notifications to minimize interruptions during immersive work, with options to silence non-essential alerts, summarize grouped notifications, and schedule sessions for deep concentration, such as allowing only work-related communications while engaging in spatial multitasking. Although native time tracking for individual sessions is handled through Screen Time analytics for overall usage patterns, Focus integration enables automated logging of focused periods to help users review productivity over time.[103] The headset employs advanced sensor fusion, combining data from its LiDAR scanner, cameras, and inertial measurement units to enable precise environmental mapping and hand-eye tracking, which supports health apps in creating responsive experiences like adaptive workouts. Introduced in later updates, developer tools allow for fatigue-aware features through motion and gaze analysis, though no built-in EEG-like monitoring exists; instead, apps can use sensor data for indirect detection of user strain.[3] Enterprise tools in visionOS include robust VPN support for secure spatial access to corporate networks, configurable via device management profiles to enable per-app tunneling and remote workflows in mixed reality. The LiDAR scanner facilitates document scanning and AR measurements natively through the Notes app for capturing and annotating physical documents in 3D space, while ARKit APIs power precise measurements of objects and rooms for professional applications like architecture or inventory.[104][102][105]Software ecosystem
Native and third-party apps
visionOS includes a suite of native applications optimized for spatial computing, enhancing user interaction within mixed reality environments. The Photos app supports spatial Memories, allowing users to relive captured moments in 3D by transforming standard photos into immersive scenes with depth and parallax effects, introduced in visionOS 2.[106] Apple TV+ delivers immersive video content through 180-degree 8K recordings with Spatial Audio, enabling viewers to experience films and series as if positioned inside the scene, with series spanning genres like documentary and music premiering regularly.[107] The Fitness app integrates guided workouts with augmented reality elements, overlaying instructional visuals and progress tracking in the user's physical space to support activities like yoga and strength training.[3] Third-party developers have contributed notable applications that leverage visionOS capabilities for productivity and creativity. Microsoft offers apps such as OneNote and Teams, enabling 3D spatial layouts for note-taking and collaboration, where users can arrange multiple documents or meeting participants in volumetric windows around their environment.[108] Adobe provides native tools like Lightroom with Firefly AI integration, allowing spatial editing of photos through generative features that extend into 3D canvases, though full creative suites like Photoshop remain limited to compatible modes.[109][110] The visionOS ecosystem spans diverse app categories, fostering engagement across entertainment, education, and utilities. In gaming, spatial puzzles such as Loóna and Puzzling Places immerse players in 3D dioramas and photorealistic assemblies, using hand gestures to manipulate pieces in shared or full immersion modes. Entertainment options include VR concerts via AmazeVR, delivering 8K interactive performances from artists like Megan Thee Stallion, where users engage with holographic stages and effects in personal spatial audio.[111] Utility apps feature spatial calculators like Spatial Calculation, which project volumetric math interfaces for multi-step computations floating in the user's view, enhancing precision for engineering tasks.[112] Educational applications offer AR dissections through tools like Visible Body, providing interactive 3D anatomical models for biology studies, compatible with visionOS for layered overlays on real-world references.[113] The App Store for visionOS launched with over 600 native spatial apps in February 2024, growing to more than 2,000 by mid-2024 and reaching more than 3,000 native titles as of October 2025, reflecting robust developer adoption across categories like education and gaming.[9][106][8] To optimize performance and battery life, many apps incorporate foveated rendering, dynamically adjusting resolution based on eye gaze to reduce computational load while maintaining high-fidelity visuals in the user's focal area.App development tools
App development for visionOS primarily utilizes Xcode 15 and later versions, which integrate the visionOS SDK to enable building, testing, and deployment of spatial applications.[12] This SDK provides comprehensive support for creating immersive experiences, including tools for integrating 3D content and handling spatial interactions directly within the Xcode environment. Developers can leverage Xcode's built-in editor, debugger, and asset management features to streamline the process of authoring apps that run on Apple Vision Pro.[114] For 3D content creation and import, Unity offers official support through its PolySpatial package, available in Unity 2022 LTS and later for Pro, Enterprise, and Industry subscribers, allowing developers to export projects to visionOS while utilizing Unity's authoring tools for games and interactive experiences.[115] Similarly, Unreal Engine provides visionOS support starting from version 5.4 in experimental form, with full integration in 5.5 via C++ projects, enabling the import of high-fidelity 3D assets and rendering pipelines optimized for spatial computing.[116] Key frameworks underpinning visionOS app development include ARKit, which facilitates anchoring virtual content to the real world through asynchronous data providers for scene understanding, object tracking, and environmental sensing in immersive spaces.[117] RealityKit employs an Entity-Component-System (ECS) architecture, where entities serve as containers for components that define behaviors and properties, and systems process updates across multiple entities to simulate physics, animations, and interactions efficiently in 3D scenes.[118] This ECS paradigm promotes modularity, allowing developers to build scalable simulations without tightly coupling data and logic. Testing visionOS apps occurs via the Xcode Simulator on macOS, which emulates spatial environments, room layouts, lighting conditions, and input methods like eye tracking and gestures to validate app behavior without hardware.[119] For hardware validation, apps deploy directly to Apple Vision Pro devices, supporting on-device debugging and real-time iteration. Performance analysis relies on Instruments, Apple's profiling tool, featuring the RealityKit Trace template to measure rendering efficiency, frame rates, and latency in spatial rendering pipelines, helping developers optimize for smooth 90Hz experiences.[83] Design guidelines for visionOS emphasize spatial principles outlined in Apple's Human Interface Guidelines, which recommend layouts that respect users' physical surroundings, intuitive gesture mappings, and depth-based content placement to create natural, immersive interfaces.[81] The App Review process, governed by Apple's guidelines, prioritizes privacy by requiring developers to disclose data collection practices, obtain explicit user consents for sensitive features like camera access, and avoid manipulative tracking to ensure apps align with platform security standards.[120] In 2024, visionOS 2 introduced new volumetric APIs, enhancing SwiftUI with depth-aligned layouts and tools for capturing and rendering spatial video, enabling developers to integrate immersive media more seamlessly into apps. With visionOS 2.6 in June 2025, additional spatial APIs were added, supporting advanced experiences on the upgraded Apple Vision Pro with M5 chip.[106][7] These updates, previewed in developer betas alongside Xcode betas since the initial SDK release, continue to evolve the toolkit for richer spatial content creation.[15]App Store and distribution
The visionOS App Store serves as an integrated spatial storefront within the operating system, allowing users to discover and download native visionOS applications designed for spatial computing experiences on Apple Vision Pro. It features immersive navigation options, including search initiated through gaze-based selection combined with hand gestures or voice commands via Siri, enabling seamless interaction in a three-dimensional environment. App previews incorporate 3D elements, such as volumetric representations of icons and interactive demonstrations of spatial content, to give users a sense of the app's immersive potential before purchase or download.[121][122][3] Distribution models for visionOS apps mirror those of the broader Apple ecosystem, encompassing free downloads, one-time paid purchases, recurring subscriptions, and in-app purchases tailored to spatial elements like custom environments or 3D assets. Developers can offer subscriptions for ongoing access to updated spatial content, while in-app purchases enable monetization of premium features such as enhanced immersive scenes or virtual objects. These models are managed through App Store Connect, ensuring secure transactions and compliance with Apple's guidelines for digital goods.[123][124] For enterprise use, visionOS supports the Volume Purchase Program (VPP) via Apple Business Manager, allowing organizations to bulk acquire and distribute apps at scale for business purposes without individual user accounts. Additionally, developers and testers can sideload apps directly onto registered Vision Pro devices using Xcode for development and beta testing, bypassing the public App Store while adhering to provisioning profiles and certificates.[125][126] App curation in the visionOS App Store emphasizes featured sections highlighting exclusive spatial experiences optimized for extended reality (XR), such as volumetric apps and immersive productivity tools. Age ratings, updated in 2025 to include categories like 4+, 9+, 13+, 16+, and 18+, account for immersive content by evaluating factors including realistic violence, user-generated interactions, and psychological immersion, helping parents and users assess suitability in a spatial context.[127][128] Developers gain access to App Analytics in App Store Connect, which provides over a dozen metrics specific to visionOS apps, covering user acquisition, engagement—such as session duration and interactions within spatial environments—and monetization performance. These tools help optimize apps by revealing patterns in how users explore 3D spaces, though detailed spatial-specific data like precise dwell times may require integration with in-app tracking.[129]Version history
visionOS 1 (initial release)
visionOS 1.0, the inaugural version of Apple's spatial operating system, was released on February 2, 2024, alongside the launch of the Apple Vision Pro headset, marking the minimum and only compatible hardware at the time.[2] The build number for this initial release was 21N305, derived from the final pre-release beta.[130] Designed to blend digital content with the user's physical environment, visionOS 1.0 emphasized spatial computing principles, enabling seamless interactions through eye tracking, hand gestures, and voice commands without traditional controllers. Central to visionOS 1.0 were its core spatial user interface (UI) elements, which positioned windows and apps in a three-dimensional space around the user, adjustable via the Digital Crown for immersion levels. The system launched with over 600 native apps optimized for spatial experiences, including first-party titles like Photos, Safari, and FaceTime, alongside third-party offerings such as PGA Tour and Disney+ adaptations that leveraged volumetric video and immersive environments.[81][9] The initial Persona system introduced photorealistic digital avatars, scanned using the Vision Pro's TrueDepth camera during setup, to represent users in video calls and collaborative sessions while preserving eye contact and facial expressions.[85] Post-launch updates rapidly addressed stability concerns. visionOS 1.0.1, released on January 23, 2024, provided early bug fixes for developers testing ahead of availability.[131] This was followed by visionOS 1.0.2 on January 31, 2024, which patched a critical WebKit vulnerability potentially exploited in the wild and included general security enhancements.[132] In March 2024, visionOS 1.1 introduced improvements to Persona rendering for better hair, makeup, and EyeSight display accuracy, alongside fixes for virtual keyboard cursor positioning and Mac Virtual Display connectivity issues.[133] It also added an option to reset the device passcode if forgotten, enhancing user recovery without external tools.[134] visionOS 1 deprecated several legacy features to streamline spatial development, including gesture-based presentation inUISplitViewController and support for non-PNG/JPEG textures in USDZ models, limiting compatibility to standard formats. Older ARKit versions were unsupported, requiring developers to migrate to visionOS-native RealityKit and ARKit integrations for optimal performance.[18]
The system delivered stable performance benchmarks, supporting 4K resolution per eye at a 90 Hz refresh rate on the Vision Pro's micro-OLED displays, ensuring smooth rendering of spatial content with low latency hand and eye tracking. Launch-period issues, such as intermittent app crashes, window glitching, and unresponsiveness after extended use, emerged in early user reports starting February 2024, often tied to high computational loads or initial calibration. These were systematically resolved across the 1.x update timeline through April 2024, with patches for third-party app stability and system-level memory management, culminating in more reliable multitasking and reduced crash frequency.[135]