Fact-checked by Grok 2 weeks ago

Pose tracking

Pose tracking is the process of determining and continuously updating the and , collectively known as the pose, of objects, devices, or body parts in over time. Typically involving (6DoF)—three for translation and three for rotation—this technique enables real-time monitoring of movements using various sensing technologies, such as optical, inertial, acoustic, and electromagnetic systems. It addresses challenges like accuracy in dynamic environments, , and through algorithms to produce smooth trajectories. The field traces its origins to the with early systems, such as Ivan Sutherland's featuring mechanical tracking arms for basic pose estimation. Advancements in the and 1990s introduced optical and magnetic tracking for improved precision in VR and , while the 2000s saw widespread adoption of inertial measurement units (IMUs). More recently, since the 2010s, integration of and has enhanced vision-based pose tracking, particularly for human motion and markerless systems. Contemporary developments include inside-out tracking in consumer VR headsets and AI-driven for robust performance in complex scenes. Pose tracking is essential in virtual and for immersive user experiences, in for precise and , and in motion for applications like performance evaluation and medical rehabilitation. Despite progress, challenges persist in achieving low-latency, high-accuracy tracking under occlusions, fast movements, and varying environmental conditions, with ongoing research focusing on hybrid systems and emerging technologies like ().

Fundamentals

Definition and Principles

Pose tracking, in the context of human pose estimation in , is the process of detecting the locations of joints or keypoints in each frame of a video and associating them across frames to produce smooth, temporally consistent trajectories of body movements. This builds on single-frame pose estimation by incorporating temporal information to handle dynamics, such as linking the same person's joints despite partial occlusions or pose variations. Human poses are typically represented in 2D (image coordinates) or 3D (world or camera coordinates), using a skeletal model with a fixed number of keypoints. Common datasets like COCO define 17 keypoints (e.g., , shoulders, elbows, wrists, hips, knees, ankles), while others like MPII use 14. The pose can be denoted as a set of keypoint positions: \mathbf{p} = [ \mathbf{p}_1, \mathbf{p}_2, \dots, \mathbf{p}_K ] where K is the number of keypoints and each \mathbf{p}_i is a 2D vector (x_i, y_i) or 3D vector (x_i, y_i, z_i). For 3D, depths are often estimated from or multi-view inputs. Key principles include per-frame keypoint detection (bottom-up or top-down approaches) and temporal tracking, using techniques like Kalman filters for prediction or deep networks for direct spatio-temporal modeling. A reference frame, such as the camera's , anchors the estimates, requiring camera calibration to map detections to if needed. Real-time performance is vital for interactive uses, with processing latencies under 33 ms for 30 video. Performance metrics differ for estimation and tracking: for estimation, average precision (AP) or percentage of correct keypoints (PCK) assess detection accuracy; for 3D, mean per-joint position error (MPJPE) measures error in millimeters (state-of-the-art around 40-60 mm as of 2023). Tracking uses metrics like multiple object tracking accuracy (MOTA) to evaluate identity preservation. Challenges include occlusions, viewpoint changes, and multi-person scenarios, addressed by robust models.

Historical Development

The field of human pose tracking in computer vision emerged in the late 1990s and early 2000s, building on earlier biomechanical motion analysis but focusing on automated image-based methods. Classical approaches relied on model-based techniques, such as models to fit body parts to detected features like edges or silhouettes, and discriminative methods using hand-crafted features (e.g., histograms of oriented gradients) for part detection. Pioneering work included part-based models like the pictorial structures framework by Felzenszwalb and Huttenlocher in 2005, which modeled spatial relationships between body parts for efficient inference. The 2010s marked a with . The 2014 DeepPose paper introduced convolutional neural networks (CNNs) for direct regression of pose from images, treating it as a detection problem and achieving significant accuracy gains on benchmarks. This spurred extensions: stacked hourglass networks (2016) for multi-scale feature capture, and recurrent neural networks (RNNs) for incorporating temporal consistency in tracking. Graph convolutional networks (GCNs) emerged around 2018 to model body joints as graphs, capturing spatial dependencies. The PoseTrack dataset, introduced in 2017, standardized benchmarking for multi-person pose tracking in videos, fostering competitions and improvements. By the 2020s, transformer-based models like PoseFormer (2021) enabled end-to-end spatio-temporal processing, while diffusion models began addressing complex poses and occlusions. As of 2025, research integrates multimodal data (e.g., RGB + depth) and lightweight architectures for on-device tracking, enhancing accessibility. These advances transitioned from frame-by-frame post-processing to unified models that jointly estimate and track poses.

Applications

In Virtual and Augmented Reality

Pose tracking plays a pivotal role in (VR) and (AR) systems, enabling immersive experiences by capturing the (6DoF)—position and orientation—of users' heads, hands, and devices in real time. In head-mounted displays (HMDs), such as the , inside-out tracking uses onboard cameras to map the environment without external sensors, supporting room-scale VR where users can move freely within a defined play area. This approach democratizes VR by eliminating the need for dedicated tracking hardware, allowing seamless integration of head and hand movements for natural interaction with virtual worlds. Similarly, hand tracking in HMDs like the Oculus Quest 2 employs algorithms to interpret finger gestures, replacing or augmenting physical controllers for more intuitive control in applications ranging from social VR to creative tools. For precise interaction, controller tracking is essential, as exemplified by the system, which relies on outside-in tracking via base stations that emit infrared lasers to triangulate the position of controllers and HMDs with sub-millimeter accuracy. This setup ensures low-latency feedback, critical for activities like precise object manipulation in simulations. In , pose tracking facilitates environmental mapping through (SLAM) techniques, as implemented in Apple's ARKit and Google's , both introduced in 2017, which leverage mobile device cameras and sensors to overlay digital content onto the real world by estimating device pose relative to surroundings. These frameworks enable stable experiences on smartphones and tablets, anchoring virtual objects to physical spaces for applications like aids and interactive furniture . Reducing latency in pose tracking is vital for preventing in /, with industry targets aiming for end-to-end latency below 20 milliseconds to align visual updates with head movements. Optical and inertial methods serve as core enablers for achieving this low-latency performance in consumer HMDs. Case studies highlight pose tracking's impact: in gaming, (released in 2018) uses precise hand and controller tracking to synchronize sword slashes with music beats, enhancing rhythmic immersion and achieving widespread popularity with over 10 million units sold on the Quest platform alone as of January 2025. In training simulations, military overlays, such as those developed for soldier training by systems like the U.S. Army's (IVAS), employ pose tracking to superimpose tactical data, enemy positions, and instructions onto the user's , improving and in simulated scenarios.

In Robotics and Motion Analysis

In , pose tracking via optical systems like Vicon supports detailed analysis of human movement, particularly in studies for . These systems capture marker-based trajectories with accuracies typically below 1 mm, enabling the quantification of during walking cycles to assess asymmetries or progress in patients with neurological disorders. Markerless pose tracking has advanced sports performance analysis by providing non-intrusive monitoring of athlete movements. Systems like Theia3D utilize video-based estimation to track up to 124 keypoints on the body, deriving skeletal models for evaluating techniques in sports such as running or jumping, with validation studies showing joint angle errors comparable to traditional marker-based methods (e.g., mean absolute errors of 2-5 degrees). In medical applications, pose tracking through hybrid myoelectric and (IMU) systems enables intuitive control of prosthetic limbs via gesture estimation. These setups fuse electromyographic signals with IMU-derived pose data to predict user intentions, improving grasp accuracy and natural arm motion in upper-limb amputees, as demonstrated in real-time control frameworks for multi-degree-of-freedom prosthetics. The primary outputs of pose tracking in these domains include , profiles, and full kinematic chains, which inform quantitative motion analysis. For example, inertial-based trackers like JointTracker compute real-time trajectories and from sensor data, supporting the reconstruction of biomechanical models for human performance evaluation.

Optical Tracking

Outside-in Systems

Outside-in systems in optical pose tracking employ an array of external cameras, typically (IR)-sensitive, to observe and the positions of markers attached to the tracked object or subject. These systems utilize multiple cameras—often 6 to over 100 in large setups—arranged around the capture volume to detect retro-reflective passive markers or active LED-based markers that reflect or emit IR light. algorithms compute 3D positions by intersecting rays from at least two cameras viewing the same marker, enabling precise pose estimation. For instance, OptiTrack systems can deploy up to 100+ cameras to achieve sub-millimeter accuracy in controlled environments, supporting tracking at high rates. Calibration is essential for aligning the cameras and ensuring accurate marker identification. The process begins with determining extrinsic parameters, such as each camera's and relative to a global , often using a specialized with known marker swept through the volume. This allows the software to compute distortions and 3D reconstructions via of the wand's markers, typically collecting thousands of samples per camera for robustness. Marker ID assignment follows, where asymmetrical marker constellations on rigid bodies or wands enable unique identification, preventing swaps during tracking. Systems like Vicon and OptiTrack provide automated tools in software such as or Motive to refine these parameters, achieving sub-millimeter residuals post-calibration. These systems offer high precision in enclosed, controlled settings, with accuracy reaching 0.1 mm and accuracy of 0.1° within calibrated volumes, making them ideal for applications requiring minimal latency and noise. This level of fidelity supports complex multi-object tracking without cumulative errors, unlike inertial methods prone to drift. However, outside-in systems require direct line-of-sight between cameras and markers, limiting their use to fixed setups where occlusions from body parts or objects can disrupt tracking. To mitigate this, predictive algorithms, such as Kalman filters integrated into the tracking software, extrapolate poses during brief losses of visibility based on prior trajectories and velocities. In contrast to inside-out systems, which prioritize portability through on-device sensors, outside-in configurations excel in stationary, high-accuracy scenarios but demand extensive setup. Prominent examples include Vicon systems, widely used in film motion capture for productions like (2009), where they facilitated precise performance capture of actors in simulated environments. Similarly, HTC Vive base stations employ a laser-swept variant of outside-in tracking since 2016, using external emitters and photodiodes on trackers for room-scale with millimeter-level accuracy.

Inside-out Systems

Inside-out systems in optical pose tracking rely on sensors mounted on the tracked device to observe and map the surrounding , enabling self-contained pose estimation without external infrastructure. The core technique is visual-inertial (VIO), which fuses data from onboard cameras and inertial measurement units () to compute the device's 6-degree-of-freedom pose. This approach typically employs structure-from-motion principles, where visual features in the camera feed are tracked across to reconstruct the environment and estimate motion, providing robustness in dynamic settings. Hardware in these systems often includes compact RGB-D cameras for depth perception or wide-angle fisheye lenses for broad field-of-view coverage. A prominent example is the , introduced in 2016, which integrates time-of-flight depth sensors alongside RGB cameras to perform environmental mapping and pose tracking. Algorithms central to VIO include feature detection methods like (ORB) descriptors, which identify and match keypoints efficiently in image sequences, while loop closure techniques detect revisited locations to correct cumulative drift and maintain global consistency. IMU data is briefly integrated to provide short-term rotational during visual occlusions or rapid movements. These systems offer key advantages, including wireless operation and scalability to large areas like room-scale environments, without the need for base stations. For instance, the Oculus Quest 2 (2020) uses a set of embedded tracking cameras to enable standalone experiences through inside-out visual , supporting untethered 6DoF tracking. Similarly, Google's ARCore framework (2018) leverages cameras for visual -based pose estimation in applications, allowing developers to overlay digital content on the physical world with minimal hardware requirements. More recent examples include the (2023), which improves inside-out tracking with higher-resolution cameras and enhanced processing for better environmental understanding, and the (2024), utilizing advanced VIO for precise in .

Inertial Tracking

Measurement Principles

Inertial measurement units (IMUs) form the core of inertial tracking systems, enabling pose estimation through self-contained without reliance on external references. These units typically comprise three main sensor types: accelerometers, which measure linear acceleration along orthogonal axes (x, y, z); gyroscopes, which detect or rotational rates around those same axes; and magnetometers, which sense the to aid in determining absolute orientation relative to magnetic north. The fundamental operation involves of sensor outputs to derive pose over time. data, after gravity compensation, undergoes double to yield : first to obtain , then to itself. measurements of (ω) are singly integrated to update , as exemplified by the basic Euler step θ_t = θ_{t-1} + ω Δt, where θ represents the and Δt is the time . Pose estimation requires handling multiple coordinate frames, including the body-fixed frame aligned with the axes and the world or frame (e.g., local-level East-North-Up). Transformations between frames are computed using rotation representations, with quaternions preferred for their compactness and ability to avoid singularities like that plague Euler angle methods. IMUs operate at high sampling rates to capture low-noise data for accurate integration, typically ranging from 100 Hz for general motion tracking to 1000 Hz in precision applications. Early applications of inertial sensors for attitude control emerged in during the 1950s, with systems like those developed for and missiles providing stable orientation tracking. Today, miniaturized are integral to wearables for human motion analysis, and they enable 3DoF rotational tracking in controllers. Recent advances as of 2025 include models for IMU-based human pose estimation, such as neural networks that reconstruct poses from sparse data, improving accuracy in occlusions and complex motions.

Drift Compensation Techniques

Drift in inertial pose tracking arises primarily from sensor noise, biases, and errors accumulated during the integration of raw measurements, leading to progressively inaccurate estimates of velocity and position. Accelerometer biases, for instance, result in linear errors in velocity and quadratic growth in position error over time due to double integration. Gyroscope biases similarly cause orientation drift through angular velocity integration. These effects are exacerbated in low-cost inertial measurement units (IMUs), where uncompensated errors can render long-term tracking unreliable without corrective measures. One widely adopted technique for mitigating velocity and position drift is the zero-velocity update (ZUPT), which detects periods of near-stationary motion to reset integrated velocity estimates to zero. In foot-mounted for , ZUPT leverages the stance phase of walking—when the foot is momentarily at rest—to correct accumulated errors, significantly reducing position drift over extended periods. This method, pioneered in pedestrian systems, substantially improves accuracy in indoor tracking scenarios compared to uncorrected inertial navigation. Complementary filtering addresses orientation drift by fusing gyroscope and accelerometer data, applying a low-pass filter to the accelerometer-derived gravity vector for long-term stability and a high-pass filter to gyroscope angular rates to capture short-term dynamics. This approach effectively compensates for gyroscope bias accumulation while filtering accelerometer noise, yielding robust attitude estimates suitable for real-time applications. A nonlinear variant on the special orthogonal group further enhances performance by preserving geometric constraints in 3D rotations. For more comprehensive 9-degree-of-freedom (9DoF) pose estimation, gradient-descent-based filters like the Madgwick algorithm and extended Kalman filters (EKFs) integrate , , and data to estimate while explicitly modeling and correcting drift. The Madgwick filter uses an objective function to minimize errors from sensor measurements, providing efficient drift compensation with low computational overhead, often achieving sub-degree accuracy in over minutes of motion. Simplified EKFs extend this by propagating state uncertainties, particularly for bias estimation in 9DoF setups, and significantly reduce errors relative to raw integration in dynamic environments. Hardware aids, such as fusion, complement these software methods by providing absolute altitude references to counteract vertical drift from IMU integration. In applications, Kalman-based fusion of barometric pressure with IMU data corrects altitude errors accumulating from biases, maintaining centimeter-level precision during hover or ascent despite environmental variations. These techniques also play a role in hybrid systems, where inertial drift compensation ensures stable pose updates between optical frames.

Acoustic Tracking

Time-of-Flight Methods

Time-of-flight (TOF) methods in acoustic pose tracking determine distances by measuring the time of sound waves between emitters and receivers, enabling estimation through ranging. The core principle involves calculating the distance d = \frac{c \cdot \Delta t}{2}, where c is the in air (approximately 343 m/s at 20°C) and \Delta t is the round-trip time of the acoustic signal. High-frequency ultrasonic waves are typically employed to achieve sufficient resolution for pose applications, as lower frequencies would limit accuracy due to longer wavelengths. System setups commonly feature arrays of ultrasonic transducers mounted as fixed beacons, with mobile receivers attached to the tracked object. These transducers operate at frequencies around 40 kHz, providing centimeter-level distance accuracy in controlled indoor environments. For pose reconstruction, at least three beacons are required to compute the object's position via of the measured distances, often supplemented by signals for synchronization to enable one-way TOF measurements. A pioneering implementation is the system developed at AT&T Laboratories Cambridge in the late 1990s, which uses ultrasonic pulses from handheld "Bats" to ceiling-mounted receivers for location and orientation tracking with about 3 cm precision. TOF methods offer advantages such as low cost (under $10 per in early designs) and reliable operation in low-light conditions, where optical alternatives may fail. However, the varies with environmental factors like (changing by about 0.6 m/s per °C), , and , which can introduce errors up to several centimeters without correction; compensation is thus essential, often achieved via integrated sensors to dynamically adjust c. These systems have found early applications in and sentient computing environments, such as the system's integration into context-aware setups for tracking user interactions in smart rooms. Unlike purely optical approaches, acoustic TOF supports limited non-line-of-sight capabilities through sound .

Phase-Coherent Methods

Phase-coherent methods in acoustic tracking utilize continuous-wave ultrasonic signals to measure the phase difference between transmitted and received waves, enabling precise distance estimation for pose determination. The core principle relies on the relationship \Delta \phi = \frac{2\pi f d}{c}, where \Delta \phi is the phase difference, f is the signal frequency, d is the distance to the target, and c is the speed of sound (approximately 343 m/s in air). This phase shift corresponds to the portion of the wavelength \lambda = c/f traveled by the wave; for typical ultrasonic frequencies of 20-50 kHz, \lambda ranges from 6.9 mm to 17.2 mm, allowing resolutions down to \lambda/100 or better, often achieving sub-centimeter accuracy. The setup involves coherent transmitters and receivers, such as ultrasonic transducers paired with phase-locked loops (PLLs) to maintain synchronization and detect phase shifts in . These systems emit a steady sinusoidal wave, and the receiver locks onto the incoming signal's phase relative to a reference, typically using to extract the difference. Frequencies in the 20-50 kHz range are common to balance with in air, while PLLs ensure stable tracking even with minor Doppler shifts from motion. This configuration supports absolute positioning when combined with initial , providing external references for in indoor environments. A key challenge is phase ambiguity, as measurements are modulo $2\pi, limiting unambiguous range to one wavelength without additional techniques. Ambiguity resolution employs multi-frequency unwrapping, where signals at two or more frequencies (e.g., 40 kHz and 41 kHz) are transmitted sequentially or simultaneously; the beat phase from their difference allows computation of the integer number of full cycles, extending the effective range to meters while preserving precision. Algorithms iteratively unwrap the phase by comparing low-resolution (longer \lambda) and high-resolution (shorter \lambda) measurements to determine the correct integer multiple. These methods offer sub-centimeter accuracy in controlled indoor settings, outperforming pulsed time-of-flight approaches in over short ranges due to continuous monitoring, which minimizes . However, they are susceptible to multipath from reflections off walls or objects, which can distort phase readings and degrade performance, and are constrained by narrow bandwidths that limit update rates and robustness to noise. Research prototypes demonstrate practical applications, such as the LLAP system, which uses speakers and microphones at around 20 kHz to track hand and finger poses with millimeter accuracy via phase analysis, enabling without wearables. Similarly, the MilliSonic prototype achieves sub-millimeter 1D tracking on mobile devices by compensating for multipath through advanced , highlighting potential for fine-grained pose estimation in human-computer . Recent advancements as of 2025 integrate acoustic methods with visual and inertial sensors for enhanced . For example, VibeMesh fuses active acoustic sensing with vision for dense hand pose and contact tracking, while UltraPoser combines acoustics with for full-body pose reconstruction.

Electromagnetic Tracking

Field Generation and Detection

Electromagnetic tracking systems generate low-frequency alternating s using one or more transmitter coils driven by oscillating currents, which induce voltages in receiver coils attached to the tracked object via Faraday's law of . The spatial distribution of the \mathbf{B} from these coils follows the Biot-Savart law, with a approximation yielding |\mathbf{B}| \propto \mu_0 I / (4\pi r^3), where \mu_0 is the permeability of free space, I is the current, and r is the distance from the coil. This near-field regime (quasi-static ) ensures the field strength decays rapidly, localizing the tracking volume while allowing penetration through non-magnetic occluders. Receiver sensors typically consist of triaxial orthogonal coils that measure the three components of the , enabling full 6 degrees-of-freedom (6DoF) pose estimation ( and ). For instance, the Ascension Flock of Birds system, introduced in 1991, employed such compact triaxial search coils in pulsed-DC mode to achieve simultaneous tracking of multiple sensors. The induced voltages are proportional to the field components and the coil's angular sensitivity, providing raw signals that encode the receiver's relative pose to the transmitter. To compute the 6DoF pose, the system solves the nonlinear : given the measured field strengths (or induced voltages), determine the position and that best fit the forward model of field propagation. This typically involves iterative nonlinear optimization, such as , minimizing the error between observed and predicted signals from at least three non-coplanar field measurements. Operating frequencies are selected in the low kilohertz range (typically 1–30 kHz) to minimize distortions while supporting effective ranges up to 3 m and accuracies of approximately 1 mm RMS in position and 0.5° RMS in orientation within the optimal volume. Early commercial adoption in medical navigation, such as the NDI Aurora system launched in 2003, leveraged these principles for precise instrument tracking in procedures like biopsies and ablations. This configuration facilitates untethered operation, unlike cable-bound alternatives.

Distortion Mitigation

Distortions in electromagnetic pose tracking primarily arise from metals, which alter lines due to their high permeability, and from electronics or conductive materials that induce currents, leading to position errors exceeding 5 even at distances greater than 10 cm from the . For instance, instruments positioned 10 cm from tracking components can introduce significant localization inaccuracies attributable to these effects. Such distortions are particularly pronounced in environments like operating rooms, where metallic tools and electronic equipment perturb the generated fields, degrading overall tracking precision. To counteract these issues, compensation methods often employ calibration through matrices that model and correct field perturbations, converting raw magnetic measurements into accurate spatial coordinates. Lookup tables, constructed from precomputed distortion maps, enable rapid interpolation for error correction during operation, particularly effective for static environmental distortions. These approaches can incorporate optimization techniques, such as , to iteratively refine calibration parameters and minimize residual errors in dynamic settings. Additionally, hybrid integration with inertial measurement units () provides short-term stability by fusing and data to bridge gaps during transient EM distortions, maintaining pose estimates with sub-millimeter accuracy in affected regions. Advanced techniques leverage pulsed magnetic fields to mitigate interference, as the intermittent transmission allows separation of primary signals from induced distortions, reducing cross-talk and effects compared to continuous fields. Systems like the Polhemus exemplify this by employing modulated fields that enhance signal-to-noise ratios and enable multiple trackers to operate without mutual interference. However, operational trade-offs exist, as higher frequencies can diminish certain distortion influences by improving signal discrimination but constrain the effective tracking range due to increased and limitations in coils. In contrast to acoustic tracking's sensitivity to air density variations, these EM-specific mitigations prioritize field integrity over medium propagation.

Sensor Fusion

Fusion Algorithms

Fusion algorithms integrate data from multiple sensors, such as inertial measurement units () worn on body segments, cameras, and depth sensors, to produce robust estimates of joint positions and s by leveraging complementary strengths—high-frequency inertial for motion with absolute positioning from visual measurements. These methods model the human pose as a of s, often using probabilistic frameworks to handle noise, uncertainties, and nonlinear articulations in multi-joint tracking. Seminal approaches include variants for Gaussian assumptions, particle filters for non-Gaussian distributions, complementary filters for , and optimization-based techniques for global consistency in pose trajectories. The (EKF) and its variants, such as the Error-State EKF (ES-EKF), are used for nonlinear pose estimation in multi-sensor fusion, linearizing dynamics around current estimates to track joint states over time. While originally applied in , adaptations extend to pose by fusing IMU data from body-worn sensors with visual keypoints, correcting for drift in occluded scenarios. The prediction-update cycle propagates states using IMU measurements and updates with camera observations, minimizing innovation via Jacobians. This has been adapted for , improving accuracy in dynamic environments. Particle filters represent the posterior distribution with weighted particles for handling non-Gaussian uncertainties in human pose tracking, suitable for multimodal distributions from sensor noise. The process involves sampling particles guided by motion models, weighting by likelihoods (e.g., from camera reprojections), and resampling to avoid degeneracy, enabling robust tracking during partial occlusions or fast movements. Complementary filters offer lightweight real-time fusion for IMU-based orientation estimation of body segments, combining short-term accuracy with gravity references. The fused estimate for angles is: \theta_{fused} = \alpha \cdot (\theta_{gyro} + \omega \cdot \Delta t) + (1 - \alpha) \cdot \theta_{accel}, with \alpha typically 0.98; this corrects gyro drift while filtering noise, extended to quaternions for 3D orientations in wearable systems. Optimization-based methods, such as , formulate fusion as least-squares minimization over a of joint poses, with edges from constraints (e.g., IMU preintegration or visual landmarks). The objective is: \hat{\mathbf{x}} = \arg\min_{\mathbf{x}} \sum_{i} \|\mathbf{r}_i(\mathbf{x})\|^2_{\Sigma_i}, solved iteratively with tools like g2o or for consistent multi-frame human pose trajectories, reducing drift in long sequences. Fused estimates reduce error covariance over single s; for example, in precision motion tasks, enhanced Kalman can halve position errors. Improvements arise from adaptive weighting, ensuring consistency in multi- pose tracking.

Hybrid System Examples

Hybrid systems integrate modalities for pose tracking, combining visual for detection with inertial for motion continuity. One example is visual-inertial in suits, where body-worn IMUs augment camera-based systems like OptiTrack to maintain tracking during occlusions, achieving sub-millimeter accuracy in studio environments. In rehabilitation, IMU-camera hybrids track patient gait and poses; systems like Xsens MVN fuse 17+ IMUs with optional cameras for full-body 3D kinematics, providing drift-free estimates with <1° orientation and <5 mm position errors for joint trajectories. Acoustic-visual fusion appears in experimental AR for human pose, but more commonly, depth-camera-IMU setups like Microsoft's Kinect with added wearables enable 6DoF body segment tracking for interactive applications. A recent example is MobilePoser (2024), which fuses cameras with user-worn for real-time full-body pose estimation and , supporting subsets of sensors for in mobile health monitoring. These hybrids improve robustness in dynamic scenes; visual-inertial fusion can reduce trajectory errors by up to 83% compared to vision-only in occluded human motion, via reliability-weighted Kalman filtering.

Challenges and Future Directions

Accuracy and Latency Limitations

Pose tracking systems are inherently limited by sensor noise, which introduces inaccuracies in position and orientation estimates. Inertial measurement units (), commonly used in such systems, suffer from gyroscope bias drift, with low-cost MEMS gyroscopes exhibiting rates around 18° per hour, leading to cumulative orientation errors over time. Accelerometers in IMUs also contribute noise from vibrations and gravitational variations, further degrading pose accuracy without external corrections. Environmental factors exacerbate these issues, particularly in optical-based tracking where occlusions—such as body parts blocking markers or keypoints—can cause sudden jumps in estimated positions. For instance, in systems like the Pro, partial occlusions or lighthouse positioning changes result in positional deviations up to 40 mm. These errors are more pronounced in dynamic, real-world settings compared to controlled labs, where mean position errors are typically 3.5 mm. Latency in pose tracking arises primarily from processing delays and transport lag. In 60 Hz systems, the frame interval is approximately 16.7 ms, and additional can impact performance, with GPU-based enabling low-latency estimation. Transport lag from data transmission further compounds this, impacting responsiveness in applications like . Jitter and limited pose additional challenges, especially for capturing high-frequency motions. Human hand tremors occur at frequencies of 2-12 Hz, which low-sampling-rate IMUs (e.g., below 20 Hz effective ) struggle to resolve accurately, resulting in smoothed or erroneous tracking. This is evident in hand tracking systems, where even minor (0.5°-1.5°) significantly reduces precision. The lack of standardized benchmarks hinders consistent evaluation across systems. Quantification often relies on root mean square error (RMSE) for position, achieving 1-5 mm in lab environments but degrading to several centimeters (e.g., 4 cm) in unconstrained real-world scenarios due to combined factors. techniques can partially mitigate these limitations by integrating complementary data sources.

Emerging Technologies

Recent advancements in have significantly enhanced markerless pose tracking through models that estimate human poses directly from RGB video streams. Google's MediaPipe framework, introduced in 2019, enables pose estimation on mobile devices by detecting 33 body keypoints with low computational overhead, achieving over 30 frames per second on standard smartphones using lightweight convolutional neural networks. This approach leverages from large datasets to infer poses from monocular video, reducing the need for specialized hardware and enabling applications in fitness tracking and without physical markers. Millimeter-wave (mmWave) radar technologies offer contactless pose tracking by detecting subtle movements through signals, bypassing limitations of optical methods in low-light or occluded environments. Google's Project Soli, announced in 2015, utilizes mmWave radar to recognize and track sub-millimeter motions at high frame rates exceeding 10,000 per second, with applications in gesture and fine motion tracking in radar-based systems. Recent implementations achieve per-joint position errors around 9.85 cm for egocentric body tracking when integrated into headsets, demonstrating robustness for immersive interactions. Brain-computer interfaces (BCIs) provide an indirect method for pose control by decoding neural signals to infer intended movements, particularly for users with mobility impairments. Neuralink's prototypes, tested in human trials starting in 2024, translate brain activity into cursor control and device manipulation at speeds of several bits per second, laying groundwork for neural-driven pose estimation in virtual environments. These systems process high-channel-count signals from implanted electrodes to map motor intentions to 3D poses, though current accuracy remains limited to basic actions without full-body fidelity. Edge computing advancements facilitate on-device for pose tracking, minimizing latency in applications. Qualcomm's Snapdragon platforms, enhanced in 2022, support real-time through optimized AI models that perform depth and pose inference locally, reducing end-to-end latency to under 20 milliseconds compared to cloud-based processing. This enables seamless integration in wearables and AR glasses by pruning and quantizing models for resource-constrained hardware. Emerging trends in 5G-enabled distributed tracking leverage ultra-low latency networks to synchronize pose data across multiple devices, enhancing collaborative / experiences. By 2025, Advanced supports distributed rendering and pose fusion with latencies below 10 milliseconds, allowing real-time body tracking in multi-user scenarios without centralized computation. Quantum sensors, still in research stages as of 2025, promise ultra-precision pose tracking by exploiting quantum effects for sub-micron accuracy in inertial and positional measurements, potentially revolutionizing navigation-denied environments like indoor .

References

  1. [1]
    None
    Summary of each segment:
  2. [2]
    [PDF] Vision-based Human Pose Estimation via Deep Learning: A Survey
    This survey covers deep learning approaches for 2D and 3D human pose estimation, including challenges, trends, and bibliometrics.
  3. [3]
    [PDF] DeepPose: Human Pose Estimation via Deep Neural Networks
    In this paper we attempt to cast a light on this question and present a simple and yet powerful formulation of holistic human pose esti- mation as a DNN. We ...Missing: tracking survey
  4. [4]
    [PDF] 6-DOF Pose Tracking with the VRduino - Stanford University
    The goal of 6-DOF pose estimation is to recover the relative position (three degrees of freedom) and rotation (another three degrees of freedom) of some rigid ...
  5. [5]
    A Survey of 6DoF Object Pose Estimation Methods for Different ... - NIH
    Object pose estimation is a key task in the field of computer vision, whose main goal is to accurately obtain a 6DoF (6 degrees of freedom) representation of ...
  6. [6]
    [PDF] Multiple View Geometry in Computer Vision, Second Edition
    PART 0: The Background: Projective Geometry, Transformations and Esti- mation. 23. Outline. 24. 2. Projective Geometry and Transformations of 2D.
  7. [7]
    [PDF] A Model of Real-time Pose Estimation Fusing Camera and LiDAR in ...
    Because LiDAR and the camera use two different coordinate reference systems, it is necessary to define a world coordinate system to realize the synchronization ...
  8. [8]
    [PDF] Accurate Catadioptric Calibration for Real-time Pose Estimation in ...
    In this paper, we develop a novel calibration model that we combine with a beacon-based pose estimation algorithm. Our approach relaxes the assumption of an ...
  9. [9]
    [PDF] Real-Time Human Pose Tracking from Range Data - Google Research
    Our system tracks the human motion in real-time from just the stream of depth images. 5 Experiments. The full tracking system has been implemented in C/C++ and ...
  10. [10]
    5 Position Tracking and Mapping | Virtual Reality
    IREDs are multiplexed at high rates (3,000 Hz) and sensed by two cameras, which triangulate the markers to an accuracy of about 5 mm at a distance of 2.5 m. ...Missing: metrics | Show results with:metrics
  11. [11]
    Determining the Accuracy of Oculus Touch Controllers for Motor ...
    Jun 6, 2019 · Results: It was found that the maximum position accuracy error of the system was 3.5 ± 2.5 mm at the largest step size of 500 mm along the z- ...<|control11|><|separator|>
  12. [12]
    MEASUREMENT of RANGE of MOTION and MUSCLE LENGTH
    Aug 10, 2016 · Although reliable goniometers were available for measuring joint ROM early in the 20th century, examiners did not agree on correct procedures ...
  13. [13]
    A historical review of gait analysis - Neurosciences Journal
    Research into human gait advanced significantly during the 1940s and 1950s. Pioneering work by Verne. Inman and his team at the College of Engineering and.
  14. [14]
    Polhemus History - Over 50 Years of Experience
    Polhemus is known as the true pioneer and leader in the motion tracking industry, first introducing our proprietary electromagnetic technology in 1969.Missing: 1972 | Show results with:1972
  15. [15]
    About Us | The Vicon Difference | Motion Capture Systems
    Vicon is the world's number one motion capture choice for over 40 years, providing tailored systems and is a recognized leader since 1984.
  16. [16]
    At the Heart of the Wii, Micron-Size Machines - The New York Times
    Dec 21, 2006 · The Wii remote also uses infrared, the same technology that links television sets with their remote controllers, to track where the controller ...
  17. [17]
    Global Localization from Monocular SLAM on a Mobile Phone
    Mar 24, 2014 · The SLAM system runs locally on a camera-equipped mobile client and provides continuous, relative 6DoF pose estimation as well as keyframe ...
  18. [18]
    See how Google's new 'Project Tango' smartphones sense the world
    Feb 25, 2014 · Google says the Tango smartphone can capture a wealth of data never before available to app developers, including depth- and object-tracking and real-time 3D ...Missing: acoustic AR
  19. [19]
    The Automatization of the Gait Analysis by the Vicon Video System
    Sep 21, 2022 · Optoelectronic motion capture is widely used in biomechanics. Vicon is a system consisting of several Vicon Vero 2.2 digital infrared ...
  20. [20]
    Motion Capture for Clinical Gait Analysis - Vicon
    Recognized as the best-in-class for clinical gait and biomechanics, our custom-built gait analysis software platform enables the creation of the most advanced, ...Clinical Gait Analysis · Motion Capture For Gait · Watch
  21. [21]
    Accuracy of a markerless motion capture system in estimating upper ...
    Jul 25, 2022 · This study aimed to assess whether a markerless approach (use of video cameras + commercial pose estimation software Theia3D) can be used to ...Missing: tracking | Show results with:tracking
  22. [22]
    Concurrent validity of Theia3D markerless motion capture for ...
    This study determined the capacity of Theia3D (Theia Markerless Inc.) MMC to detect sagittal-plane kinematic responses to different gait speeds.
  23. [23]
    Myoelectric control of robotic lower limb prostheses - PubMed Central
    The objective of this paper is to review robotic lower limb Prosthesis control via EMG signals recorded from residual muscles in individuals with lower limb ...2. Literature Review · 2.2. Neural Interfaces · 2.3. Emg Control Paradigms
  24. [24]
    JointTracker: Real-time inertial kinematic chain tracking with joint ...
    The proposed algorithm can be used to track joint positions, i.e. joint trajectories, for motion analysis, and it delivers consistent information with camera- ...
  25. [25]
    Kinematic Chain - an overview | ScienceDirect Topics
    A kinematic model is a mathematical representation of a robot's motion that overlooks the forces that affect motion and focuses on the geometrical relationship ...
  26. [26]
    OptiTrack - Motion Capture Systems
    Industry leading precision motion capture and 3D tracking systems for video game design, animation, virtual reality, robotics, and movement sciences.Cameras · Build My System · Flex 3 · Software
  27. [27]
    Quick Start Guide: Precision Capture - OptiTrack Documentation
    A well-tracked marker will have a sub-millimeter average residual value. In Motive, the tolerable residual distance is defined from the Reconstruction Settings ...Residual Value · Camera Placement · Live-Reconstruction Settings · Wands
  28. [28]
    Vicon: Award Winning Motion Capture Systems
    With industry-leading accuracy and real-time tracking, Vicon empowers researchers and clinicians to drive objective treatment planning, optimize performance ...What is Motion Capture? · Hardware · Contact Us · About Us<|separator|>
  29. [29]
    Calibration | EXTERNAL OptiTrack Documentation
    ### Calibration Process Summary: Camera Extrinsic Parameters and Marker ID Assignment for Outside-In Optical Tracking
  30. [30]
    Motion Capture Calibration Guide | Vicon Life Sciences
    In this guide, we'll walk through motion capture calibration from start to finish, covering the tools you need, the workflow inside Vicon Nexus, and the common ...
  31. [31]
    Agreement Analysis between Vive and Vicon Systems to Monitor ...
    As a traditional method of motion tracking, Vicon has established an accuracy of 0.1 mm and 0.1° within a calibrated volume space. This analysis shows that ...
  32. [32]
    Optical Motion Capture - an overview | ScienceDirect Topics
    An outside-in system uses external sensors to collect data from sources placed on the body, an inside-out systems have sensors placed on the body that collect ...
  33. [33]
  34. [34]
    THE FUTURE OF MOTION CAPTURE - Post Magazine
    Sep 1, 2008 · The future of mocap will be a seamlessly integrated motion capture experience where the limitations imposed by current technologies are overcome.
  35. [35]
    Analysis of Valve's 'Lighthouse' Tracking System Reveals Accuracy
    Jul 17, 2016 · Based on these measurements, Kreylos estimated the precision of Lighthouse tracking to be around RMS 1.5mm and the accuracy around RMS 1.9mm.
  36. [36]
    Inertial Measurement Unit (IMU) | An Introduction
    Feb 13, 2023 · Advanced Navigation IMUs feature a 1000 Hz (1 kHz) sampling rate and typically utilize angular rate and acceleration sensors with a 400 Hz ...
  37. [37]
    [PDF] An introduction to inertial navigation - University of Cambridge
    Inertial navigation is a self-contained navigation technique in which measurements provided by accelerom- eters and gyroscopes are used to track the position ...
  38. [38]
    The Inertialist: Fundamentals of Inertial Navigation - Inside GNSS
    Mar 31, 2022 · To integrate accelerometer measurements into velocity and position, the INS first needs to relate orientation of sensors' sensitive axes to a ...Missing: double | Show results with:double
  39. [39]
    How does an INS work? - Inertial Labs
    Oct 10, 2024 · Evolution of technology (1950s-1960s). In the 1950s and 60s, inertial navigation systems became more compact and accurate with the development ...
  40. [40]
    [PDF] IMU Errors and Their Effects - NET
    Feb 21, 2014 · In this case, a constant bias (error) in acceleration becomes a linear error in velocity and a quadratic error in position.
  41. [41]
    Quantification of Error Sources with Inertial Measurement Units in ...
    Dec 13, 2022 · The goal of this study was to present a method to quantify each of the sources of error of an IMU separately.
  42. [42]
    [PDF] Fifteen Years of Progress at Zero Velocity: A Review - arXiv
    Aug 20, 2020 · Zero-velocity updates are used to reduce the error drift of foot-mounted inertial navigation systems. FOOT-MOUNTED INERTIAL NAVIGATION.
  43. [43]
    [PDF] Nonlinear Complementary Filters on the Special Orthogonal Group
    Jun 1, 2010 · This paper introduces direct, passive, and explicit complementary filters on the special orthogonal group SO(3) for attitude estimation from ...
  44. [44]
    Kalman-Filter-Based Orientation Determination Using Inertial ... - NIH
    In this paper we present a quaternion-based Extended Kalman Filter (EKF) for estimating the three-dimensional orientation of a rigid body.
  45. [45]
    9-DOF IMU-Based Attitude and Heading Estimation Using an ... - MDPI
    Apr 29, 2022 · A two-step extended Kalman Filter (EKF) algorithm is used in this study to estimate the orientation of an IMU. A 9-DOF device is used for this purpose.
  46. [46]
    Kalman Filter Based Sensor Fusion for Altitude Estimation of Aerial ...
    Jun 12, 2020 · The proposed method is divided into two parts: attitude estimation and altitude estimation. The method is implemented using IMU and barometer.
  47. [47]
    A barometer-IMU fusion method for vertical velocity and height ...
    A sensor fusion method was developed for vertical channel stabilization by fusing inertial measurements from an Inertial Measurement Unit (IMU) and pressure ...
  48. [48]
    The Cricket location-support system - ACM Digital Library
    This paper presents the design, implementation, and evaluation of Cricket, a location-support system for in-building, mobile, location-dependent applications.
  49. [49]
    The Bat Ultrasonic Location System - University of Cambridge
    Jan 7, 2005 · A 3D ultrasonic location system (known here as the Bat System) which is low-power, wireless and relatively inexpensive.
  50. [50]
    The Cricket indoor location system - ResearchGate
    The cricket 11 system, developed by the Massachusetts Institute of Technology (MIT), uses a combination of RF and ultrasound technologies to provide ...
  51. [51]
    [PDF] Motion Tracking: No Silver Bullet, but a Respectable Arsenal
    Meyer and colleagues point out that this “phase-coherent” method enables continuous measurement without latency but can only measure relative distance changes ...
  52. [52]
    A digital, constant-frequency pulsed phase-locked-loop instrument ...
    May 8, 2018 · The phase comparison technique has since spawned several variations of pulsed phase-locked loop (PPLL) ultrasonic phase measurement methods.
  53. [53]
  54. [54]
    Device-Free Gesture Tracking Using Acoustic Signals
    Oct 3, 2016 · LLAP demonstrates that commercial mobile devices can use acoustic phase information to track hands/fingers with millimeter- level accuracy.
  55. [55]
    [PDF] MilliSonic: Pushing the Limits of Acoustic Motion Tracking
    State-of-the-art acoustic motion tracking systems however do not achieve millimeter accuracy and require large separation between microphones and speakers, and ...
  56. [56]
    (PDF) Electromagnetic Tracking in Medicine-A Review of ...
    Aug 5, 2025 · PDF | Object tracking is a key enabling technology in the context of computer-assisted medical interventions.Missing: history | Show results with:history
  57. [57]
    Electromagnetic tracking - VR & AR Wiki - Virtual Reality ...
    May 7, 2025 · ... Flock of Birds'' (1991). During the 1990s EMT migrated from ... Pulsed-DC trackers (Ascension “Bird”) reduce such distortion at ...
  58. [58]
    Electromagnetic tracking in the clinical environment - PMC - NIH
    In this work we assess the 3D Guidance and Aurora EMTSs. Both systems consist of three basic components, the field generator, a system control unit that ...
  59. [59]
    Aurora - NDI's premier electromagnetic tracking solution
    Controls the field generator, collects information from the SIUs, calculates the position and orientation of each sensor, and interfaces with the host computer.
  60. [60]
    EM tracking with conventional CT tables: Quantifying and correcting ...
    Jun 5, 2025 · The metal components in conventional CT scanner tables produce EM distortion that can result in EM tracking localization errors >5 mm for ...Missing: 10cm computers
  61. [61]
    Systematic distortions in magnetic position digitizers - Birkfellner
    Oct 28, 1998 · The results show that surgical instruments not containing conductive material are to be preferred when using an ac tracker. Nonferromagnetic ...
  62. [62]
    Minimizing Electromagnetic Interference from Surgical Instruments ...
    A stainless steel mallet was positioned 10 cm from the stylus, and then 10 cm from the localizer to create errors attributable to electromagnetic interference.
  63. [63]
    Assessment of metallic distortions of a electromagnetic tracking system
    Aug 9, 2025 · Ferrous metals are known to produce distortion of the electromagnetic (EM) field, causing a decrease in accuracy of EMT (2)(3) (4) . Thus ...
  64. [64]
    [PDF] Calibration and characterization of electromagnetic position and ...
    Apr 2, 2025 · Typical position errors seen in an ideal environment are 0.5 mm − 1 mm RMS at a range of. 200 mm − 400 mm. In practice EMT pose error is ...
  65. [65]
    (PDF) A survey of electromagnetic position tracker calibration
    May 26, 2025 · It then explains a method based on creating tetrahedrons to create a lookup table used in compensating for the positional data component quickly ...
  66. [66]
    [PDF] Automatic Calibration of Magnetic Tracking - Dongyao Chen
    The soft/hard-iron parameters are calculated iter- atively using the gradient descent method. It can achieve 6-DoF soft-iron and hard-iron calibration. Wu ...
  67. [67]
    Improving Tracking Accuracy via Fusion of Electromagnetic and ...
    This study presents a hybrid electromagnetic. (EM) and inertial (IMU) tracking system that maintains its accuracy even when the EM measurements are disturbed ...
  68. [68]
    [PDF] Electromagnetic Tracking for Medical Imaging
    The NDI Aurora system has made the transition from virtual reality to medical applications through shrinking down the size of the sensor coils to make them ...Missing: history | Show results with:history
  69. [69]
    Polhemus Fastrak
    Polhemus FASTRAK is proven technology with real world credentials. Offering real time six-degrees-of-freedom tracking with virtually no latency.
  70. [70]
    [PDF] An Algorithm for Transmitter Optimization in Electromagnetic ...
    Commercially available tracking systems use a low frequency signal (e.g. as 25 Hz [26], 120 Hz [21] or 144 Hz [22]). These frequencies minimize any distortion.
  71. [71]
    (PDF) Electromagnetic tracking system with reduced distortion using ...
    Electromagnetic tracking with quadratic excitation is immune to the effects of a conductive distorter, especially compared with sinusoidal excitation of the ...
  72. [72]
    [PDF] Error State Extended Kalman Filter Multi-Sensor Fusion for ... - arXiv
    Sep 10, 2021 · Its implementation is adapted to fuse measurements from multiple sensor sources and the state model is extended to account for sensor drift and ...
  73. [73]
  74. [74]
    On the Functional and Extra-Functional Properties of IMU Fusion ...
    Apr 13, 2021 · The complementary filter used an α value of 0.97 . This parameter is close to an α of 0.98 often used for a complementary filter. The Mahony ...
  75. [75]
    A General Optimization-based Framework for Global Pose ... - arXiv
    Jan 11, 2019 · In this paper, we propose a sensor fusion framework to fuse local states with global sensors, which achieves locally accurate and globally drift-free pose ...Missing: tracking | Show results with:tracking
  76. [76]
    Enhanced Kalman filter methods for end pose measurement of ...
    Jan 15, 2025 · Results demonstrate a significant reduction in mean position error from 0.569 mm to 0.293 mm and mean angular error from 0.251° to 0.144° ...Missing: percentage | Show results with:percentage
  77. [77]
    The Oculus Insight Positional Tracking System - AI Accelerator Institute
    Mar 19, 2020 · Oculus Insight is the "inside-out" positional tracking system that enables full 6DoF (degree of freedom) tracking on the these VR headsets, as ...
  78. [78]
    Quantitative analysis of the Oculus Rift S in controlled movement
    Nov 15, 2019 · To assess the translational and rotational tracking performance of the Oculus Rift S using controlled robotic motion and a gold-standard motion tracking system.<|separator|>
  79. [79]
    3D Guidance: Quality OEM Tracking Solution - NDI
    The 3D Guidance® electromagnetic tracking solution delivers exceptional performance and value for OEM image-guided interventional and therapy systems and ...Missing: inertial | Show results with:inertial
  80. [80]
    Recent Advances in Tracking Devices for Biomedical Ultrasound ...
    Tracking systems using electromagnetic signals can also provide sub-millimeter accuracy in dynamic and real-time 6 DOF tracking. Its advantages include being ...
  81. [81]
    Project Tango: Introduction to Motion Tracking - YouTube
    Oct 6, 2015 · Project Tango is a computer vision platform for mobile devices that can track their motion, recognize spaces, and scan their surroundings in ...Missing: ultrasonic pose
  82. [82]
    Spring 2016 Pathfinder: Project Tango Preliminary Research - Arxterra
    Apr 21, 2016 · Motion Tracking. The IMU is used to give feedback on the device orientation and spatial displacement from a reference point. The Motion Tracking ...
  83. [83]
    Project Tango Data Handling - Google Help
    Motion Tracking adds capabilities to understand the movement of your device while you use an app. Motion Tracking uses cameras and sensors such as the ...Missing: ultrasonic AR
  84. [84]
    Apple Vision Pro - Technical Specifications
    Four eye‑tracking cameras; TrueDepth camera; LiDAR Scanner; Four inertial measurement units (IMUs); Flicker sensor ... Hands; Eyes; Voice. Supported Input ...Missing: fusion | Show results with:fusion
  85. [85]
    Apple Vision Pro
    Responsive, precision eye tracking. A high‑performance eye‑tracking system of LEDs and infrared cameras projects invisible light patterns onto each eye. This ...Apple (AU) · Apple (CA) · Apple (SG) · Apple (UK)Missing: fusion IMU
  86. [86]
    Apple Vision Pro sensors: What they all do, and where they are on ...
    Jan 22, 2024 · Here, we're going to take a deep dive into how the Vision Pro works its magic using sensors to track you and what they are all capable of doing.
  87. [87]
    Minimal-Drift Heading Measurement using a MEMS Gyro for Indoor ...
    4.1. Experimental Setup. The technique proposed in this paper was tested using an Epson XV3500, a low-cost MEMS gyro with 17.99 deg/h bias drift, and 2.5°/√hr ...Missing: per tracking
  88. [88]
    Accuracy Investigation of the Pose Determination of a VR System
    Feb 25, 2021 · This paper investigates the system behaviour of the VR system HTC Vive Pro at a testing facility that is designed for the calibration of highly accurate ...
  89. [89]
    Real-time, low-latency closed-loop feedback using markerless ...
    We provide a new DeepLabCut-Live! package that achieves low-latency real-time pose estimation (within 15 ms, >100 FPS), with an additional forward-prediction ...
  90. [90]
    Validity of tremor analysis using smartphone compatible computer ...
    Apr 18, 2025 · This study examines the validity of the Mediapipe (by Google) and Vision (by Apple) smartphone-enabled hand detection frameworks for tremor analysis.Missing: bandwidth issues
  91. [91]
    Development and Evaluation of a Low-Jitter Hand Tracking System ...
    Jitter refers to image rendering issues that cause virtual hands to shake, resulting in inconsistency between the movements of virtual hands and the responses ...
  92. [92]
    Visual performance standards for virtual and augmented reality
    Jul 15, 2025 · This article presents an overview of the available international visual performance standards for virtual and augmented reality devices.
  93. [93]
    On-Device, Real-Time Hand Tracking with MediaPipe
    Aug 19, 2019 · This approach provides high-fidelity hand and finger tracking by employing machine learning (ML) to infer 21 3D keypoints of a hand from just a single frame.
  94. [94]
    BlazePose: On-device Real-time Body Pose tracking - arXiv
    Jun 17, 2020 · We present BlazePose, a lightweight convolutional neural network architecture for human pose estimation that is tailored for real-time inference on mobile ...Missing: Blog | Show results with:Blog<|separator|>
  95. [95]
    Pose landmark detection guide | Google AI Edge
    Jan 13, 2025 · The MediaPipe Pose Landmarker task lets you detect landmarks of human bodies in an image or video. You can use this task to identify key body locations.Python · Web · Android · iOSMissing: 2019 | Show results with:2019
  96. [96]
    Soli: ubiquitous gesture sensing with millimeter wave radar
    We demonstrate that Soli can be used for robust gesture recognition and can track gestures with sub-millimeter accuracy, running at over 10,000 frames per ...
  97. [97]
    Egocentric Body Tracking using mmWave Radars on an MR Headset
    Sep 3, 2025 · We tackle the sparsity issue of mmWave radar by processing the raw signal into high-resolution range profiles to predict fine-grained 3D ...
  98. [98]
    What Is Neuralink? What We Know So Far. - Built In
    May 25, 2023 · The implant device enables users with paralysis to control computers with their thoughts and aims to restore mobility, treat neurological ...<|separator|>
  99. [99]
    Neuralink Updates
    In January, we conducted the first human implantation of our brain-computer interface (BCI). We were able to detect the participant's neural signals shortly ...Neuralink raises $650 million... · Datarepo - Neuralink's... · A Year of TelepathyMissing: pose | Show results with:pose
  100. [100]
    Speech motor cortex enables BCI cursor control and click - PMC
    Nov 22, 2024 · Decoding neural activity from ventral (speech) motor cortex is known to enable high-performance speech brain-computer interface (BCI) control.
  101. [101]
    [PDF] 3D perception: Cutting-edge AI research to understand the 3D world
    Jul 19, 2022 · Qualcomm AI Research is an initiative of Qualcomm Technologies, Inc. Pose estimation. Depth estimation & 3DR. Novel AI techniques for 3D ...
  102. [102]
    Optimizing Your AI Model for the Edge - Qualcomm
    Jun 12, 2025 · We talk about five techniques—compiling to machine code, quantization, weight pruning, domain-specific fine-tuning, and training small models ...
  103. [103]
    5G Advanced, your network for the next wave of 5G - Ericsson
    5G Advanced unleashes new powerful network features and capabilities that makes it possible to automate, innovate, and monetize with 5G in entirely new ways.Missing: tracking | Show results with:tracking<|separator|>
  104. [104]
    Scientists sidestep Heisenberg uncertainty principle in precision ...
    Sep 25, 2025 · Ultra-precise quantum sensors could sharpen navigation in environments where GPS doesn't work (such as submarines, underground or spaceflight); ...
  105. [105]
    Towards an Evolved Immersive Experience: Exploring 5G - MDPI
    With the advancement of 5G technology, AR/VR requires significant resources for real-time processing, i.e., rendering, vision, and physics engines, at end ...