Fact-checked by Grok 2 weeks ago

Point cloud

A point cloud is a collection of data points in a three-dimensional that represents the external surface of an object or , typically consisting of unstructured vectors with spatial coordinates (x, y, z) and optional attributes such as color, , or surface normals. These points are unordered and lack predefined connectivity, making them a fundamental yet primitive representation for data in fields like and . Point clouds can contain millions to billions of points, capturing geometric, colorimetric, and radiometric information to model , , , and orientation. Point clouds are primarily acquired through techniques such as (Light Detection and Ranging) scanners, which emit laser pulses to measure distances and generate dense point sets at rates up to 2.2 million points per second, or using structure-from-motion (SfM) algorithms on overlapping images from cameras or UAVs. Depth sensors like RGB-D cameras (e.g., Kinect) also produce point clouds by combining color images with depth maps, while hybrid methods integrate with to mitigate issues like sparsity and occlusions. Processing point clouds involves challenges such as handling noise, varying density, and the absence of semantic structure, often requiring segmentation (grouping points into clusters) and (labeling for meaning) to enable further or into meshes or models. In applications, point clouds support for of historical sites, in , for autonomous vehicles, and canopy analysis in via derived models like Digital Surface Models (DSMs) and Canopy Height Models (CHMs). They are also integral to , , , and , where techniques have advanced tasks like and segmentation despite the data's irregularity. Advances in point cloud processing continue to address computational demands, enabling broader use in and environmental modeling.

Fundamentals

Definition and Characteristics

A point cloud is a set of data points in , where each point is defined by its Cartesian coordinates (x, y, z) to represent the surface or geometry of an object or environment. These points may also include additional attributes, such as color (RGB values), intensity (reflectance measure), surface normals (directional vectors perpendicular to the surface), or classification labels, which provide contextual information beyond mere position. Point clouds can be organized, retaining a structured arrangement like a grid from acquisition methods such as depth sensors, or unorganized, consisting of a simple list of points without inherent order. The concept of point clouds originated in the 1960s through early techniques, which involved manual stereo compilation from aerial imagery to generate sparse data points representing surfaces. It gained widespread popularity in the with the advent of technologies, including , which enabled the automated capture of denser point distributions for applications in and modeling. Key characteristics of point clouds include their sparsity, where point density varies unevenly across the dataset due to factors like distance from the source or occlusions, leading to irregular sampling. They are prone to from inaccuracies and outliers as anomalous points that deviate significantly from the true surface, often requiring preprocessing for reliable use. Additionally, point clouds exhibit high challenges, as datasets can encompass millions to billions of points, demanding efficient storage and computational methods to handle large-scale processing. Unorganized point clouds are typically unordered collections lacking predefined topological relationships. This structure makes them flexible for raw geometric representation but necessitates additional algorithms to infer surfaces or features.

Mathematical Representation

A point cloud is formally defined as a P = \{ \mathbf{p}_i \mid i = 1, \dots, N \}, where N is the number of points and each \mathbf{p}_i is a vector in three-dimensional \mathbb{R}^3, expressed as \mathbf{p}_i = (x_i, y_i, z_i)^T. For unorganized point clouds, this representation captures the spatial positions of sampled points from an object's surface or environment, treating the cloud as an unordered collection without inherent between points. The basic positional data can be augmented with additional attributes to enrich the geometric and semantic information. For instance, each point may include a surface vector \mathbf{n}_i \in \mathbb{R}^3 to indicate local orientation or RGB color values \mathbf{c}_i \in \mathbb{R}^3 for visual properties, yielding an extended form \mathbf{p}_i = (x_i, y_i, z_i, \mathbf{n}_i, \mathbf{c}_i). Such attributes support downstream tasks like rendering and analysis while maintaining the core set-based structure. Point clouds are primarily represented in a , which aligns naturally with for most processing algorithms. However, for applications involving radial acquisition like , spherical or polar coordinates—defined by r, \theta, and \phi—may be used to better match geometries. Transformations between these systems, or between different frames, rely on motions to preserve distances and angles: \mathbf{p}' = R \mathbf{p} + \mathbf{t}, where R is an orthogonal 3×3 (R^T R = I, \det(R) = 1) and \mathbf{t} \in \mathbb{R}^3 is the translation vector. This formulation enables of clouds from multiple viewpoints. To quantify point distribution and identify variations in sampling uniformity, local density metrics are computed. A straightforward k-nearest neighbors (k-NN) approach measures the average distance from a point \mathbf{p}_i to its k closest neighbors: \rho(\mathbf{p}_i) = \frac{1}{k} \sum_{j \in \mathcal{N}_k(i)} d(\mathbf{p}_i, \mathbf{p}_j), where \mathcal{N}_k(i) denotes the set of k nearest indices and d(\cdot, \cdot) is the Euclidean distance; lower \rho values signal higher local density. Kernel density estimation offers a smoother alternative, convolving the points with a kernel function (e.g., Gaussian) to approximate the underlying probability density. Sampling theory addresses the generation or of point clouds to achieve desired properties like uniformity. disk sampling ensures a blue-noise by enforcing a minimum separation \delta between any two points, which helps maintain detail without clustering or gaps during reduction of N. This method is particularly valuable for preserving geometric fidelity in large-scale clouds.

Acquisition Methods

Sensor Technologies

Point clouds are primarily generated using active and passive technologies that capture three-dimensional spatial data through various physical principles. Among the most prevalent active sensors is (Light Detection and Ranging), which employs pulses to measure distances via the time-of-flight method. In this approach, a emitter sends out short pulses of toward a target surface, and a detects the reflected signals; the time delay between emission and return, combined with the , calculates the distance to each point, enabling the construction of dense point clouds with sub-millimeter accuracy in controlled settings. systems are categorized into terrestrial (ground-based, tripod-mounted for static scanning), (mounted on or drones for large-area coverage), and (vehicle-integrated for dynamic environments like urban mapping). Terrestrial achieves high precision for localized objects, while variants cover expansive terrains, and setups facilitate acquisition during motion. Structured light scanners represent another key active technology, projecting known —such as stripes, grids, or speckle—onto an object and capturing the deformation of these patterns with a camera to compute coordinates via . The principle relies on the geometric relationship between the , camera, and surface: by analyzing the shifted pattern, the triangulates the of projected rays and viewing lines to generate point clouds, often augmented with RGB data for textured representations. A prominent example is the Kinect sensor, which uses structured light to produce RGB-D (color and depth) point clouds suitable for indoor and short-range applications. These scanners excel in capturing fine surface details but are typically limited to close-range scenarios due to pattern visibility constraints. Photogrammetry systems, in contrast, are passive sensors that derive point clouds from photographic images captured by cameras or multi-view setups, employing feature matching and algorithms. photogrammetry uses pairs of images from slightly offset viewpoints to identify corresponding (e.g., edges or corners) and compute disparities, which are converted to depth via and camera calibration parameters. Multi-view extensions overlapping images from various angles to reconstruct denser clouds through structure-from-motion techniques, estimating both camera poses and points iteratively. These methods are cost-effective for large-scale but depend on image quality and scene for reliable matching. Each sensor technology has inherent limitations that influence point cloud quality and applicability. offers long-range capabilities, extending up to several kilometers in configurations, but its performance degrades in adverse weather like or , which scatter pulses, and it struggles with occlusions in dense . Resolution varies by system, with point densities reaching hundreds of points per square meter for high-end setups, though voxel-based representations may achieve resolutions with voxel sizes on the order of 0.5 meters (or smaller in high-resolution processing). Structured light scanners provide high resolution for small objects but are constrained to short ranges (typically under 5 meters) and sensitive to , which can wash out projected patterns, leading to incomplete clouds in reflective or transparent surfaces. excels in texture-rich environments but fails on featureless or low-contrast areas, with accuracy dropping below 1 cm in poor or motion-blurred images, and it inherently suffers from occlusions where viewpoints cannot access hidden surfaces. As of 2025, emerging advancements include solid-state , which replaces mechanical rotating components with integrated photonic chips for compact, reliable integration into consumer devices like smartphones and autonomous vehicles, enabling widespread point cloud generation at lower costs. As of 2025, AI-enhanced algorithms are increasingly integrated with for improved real-time performance in GPS-denied environments, such as indoor . Additionally, hyperspectral sensors are increasingly fused with to create attribute-rich point clouds, capturing not only but also spectral signatures across hundreds of wavelengths for enhanced material classification in . These developments address traditional limitations in portability and data dimensionality, broadening point cloud applications.

Data Capture Techniques

Point cloud data capture techniques encompass a range of procedural methods designed to collect spatial data efficiently and accurately, often tailored to the environment and application requirements. These techniques prioritize systematic scanning and integration strategies to generate dense, representative point sets while mitigating inherent limitations such as incomplete coverage or environmental interference. Active methods, which emit controlled energy sources like pulses to directly measure distances, provide precise depth information independent of , making them suitable for controlled or low-light settings. In contrast, passive methods infer indirectly from natural or ambient light, typically through multi-image analysis like structure-from-motion, which reconstructs point clouds from overlapping photographs but requires sufficient visual features and illumination for reliable matching. Active approaches, such as those using , achieve sub-centimeter accuracy in direct ranging, while passive techniques like can scale to large areas but often introduce higher variability due to inference-based estimation. Scanning protocols vary between single-scan and multi-scan setups to balance coverage and efficiency. Single-scan protocols involve a stationary sensor capturing a complete view from one position, ideal for small, unobstructed objects where full visibility is feasible, but they limit data density for complex geometries. Multi-scan setups, conversely, employ sequential acquisitions from multiple viewpoints to compile comprehensive datasets, often using terrestrial or mobile platforms to circumnavigate the scene. (SLAM) extends multi-scan protocols for real-time capture in dynamic environments, such as indoor or urban navigation, by iteratively estimating sensor pose and building incremental point clouds without external positioning aids. algorithms, like those integrating with inertial measurements, enable handheld or vehicle-mounted scanning with loop-closure optimizations to correct drift, achieving global accuracies on the order of 1-5 cm in GPS-denied spaces. Multi-view fusion integrates data from these scans by aligning point clouds across sensor positions, commonly via to minimize reprojection errors and enforce geometric consistency. This process optimizes camera or poses and point positions jointly, reducing accumulated misalignment in large-scale reconstructions; for instance, it has been shown to improve registration accuracy in multi-frame datasets compared to pairwise methods. treats the fusion as a non-linear least-squares problem, incorporating constraints from overlapping views to produce a unified, dense point cloud suitable for applications like documentation. Quality control during capture ensures georeferencing accuracy and data reliability through ground control points (GCPs), which are precisely surveyed markers used to anchor point clouds to a global . GCPs facilitate transformation computations, with error metrics like Error (RMSE) quantifying positional deviations—typically targeting sub-centimeter RMSE for high-fidelity surveys by distributing 6-10 points evenly across the scene. Additional protocols include on-site validation scans and reflectance calibration to account for surface properties affecting signal return. Challenges in data capture, particularly occlusions from self-shadowing or obstructing elements, are addressed through multi-angle acquisition strategies that ensure redundant viewpoints, or aerial surveys using drones to access elevated perspectives. Drone-based methods, for example, have demonstrated significantly improved completeness in vegetated terrains by capturing top-down , relative to ground-based single scans. These approaches demand careful to manage computational load from high-volume data, but they enhance overall point cloud integrity for downstream analysis.

Data Representation

Storage Formats

Point cloud data is stored in a variety of formats designed to balance human readability, compactness, and support. These formats range from simple text-based structures to sophisticated standards, each optimized for specific applications such as , geospatial , or industrial measurement. ASCII-based formats, such as .xyz, .pts, or .asc s, represent the simplest approach to point cloud storage. These are files where each line typically contains delimited values for point coordinates (e.g., X Y Z separated by spaces or commas), and optionally additional attributes like intensity or color. For instance, a basic .xyz might list coordinates as floating-point numbers without a formal header, making it straightforward to generate or parse with standard tools. However, their human-readable nature comes at a cost: they produce large sizes for datasets with millions of points—often several gigabytes for moderate scans—and require time-intensive parsing, rendering them inefficient for large-scale processing. Binary formats address these limitations by offering compactness and faster I/O operations. The Polygon File Format (PLY), originally developed at , supports both ASCII and encodings and is widely used for storing graphical objects, including point clouds as collections of . A PLY file begins with a header specifying elements like (with core properties such as x, y, z coordinates as floats) and optional scalar or list properties (e.g., colors as unsigned chars or normals as floats), followed by the data section. It also accommodates faces defined by indices, enabling representations alongside points, though for pure point clouds, only data is utilized. Binary PLY files achieve smaller sizes and quicker loading compared to ASCII equivalents, making them suitable for graphics applications. The format, native to the Point Cloud Library (PCL), provides flexible binary or ASCII storage tailored for 2D/3D point cloud processing. Its header includes metadata like the data version (e.g., 0.7), field names (e.g., x, y, z, rgb), data types (e.g., float32), point count, and viewpoint; the data follows in either unorganized (flat list, height=1) or organized (structured like an image, with width and height) layouts. PCD supports additional per-point properties beyond coordinates, such as normals or curvatures, stored contiguously for efficient access. While PCD files themselves do not embed spatial indexing, PCL's runtime structures like can be applied to PCD data for accelerated queries and downsampling, enhancing scalability for datasets with billions of points. Standardized formats ensure in specialized domains. The () format, defined by the American Society for Photogrammetry and (ASPRS), is a standard for LiDAR-derived point clouds in geospatial applications. It features a fixed-size public header block (375 bytes in version 1.4) for file , followed by variable-length (VLRs) for (e.g., via WKT or ) and point data (20-67 bytes each, supporting up to 10 formats with attributes like , return number, and ). LAS enables storage of up to 15 returns per pulse and scales to massive aerial surveys, with its structure allowing efficient reading over ASCII alternatives. The E57 format, an standard (E2807), targets 3D imaging systems for industrial and workflows. It uses a : an XML root file describes the hierarchical organization (e.g., scans and images), while binary sections store raw point data (Cartesian or spherical coordinates, with flexible fields like or color) and imagery (e.g., embeds). Supporting unorganized or gridded point clouds up to exabyte scales, E57 includes comprehensive such as creation timestamps, sensor poses, and geodetic references, promoting vendor-neutral exchange without proprietary extensions. Its design facilitates integration of points with associated images, though it results in larger files than pure binary formats like LAS due to XML overhead. Efficiency trade-offs among these formats depend on dataset size and : ASCII options excel in and editability for small datasets but falter in storage (e.g., a 1 binary file might expand to 5-10 in ASCII) and performance, while binary formats like PLY, , , and E57 offer 4-10x compression advantages and faster processing, with indexing in libraries like PCL further enabling handling of billion-point clouds via structures such as octrees. Open-source tools, particularly the Point Cloud Library (PCL), support reading and writing across these formats—including PLY, , , and E57—facilitating seamless conversion and integration in workflows.

Geometric and Attribute Data

Point clouds fundamentally consist of geometric data that defines the spatial arrangement of sampled points in . The core geometric attribute is the position of each point, typically represented as a triplet (x, y, z) in Cartesian coordinates, which captures the location relative to a global or . Beyond positions, surface normals provide for each point, serving as unit vectors perpendicular to the estimated local surface tangent plane; these are essential for tasks like and in rendering. Curvature metrics further describe local surface geometry, with measuring intrinsic bending (product of principal curvatures) and indicating average bending, both derived from approximations of the second form on discrete points. Attribute data augments geometric information to enhance interpretability and utility. Intensity values, common in LiDAR-acquired clouds, quantify the returned signal strength, reflecting surface reflectivity and enabling material differentiation. Color attributes, often as RGB triplets, are fused from co-registered imagery to add visual fidelity, particularly useful in photogrammetric reconstructions. Semantic labels assign categorical identifiers to points (e.g., "wall" or "furniture"), facilitating understanding in applications like indoor . For dynamic point clouds, timestamps record acquisition times, supporting motion analysis in time-varying environments such as . Point cloud data can be organized as unstructured sets of independent points or structured into grids for efficient querying. Spatial indexing structures like k-d trees, which recursively partition space along alternating axes, and octrees, which hierarchically subdivide into cubic voxels, accelerate neighbor searches and reduce computational overhead in large datasets. Enrichment methods, such as normal estimation, often employ on local neighborhoods: for a point p_i, compute the covariance matrix C of its k-nearest neighbors, then select the normal \mathbf{n}_i as the eigenvector corresponding to the smallest eigenvalue of C, minimizing variance along the surface normal direction. \mathbf{n}_i = \arg\min_{\mathbf{v}} \mathbf{v}^T C \mathbf{v}, \quad \|\mathbf{v}\| = 1 This approach provides a robust approximation of surface orientation from raw positions alone. In analysis, attributes enable targeted operations; for instance, intensity thresholds can filter points by reflectivity to isolate vegetation from bare earth in environmental surveys.

Processing Techniques

Alignment and Registration

Alignment and registration in point cloud processing involve aligning multiple point clouds acquired from different viewpoints or sensors into a unified , enabling the creation of comprehensive 3D models. The core problem is to estimate a , consisting of a R and translation vector t, that minimizes the distance between a source point cloud S and a target point cloud T. This transformation aligns corresponding points across the clouds while preserving geometric structure, typically assuming initial overlap and no significant deformations. The (ICP) algorithm serves as a foundational method for this task, iteratively refining the alignment by establishing correspondences between points in S and T, then computing the optimal . Introduced in its seminal form for 3D shape registration, ICP operates in two alternating steps: finding the closest point in T for each point in the transformed S, followed by least-squares minimization to update R and t. The objective minimizes the error metric E = \sum_i \| p_i - (R q_i + t) \|^2, where p_i are points in T and q_i their corresponding points in S. Common variants include point-to-point ICP, which directly matches points, and point-to-plane ICP, which aligns points to the tangent planes of the target surface for improved robustness to sparse data. Feature-based methods enhance registration by first extracting robust descriptors to identify correspondences, particularly useful when initial alignments are poor or overlaps are partial. Descriptors such as Fast Point Feature Histograms (FPFH) capture local geometric properties around each point using simplified histograms of angular and distance relations among neighboring points, enabling efficient matching via nearest-neighbor search. These features guide initial correspondence estimation, often followed by for refinement, reducing sensitivity to outliers and noise compared to pure . Registration approaches distinguish between global and local strategies to handle varying initial poses. Global methods, such as those employing RANSAC, provide coarse alignment by randomly sampling point correspondences and fitting transformations, rejecting outliers to estimate an initial R and t robustly across non-overlapping or noisy clouds. Local fine-tuning then applies or its variants iteratively for precision. For scenarios involving deformable objects, non-rigid variants extend this framework by incorporating flexible transformations, such as piecewise affine mappings or deformation fields, to account for elastic changes while maintaining overall rigidity where possible. Alignment quality is evaluated using metrics that quantify geometric fidelity between registered clouds. The Chamfer distance measures average nearest-neighbor distances bidirectionally, providing a symmetric assessment of point discrepancies suitable for dense clouds. The , conversely, captures the maximum deviation between sets, highlighting worst-case misalignments and thus emphasizing accuracy. These metrics guide selection and validate results, with lower values indicating better registration.

Segmentation and Feature Extraction

Segmentation of point clouds involves partitioning the into meaningful subsets, such as surfaces, objects, or semantic classes, to facilitate further . Traditional methods rely on geometric properties like or planarity to group points. Region-growing algorithms initiate from points and expand regions by incorporating neighboring points that satisfy criteria such as surface normal similarity or thresholds, ensuring smooth . This approach is particularly effective for identifying planar or curved surfaces in scanned environments. Plane fitting techniques detect flat regions by estimating dominant within the point cloud. The (RANSAC) algorithm samples minimal point sets to hypothesize plane parameters, iteratively refining the model by inlier consensus while rejecting outliers, making it robust for segmenting large, noisy datasets into primitive shapes like walls or floors. An efficient variant accelerates this process by prioritizing shape primitives and adaptive sampling, achieving performance on unorganized point clouds. Semantic segmentation assigns class labels to individual points or regions, enabling high-level understanding such as distinguishing ground from in outdoor scans. Machine learning methods, particularly deep neural networks, process raw point coordinates directly without voxelization or projection. PointNet, a pioneering architecture, uses shared multilayer perceptrons and max-pooling to extract permutation-invariant features, followed by classification layers for per-point labeling, demonstrating superior performance on benchmarks like ShapeNet for part segmentation. Following PointNet, advanced architectures like PointNet++ have introduced hierarchical for multi-scale analysis, while models such as KPConv and Point Transformer, developed up to 2021, further enhance performance on large-scale outdoor scenes through kernel-based convolutions and attention mechanisms, as surveyed in reviews through 2025. Feature extraction identifies salient points and descriptors to capture local geometry for tasks like matching or recognition. Keypoint detection methods locate stable interest points robust to noise and transformations. The 3D Harris operator extends the corner detector by analyzing eigenvalue ratios of the derived from surface normals in a local neighborhood, highlighting corners or high-curvature regions in point clouds. Similarly, Intrinsic Shape Signatures (ISS) compute keypoints based on principal curvatures from the , selecting points where eigenvalues differ significantly to ensure uniqueness and repeatability. Descriptors encode neighborhood shapes around these keypoints; spin images represent points in a relative to a oriented base point, accumulating densities in a histogram for rotation-invariant matching, originally developed for cluttered scene recognition. Clustering algorithms group points based on density or proximity without assuming predefined shapes. DBSCAN (Density-Based Spatial Clustering of Applications with Noise) identifies clusters as dense regions separated by sparse areas, using two parameters: ε, the maximum distance between points in a cluster, and MinPts, the minimum number of points required to form a core. This method excels in handling arbitrary shapes and outliers in unevenly distributed point clouds, such as those from scans, by expanding clusters from core points within ε neighborhoods. An improved variant adapts ε locally for varying densities in data, enhancing segmentation accuracy for urban scenes. Point cloud segmentation faces challenges from noise introduced by sensor inaccuracies and varying point densities due to distance or occlusion, which can lead to fragmented regions or misgrouped points. Over-segmentation occurs when minor density variations create spurious boundaries, requiring adaptive thresholds or preprocessing like denoising to maintain coherence. Extracted features from segmentation aid in aligning multiple point clouds by providing robust correspondences.

Surface Reconstruction

Surface reconstruction from point clouds involves algorithms that convert discrete point samples into continuous representations, such as triangle meshes or implicit surfaces, to model underlying . These methods typically require oriented point clouds, often obtained from aligned scans, to infer local surface normals and ensure coherent . Common approaches prioritize robustness to noise, preservation of sharp features, and generation of watertight or manifold surfaces suitable for downstream applications like rendering and simulation. Poisson surface reconstruction formulates the problem as solving a for an \chi that implicitly defines the surface as its zero . The core is \nabla^2 \chi = \nabla \cdot \mathbf{N}, where \mathbf{N} is a of smoothed point normals, solved efficiently using multigrid techniques to produce watertight meshes even from noisy inputs. This method excels in filling holes and handling non-uniform sampling densities, as demonstrated on range scans of complex objects like statues, yielding surfaces with low to (typically under 1% of bounding box diagonal). Introduced by Kazhdan et al., it has become a for implicit due to its theoretical guarantees on quality for manifolds. Delaunay triangulation-based methods construct meshes by filtering the 3D Delaunay complex of the points, with alpha shapes providing a parameterized way to extract boundary facets. The alpha shape is defined as the subset of the Delaunay triangulation where circumspheres of simplices have radius at most \alpha, controlling the tightness of the surface by excluding large voids (small \alpha) or including concave regions (larger \alpha). This convex hull-inspired approach guarantees a manifold triangulation for sufficiently dense samples on smooth surfaces, with \alpha often tuned via critical values from the filtration. Edelsbrunner and Mücke formalized alpha shapes as a generalization of convex hulls, enabling reconstruction of genus-zero objects from unorganized points with minimal parameters. Limitations include sensitivity to outliers, which can introduce spurious triangles, though post-processing like edge flipping improves aspect ratios. Moving least squares (MLS) reconstruction defines an implicit surface by projecting points onto local approximations fitted via weighted least squares. For a query point r, first fit a local plane H by minimizing \sum_i (\langle n, p_i \rangle - D)^2 \theta(\|p_i - q\|), where q is the foot of the perpendicular from r to H, n its normal, and \theta a compactly supported weight function (e.g., Gaussian). Then, fit a bivariate polynomial g in local coordinates by minimizing \sum_i (g(x_i, y_i) - f_i)^2 \theta(\|p_i - q\|), where f_i is the signed distance of p_i to H. The projected point on the surface is q + g(0, 0) n. This yields a smooth, interpolating surface without explicit meshing, ideal for denoising sparse clouds, and can be rendered via ray tracing or triangulated afterward. Alexa et al. pioneered MLS for point-set surfaces, showing superior feature preservation compared to algebraic methods on scanned models. The approach handles boundaries by weighting, but may over-smooth thin structures unless higher-degree polynomials are used. The ball pivoting algorithm generates a by iteratively "rolling" a virtual ball of fixed radius over seed edges formed by point pairs, a third point when the ball touches it without intersecting others. Starting from a or random edges, it propagates facets across the cloud, naturally respecting local and avoiding intersections. Bernardini et al. developed this for multi-view range data, demonstrating efficient reconstruction of models like the Michelangelo with fewer than 100k triangles and runtime scaling linearly with points. It performs well on uniform densities but struggles with varying sampling rates, often requiring adaptive radii or pre-alignment to minimize holes. Quality assessment of reconstructed surfaces focuses on geometric fidelity and mesh validity, using metrics like triangle aspect ratio (ideally close to 1 for equilateral triangles, computed as r = \frac{2h}{\sqrt{3}a} where h is and a base) to detect skinny elements that distort . Hole detection involves identifying boundary loops longer than a threshold (e.g., 5% of edge length) or deviations from expected , often via checks. Handling thin structures remains challenging; methods like mitigate shrinkage through , while local approaches like ball pivoting may require estimation to avoid collapsing protrusions. et al. surveyed these metrics, noting that aspect ratios below 0.5 correlate with poor consistency in benchmarks like the Princeton Shape Benchmark.

Applications

Computer Graphics and Visualization

Point clouds serve as a fundamental representation in and , enabling the rendering of complex 3D scenes directly from discrete spatial samples without requiring intermediate . This approach is particularly advantageous for handling massive datasets from sources like or , where traditional polygonal models may be inefficient due to high triangle counts. Rendering pipelines for point clouds typically involve projecting points onto the and applying filtering to achieve smooth, anti-aliased visuals, while emphasizes interactive exploration and attribute-based coloring. One prominent rendering technique is splatting, which projects each point as an elliptical Gaussian kernel onto the screen to create a smooth, continuous appearance and mitigate artifacts inherent in point sampling. Developed in the early 2000s, elliptical weighted average (EWA) splatting extends basic point splatting by incorporating , ensuring high-quality and even for sparse or irregularly distributed points. This method has been widely adopted for its balance of visual fidelity and computational efficiency, with implementations leveraging GPU acceleration for real-time performance. Point-based further enhance this by integrating point clouds into GPU pipelines, such as through OpenGL's programmable shaders, allowing for hardware-accelerated rasterization of billions of points per . Early frameworks vertex and fragment shaders to perform splatting operations directly on the GPU, achieving interactive frame rates for complex models by avoiding CPU bottlenecks in . Modern extensions, including compute shaders, have pushed boundaries, enabling rendering of hundreds of millions of points in a few milliseconds on consumer hardware like GPUs. For interactive visualization, open-source tools like and provide robust platforms for exploring point clouds, supporting operations such as cross-sectional slicing to reveal internal structures and attribute-based coloring to highlight properties like intensity or normals. , in particular, incorporates level-of-detail () mechanisms that automatically decimate large clouds during viewport interactions, ensuring fluid navigation of datasets exceeding hundreds of millions of points without sacrificing detail in focused views. complements this with advanced rendering options, including shader-based effects for enhanced and multi-light setups to emphasize geometric features. These tools facilitate qualitative analysis and preparation for further graphical processing. In real-time applications such as video games and (), point clouds are rendered using surfels—surfaced elements that approximate local with oriented disks or ellipses for efficient drawing and . Introduced as a in the early , surfels enable interactive rates by storing only per-point attributes like position, , and radiance, bypassing connectivity overheads of meshes and supporting dynamic adjustments based on viewer distance. This approach has been integrated into pipelines for immersive walkthroughs of scanned environments, where continuous ensures seamless transitions from high-detail close-ups to sparse overviews, maintaining 60+ frames per second on mid-range hardware. Rendering point clouds presents challenges including from , which splatting techniques address through kernel filtering, and handling to prevent visual artifacts in dense scenes. Efficient ray tracing of point clouds often relies on acceleration structures like kd-trees, which partition space hierarchically to prune invisible regions and accelerate intersection queries, reducing traversal costs for photorealistic effects like shadows. Quantized variants of kd-trees further optimize storage for compressed clouds, enabling anti-aliased rendering of massive datasets with minimal . As of 2025, advances in point cloud rendering include the integration of within GPU shaders, particularly using Gaussian representations for volumetric effects that simulate and transparency in scanned objects. Techniques like Gaussian-based accelerate novel view synthesis by marching rays through implicit density fields derived from point attributes, achieving in applications from overlays to dynamic simulations. These methods build on 3D Gaussian splatting foundations, enhancing efficiency for sparse inputs while preserving high-fidelity outputs.

Robotics and Autonomous Systems

In robotics and autonomous systems, point clouds are essential for real-time environmental perception, supporting motion planning and obstacle avoidance in dynamic settings. Perception pipelines typically process raw point clouds from onboard LiDAR sensors by first applying clustering techniques, such as Euclidean distance-based grouping, to detect objects as distinct clusters within the scene. This is followed by 6D pose estimation, often using deep learning models on segmented point clouds, to compute the position and orientation of obstacles, enabling robots to predict trajectories and execute avoidance maneuvers with high precision. Point clouds are also integrated into (SLAM) frameworks to provide robust and in unstructured environments. In visual-inertial systems, algorithms like ( Odometry and ) leverage point cloud features—such as edges and surfaces—to estimate sensor velocity, undistort scans in real time, and register successive clouds for accurate pose tracking. LOAM's mapping module further supports loop closure by matching global features across scans, reducing cumulative drift and enabling consistent localization over long traversals, as demonstrated in ground vehicle and aerial applications. For path , point clouds are downsampled and voxelized to create efficient occupancy representations, which inform sampling-based or grid-search algorithms. Voxel grids derived from these clouds allow planners like A* for global routing on discretized spaces or RRT* for probabilistic exploration in high-dimensional configurations, optimizing paths while respecting kinematic constraints. is accelerated by enclosing clustered points in bounding volumes, such as oriented bounding boxes, to query potential intersections during planning iterations, ensuring collision-free trajectories in cluttered tasks. Prominent case studies highlight these applications: in autonomous driving, systems generate dense point clouds for 3D environmental reconstruction, supporting and path prediction in urban settings, as seen in early self-driving prototypes. For drone navigation, point cloud-based enables micro aerial vehicles (MAVs) to achieve speeds up to 4 m/s (approximately 9 mph) in forested or indoor environments, using to generate safe trajectories directly from onboard data. By 2025, edge advancements have further enabled on-device point cloud processing in mobile robots, with lightweight models performing clustering and locally to significantly reduce latency compared to cloud-dependent systems, enhancing responsiveness in time-critical operations.

Cultural Heritage and Archaeology

Point clouds have revolutionized the documentation of cultural heritage sites by enabling high-resolution, non-invasive scanning that captures intricate details of monuments for digital preservation. For instance, terrestrial laser scanning was employed in the Petra documentation project to generate dense point clouds of the rock walls in the Siq canyon and over 30 major structures, creating a comprehensive digital archive that supports long-term conservation planning. Similarly, the Scanning of the Pyramids Project utilized state-of-the-art RIEGL LMS terrestrial laser scanners to produce accurate 3D point cloud models of Egyptian pyramids, facilitating detailed analysis of their architectural features and structural integrity for archival purposes. The Giza Laser Scanning Survey further extended this approach by generating point clouds of monuments like Queen Khentkawes' tomb, allowing researchers to interpret landforms and historical modifications with millimeter precision. In archaeological analysis, point clouds support advanced techniques such as deviation mapping to monitor and degradation over time. By aligning sequential scans and computing color-coded deviation maps, researchers can quantify surface changes, as demonstrated in studies of historical constructions where point cloud-based damage detection identified patterns with sub-millimeter accuracy. For the Ming Great Wall, high-precision point cloud models integrated with imagery enabled comprehensive monitoring of damage progression, revealing localized wear rates that inform targeted efforts. restoration techniques, including point cloud , further aid in reconstructing missing elements; a self-supervised method projects incomplete point clouds into multi-channel occupancy probability images for pixel-level inpainting, yielding finer reconstructions than traditional approaches and supporting hypothetical site restorations. Notable case studies illustrate the practical impact of point cloud-derived 3D models in enabling remote access and simulations. The has been digitally reconstructed using point clouds to create immersive visualizations, allowing global audiences to explore its original sculptural decorations, columns, and friezes through virtual platforms that simulate ancient . For the , point cloud data from fragmented warriors facilitated classification and completion models, such as the CPDC-MFNet diffusion-based approach, which infers missing parts probabilistically to enable simulations of assembly and facial reconstructions accessible remotely for research and education. These models not only preserve fragile artifacts but also support virtual simulations of historical events, enhancing public engagement without physical handling. Collaborative EU-funded initiatives like Scan4Reco exemplify integrated point cloud applications for automated damage assessment in . The developed a portable multimodal scanning platform that generates layered point cloud models from and , predicting future deterioration states through spatiotemporal analysis to guide preventive conservation. This approach automates the detection of cracks and material loss, as tested on diverse assets, reducing manual intervention and enabling cost-effective for institutions. Ethical considerations in point cloud applications emphasize ownership and accessibility, particularly for sites, where must respect to prevent exploitation. The CARE Principles for Governance advocate for collective benefit and authority over derived from , ensuring that point cloud archives of sacred sites are managed with input to avoid unauthorized use. As of 2025, technologies have emerged to enhance tracking, with platforms like Ethereum-based systems securing for cultural assets, verifying and enabling transparent access controls that align with ethical standards. Such implementations address risks of in contexts by providing immutable records of and custodianship.

Compression and Standards

Compression Algorithms

Point cloud compression algorithms aim to reduce the storage and transmission requirements of large datasets while maintaining essential geometric and attribute information. These methods exploit spatial redundancies, predictive patterns, and statistical correlations inherent in point distributions. Broadly, they are categorized into lossless and lossy approaches, with the former preserving exact data fidelity and the latter allowing controlled distortion for higher compression ratios. Lossless compression techniques, such as those based on partitioning, enable by recursively subdividing space into voxels and encoding occupancy patterns with coders like or Golomb-Rice. A seminal method uses local surface predictions to estimate occupied cells, achieving compression ratios of 1.85–2.81 for without . In contrast, lossy methods introduce approximation, often via quantization of coordinates, to attain ratios exceeding 100:1 in applications like scanning, where minor geometric errors are tolerable. Geometry compression frequently employs octree-based to traverse and encode point positions efficiently, leveraging neighborhood correlations for intra-frame prediction. While traversal techniques inspired by connectivity, such as spiral or edge-based walks, have been adapted for sparse point sets, modern variants integrate for refined occupancy modeling. Attribute compression targets associated data like colors and normals through quantization, which reduces (e.g., from 10 to 8 bits per channel) while preserving perceptual quality. Transform coding, such as (DWT) applied to local point neighborhoods, decomposes attributes into frequency components for selective encoding, outperforming graph transforms in certain scenarios with up to 20% bitrate savings. Sparse convolution methods voxelize the point cloud into a hierarchical , then apply al neural networks (CNNs) to encode sparse occupancy and features, enabling end-to-end learning of artifacts. These approaches, often building on structures, achieve ratios up to 100:1 by focusing on non-empty voxels, with applications in dynamic scenes where motion prediction further enhances efficiency. Scalable coding supports progressive transmission by organizing data into hierarchical levels, allowing decoding at varying quality based on available —ideal for streaming in or autonomous systems. Techniques like layered entropy models enable base-layer lossless with enhancement layers for attributes, facilitating adaptive bitrate control. Such methods align with standards like MPEG's G-PCC for interoperable implementations. Performance is evaluated using metrics like BD-rate, which quantifies bitrate savings at equivalent distortion levels (negative values indicate gains, e.g., 30–40% over baselines), and PSNR variants such as point-to-point (D1) or point-to-plane (D2) for geometric fidelity, typically targeting 40–60 dB in high-quality scenarios.

Industry Standards and Formats

The Moving Picture Experts Group (MPEG) has developed the Point Cloud Compression (PCC) standards under ISO/IEC 23090 to enable efficient storage, transmission, and rendering of point cloud data, particularly for immersive applications. These standards, published starting in 2021 for V-PCC and 2023 for G-PCC, and updated through 2025 including the third edition of V-PCC published in 2025, address both static and dynamic point clouds by leveraging established video coding technologies and geometry-specific methods. MPEG PCC includes Video-based Point Cloud Compression (V-PCC, ISO/IEC 23090-5), which projects 3D points onto 2D frames for encoding via video pipelines, enabling compatibility with existing HEVC or infrastructure and facilitating dynamic sequences for real-time streaming, and Geometry-based Point Cloud Compression (G-PCC, ISO/IEC 23090-9), which employs voxelization and structures for encoding, while attributes like color are handled separately to support high-fidelity . These approaches cater to different use cases: G-PCC for sparse or irregular clouds and V-PCC for dense, attribute-rich data suitable for broadcast or AR/VR. Beyond MPEG, Google's Draco provides an open-source format and library for compressing point clouds and meshes, optimized for web transmission and integration with environments, achieving up to 90% size reduction without significant quality loss. Draco uses edgebreaker and predictive quantization techniques, making it ideal for browser-based rendering. Similarly, Potree is an open-source viewer that supports visualization of large-scale point clouds stored in compressed LAS (LAZ) formats, converting them into hierarchical octrees for efficient streaming and interaction over the web. Adoption of these standards has grown in AR/VR ecosystems, with glTF 2.0 incorporating Draco compression via the KHR_draco_mesh_compression extension to enable compact 3D asset delivery, as endorsed by the Khronos Group. Recent ISO/IEC updates from 2023 to 2025, including enhancements in MPEG's 152nd meeting, have improved support for high-resolution streaming up to 8K, integrating PCC with VVC for volumetric media in immersive video pipelines. Despite these advances, remains a challenge, with proprietary implementations potentially leading to and inconsistent decoding across devices. Validation tools like the G-PCC Test Model (TMC13) and V-PCC Test Model (TMC2) software provide reference encoders and decoders to ensure compliance and facilitate cross-vendor testing.

References

  1. [1]
    Deep Learning on Point Clouds and Its Application: A Survey - PMC
    Point cloud is a widely used 3D data form, which can be produced by depth sensors, such as Light Detection and Ranging (LIDAR) and RGB-D cameras.
  2. [2]
    [PDF] Comparing Point Clouds - Stanford Computer Graphics Laboratory
    Point clouds are one of the most primitive and fundamental surface representations. A popular source of point clouds are three dimensional shape acquisition ...
  3. [3]
    6.1.4. Point cloud and Digital Surface Models
    Jan 13, 2023 · As discussed in the lidar section, point clouds are the result of numerous registered points that can represent shape, size, position and ...
  4. [4]
    Point Cloud - an overview | ScienceDirect Topics
    A point cloud is basically a set of data points in a 3D coordinate system, commonly defined by x, y, and z coordinates. They are used to represent the surface ...
  5. [5]
    Structured vs. Unstructured Point Cloud Data - RCDOCS
    Point clouds are a type of 3D reality capture data consisting of individual points ... attributes like colour, intensity, surface normal, classification, etc.
  6. [6]
    Evolution of Point Cloud - LIDAR Magazine
    Jul 16, 2018 · Wikipedia has defined a “Point Cloud” as a set of data points in some coordinate system. In a three-dimensional coordinate system, these points ...
  7. [7]
    [PDF] Learning to Denoise and Remove Outliers from Dense Point Clouds
    Point clouds obtained with 3D scanners or by image-based reconstruction techniques are often corrupted with significant amount of noise and outliers.
  8. [8]
    Comprehensive Guide to Point Cloud Data in Computer Vision
    Jul 25, 2025 · Large point cloud datasets can contain billions of points, creating storage and processing bottlenecks. FiftyOne's efficient data structures ...
  9. [9]
    [PDF] Representations for 3D Geometry: Voxels, Point Clouds, Meshes ...
    May 2, 2018 · Points = unordered set of 3-tuples. Often ... No sharp features. 91. Page 92. Recap: Points→Implicit→Mesh. 92. Page 93. MESH-> POINT CLOUD.
  10. [10]
    Beyond the pixel plane: sensing and learning in 3D - The Gradient
    Aug 24, 2018 · "Voxels" are like pixels in 3D; think of voxel grids as quantized, fixed-sized point clouds. Whereas point clouds can have an infinite number of ...
  11. [11]
    [PDF] PointNet: Deep Learning on Point Sets for 3D Classification and ...
    A point cloud is represented as a set of 3D points {Pi| i = 1, ..., n}, where each point Pi is a vector of its (x, y, z) coordinate plus extra feature channels ...
  12. [12]
    Spherical coordinate transformation-embedded deep network for ...
    Ablation studies show that spherical coordinate transformation significantly improves the standard point cloud feature extraction network. ... Cartesian ...
  13. [13]
    A Tutorial Review on Point Cloud Registrations: Principle ...
    Jul 29, 2021 · In this section, we first briefly introduce the source of the point cloud, namely, the point cloud acquisitions. ... ℛ3 represent the rotation and ...<|control11|><|separator|>
  14. [14]
    [PDF] Adaptive ore point cloud filtering algorithm based on the K-nearest ...
    The density of the neighborhood of this point is represented by the inverse of the average distance of the K-neighborhood. The larger the density, the denser ...
  15. [15]
    Revisiting Poisson-disk Subsampling for Massive Point Cloud ...
    Nov 29, 2023 · In this paper, we revisit Poisson-disk sampling for point clouds. Instead of optimizing for equal densities, we propose to maximize the distance to the closest ...
  16. [16]
    LiDAR: What Is It and How Does It Work? - YellowScan
    LiDAR principle is pretty easy to understand: it emits a laser pulse on a surface. catches the reflected laser back to the LiDAR pulse source with sensors.
  17. [17]
    Understanding Lidar: A comprehensive guide to its functionality
    Jul 2, 2025 · Lidar technology is a cutting-edge remote sensing method that uses rapid laser pulses to measure distances with exceptional accuracy.
  18. [18]
    LiDAR Visualization - Computer Science | UC Davis Engineering
    Mar 26, 2023 · LiDAR's main benefits are accuracy of each sample point, and ... This allows the measurement of features with sub-millimeter accuracy.<|control11|><|separator|>
  19. [19]
    [PDF] Lidar 101: An Introduction to Lidar Technology, Data, and Applications
    Nov 6, 2012 · Tripod-based lidar systems produce point data with centimeter accuracy and are often used for localized terrain-mapping applications that ...
  20. [20]
    Accuracy comparison of terrestrial and airborne laser scanning and ...
    Airborne and terrestrial laser scanning (ALS, TLS) are remote sensing techniques based on highly-accurate range and angle measurements that enable accurate ...Accuracy Comparison Of... · 2. Material And Methods · 3. Results And Discussion
  21. [21]
    3D Scanning 101: Basics of Structured Light 3D Scanning - Polyga
    Dec 7, 2023 · The core principle of structured light 3D scanning revolves around triangulation. By projecting a known pattern onto an object and capturing it ...Missing: cloud Microsoft Kinect RGB-
  22. [22]
    A Structured Light RGB‐D Camera System for Accurate Depth ...
    Nov 1, 2018 · 3D reconstruction of the object surface is based on the ray triangulation principle. We construct an RGB-D sensing system with an industrial ...
  23. [23]
    [PDF] High-speed 3D shape measurement with structured light methods
    This section briefly explains the basic principles behind structured light techniques that use triangulation for 3D reconstruction, the epipo- lar geometry that ...
  24. [24]
    Photogrammetry Explained: From Multi-View Stereo to Structure ...
    Oct 14, 2024 · Multi-View Stereo is the geometric way to reconstruct a scene from multiple pairs of images using depth triangulation and digital camera ...
  25. [25]
    [PDF] Course Notes 5: Multi-View Stereo - TU Delft 3D Geoinformation
    Multi-View Stereo (MVS) uses images from different viewpoints to create 3D reconstructions by finding corresponding points to estimate depth.
  26. [26]
    Photogrammetric Multi-View Stereo and Imaging Network Design
    This thesis proposes a new approach, photogrammetric multi-view stereo, for accurate and dense 3D reconstruction, including scale recovery from images.
  27. [27]
    Demonstration of >1km LiDAR using 1.3µm Nd:YAG/vanadium:YAG ...
    Mar 21, 2025 · Single pulse return signals could readily be detected from objects having a range greater than 1500 m and point cloud maps of various scenes ...
  28. [28]
    Monitoring of urban forests using 3D spatial indices based on LiDAR ...
    The vegetation volume was calculated using airborne laser scanning point clouds (ALS LiDAR) processed into voxels (0.5 m). The volume of the buildings was ...
  29. [29]
    LiDAR vs Photogrammetry: Differences & Use Cases - YellowScan
    Jul 2, 2024 · Possible photogrammetry limitations and drawbacks:​​ Photogrammetry relies on good lighting conditions to capture high-quality images. Weather ...Missing: range | Show results with:range
  30. [30]
    Solid State LiDAR Sensor Market - Future Market Insights
    Aug 29, 2025 · Global solid state lidar sensor market to garner sales of USD 1,971.8 million in 2025 and further expected to total a size of USD 8,787.2 ...Missing: hyperspectral clouds
  31. [31]
    LiDAR Technology in Mapping 2025-2033 Analysis
    Rating 4.8 (1,980) Mar 15, 2025 · Emerging Trends in LiDAR Technology in Mapping. Emerging trends include: Integration with other sensor modalities (e.g., hyperspectral imaging) ...Missing: devices | Show results with:devices
  32. [32]
    Towards 3D Hyperspectral Imaging | IEEE Conference Publication
    We propose that 3D hyperclouds, in which Lidar or photogrammetric point clouds are augmented with hyperspectral attributes, can address numerous of these ...
  33. [33]
    (PDF) A comparison between active and passive techniques for ...
    Aug 6, 2025 · In particular, the point clouds obtained are compared in terms of number of acquired 3D points and geometrical deviation.
  34. [34]
    Computational Methods of Acquisition and Processing of 3D Point ...
    Feb 20, 2019 · This paper aims to review the state-of-the-art methods to acquire and process 3D point cloud data for construction applications.
  35. [35]
    SLAM-driven robotic mapping and registration of 3D point clouds
    This study proposes a framework to build 3D high-resolution point clouds registered in real time using a hybrid laser scanning system with a mobile robot.
  36. [36]
  37. [37]
    [PDF] On Bundle Adjustment for Multiview Point Cloud Registration
    Despite the success of pair- wise registration and pose synchronization, the concept of Bun- dle Adjustment (BA) has been proven to better maintain global.
  38. [38]
    Optimization of Ground Control Point Distribution for Unmanned ...
    Ground control point (GCP) is an important calibration factor when correcting position information during unmanned aerial vehicle (UAV) remote sensing.2. Materials And Methods · 2.3. 3. Gcp Distribution... · 3. Results
  39. [39]
    The PCD (Point Cloud Data) file format
    This document describes the PCD (Point Cloud Data) file format, and the way it is used inside Point Cloud Library (PCL).Missing: LAS E57
  40. [40]
    Knowledge: Point cloud formats: overview and practical use - Qbitec
    Apr 8, 2025 · Example: ASCII files such as .XYZ, .CSV, or .PTS. These formats originate from the early days of laser scanning. They store point data in ...
  41. [41]
    [PDF] PLY polygon file format - GAMMA
    This document presents the PLY polygon file format, a format for storing graphical objects that are described as a collection of polygons. Our goal is.
  42. [42]
    Module octree - Point Cloud Library (PCL)
    The pcl_octree library creates a hierarchical tree from point cloud data for spatial partitioning, downsampling, and search operations. It also provides ...
  43. [43]
    [PDF] LAS Specification 1.4 - R15 | ASPRS
    Jul 9, 2019 · The LASer (LAS) file is intended to contain point cloud data records, including those derived from lidar or other sources.
  44. [44]
    [PDF] The ASTM E57 File Format for 3D Imaging Data Exchange
    The E57 File Format for 3D Imaging Data Exchange (E57 format hereafter) is capable of storing point cloud data from laser scanners and other 3D imaging systems ...Missing: PLY PCD
  45. [45]
  46. [46]
    The Basics of LiDAR - Light Detection and Ranging - Remote Sensing
    Sep 13, 2024 · LiDAR or Light Detection and Ranging is an active remote sensing system that can be used to measure vegetation height across wide areas.Missing: kilometers | Show results with:kilometers
  47. [47]
    Estimating Surface Normals in a PointCloud - Read the Docs
    This tutorial will address the latter, that is, given a point cloud dataset, directly compute the surface normals at each point in the cloud.
  48. [48]
    [PDF] Curvature Estimation of 3D Point Cloud Surfaces Through the Fitting ...
    Experiments are designed on Gaussian curvature, mean curvature and principal directions for an analytical surface and discrete surfaces of point cloud data.
  49. [49]
    Point cloud scene layer symbology—ArcGIS Pro | Documentation
    A point cloud sourced from lidar can be attributed with RGB (red, green, and blue) bands. This attribution often comes from imagery collected at the same time ...
  50. [50]
    [PDF] Semantic Labeling of 3D Point Clouds for Indoor Scenes
    This paper uses 3D point clouds from RGB-D cameras to perform semantic labeling of indoor scenes, using a graphical model to capture geometric relationships.
  51. [51]
    [PDF] Efficient Radius Neighbor Search in Three-dimensional Point Clouds
    In robotics, kD-trees [9] and octrees [10] are widely adopted to search for nearest neighbors in three- dimensional data and research concentrated on memory-.
  52. [52]
    What Are All These Attributes In My Lidar Data?
    Jul 2, 2024 · Intensity. The intensity value of a point in a point cloud is the strength of the pulse returned to the sensor. Values for intensity range ...
  53. [53]
  54. [54]
    A review of non-rigid transformations and learning-based 3D point ...
    The goal of point cloud registration is to find a transformation between two or more corresponding point sets, which minimizes the distance between the ...
  55. [55]
    (PDF) Segmentation of point clouds using smoothness constraint
    We present a method for segmentation of point clouds using smoothness constraint, which finds smoothly connected areas in point clouds.
  56. [56]
    Efficient RANSAC for Point‐Cloud Shape Detection - Schnabel - 2007
    May 21, 2007 · In this paper we present an automatic algorithm to detect basic shapes in unorganized point clouds. The algorithm decomposes the point cloud into a concise, ...
  57. [57]
    Deep Learning on Point Sets for 3D Classification and Segmentation
    Dec 2, 2016 · Our network, named PointNet, provides a unified architecture for applications ranging from object classification, part segmentation, to scene semantic parsing.
  58. [58]
    A robust extension of the Harris operator for interest point detection ...
    Aug 6, 2025 · In this paper, we present an interest points detector for 3D objects based on Harris operator, which has been used with good results in computer vision ...
  59. [59]
    Intrinsic shape signatures: A shape descriptor for 3D object recognition
    Abstract: This paper presents a new approach for recognition of 3D objects that are represented as 3D point clouds. We introduce a new 3D shape descriptor ...
  60. [60]
    [PDF] Using Spin Images for Efficient Object Recognition in Cluttered 3D ...
    Abstract—We present a 3D shape-based object recognition system for simultaneous recognition of multiple objects in scenes containing clutter and occlusion.
  61. [61]
    DBSCAN Clustering in ML - Density based clustering - GeeksforGeeks
    Oct 30, 2025 · DBSCAN is a density-based clustering algorithm that groups data points that are closely packed together and marks outliers as noise based on their density in ...
  62. [62]
    An Improved DBSCAN Method for LiDAR Data Segmentation with ...
    Jan 5, 2019 · The results demonstrate that the proposed algorithm can segment different types of LiDAR point clouds with higher accuracy in a robust manner.
  63. [63]
    Rethinking Design and Evaluation of 3D Point Cloud Segmentation ...
    In this paper, an in-depth analysis of five fundamental and representative deep learning models for 3D point cloud segmentation is presented.
  64. [64]
    [PDF] Poisson Surface Reconstruction - People @EECS
    We show that surface reconstruction from oriented points can be cast as a spatial Poisson problem. This Poisson formulation considers all the points at once ...
  65. [65]
    [PDF] Point Set Surfaces - TAU
    Lee [24] uses a moving-least squares approach to the reconstruc- tion of curves from unorganized and noisy points. He proposes a solution for reconstructing ...
  66. [66]
    [PDF] The Ball-Pivoting Algorithm for Surface Reconstruction
    This paper focuses on the data integration phase. We present a new method for finding a triangle mesh that interpolates an unorganized set of points. Fig. 1 ...Missing: seminal | Show results with:seminal
  67. [67]
    [PDF] State of the Art in Surface Reconstruction from Point Clouds - Hal-Inria
    Contouring a triangulation can be done via marching tetrahedra, but if a mesh with a lower number of triangles and well-conditioned (i.e. good aspect ratio) ...
  68. [68]
    [PDF] EWA Splatting - UMD Computer Science
    Abstract—In this paper, we present a framework for high quality splatting based on elliptical Gaussian kernels. To avoid aliasing.
  69. [69]
    Surface splatting | Proceedings of the 28th annual conference on ...
    Surface splatting makes the benefits of EWA texture filtering available to point-based rendering. It provides high quality anisotropic texture filtering, hidden ...
  70. [70]
    [PDF] High-Quality Point-Based Rendering on Modern GPUs
    In this paper we propose a rendering framework for point-based geometry providing high visual quality as well as efficient rendering. Our approach is based on a ...Missing: seminal | Show results with:seminal
  71. [71]
    [PDF] Rendering Point Clouds with Compute Shaders - arXiv
    Aug 7, 2019 · ABSTRACT. We propose a compute shader based point cloud rasterizer with up to 10 times higher performance than classic point-based rendering.
  72. [72]
    MeshLab
    Visualization and Presentation. The visualization features of MeshLab (including Decorators and Shaders) can help in graphically present the peculiar ...Missing: LOD | Show results with:LOD
  73. [73]
    [PDF] Surfels: Surface Elements as Rendering Primitives
    Surfels are point primitives without explicit connectivity, used to efficiently render complex objects, unlike triangles which store connectivity data.
  74. [74]
    [PDF] Real-Time Continuous Level of Detail Rendering of Point Clouds
    In this paper, we propose a continuous level-of-detail (CLOD) method for point clouds that addresses the challenges of real-time point cloud rendering in VR ...Missing: video | Show results with:video
  75. [75]
    The Quantized kd-Tree: Efficient Ray Tracing of Compressed Point ...
    This paper introduces a novel acceleration structure called the quantized kd-tree, which offers both efficient traversal and storage.
  76. [76]
    RayGaussX: Accelerating Gaussian-Based Ray Marching for Real ...
    Sep 9, 2025 · RayGauss has achieved state‑of‑the‑art rendering quality for novel‑view synthesis on synthetic and indoor scenes by representing radiance and ...
  77. [77]
    PFGS: High Fidelity Point Cloud Rendering via Feature Splatting
    Jul 4, 2024 · We propose a novel framework to render high-quality images from sparse points. This method first attempts to bridge the 3D Gaussian Splatting and point cloud ...
  78. [78]
    Instance segmentation based 6D pose estimation of industrial ...
    This article proposes a deep learning-based pose estimation using point cloud as input, which consists of instance segmentation and instance point cloud pose ...
  79. [79]
    [PDF] LOAM: Lidar Odometry and Mapping in Real-time
    An odometry algorithm estimates velocity of the lidar and corrects distortion in the point cloud, then, a mapping algorithm matches and registers the point ...
  80. [80]
    [PDF] Exploration-RRT: A multi-objective Path Planning and ... - DiVA portal
    At every call to the algorithm the grid map G is initialized based on point clouds of occupied and unknown voxels, and the user is free to change resolution ...
  81. [81]
    [PDF] SIMD-Amenable Nearest Neighbors for Fast Collision Checking
    Jun 4, 2024 · In this paper, we present a novel data structure for motion planning with observed point clouds, demonstrating an order-of-magnitude speedup ...<|separator|>
  82. [82]
    Mastering data complexity for autonomous driving with adaptive ...
    By firing high-rate laser beams, a LiDAR device is able to project its surroundings as 2D or 3D point cloud, which can be used for different purposes such as ...
  83. [83]
    Flying on Point Clouds with Reinforcement Learning - arXiv
    Mar 1, 2025 · In this paper, we combine onboard 3D lidar sensing and sim-to-real reinforcement learning (RL) to enable autonomous flight in cluttered environments.
  84. [84]
    The Role of Edge AI and Tiny ML in Modern Robots | IoT For All
    Sep 9, 2025 · Edge AI and Tiny ML in robotics revolutionize automation with real-time decision-making, enhanced privacy, and efficiency across industries.
  85. [85]
    [PDF] Petra documentation project using Laser Scanning
    The point cloud includes scans of the rock walls of the Siq, which is the narrow access canyon to the site, the outside and inside of some 30 major structures ...
  86. [86]
    (PDF) The Scanning of the Pyramids Project - ResearchGate
    The aim of the project is to apply and test latest state-of-the-art terrestrial laser scanners produced by the Austrian company RIEGL LMS for high accuracy.
  87. [87]
    [PDF] 3D DATA OF PYRAMIDS AND THEIR INTERPRETATIONS
    The Giza Laser Scanning Survey (GLSS) created 3D models of monuments, including Queen Khentkawes' tomb, to understand its structure and the landforms.
  88. [88]
    Detection of damage in heritage constructions based on 3D point ...
    Oct 15, 2023 · A systematic review of 188 publications focused on damage detection in historical construction by means of 3D point clouds.
  89. [89]
    Survey of the Erosion Degree of the Rammed Earth Ming Great Wall ...
    Therefore, high-precision digital 3D point-cloud models of cultural heritage ... erosion images and 3D point-cloud data for complete erosion damage monitoring.2. Methods And Materials · 3. Results · 4. Discussion
  90. [90]
    [PDF] Self-Supervised Large Scale Point Cloud Completion for ...
    This paper presents a self-supervised method for restoring large, incomplete point clouds, using a MCOP image and inpainting, for archaeological sites with ...
  91. [91]
    Peek Inside the Parthenon in 3D - Google Arts & Culture
    Beautifully decorated with sculptures, columns, architraves, friezes and pediments, the Parthenon is considered one of the finest examples of Greek architecture ...
  92. [92]
    CPDC-MFNet: conditional point diffusion completion network with ...
    Apr 9, 2024 · We propose a probabilistic model based on diffusion for Terracotta Warriors point cloud completion. The model can infer the conditional ...Missing: Parthenon remote<|separator|>
  93. [93]
    Multimodal Scanning of Cultural Heritage Assets for their ... - CORDIS
    Aug 11, 2022 · Scan4Reco will develop a novel portable, integrated and modular solution for customized and thus cost-effective, automatic digitization and analysis of ...Missing: point clouds
  94. [94]
    [PDF] Scan4Reco: Towards the digitized conservation of Cultural Heritage ...
    To bridge this gap the Scan4Reco project aims to automatically produce digital replicas of cultural heritage assets that will be displayed through a virtual ...
  95. [95]
    Data Ethics in Cultural Heritage: FAIR and CARE Principles
    Dec 31, 2024 · The CARE Principles for Indigenous Data Governance outline key considerations for the ethical and responsible use of Indigenous data. These ...Missing: cloud | Show results with:cloud
  96. [96]
    Digital Protection and Inheritance of Cultural Heritage based on ...
    A novel system is proposed that utilizes Ethereum blockchain technology for securing and managing cultural heritage assets.
  97. [97]
    Blockchain in digital cultural heritage resources - Nature
    Jun 5, 2025 · This paper investigates the current challenges in managing DCHR and analyzes the corresponding solutions. We discuss outstanding technical implementations.
  98. [98]
    A Survey on Data Compression Techniques for Automotive LiDAR ...
    This article presents a survey of existing methods used to compress point cloud data for automotive LiDAR sensors.Missing: seminal papers
  99. [99]
    [PDF] Octree-based Point-Cloud Compression - Eurographics Association
    To compress the octree we employ novel prediction techniques that were specifically designed for point sampled geometry and are based on local surface ...
  100. [100]
  101. [101]
    [PDF] POINT CLOUD ATTRIBUTE COMPRESSION WITH GRAPH ...
    Our goal is to compress the attributes on the point cloud, such as colors and normals. Without loss of generality, we use color attributes as an example in this ...
  102. [102]
    A efficient predictive wavelet transform for LiDAR point cloud ...
    In this paper, a new predictive wavelet transform (PWT) is proposed to solve LiDAR point clouds attribute compression.
  103. [103]
    [PDF] Deep Dynamic Point Cloud Compression via 3D Motion Prediction
    This paper proposes a novel 3D sparse convolution- based Deep Dynamic Point Cloud Compression. (D-DPCC) network to compensate and compress the DPC geometry with ...
  104. [104]
    Scalable Point Cloud Attribute Compression - ACM Digital Library
    Nov 9, 2023 · This article develops a Scalable Point Cloud Attribute Compression solution, termed ScalablePCAC . In a two-layer example, ScalablePCAC uses ...
  105. [105]
    Evaluating the effect of sparse convolutions on point cloud ...
    Results show that the use of sparse convolutions allows for an average BD-rate reduction of approximately 9% for both D1 and D2 PSNR metrics based on similar ...
  106. [106]
    Visual Volumetric Video-based Coding (V3C) and Video ... - MPEG.org
    MPEG-I: Visual Volumetric Video-based Coding (V3C) and Video-based Point Cloud Compression (V-PCC) ... Text of ISO/IEC DIS 23090-5 Video-based Point Cloud ...
  107. [107]
    MPEG-I — Point Cloud Compression (V-PCC, G-PCC & branches)
    The G-PCC Test Model (TMC13) implements the Geometry-based Point Cloud Compression algorithms standardized in ISO/IEC 23090-9. It includes geometry and ...
  108. [108]
    ISO/IEC 23090-5:2021 - Information technology
    This document specifies the syntax, semantics and decoding for visual volumetric media using video‐based coding methods.
  109. [109]
    An Introduction to Point Cloud Compression Standards | GetMobile
    May 22, 2023 · In this article, we provide an overview of some topical and popular MPEG point cloud compression (PCC) standards.
  110. [110]
    Draco 3D Graphics Compression - Google
    Draco is an open-source library for compressing and decompressing 3D geometric meshes and point clouds. It is intended to improve the storage and transmission ...
  111. [111]
    potree/potree: WebGL point cloud viewer for large datasets - GitHub
    Potree is a free open-source WebGL based point cloud renderer for large point clouds. It is based on the TU Wien Scanopy project and research projects ...Potree · potree/PotreeConverter · Releases · potree/PotreeDesktop · Releases 14
  112. [112]
    Khronos Announces glTF Geometry Compression Extension Using ...
    Feb 15, 2018 · In sample glTF models, up to 12X compression has been demonstrated with no change in visual fidelity. Smaller glTF files will enable an ...
  113. [113]
  114. [114]
    MPEGGroup/mpeg-pcc-tmc2: Video codec based point ... - GitHub
    Video codec based point cloud compression (V-PCC) Test Model (encoder and decoder) - MPEGGroup/mpeg-pcc-tmc2.Missing: interoperability vendor lock-