SimULi: Real-Time LiDAR and Camera Simulation with Unscented Transforms

ICLR 2026 Conference SubmissionAnonymous Authors
neural rendering3d gaussians3d vision
Abstract:

Rigorous testing of autonomous robots, such as self-driving vehicles, is essential to ensure their safety in real-world deployments. This requires building high-fidelity simulators to test scenarios beyond those that can be safely or exhaustively collected in the real-world. Existing neural rendering methods based on NeRF and 3DGS hold promise but suffer from low rendering speeds or can only render pinhole camera models, hindering their suitability to applications that commonly require high-distortion lenses and LiDAR data. Multi-sensor simulation poses additional challenges as existing methods handle cross-sensor inconsistencies by favoring the quality of one modality at the expense of others. To overcome these limitations, we propose SimULi, the first method capable of rendering arbitrary camera models and LiDAR data in real-time. Our method extends 3DGUT, which natively supports complex camera models, with LiDAR support, via an automated tiling strategy for arbitrary spinning LiDAR models and ray-based culling. To address cross-sensor inconsistencies, we design a factorized 3D Gaussian representation and anchoring strategy that reduces mean camera and depth error by up to 40% compared to existing methods. SimULi renders 10-20×\times faster than ray tracing approaches and 1.5-14×\times faster than prior rasterization-based work (and handles a wider range of camera models). When evaluated on two widely benchmarked autonomous driving datasets, SimULi matches or exceeds the fidelity of existing state-of-the-art methods across numerous camera and LiDAR metrics.

Disclaimer
This report is AI-GENERATED using Large Language Models and WisPaper (A scholar search engine). It analyzes academic papers' tasks and contributions against retrieved prior work. While this system identifies POTENTIAL overlaps and novel directions, ITS COVERAGE IS NOT EXHAUSTIVE AND JUDGMENTS ARE APPROXIMATE. These results are intended to assist human reviewers and SHOULD NOT be relied upon as a definitive verdict on novelty.
NOTE that some papers exist in multiple, slightly different versions (e.g., with different titles or URLs). The system may retrieve several versions of the same underlying work. The current automated pipeline does not reliably align or distinguish these cases, so human reviewers will need to disambiguate them manually.
If you have any questions, please contact: mingzhang23@m.fudan.edu.cn

Overview

Overall Novelty Assessment

The paper proposes SimULi, a real-time multi-sensor rendering method supporting arbitrary camera models and LiDAR data for autonomous driving simulation. It resides in the Real-Time Sensor Rendering leaf, which contains only three papers total, including the original work. This sparse population suggests the specific combination of real-time performance, multi-sensor support, and arbitrary camera models represents a relatively underexplored niche within the broader simulation landscape. The sibling papers (Unisim and Neural Point Rendering) share the real-time rendering goal but differ in sensor coverage or representation strategies.

The taxonomy reveals that Real-Time Sensor Rendering sits within the larger Simulation and Data Generation branch, which also includes Scene and Scenario Synthesis (four papers), Physical Sensor Modeling (three papers), and Hardware-in-the-Loop Simulation (one paper). These neighboring directions emphasize offline scenario generation or physical sensor characteristics rather than real-time multi-sensor rendering. The broader field shows substantial activity in Sensor Fusion Architectures (seventeen papers across multiple leaves) and End-to-End Driving Systems (four papers), indicating that while perception and control receive extensive attention, real-time simulation infrastructure remains comparatively less crowded.

Among sixteen candidates examined, the first contribution (real-time LiDAR and camera rendering with arbitrary sensor models) shows one refutable candidate out of ten examined, suggesting some prior work addresses overlapping goals within this limited search scope. The second contribution (factorized 3D Gaussian representation with anchoring strategy) examined five candidates with none clearly refuting it, indicating potential novelty in the representation design. The third contribution (automated tiling strategy for irregular LiDAR sampling patterns) examined only one candidate without refutation, though the small sample size limits confidence in assessing its novelty.

Based on the top-sixteen semantic matches examined, the work appears to occupy a sparsely populated research direction where real-time multi-sensor rendering remains an active challenge. The limited search scope means substantial prior work could exist beyond these candidates, particularly in computer graphics or robotics venues not captured here. The factorized representation and tiling strategy show fewer overlaps within the examined set, though broader literature searches would strengthen confidence in their novelty claims.

Taxonomy

Core-task Taxonomy Papers
50
3
Claimed Contributions
16
Contribution Candidate Papers Compared
1
Refutable Paper

Research Landscape Overview

Core task: Real-time multi-sensor simulation for autonomous driving. The field organizes around several complementary branches that together address the challenge of developing and validating autonomous systems. Sensor Fusion Architectures and Algorithms explores how to combine data from cameras, LiDAR, radar, and other modalities, with works like Sensor Fusion Survey[4] and Transformer Fusion Review[26] examining both classical and learning-based integration strategies. Data Collection and Calibration focuses on acquiring high-quality datasets and ensuring spatial-temporal alignment across sensors, as seen in Multisensor Calibration[1] and Multimodal Dataset Framework[5]. Simulation and Data Generation addresses the creation of synthetic training and testing environments, including real-time rendering approaches like Unisim[33] and Neural Point Rendering[36]. End-to-End Driving Systems investigates holistic perception-to-control pipelines such as TransFuser[2], while Sensor Technologies and System Integration examines hardware choices and deployment considerations covered in reviews like Sensors Overview[10] and Perception Technology Review[30]. Within the simulation landscape, a central tension exists between photorealism and computational efficiency for real-time applications. Many recent efforts pursue neural rendering and generative models to produce high-fidelity sensor data, exemplified by OmniGen[8] and Critical Data Generation[9], yet these can be computationally demanding. SimULi[0] sits within the Real-Time Sensor Rendering cluster, emphasizing efficient multi-sensor simulation that balances visual quality with the strict latency requirements of closed-loop testing. Compared to Unisim[33], which also targets unified sensor simulation, and Neural Point Rendering[36], which leverages neural representations for LiDAR rendering, SimULi[0] appears to prioritize real-time performance across multiple modalities simultaneously. This positioning reflects broader questions in the field about whether to invest in offline high-fidelity generation or develop faster methods suitable for hardware-in-the-loop validation, a trade-off that remains actively debated as autonomous driving systems mature.

Claimed Contributions

Real-time LiDAR and camera rendering with arbitrary sensor models

The authors extend 3DGUT to support both complex camera models (e.g., fisheye lenses) and LiDAR sensors simultaneously, enabling real-time rendering of multi-sensor data with time-dependent effects like rolling shutter. This is achieved through an automated tiling strategy for spinning LiDAR models and ray-based culling.

10 retrieved papers
Can Refute
Factorized 3D Gaussian representation with anchoring strategy

The method encodes camera and LiDAR information into separate sets of 3D Gaussians coupled via a nearest-neighbor anchoring loss. This factorization addresses cross-sensor inconsistencies more effectively than prior unified representations, improving both camera and LiDAR reconstruction quality without prioritizing one modality over the other.

5 retrieved papers
Automated tiling strategy for irregular LiDAR sampling patterns

The authors develop a histogram-equalization-based algorithm that automatically computes optimal tiling patterns for LiDAR sensors with non-uniform sampling, eliminating the need for manual heuristics. This is combined with ray-based culling to accelerate LiDAR rendering by 2-3× compared to prior methods.

1 retrieved paper

Core Task Comparisons

Comparisons with papers in the same taxonomy category

Contribution Analysis

Detailed comparisons for each claimed contribution

Contribution

Real-time LiDAR and camera rendering with arbitrary sensor models

The authors extend 3DGUT to support both complex camera models (e.g., fisheye lenses) and LiDAR sensors simultaneously, enabling real-time rendering of multi-sensor data with time-dependent effects like rolling shutter. This is achieved through an automated tiling strategy for spinning LiDAR models and ray-based culling.

Contribution

Factorized 3D Gaussian representation with anchoring strategy

The method encodes camera and LiDAR information into separate sets of 3D Gaussians coupled via a nearest-neighbor anchoring loss. This factorization addresses cross-sensor inconsistencies more effectively than prior unified representations, improving both camera and LiDAR reconstruction quality without prioritizing one modality over the other.

Contribution

Automated tiling strategy for irregular LiDAR sampling patterns

The authors develop a histogram-equalization-based algorithm that automatically computes optimal tiling patterns for LiDAR sensors with non-uniform sampling, eliminating the need for manual heuristics. This is combined with ray-based culling to accelerate LiDAR rendering by 2-3× compared to prior methods.