Vision-Language-Action Models for Autonomous Driving: Past, Present, and Future Paper • 2512.16760 • Published 9 days ago • 12
Multi-Modal Data-Efficient 3D Scene Understanding for Autonomous Driving Paper • 2405.05258 • Published May 8, 2024
Beyond One Shot, Beyond One Perspective: Cross-View and Long-Horizon Distillation for Better LiDAR Representations Paper • 2507.05260 • Published Jul 7
An Empirical Study of Training State-of-the-Art LiDAR Segmentation Models Paper • 2405.14870 • Published May 23, 2024
SuperFlow++: Enhanced Spatiotemporal Consistency for Cross-Modal Data Pretraining Paper • 2503.19912 • Published Mar 25
Point Transformer V3 Extreme: 1st Place Solution for 2024 Waymo Open Dataset Challenge in Semantic Segmentation Paper • 2407.15282 • Published Jul 21, 2024
SEE4D: Pose-Free 4D Generation via Auto-Regressive Video Inpainting Paper • 2510.26796 • Published Oct 30
RewardMap: Tackling Sparse Rewards in Fine-grained Visual Reasoning via Multi-Stage Reinforcement Learning Paper • 2510.02240 • Published Oct 2 • 17
Open-o3 Video: Grounded Video Reasoning with Explicit Spatio-Temporal Evidence Paper • 2510.20579 • Published Oct 23 • 55
VideoLucy: Deep Memory Backtracking for Long Video Understanding Paper • 2510.12422 • Published Oct 14 • 1
MERIT: Multilingual Semantic Retrieval with Interleaved Multi-Condition Query Paper • 2506.03144 • Published Jun 3 • 7
Talk2Event: Grounded Understanding of Dynamic Scenes from Event Cameras Paper • 2507.17664 • Published Jul 23 • 1
Can MLLMs Guide Me Home? A Benchmark Study on Fine-Grained Visual Reasoning from Transit Maps Paper • 2505.18675 • Published May 24 • 26
FRNet: Frustum-Range Networks for Scalable LiDAR Segmentation Paper • 2312.04484 • Published Dec 7, 2023 • 1