Do you want to publish a course? Click here

Range Image-based LiDAR Localization for Autonomous Vehicles

169   0   0.0 ( 0 )
 Added by Xieyuanli Chen
 Publication date 2021
and research's language is English




Ask ChatGPT about the research

Robust and accurate, map-based localization is crucial for autonomous mobile systems. In this paper, we exploit range images generated from 3D LiDAR scans to address the problem of localizing mobile robots or autonomous cars in a map of a large-scale outdoor environment represented by a triangular mesh. We use the Poisson surface reconstruction to generate the mesh-based map representation. Based on the range images generated from the current LiDAR scan and the synthetic rendered views from the mesh-based map, we propose a new observation model and integrate it into a Monte Carlo localization framework, which achieves better localization performance and generalizes well to different environments. We test the proposed localization approach on multiple datasets collected in different environments with different LiDAR scanners. The experimental results show that our method can reliably and accurately localize a mobile system in different environments and operate online at the LiDAR sensor frame rate to track the vehicle pose.



rate research

Read More

Reliable and accurate localization is crucial for mobile autonomous systems. Pole-like objects, such as traffic signs, poles, lamps, etc., are ideal landmarks for localization in urban environments due to their local distinctiveness and long-term stability. In this paper, we present a novel, accurate, and fast pole extraction approach that runs online and has little computational demands such that this information can be used for a localization system. Our method performs all computations directly on range images generated from 3D LiDAR scans, which avoids processing 3D point cloud explicitly and enables fast pole extraction for each scan. We test the proposed pole extraction and localization approach on different datasets with different LiDAR scanners, weather conditions, routes, and seasonal changes. The experimental results show that our approach outperforms other state-of-the-art approaches, while running online without a GPU. Besides, we release our pole dataset to the public for evaluating the performance of pole extractor, as well as the implementation of our approach.
135 - Qin Shi , Xiaowei Cui , Sihao Zhao 2019
High-accuracy absolute localization for a team of vehicles is essential when accomplishing various kinds of tasks. As a promising approach, collaborative localization fuses the individual motion measurements and the inter-vehicle measurements to collaboratively estimate the states. In this paper, we focus on the range-only collaborative localization, which specifies the inter-vehicle measurements as inter-vehicle ranging measurements. We first investigate the observability properties of the system and derive that to achieve bounded localization errors, two vehicles are required to remain static like external infrastructures. Under the guide of the observability analysis, we then propose our range-only collaborative localization system which categorize the ground vehicles into two static vehicles and dynamic vehicles. The vehicles are connected utilizing a UWB network that is capable of both producing inter-vehicle ranging measurements and communication. Simulation results validate the observability analysis and demonstrate that collaborative localization is capable of achieving higher accuracy when utilizing the inter-vehicle measurements. Extensive experimental results are performed for a team of 3 and 5 vehicles. The real-world results illustrate that our proposed system enables accurate and real-time estimation of all vehicles absolute poses.
LiDARs are usually more accurate than cameras in distance measuring. Hence, there is strong interest to apply LiDARs in autonomous driving. Different existing approaches process the rich 3D point clouds for object detection, tracking and recognition. These methods generally require two initial steps: (1) filter points on the ground plane and (2) cluster non-ground points into objects. This paper proposes a field-tested fast 3D point cloud segmentation method for these two steps. Our specially designed algorithms allow instantly process raw LiDAR data packets, which significantly reduce the processing delay. In our tests on Velodyne UltraPuck, a 32 layers spinning LiDAR, the processing delay of clustering all the $360^circ$ LiDAR measures is less than 1ms. Meanwhile, a coarse-to-fine scheme is applied to ensure the clustering quality. Our field experiments in public roads have shown that the proposed method significantly improves the speed of 3D point cloud clustering whilst maintains good accuracy.
Recent years have witnessed an increasing interest in improving the perception performance of LiDARs on autonomous vehicles. While most of the existing works focus on developing novel model architectures to process point cloud data, we study the problem from an optimal sensing perspective. To this end, together with a fast evaluation function based on ray tracing within the perception region of a LiDAR configuration, we propose an easy-to-compute information-theoretic surrogate cost metric based on Probabilistic Occupancy Grids (POG) to optimize LiDAR placement for maximal sensing. We show a correlation between our surrogate function and common object detection performance metrics. We demonstrate the efficacy of our approach by verifying our results in a robust and reproducible data collection and extraction framework based on the CARLA simulator. Our results confirm that sensor placement is an important factor in 3D point cloud-based object detection and could lead to a variation of performance by 10% ~ 20% on the state-of-the-art perception algorithms. We believe that this is one of the first studies to use LiDAR placement to improve the performance of perception.
376 - Xin Zheng , Jianke Zhu 2021
LiDAR odometry plays an important role in self-localization and mapping for autonomous navigation, which is usually treated as a scan registration problem. Although having achieved promising performance on KITTI odometry benchmark, the conventional searching tree-based approach still has the difficulty in dealing with the large scale point cloud efficiently. The recent spherical range image-based method enjoys the merits of fast nearest neighbor search by spherical mapping. However, it is not very effective to deal with the ground points nearly parallel to LiDAR beams. To address these issues, we propose a novel efficient LiDAR odometry approach by taking advantage of both non-ground spherical range image and birds-eye-view map for ground points. Moreover, a range adaptive method is introduced to robustly estimate the local surface normal. Additionally, a very fast and memory-efficient model update scheme is proposed to fuse the points and their corresponding normals at different time-stamps. We have conducted extensive experiments on KITTI odometry benchmark, whose promising results demonstrate that our proposed approach is effective.
comments
Fetching comments Fetching comments
Sign in to be able to follow your search criteria
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا