ترغب بنشر مسار تعليمي؟ اضغط هنا

An Efficient L-Shape Fitting Method for Vehicle Pose Detection with 2D LiDAR

123   0   0.0 ( 0 )
 نشر من قبل Canbo Ye
 تاريخ النشر 2018
  مجال البحث الهندسة المعلوماتية
والبحث باللغة English




اسأل ChatGPT حول البحث

Detecting vehicles with strong robustness and high efficiency has become one of the key capabilities of fully autonomous driving cars. This topic has already been widely studied by GPU-accelerated deep learning approaches using image sensors and 3D LiDAR, however, few studies seek to address it with a horizontally mounted 2D laser scanner. 2D laser scanner is equipped on almost every autonomous vehicle for its superiorities in the field of view, lighting invariance, high accuracy and relatively low price. In this paper, we propose a highly efficient search-based L-Shape fitting algorithm for detecting positions and orientations of vehicles with a 2D laser scanner. Differing from the approach to formulating LShape fitting as a complex optimization problem, our method decomposes the L-Shape fitting into two steps: L-Shape vertexes searching and L-Shape corner localization. Our approach is computationally efficient due to its minimized complexity. In on-road experiments, our approach is capable of adapting to various circumstances with high efficiency and robustness.



قيم البحث

اقرأ أيضاً

LiDARs are usually more accurate than cameras in distance measuring. Hence, there is strong interest to apply LiDARs in autonomous driving. Different existing approaches process the rich 3D point clouds for object detection, tracking and recognition. These methods generally require two initial steps: (1) filter points on the ground plane and (2) cluster non-ground points into objects. This paper proposes a field-tested fast 3D point cloud segmentation method for these two steps. Our specially designed algorithms allow instantly process raw LiDAR data packets, which significantly reduce the processing delay. In our tests on Velodyne UltraPuck, a 32 layers spinning LiDAR, the processing delay of clustering all the $360^circ$ LiDAR measures is less than 1ms. Meanwhile, a coarse-to-fine scheme is applied to ensure the clustering quality. Our field experiments in public roads have shown that the proposed method significantly improves the speed of 3D point cloud clustering whilst maintains good accuracy.
In this paper, we present a novel end-to-end learning-based LiDAR relocalization framework, termed PointLoc, which infers 6-DoF poses directly using only a single point cloud as input, without requiring a pre-built map. Compared to RGB image-based re localization, LiDAR frames can provide rich and robust geometric information about a scene. However, LiDAR point clouds are unordered and unstructured making it difficult to apply traditional deep learning regression models for this task. We address this issue by proposing a novel PointNet-style architecture with self-attention to efficiently estimate 6-DoF poses from 360{deg} LiDAR input frames.Extensive experiments on recently released challenging Oxford Radar RobotCar dataset and real-world robot experiments demonstrate that the proposedmethod can achieve accurate relocalization performance.
376 - Xin Zheng , Jianke Zhu 2021
LiDAR odometry plays an important role in self-localization and mapping for autonomous navigation, which is usually treated as a scan registration problem. Although having achieved promising performance on KITTI odometry benchmark, the conventional s earching tree-based approach still has the difficulty in dealing with the large scale point cloud efficiently. The recent spherical range image-based method enjoys the merits of fast nearest neighbor search by spherical mapping. However, it is not very effective to deal with the ground points nearly parallel to LiDAR beams. To address these issues, we propose a novel efficient LiDAR odometry approach by taking advantage of both non-ground spherical range image and birds-eye-view map for ground points. Moreover, a range adaptive method is introduced to robustly estimate the local surface normal. Additionally, a very fast and memory-efficient model update scheme is proposed to fuse the points and their corresponding normals at different time-stamps. We have conducted extensive experiments on KITTI odometry benchmark, whose promising results demonstrate that our proposed approach is effective.
70 - Li Wang , Tao Zhang , Lin Ye 2021
This paper presents an efficient servomotor-aided calibration method for the triaxial gyroscope. The entire calibration process only requires approximately one minute, and does not require high-precision equipment. This method is based on the idea th at the measurement of the gyroscope should be equal to the rotation speed of the servomotor. A six-observation experimental design is proposed to minimize the maximum variance of the estimated scale factors and biases. In addition, a fast converging recursive linear least square estimation method is presented to reduce computational complexity. The simulation results reflect the robustness of the calibration method under normal and extreme conditions. We experimentally demonstrate the feasibility of the proposed method on a robot arm, and implement the method on a microcontroller. We verify the calibration results of the proposed method by comparing with a traditional turntable approach, and the experiment indicates that the results of these two methods are comparable. By comparing the calibrated low-cost gyroscope reading with the reading from a high-precision gyroscope, we can conclude that our method significantly increases the gyroscopes accuracy.
Globally localizing in a given map is a crucial ability for robots to perform a wide range of autonomous navigation tasks. This paper presents OneShot - a global localization algorithm that uses only a single 3D LiDAR scan at a time, while outperform ing approaches based on integrating a sequence of point clouds. Our approach, which does not require the robot to move, relies on learning-based descriptors of point cloud segments and computes the full 6 degree-of-freedom pose in a map. The segments are extracted from the current LiDAR scan and are matched against a database using the computed descriptors. Candidate matches are then verified with a geometric consistency test. We additionally present a strategy to further improve the performance of the segment descriptors by augmenting them with visual information provided by a camera. For this purpose, a custom-tailored neural network architecture is proposed. We demonstrate that our LiDAR-only approach outperforms a state-of-the-art baseline on a sequence of the KITTI dataset and also evaluate its performance on the challenging NCLT dataset. Finally, we show that fusing in visual information boosts segment retrieval rates by up to 26% compared to LiDAR-only description.
التعليقات
جاري جلب التعليقات جاري جلب التعليقات
سجل دخول لتتمكن من متابعة معايير البحث التي قمت باختيارها
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا