Loading...
Thumbnail Image
Publication

Multi-sensor fusion for efficient and robust UAV state estimation

Date
2025
Abstract
UAV State estimation is fundamental aspect across a wide range of applications, including robot navigation, autonomous driving, virtual reality, and augmented reality (AR). The proposed research emphasizes the vital role of robust state estimation in ensuring the safe navigation of autonomous UAVs. In this paper, we developed an optimization-based odometry state estimation framework that is compatible with multiple sensor setups. Our evaluation of the system is conducted using inhouse integrated UAV platform outfitted with multiple sensors including stereo cameras, an IMU, LiDAR sensors and GPS-RTK for ground truth comparison. The algorithm delivers robust and consistent UAV state estimation in various conditions including illumination changes, feature or structureless environment or even during degraded Global Positioning System (GPS) signals or total signal loss, where single sensor SLAM mostly fails. The experimental findings demonstrate that the proposed method is superior in compare to current state-of-the-art techniques.
Supervisor
Description
Publisher
Institute of Electrical and Electronics Engineers
Citation
12th International Conference on Control, Mechatronics and Automation (ICCMA), London, United Kingdom, 2024, pp. 35-40
Funding code
Funding Information
This work was supported in part by the European Commission’s Horizon 2020 Project RAPID under Grant 861211, and in part by the Enterprise Ireland’s Disruptive Technologies Innovation Fund (DTIF) Project GUARD under Grant DT2020 0286B (Corresponding author: Mahammad Irfan.)
Sustainable Development Goals
External Link
Type
Meetings and Proceedings
Rights
https://creativecommons.org/licenses/by-nc-sa/4.0/
License