Research Article
Open access
Published on 8 November 2024
Download pdf
Wu,D. (2024). Object detection and tracking for drones: A system design using dynamic visual SLAM. Applied and Computational Engineering,81,71-82.
Export citation

Object detection and tracking for drones: A system design using dynamic visual SLAM

Dongli Wu *,1,
  • 1 College of Design and Engineering, National University of Singapore, Singapore, Singapore

* Author to whom correspondence should be addressed.

https://doi.org/10.54254/2755-2721/81/20241013

Abstract

Drones can play a quite crucial role in many walks of life today. Enhancing the visual perception ability of drones is crucial to their intelligence level. Among them, it is necessary to focus on strengthening the detection, tracking and mapping capabilities of drones for dynamic objects. However, the existing visual SLAM systems carried by drones do not perform well in dynamic environments. This project designs a monocular visual SLAM system specifically for drones, aiming to achieve efficient three-dimensional mapping and target tracking, surpassing the limitations of simple static mapping and positioning. Besides, this project constructs a drone dynamic SLAM system developed on the ORB-SLAM3 structure, uses drone images to detect, track and map object motion models, and reconstructs environmental maps to obtain motion parameters with real physical scales. This project strives to optimize the input pre-processing module, improve the validity of data and output environmental maps and raster maps. The outcomes demonstrate the system's strong accuracy and adaptability in dynamic installation procedures.

Keywords

Visual simultaneous localization and mapping, Drones, Dynamic objects tracking.

[1]. Xi, Y., Liu, J., Wang, Z., & Chen, C. (2022). Learning-empowered resource allocation for air slicing in drone-assisted cellular V2X communications. IEEE Systems Journal, 17, 1008–1011.

[2]. Bescós, B., Fácil, J. M., Civera, J., & Neira, J. (2018). DynaSLAM: Tracking, mapping, and inpainting in dynamic scenes. IEEE Robotics and Automation Letters, 3, 4076–4083.

[3]. Brata, K., Funabiki, N., Panduman, N., & Fajrianti, E. (2024). An enhancement of outdoor location-based augmented reality anchor precision through VSLAM and Google Street View. Sensors, 24(4), 1161.

[4]. Barros, A. M., Michel, M., Moline, Y., & Corre, G. (2022). A comprehensive survey of visual SLAM algorithms. Robotics, 11(1), 24.

[5]. Zhang, B., Dong, Y., Zhao, Y., & Qi, X. (2024). DynPL-SLAM: A robust stereo visual SLAM system for dynamic scenes using points and lines. IEEE Transactions on Intelligent Vehicles.

[6]. Chen, W., Shang, G., Ji, A., et al. (2022) An overview on visual slam: From tradition to semantic. Remote Sensing, 14(13): 3010.

[7]. Shan, M., Wang, F., Lin, F., Gao, Z., Tang, Y. Z., & Chen, B. M. (2015). Google map aided visual navigation for drones in GPS-denied environment. In Proceedings of the 2015 IEEE International Conference on Robotics and Biomimetics, 114–119.

[8]. Forster, C., Pizzoli, M., & Scaramuzza, D. (2014). SVO: Fast semi-direct monocular visual odometry. In Proceedings of the 2014 IEEE International Conference on Robotics and Automation, 15–22.

[9]. Kim, Y. (2021). Aerial map-based navigation using semantic segmentation and pattern matching. arXiv, arXiv:2107.00689.

[10]. Campos, C., Elvira, R., Gómez Rodríguez, J. J., Montiel, J. M. M., & Tardós, J. D. (2021). ORB-SLAM3: An accurate open-source library for visual, visual-inertial, and multimap SLAM. IEEE Transactions on Robotics, 37, 1874–1890.

[11]. de Jesus, K. J., Pereira, M. O. K., Emmendorfer, L. R., & Gamarra, D. F. T. (2023). A comparison of visual SLAM algorithms ORB-SLAM3 and DynaSLAM on KITTI and TUM monocular datasets. In Proceedings of the 2023 36th SIBGRAPI Conference on Graphics, Patterns and Images, 109–114.

[12]. Endres, F., Hess, J., Sturm, J., Cremers, D., & Burgard, W. (2014). 3-D mapping with an RGB-D camera. IEEE Transactions on Robotics, 30, 177–187.

[13]. [13] Wang, G., Chen, Y., An, P., Hong, H., Hu, J., & Huang, T. (2023). drone-Yolov8: A small-object-detection model based on improved Yolov8 for drone aerial photography scenarios. Sensors, 23(16), 7190.

[14]. Dong, Q., Cao, C., & Fu, Y. (2023). Rethinking optical flow from geometric matching consistent perspective. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 1337–1347.

[15]. Zhang, J., Henein, M., Mahony, R., & Ila, V. (2020). VDO-SLAM: A visual dynamic object-aware SLAM system. arXiv, arXiv:2005.11052.

Cite this article

Wu,D. (2024). Object detection and tracking for drones: A system design using dynamic visual SLAM. Applied and Computational Engineering,81,71-82.

Data availability

The datasets used and/or analyzed during the current study will be available from the authors upon reasonable request.

Disclaimer/Publisher's Note

The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of EWA Publishing and/or the editor(s). EWA Publishing and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

About volume

Volume title: Proceedings of the 2nd International Conference on Machine Learning and Automation

Conference website: https://2024.confmla.org/
ISBN:978-1-83558-563-4(Print) / 978-1-83558-564-1(Online)
Conference date: 21 November 2024
Editor:Mustafa ISTANBULLU, Xinqing Xiao
Series: Applied and Computational Engineering
Volume number: Vol.81
ISSN:2755-2721(Print) / 2755-273X(Online)

© 2024 by the author(s). Licensee EWA Publishing, Oxford, UK. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license. Authors who publish this series agree to the following terms:
1. Authors retain copyright and grant the series right of first publication with the work simultaneously licensed under a Creative Commons Attribution License that allows others to share the work with an acknowledgment of the work's authorship and initial publication in this series.
2. Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the series's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgment of its initial publication in this series.
3. Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See Open access policy for details).