UAeroNet: domain-specific dataset for automation of unmanned aerial vehicles

Authors

DOI:

https://doi.org/10.20535/2786-8729.7.2025.341779

Keywords:

unmanned aerial vehicles, UAeroNet, object detection, autonomous navigation, computer vision

Abstract

This paper addresses the challenges and key principles of designing domain-specific datasets that can be used especially for automation of unmanned aerial vehicles. Such datasets play a key role in building intelligent systems that enable autonomous operation and support data-driven decisions. The study presents approaches we used for data collection, analysis and annotation, highlighting their importance and practical impact on real-world application. The preparation of a domain-specific dataset for automating unmanned aerial vehicles operations (such as navigation and environmental monitoring) is a challenging task due to frequently low image resolution, complex weather conditions, a wide range of object scales, background noise and heterogeneous terrain landscapes. Existing open datasets typically cover only a limited variety of unmanned aerial vehicles use cases, which restricts the ability of deep learning models to perform adequately under non-standard or unpredictable conditions.

The object of the study is video data acquired by unmanned aerial vehicles for creating domain-specific datasets that enable machine learning models to perform autonomous object recognition, navigation, obstacle avoidance and interaction with an environment with minimal operator involvement. The subject focuses on the collection, preparation and annotation of video data acquired by unmanned aerial vehicles. The purpose of the study is to develop and systematize workflow for creating specialized datasets to train robust models capable of autonomously recognizing objects in real-time video captured by unmanned aerial vehicles. To achieve this goal, a workflow was designed for collecting and annotating video data, raw video data were acquired from unmanned aerial vehicles sensors and manually annotated using the Computer Vision Annotation Tool. 

As a result of this work, we developed a domain-specific dataset (UAeroNet) using an open-source annotation tool for object tracking task in real scenarios. UAeroNet consists of 456 annotated tracks and a total of 131 525 labeled instances that belong to 13 distinct classes.

Author Biographies

Yuriy Kochura, National Technical University of Ukraine “Igor Sikorsky Kyiv Polytechnic Institute”

Assistant of the Computer Engineering Department of the Faculty of informatics and Computer Technique

Yevhenii Trochun, National Technical University of Ukraine “Igor Sikorsky Kyiv Polytechnic Institute”, Kyiv

Assistant of the Computer Engineering Department of the Faculty of informatics and Computer Technique

Vladyslav Taran, National Technical University of Ukraine “Igor Sikorsky Kyiv Polytechnic Institute”, Kyiv

Senior Lecturer of the Department of Computer Engineering, Faculty of Informatics and Computer Technology, Ph.D.

Yuri Gordienko, National Technical University of Ukraine “Igor Sikorsky Kyiv Polytechnic Institute”, Kyiv

Professor of the Computer Engineering Department of the Faculty of Informatics and Computer Technique, Doctor of Sciences in Physics and Mathematics, Senior Research Fellow

Oleksandr Rokovyi, National Technical University of Ukraine “Igor Sikorsky Kyiv Polytechnic Institute”, Kyiv

Associated Professor of the Department of Software Engineering of the Faculty of informatics and Computer Technique, Candidate of Technical Sciences, Associated Professor

Sergii Stirenko, National Technical University of Ukraine “Igor Sikorsky Kyiv Polytechnic Institute”, Kyiv

Vice-Rector for Research of Igor Sikorsky Kyiv Polytechnic Institute, Doctor of Technical Sciences, Professor

References

Y. Gordienko, N. Gordienko, R. Fedunyshyn, and S. Stirenko. Multibackbone ensembling of efficientnetv2 architecture for image classification. In International Conference on Trends in Sustainable Computing and Machine Intelligence, pages 55–69. Springer, 2024, http://dx.doi.org/10.1007/978-981-96-1452-3_5.

Z. Ouyang, J. Niu, T. Ren, Y. Li, J. Cui, and J. Wu, “Mbbnet: An edge iot computing-based traffic light detection solution for autonomous bus,” Journal of Systems Architecture, vol. 109, p. 101835, 2020, https://doi.org/10.1016/j.sysarc.2020.101835.

A. R. Khouas, M. R. Bouadjenek, H. Hacid, and S. Aryal, “Training machine learning models at the edge: A survey,” arXiv preprint arXiv:2403.02619, 2024.

Y. Matsubara and M. Levorato, “Neural compression and filtering for edge-assisted real-time object detection in challenged networks,” in 2020 25th International Conference on Pattern Recognition (ICPR). IEEE, 2021, pp. 2272–2279, https://doi.org/10.48448/830a-4b92.

A. Polukhin, Y. Gordienko, G. Jervan, and S. Stirenko, “Object detection for rescue operations by high-altitude infrared thermal imaging collected by unmanned aerial vehicles,” in Iberian Conference on Pattern Recognition and Image Analysis. Springer, 2023, pp. 490–504, http://dx.doi.org/10.1007/978-3-031-36616-1_39.

R. Murphy and T. Manzini, “Improving drone imagery for computer vision/machine learning in wilderness search and rescue,” in 2023 IEEE International Symposium on Safety, Security, and Rescue Robotics (SSRR). IEEE, 2023, pp. 159–164, http://dx.doi.org/10.48550/arXiv.2309.01904.

A. Polukhin, Y. Gordienko, M. Leier, G. Jervan, O. Rokovyi, O. Alienin, and S. Stirenko, “Edge intelligence resource consumption by uav-based ir object detection,” in Proceedings of the 2023 Workshop on UAVs in Multimedia: Capturing the World from a New Perspective, 2023, pp. 57–61, https://doi.org/10.1145/3607834.3616566.

N. Gordienko, Y. Gordienko, and S. Stirenko, “Enhancing deep learning sustainability by synchronized multi augmentation with rotations and multi-backbone architectures,” Big Data and Cognitive Computing, vol. 9, no. 5, p. 115, 2025, https://doi.org/10.3390/bdcc9050115.

K. A. Hambarde, N. Mbongo, P. K. MP, S. Mekewad, C. Fernandes, G. Silahtaroğlu, A. Nithya, P. Wasnik, M. Rashidunnabi, P. Samale et al., “Detreidx: A stress-test dataset for real-world uav-based person recognition,” arXiv preprint arXiv:2505.04793, 2025.

J. Xiao and G. Loianno, “Uasthn: Uncertainty-aware deep homography estimation for uav satellite-thermal geo-localization,” arXiv preprint arXiv:2502.01035, 2025.

D. Cohn, L. Atlas, and R. Ladner, “Improving generalization with active learning,” Machine learning, vol. 15, no. 2, pp. 201–221, 1994, https://doi.org/10.1007/BF00993277.

Image and V. U. Lab, “Uav123 & uav20l,” [Online]. Available: https://cemse.kaust.edu.sa/ivul/uav123. Accessed: Mar. 16, 2024.

“Visualization tools for uavdt dataset,” [Online]. Available: https://datasetninja.com/uavdt. Accessed: Dec. 08, 2024.

“Visualization tools for dac-sdc 2022 dataset,” [Online]. Available: https://datasetninja.com/dac-sdc-2022. Accessed: Dec. 08, 2024.

P. Zhu, L. Wen, D. Du, X. Bian, H. Fan, Q. Hu, and H. Ling, “Detection and tracking meet drones challenge,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 44, no. 11, pp. 7380–7399, 2021, http://dx.doi.org/10.1109/TPAMI.2021.3119563.

CVAT.ai Team. (2024) Computer vision annotation tool (cvat). [Online]. Available: https://github.com/cvat-ai/cvat?tab=readme-ov-file#license. Accessed: Dec. 04, 2024.

Downloads

Published

2025-12-27

How to Cite

[1]
Y. Kochura, Y. Trochun, V. . Taran, Y. Gordienko, O. Rokovyi, and S. . Stirenko, “UAeroNet: domain-specific dataset for automation of unmanned aerial vehicles”, Inf. Comput. and Intell. syst. j., no. 7, pp. 83–95, Dec. 2025.