Artificial intelligence and machine learning
Research Article
Gesture control of small unmanned aerial vehicle flight
Nikolai Sergeevich Abramov1, Vita Viktorovna Sattarova2, Vitaly Petrovich Fralenko3, Mikhail Vyacheslavovich Khachumov4
Abstract. The problem of constructing gesture commands for controlling a small unmanned aerial vehicle, such as a quadcopter, is considered. Commands coming from a video camera are identified by a classifier based on a convolutional neural network, and the multimodal control interface equipped with an intelligent solver converts them into control commands for the quadcopter. Neural networks from the Ultralytics neural network library allow selecting targets in a frame in real-time. The commands are sent to a specialized program on a smartphone, developed on the basis of DJI SDK flight simulators, which then sends commands via the remote control channel.
The quality of recognition of developed gesture commands for DJI Phantom 3 standard edition quadcopters is investigated, and a brief guide in the form of operator work scenarios with unmanned vehicles is provided. The prospects of gesture control of several vehicles in extreme conditions have been revealed, considering the complex safety challenges of joint flight and interaction of aircraft in confined space. (In Russian).
Keywords: unmanned aerial vehicle, control, gestures, convolutional neural network, Ultralytics, intelligent interface, recognition
MSC-2020 68T45; 68T07, 68T40Acknowledgments: This work was financially supported by the Russian Science Foundation, project № 21\–71\–10056, https://rscf.ru/project/21-71-10056/
For citation: Nikolai S. Abramov, Vita V. Sattarova, Vitaly P. Fralenko, Mikhail V. Khachumov. Gesture control of small unmanned aerial vehicle flight. Program Systems: Theory and Applications, 2024, 15:2, pp. 21–36. (In Russ.). https://psta.psiras.ru/2024/2_21-36.
Full text of article (PDF): https://psta.psiras.ru/read/psta2024_2_21-36.pdf.
The article was submitted 09.02.2024; approved after reviewing 13.03.2024; accepted for publication 17.03.2024; published online 22.04.2024.