Download as pdf or txt
Download as pdf or txt
You are on page 1of 4

Intelligent System Concept of an IoT Cameras

Network Application for an Unmanned Aerial


Vehicle Control via a Graphical User Interface
2022 6th International Scientific Conference on Information, Control, and Communication Technologies (ICCT) | 978-1-6654-9367-3/22/$31.00 ©2022 IEEE | DOI: 10.1109/ICCT56057.2022.9976710

Alexander Dubelschikov Tatyana Tsoy Yang Bai


Tikhonov Moscow Institute of Laboratory of Intelligent Robotics College of Information Science and
Electronics and Mathematics, Systems (LIRS), Institute of Information Engineering, Ritsumeikan University
HSE University Technologies and Intelligent Systems, Biwako-Kusatsu Campus
Moscow, Russia Kazan Federal University Kusatsu, Shiga, Japan
aadubelschikov@miem.hse.ru Kazan, Russia yangbai@fc.ritsumei.ac.jp
tt@it.kfu.ru

Mikhail Svinin Evgeni Magid


College of Information Science and Laboratory of Intelligent Robotics
Engineering, Ritsumeikan University Systems (LIRS), Institute of Information
Biwako-Kusatsu Campus Technologies and Intelligent Systems,
Kusatsu, Shiga, Japan Kazan Federal University
svinin@fc.ritsumei.ac.jp Kazan, Russia
HSE University
Moscow, Russia
magid@it.kfu.ru

Abstract — an important goal of modern robotics is to operator to a control joystick. In [13] authors offered to
benefit from a growing availability of internet of things (IoT) employ control barrier functions to generate force
devices. A single IoT device and their networks could
feedback that helped a human operator to control a
significantly improve user awareness and ease a control of
teleoperated robots. This paper discusses a concept of employing quadrotor UAV in a safe manner. Moreover, haptic
stationary IoT cameras in order to improve the convenience of feedback performed well in situations, where visual
an unmanned aerial vehicle (UAV) control by displaying interface suffered from delays or errors, which occurred
additional visual information directly to a UAV control panel of due to technical imperfection and interference [14].
an operator. The concept employs robot operating system (ROS) Another approach to ease a control considers special
and PX4-LIRS UAV model. The system will be initially
sub-systems, which simplify control and automatically
modelled in Gazebo simulator and then transferred onto real
hardware devices. dodge obstacles. A visual odometry solution based on
RGB-D cameras for a short-term position estimation and
Keywords — IoT, PX4 LIRS_UAV, ESP32 CAM, OpenCV a local planner that generated trajectories and provided
obstacles’ avoidance was proposed in [15]. Simultaneous
I. INTRODUCTION localization and mapping (SLAM) algorithms could also
Unmanned aerial vehicles (UAV) based applications, be employed in order to improve teleoperation safety for a
including quadcopters, are one of the most promising UAV, e.g., FastSLAM [16].
directions in robotics today [1]. Used for civil needs, for However, one of the most popular approaches
example, in smart agricultural farming [2], warehouse concentrates on visual interfaces. In [17] authors proposed
management [3], traffic analysis [4], multi-agent coverage a visual servoing solution for a tethered UAV
of large-scale outdoor environments [5] or search and teleoperation that uses a grid map in order to solve a
rescue operations [6], [7], as well as for military purposes problem of a high video stream load, which prevents video
[8], [9], UAVs are the subject of many studies, including stutters and reduces a frame rate. In [18] authors discussed
those aimed at improving various control criteria. Ease of two novel visualizations, which were designed to obtain
control is one of the most important criteria for a UAV higher user acceptance ratings for a haptic collision
operator. According to statistics, from 21% to 67% of avoidance system by visualizing force fields. A concept of
UAV-related accidents occur due to a human factor [10]. using adaptive interfaces for a simultaneous control of
Timely information, accurate, intuitive, responsive several UAVs by a single human operator was presented
control, and additional sources of information increase in [19]. A solution for improving a control interface in [20]
operator awareness and a probability of complex tasks suggested to employ an additional accompanying
success, and reduce a number of errors that could lead to a quadcopter, whose camera images helped to control a
UAV crash. main UAV in a field of view of the additional quadcopter
Researchers offer various ways of solving UAV while creating a third-person control effect.
control convenience issues. In [11] authors proposed to This paper presents an approach of a UAV-based
control a UAV using gestures that are tracked by several environment exploration that uses a network of stationary
cameras. Tactile interface in [12] adjusted UAV’s speed Internet of Things (IoT) cameras located in an indoor
depending on a pressure force, which was applied by an environment. The exploration considers estimation of

XXX-X-XXXX-XXXX-X/XX/$XX.00 ©20XX IEEE

Authorized licensed use limited to: University of East London. Downloaded on February 26,2023 at 21:21:37 UTC from IEEE Xplore. Restrictions apply.
potentially dangerous for a human variables and situations • UAV control buttons for 3D positioning and
in the environment, which could not be fully tracked solely orientation: throttle up/down, yaw right/left, pitch up/down,
by the available indoor IoT cameras, e.g., within a typical roll left/right.
rescue operation [21]. In [22] authors estimated that a total • Return - a button that initializes an autonomous
number of IoT cameras, which are expected to be placed UAV return to its starting position. The autonomous return
in the world by 2030, could reach 13 billion. These could be triggered by an operator manually or starts
cameras could be employed in our proposed approach in a automatically if a connection between the UAV and the
similar way to IoT camera based mobile robot indoor operator is lost [26].
localization [23]. The proposed approach targets to • A sliding controller and a corresponding input
improve situation awareness for an operator and ease the window (which duplicate each other and could be selected by
UAV control while minimizing additional onboard the operator depending on his/her preferences) that allow to
equipment load of the UAV. set a UAV velocity.
• Four screens for displaying video data streams from
II. SYSTEM DESIGN a UAV onboard camera (top central screen) and three
stationary IoT cameras (or virtual stationary cameras in a
The developed by Laboratory of Intelligent Robotic simulation). The stationary cameras are labeled as Neighbor
Systems (LIRS) of Kazan Federal University UAV PX4- 1 (top left screen), Neighbor 2 (top right screen) and Indoor
LIRS [24] was chosen as a target UAV for the proposed (bottom central screen). While Indoor camera streams data
approach implementation (Fig. 1, top). At the first stage of
from a fixed IoT camera in the current location (a room, a
this work, a virtual model of the UAV (Fig. 1, bottom) in the
corridor, a hall or any other environment) of the UAV,
Gazebo simulator is used. The UAV is controlled with a set Neighbor cameras stream data from two other locations,
of control commands that are used for the real UAV [25]. The which are manually selected by the operator in Map Window
cameras that are placed in a virtual environment could tab. Neighbor IoT camera allows the operator to see the UAV
directly transfer data to a graphical user interface (GUI),
from a third person perspective.
which allows an operator to control the UAV. The GUI
• Special indicators for a battery charge level, a wi-fi
allows a manual switch between cameras in order to employ
connection and a compass.
the most suitable one for a current operational task support.
• Takeoff and land buttons
The GUI corresponds to the PX4-LIRS UAV hardware
with a single onboard camera, which is directed downwards.
The first version of the system was implemented using real
IoT cameras, which are described in details in Section 4,
virtual stationary cameras in Gazebo and the Gazebo model
of PX4-LIRS UAV. Therefore the onboard camera streams
data from the Gazebo simulation, which is connected with
GUI using a standard rQt plugin. The real IoT cameras are
connected with GUI using a custom plugin, which employs
OpenCV library and connects to a local URL to get a mjpg
stream from the camera. The GUI control buttons are linked
to the topics of the UAV controller.

Fig. 1. PX4-LIRS UAV (top) and its virtual model for the Gazebo simulator Fig. 2. Control Window.
(bottom).
Map Window (Fig.3) contains:
• Control cameras’ screens (top and bottom, on the
III. GRAPHICAL USER INTERFACE right) duplicate the corresponding screen of Control Window
GUI for the proposed system was developed in Qt and to guarantee a continuous UAV tracking while the operator
rQt, which is its special version for ROS. It was divided into sets other variables within Map Window.
two tabs: Control Window and Map Window. • Map is a screen that displays a map of the entire
Control Window (Fig.2) contains: environment using RViz (top left). The RViz map allows
tracking a UAV position in the mission area and determining

Authorized licensed use limited to: University of East London. Downloaded on February 26,2023 at 21:21:37 UTC from IEEE Xplore. Restrictions apply.
adjacent rooms. The map could be available for the operator Arduino IDE was used for the camera programing. Using
in advance or constructed in real time using SLAM a modified standard sketch, we obtained a fairly clear image
algorithms [27]. The map is connected to the GUI with the from the camera, mainly working in a range 25-30 frames
RViz plugin. (Fig.3, screens Neighbor 1,2 and Indoor). Due to a high load
• Two context drop-down menus that allow a camera and a high heat during the operation, we noticed rare cases of
selection for Neighbor 1 and Neighbor 2 screens of Control frame rate drops and stutters. A possible solution to this
Window. Each camera is labeled on the environment map so problem could be the use of more expensive microcontrollers,
that the operator could decide, which cameras to select in the including advanced models of Raspberry Pi. However, a
menus. In Fig.3 a particular example is demonstrated; the significant increase in the cost of IoT equipment is not worth
cameras’ locations are displayed with red circles and are a slight improvement in productivity since scaling the
simply labeled as “Indoor”, “Neighbor 1” and “Neighbor 2”. proposed approach to large indoor environments would
The menus are integrated using a standard rQt camera image increase costs drastically and make it unfeasible.
output widget. Using a local IP address of the camera, we plan to transfer
• Indicators for a battery charge level, a wi-fi the image from the camera directly to the UAV control panel
connection and a compass are duplicated in this window due in the simulation using a special widget.
to their high importance in navigation tasks.
V. CONCLUSION
This paper proposed an approach of employing stationary
IoT cameras to improve an operator convenience of an
unmanned aerial vehicle (UAV) control by displaying
additional visual information directly to a UAV control panel
of an operator. The described system is based on PX4-LIRS
UAV and ESP 32-CAM environmental cameras, and
employs robot operating system (ROS). The system will be
initially modelled in Gazebo simulator and then transferred
onto real hardware devices. A potential bottlenecks, which
Fig. 3. Map Window. are related to data throughput channels are important
bottlenecks will be tested in real world experiments in various
conditions.
IV. ESP32-CAM CAMERA BASED HARDWARE SETUP
A specialized module ESP-32 CAM [28] camera (Fig. 4) ACKNOWLEDGMENT
is used as an indoor environmental camera of the system. This The second and fifth authors acknowledge the support of the
module is a combination of an ESP-32 microcontroller with Kazan Federal University Strategic Academic Leadership
a low power consumption, integrated Wi-fi and Bluetooth Program ("PRIORITY-2030").
controllers, and OV 2640 camera. This module was specially
created for solving issues with a streaming video REFERENCES
broadcasting. For the convenience of flashing this module, a
special UART converter ESP32-CAM-MB was used, which [1] N. Prosvirina, “Analysis and prospects for the development of
simplified the programming process since other UART unmanned aircraft,” Moscow Economic Journal, pp. 560-575, 2021.
converters require additional terminals in order to put the [2] P. Nintanavongsa and I. Pitimon, “Impact of sensor mobility on UAV-
ESP32-CAM into a programming mode. The outer case of based smart farm communications,” International Electrical
Engineering Congress (iEECON), pp. 1-4, 2017.
the camera was assembled using plastic containers and
[3] A. Khazetdinov, A. Aleksandrov, A. Zakiev, E. Magid and K.-H. Hsia,
powered by two 3.7V batteries of 2600 mAh capacity. “RFID-based warehouse management system prototyping using a
An image from the camera will be transferred via a web heterogeneous team of robots,” International Conference on Climbing
server to a corresponding UAV GUI window. Data and Walking Robots and Support Technologies for Mobile Machines
(CLAWAR), pp. 263-270, 2020.
throughput of camera-server and server-GUI channels are
important bottlenecks of the system and therefore it is critical [4] P. Peiro, C.Q. Gómez Muñoz and F. P. GarcíaMárquez, “Use of UAVs,
computer vision, and IoT for traffic analysis,” Internet of Things,
to follow the virtual world Gazebo simulation by a real world Springer, Cham, pp. 275-296, 2021.
experiments in various conditions. [5] Y. Bai, Y. Wang, M. Svinin, E. Magid and R. Sun, “Adaptive multi-
agent coverage control with obstacle avoidance,” IEEE Control
Systems Letters, Vol. 6, pp. 944-949, 2022.
[6] S. Charalampidou, E. Lygouras, I. Dokas, A. Gasteratos and A.
Zacharopoulou, “A Sociotechnical Approach to UAV Safety for
Search and Rescue Missions,” International Conference on Unmanned
Aircraft Systems (ICUAS), pp. 1416-1424, 2020.
[7] E. Magid et al., “Artificial intelligence based framework for robotic
search and rescue operations conducted jointly by international teams,”
International Conference on Electromechanics and Robotics
“Zavalishin's Readings” Springer, Singapore, pp. 15-26, 2020.
Fig. 4. ESP32-CAM-MB (left), ESP 32-CAM (central), OV2640 (right). [8] Z. Xiaoning, “Analysis of military application of UAV swarm
technology,” International Conference on Unmanned Systems (ICUS),
pp. 1200-1204, 2020.

Authorized licensed use limited to: University of East London. Downloaded on February 26,2023 at 21:21:37 UTC from IEEE Xplore. Restrictions apply.
[9] K. W. Williams, “A Summary of Unmanned Aircraft Accident- International Conference on Human-Robot Interaction (HRI), pp. 351-
Incident Data: Human Factors Implications, ” FAA The Defence 358, 2008.
Technical Information Center Tech. Rep., 2004. [20] X. Xiao, J. Dufek and R. Murphy, “Visual servoing for teleoperation
[10] K. Hartmann and K. Giles, “UAV exploitation: A new domain for using a tethered UAV,” IEEE International Symposium on Safety,
cyber power,” International Conference on Cyber Conflict (CyCon), Security and Rescue Robotics (SSRR), pp. 147-152, 2017.
pp. 205-221, 2016. [21] N. Simakov, R. Lavrenov, A. Zakiev, R. Safin and E.A. Martínez-
[11] J. Rodríguez-Gómez, R. Tapia, A. Eguíluz, J. Martínez-de Dios and A. García, “Modeling USAR maps for the collection of information on the
Ollero, “UAV human teleoperation using event-based and frame-based state of the environment,” International Conference on Developments
cameras,” Aerial Robotic Systems Physically Interacting with the in eSystems Engineering (DeSE), pp. 918-923, 2019.
Environment (AIRPHARO), pp. 1-5, 2021. [22] A. Mohan, K. Gauen, Y. Lu, W. W. Li and X. Chen, “Internet of video
[12] A. Kanso, I. H. Elhajj, E. Shammas and D. Asmar, “Enhanced things in 2030: A world with many cameras,” IEEE International
teleoperation of UAVs with haptic feedback,” IEEE International Symposium on Circuits and Systems (ISCAS) pp. 1-4, 2017.
Conference on Advanced Intelligent Mechatronics (AIM), pp. 305- [23] K. Kononov, R. Lavrenov, T. Tsoy, E. A. Martinez-Garcia and E.
310, 2015. Magid, “Virtual experiments on mobile robot localization with an
[13] D. Zhang, G. Yang and R. P. Khurshid, “Haptic Teleoperation of external smart RGB-D camera using ROS,” IEEE Eurasia Conference
UAVs Through Control Barrier Functions (IEEE Transactions on on IOT, Communication and Engineering (IEEE ECICE), pp. 656-659,
Haptics), ” IEEE Transactions on Haptics, pp. 109-115, 2020. 2021.
[14] T. M. Lam, V. D'Amelio, M. Mulder and M. M. van Paassen, “UAV [24] A. Khazetdinov, A. Zakiev, T. Tsoy, M. Svinin and E. Magid,
Tele-operation using Haptics with a Degraded Visual Interface, ” IEEE “Embedded ArUco: a novel approach for high precision UAV
International Conference on Systems, Man and Cybernetics pp. 2440- landing,” The 15th Siberian Conference on Control and
2445, 2006. Communications, 2021.
[15] F. J. Perez-Grau, R. Ragel, F. Caballero, A. Viguria and A. Ollero, [25] S. Mghabghab, I. H. Elhajj and D. Asmar, “Adaptive gain tuning for
“Semi-autonomous teleoperation of UAVs in search and rescue teleoperation of quadrotors,” 2016 IEEE International
scenarios,” International Conference on Unmanned Aircraft Systems Multidisciplinary Conference on Engineering Technology (IMCET),
(ICUAS), pp. 1066-1074, 2017. pp. 173-178, 2016.
[16] J. Mendes, R. Ventura, “Safe teleoperation of a quadrotor using [26] D. Bereznikov and A. Zakiev, “Network failure detection and
FastSLAM, IEEE International Symposium on Safety, ” Security, pp. autonomous return for PMB-2 mobile robot,” International Conference
1-6, 2012. on Artificial Life and Robotics, pp. 444-447, 2020.
[17] M. Riestock, F. Engelhardt, S. Zug and N. Hochgeschwender, “User [27] E. Mingachev et al., “Comparison of ROS-based monocular visual
study on remotely controlled UAVs with focus on interfaces and data SLAM methods: DSO, LDSO, ORB-SLAM2 & DynaSLAM,”
link quality,” IEEE/RSJ International Conference on Intelligent Robots International Conference on Interactive Collaborative Robotics
and Systems (IROS), pp. 3394-3400, 2017. Springer, Cham, pp. 222-233, 2020.
[18] V. Ho, C. Borst, M. Van Paassen and M. Mulder, “Increasing [28] A. Kaur, A. Jadli, A. Sadhu, S. Goyal and A. Mehra, “Cloud Based
Acceptance of Haptic Feedback in UAV Teleoperation by Visualizing Surveillance using ESP32 CAM,” International Conference on
Force Fields,” IEEE International Conference on Systems, Man, and Intelligent Technology, System and Service for Internet of Everything
Cybernetics (SMC) pp. 3027-3032, 2018. (ITSS-IoE) pp. 1-5, 2021.
[19] J. Cooper and M. A. Goodrich, “Towards combining UAV and sensor
operator roles in UAV-enabled visual search,” ACM/IEEE

Authorized licensed use limited to: University of East London. Downloaded on February 26,2023 at 21:21:37 UTC from IEEE Xplore. Restrictions apply.

You might also like