Professional Documents
Culture Documents
JETIR2211571
JETIR2211571
JETIR2211571
org (ISSN-2349-5162)
Abstract - The Development of different navigation strategies to be developed. Hence, for the development and deployment
for Unmanned Aerial Vehicles has been under scientific of UAVs, a high-performance capability of autonomous
research for the last few decades and resulted in the navigation is of great significance.
construction of a variety of aerial platforms. Currently, the
main challenge that the scientific community is facing is the
design of fully autonomous unmanned aerial vehicles having A. UAV navigation
the competency of safely carrying out missions without any UAV navigation is a process of planning a path to navigate
human intervention. In order to improve the guidance and quickly and safely to the target location using only the
navigation skills of these aerial platforms, the control pipeline information about the current location and environment. In
of the UAVs is integrated with visual sensing techniques. The order to carry out the entire programmed mission successfully,
objective of this survey is to demonstrate an extensive review a UAV must be completely sensible of its position, pose,
of the literature on vision-aided techniques for autonomous
turning direction, forward speed, starting position, and location
UAV navigation. Particularly on vision-based localization
and UAV state estimation, collision avoidance, path planning, of the target. Until now, several methods for UAV navigation
and control. In addition to this, throughout this article, we will have been put forward by researchers and can be categorized
provide an insight into the challenges to be addressed, current into three types: satellite navigation, inertial navigation, and
developments, and trends in autonomous UAV navigation. vision-based navigation. Yet, these approaches alone are not
ideal for fully autonomous UAV navigation. Therefore, it is
Keywords - Unmanned aerial vehicles(UAVs), obstacle avoidance, difficult to go for the most relevant method for autonomous
path planning, visual SLAM. UAV navigation depending upon the particular task. Although,
after reviewing the proposed methods for UAV navigation
I. INTRODUCTION
under each category, vision-based navigation is a promising and
An Unmanned Aerial Vehicle (UAV) can be defined as an primary research area with the growing research and
aircraft that can navigate without the intervention of a human development in the field of computer vision. The visual sensors
pilot. Nowadays, the deployment of UAVs is increasing more that are used in vision-based autonomous navigation have
and more for regular applications, particularly for infrastructure several advantages over traditional sensors. First, the on-board
inspection, and surveillance due to its high flexibility and cameras can provide a good amount of online information about
mobility. Although, in several intricate environments, the UAV the surrounding environment; Second, they are perfect for the
is unable to sense the local environment exactly due to the perception of the rapidly changing environment because they
shortcomings of traditional sensors like poor perception ability possess a valuable anti-inference ability; Third, generally, most
and lack of stable communication with one another. In order to of the visual aided sensors are unassertive, that is, they don't
overcome these limitations a lot of effort has been made, allow the detection of sensing system. The United States and
however, a more effective and efficient method is still required Europe have already developed research institutions for
JETIR2211571 Journal of Emerging Technologies and Innovative Research (JETIR) www.jetir.org f556
© 2022 JETIR November 2022, Volume 9, Issue 11 www.jetir.org (ISSN-2349-5162)
navigation of aerial vehicles, such as Johnson Space Center of that can be steered, powered, and its shape remains preserved
NASA [32]. Massachusetts Institute of Technology [44], the by the pressure of gases within its envelope. These air vehicles
University of Pennsylvania [34] and several other top-ranked are suitable for long-duration flights as no energy is required
institutions are also rapidly developing research in the field of for lifting them, so that saved energy can be leveraged as a
vision-based autonomous UAV navigation and have embodied source of power for the movement of actuators, thus allowing
this automation technique into transport systems of next- flights of long-duration. Furthermore, these aircraft are capable
generation such as NextGen [80] and SESAR [35]. of navigating with safety at low levels, close to people and
The illustration of vision-based navigation is presented in Fig buildings of the local environment. Fig 2 represents the
1. discussed four classes of UAVs.
Through the usage of inputs from proprioceptive and
exteroceptive sensors, a UAV would be able to steer safely
towards the location of the target after completing the tasks of
state estimation, collision detection and avoidance, and motion
planning.
(a) (b)
(c) (d)
fusion of multi-sensor data [11], which integrates the Fig 3: Typical visual sensors. (a) monocular camera; (b) stereo
advantages of a variety of multiple sensors. However, restricted camera; (c) RGB-D camera; (d) fisheye camera
to particular environments, such as areas where GPS signal is
unavailable, it is assumed that the incorporation of multiple D. Previous Review Studies
Some of the researchers have carried out a survey of different
sensors in small UAVs would be unnecessary and impractical.
techniques that could be used for autonomous UAV navigation
Therefore, a more typical approach is needed for the and presented their valuable remarks and conclusions on those
enhancement of UAVs’ perception ability of the environment. methods.
After comparing visual sensors to ultrasonic sensors, laser Kanellakis and Nikolakopoulos (2017) [40] presented a survey
lightning, GPS, and other traditional sensors, we conclude that on vision-based methods that can be used for autonomous UAV
visual sensors capture much better information of surroundings navigation focusing mainly on current developments and
with texture, the color of surrounding objects, and other visual trends. They discussed three steps that are required for
details. Additionally, they can be deployed easily as well as autonomous flight: Navigation, Guidance, and Control.
they are cheaper, hence, this is the reason why vision-aided Navigation is further divided into the task of vision-based
autonomous UAV navigation is tending to be a relevant topic localization and mapping, obstacle avoidance, and aerial target
in current research. The visual sensors which are commonly tracking. Similarly, the guidance includes path planning,
used for vision-based navigation are divided into four types as mission planning, and exploration. And then in last, the control
shown by Fig 3: monocular, stereo, RGB-D, and fisheye. task constitutes estimation of Attitude, position, velocity, and
Monocular visual sensors are generally used in operations acceleration. Furthermore, the authors presented certain
where minimum weight and compactness are required. challenges that one have to face in implementing a vision-based
Furthermore, they are cheap and can be easily deployed into autonomous UAV navigation system, such as lack of solid
UAVs. However, they are not capable of obtaining a depth map experimental evaluation in the integration of visual sensors in
of the surrounding environment. When two monocular cameras the UAV ecosystem, In spite of the development of elaborated
of the same configuration are mounted on a rig then this system SLAM algorithms for applications of vision-based systems,
becomes a stereo camera. Therefore, a stereo camera is not only most of them cannot be used directly for UAV because of
capable of providing information that a single monocular certain limitations posed by their processing power and
camera can give but also offers some further advantages of dual architecture and the challenge imposed due to the incapability
views. Additionally, it can be used to obtain depth maps using of UAVs to just stop operating in the state of great uncertainty
the principle of parallax without the support of infrared sensors. like ground vehicles. Additionally, they presented their views
RGB-D cameras are capable of simultaneously generating on future trends in vision-based autonomous UAV navigation.
depth maps and visible images using infrared sensors. They are According to them in the near future UAVs could get turned
mostly used in indoor environments due to the restricted scope into the evolving elements in various applications as they
of infrared sensors. Fisheye visual sensors are adapted from possess some powerful characteristics like versatile movement,
monocular visual sensors with the additional capability of lightweight chassis, and onboard sensors. Also, they predicted
providing a broad viewing angle and additional capability of that the techniques for mapping the surroundings will be further
obstacle detection in complex environments. researched and revised for dynamic environments.
Furthermore, they talked about the ongoing research on
incorporating robotic arms/tools into UAVs so as to increase
their competencies in aerial maneuver for several tasks, such as
regular inspection, maintenance, and service. However, the
authors don't discuss the vision-based navigation techniques
that completely rely upon the front camera of UAV for self
localization, mapping, and control. Furthermore, in the section
visual localization and mapping the authors have mentioned
about the techniques that utilize both the monocular camera and
Inertial measurement unit (IMU) of the UAV. In these
(a) (b) techniques the data from IMU sensors and camera are fused
together and fed to an Extended Kalman filter (EKF) that would
further estimate the pose, attitude, and velocity of the vehicle
for flight control. The drawback in these techniques is that these
techniques use additional IMU sensors for autonomous UAV
navigation. The recent state-of-the-art works solve this issue by
using ORB SLAM for pose estimation of the UAV and creating
the map of the environment by leveraging only a monocular
(c) (d)
camera. Therefore, this review lacks the discussion about the
latest state-of-the-art techniques that utilize only the front
JETIR2211571 Journal of Emerging Technologies and Innovative Research (JETIR) www.jetir.org f558
© 2022 JETIR November 2022, Volume 9, Issue 11 www.jetir.org (ISSN-2349-5162)
camera of the UAV for the fully autonomous navigation in control law depends on the error signal ascertained from data of
GPS-denied areas. the surrounding environment obtained through visual sensors.
Lu et al. (2018)[95] present a comprehensive review of the After this, the authors discussed the next subtask, that is, vision-
vision-based UAV navigation methods. In this review the aided tracking or guidance. According to them, the control
process of UAV navigation is described in three steps: visual system of UAVs leverages relative navigation for achieving a
localization and mapping, obstacle avoidance, and path flight with respect to a target. Finally, vision-aided sense-and-
planning. In visual localization and mapping authors discussed avoid (SAA) is needed for a fully autonomous UAV system in
various methods that can estimate the position, attitude, and order to sense and avoid obstacles in both static and dynamic
velocity of the UAV. These methods may or may not use the environments. In the benchmark works reviewed by authors,
map of the environment. The methods that utilize the map of the researchers used one or more visual sensors for detecting
the environment are further divided into two types: map-based possible collisions with obstacles including other UAVs in the
and map building systems. Map-based systems require a priori environment, and for determining the necessary actions
map of the environment according to which they plan the required for achieving control over the flight which is free from
navigation and control strategy of the UAV. Map-building collisions. Then, in the end, the authors concluded that the best
systems leverage vision-based SLAM algorithms like ORB system that could be used for autonomous navigation is the
SLAM and PTAM that can build the map of the environment as monocular UAV system as it is easier to install and allow the
a UAV moves forth, and localize the UAV in the environment users to reduce the payload of UAV. Furthermore, the authors
by estimating its position coordinates. The second category of suggested that virtual reality would be an important aspect in
visual localization methods doesn’t require any map of the the development process of personal UAVs as it would help in
environment for localizing the UAV in the environment. These conducting experiments with UAVs in realistic indoor
methods include optical flow, and feature tracking techniques. environments containing different kinds of obstacles, as well as
The authors haven’t discussed the state-of-the-art deep in outdoor environments.
reinforcement learning methods under this category for
E. Motivation of this review
localizing and navigating UAV in the given environment. For
The objective of this survey is to present an outline of the most
obstacle detection and avoidance the review lacks the
important vision-based navigation methods that can be used for
discussion on the latest developments in this field. The
autonomous UAV navigation. Furthermore, a collection of
algorithms that are discussed by the authors are optical-flow
benchmark studies is provided that could act as a guideline for
based and SLAM-based. However, these algorithms take much
future research towards vision-based autonomous aerial
more processing time, hence not efficient for on-board
exploration. In [83] authors proposed a fact that with the rapidly
utilization. Instead, many recent AI-based obstacle avoidance
developing popularity of small-sized commercial UAVs and the
algorithms are proposed that can be executed during flight and
huge development of computer vision, the combination of both
take less computational time for sending response to the UAV.
of them, that is, vision-based UAV navigation has been a
Furthermore, the authors of the review have discussed only
working research area. Seeing that, the field of Computer vision
classical path planning algorithms like A-star, RRT, Ant-colony
for aerial vehicles is emerging as a popular trend in autonomous
optimization, etc. They don't present their review on the state-
navigation, hence the presented work will focus only on
of-the-art AI-based path planning algorithms that perform much
reviewing the fields of localization and mapping, obstacle
better than the traditional algorithm in terms of complexity, and
detection and avoidance, and path planning. The essence of this
computational time in unknown, dynamic, and large-scale
work is to provide rich insight into the entire task of
environments. Along with this, the authors don't mention about
autonomous UAV navigation collecting all pieces together.
the papers who have conducted real world experiments with the
UAV state estimation includes the extraction of distinct features
fully autonomous UAV by incorporating all the steps of
from the surrounding environment with the aid of visual
autonomous navigation.
sensors, so as to determine the next step of UAV in the
Later on, Belmonte et al. (2019) [6] carried out extensive
environment. We categorized this phase into three categories:
research on the vision-based autonomous navigation method.
optical flow and machine learning based systems, sensor based
He divides the UAV vision-based task into four subtasks:
systems, and Visual SLAM based systems. Where the Visual
navigation, control, tracking or guidance, and sense-and-avoid.
SLAM based methods are divided into feature based SLAM
According to their review, the navigation task includes figuring
methods, intensity based methods, hybrid methods, and Visual
out the position and orientation of an aircraft using visual
inertial SLAM based techniques. As the size of the inertial
sensing devices. Then, in the next step authors discussed the
measurement unit (IMU) is getting cheaper and smaller, hence
control of aircraft position on the basis of the data of
Shen et al. 2014 [75] and Leuteneggar et al. 2015 [45] proposed
surroundings captured by visual sensors. They further state that
and developed a navigation system in which they fused inertial
in the past few years several methods have been proposed for
measurement unit (IMU) and visual measurements together for
vision-based control of UAV. One of these methods is control
obtaining better performance results. In order to circumvent the
using visual servoing. When the visual data of the surrounding
limitations on power consumption and perception ability that
environment is directly incorporated into the control loop, this
make a single UAV incapable of completing certain types of
is known as visual servoing. Therefore, in this technique, the
tasks, In [29,57] it is shown that with the enhancement of
JETIR2211571 Journal of Emerging Technologies and Innovative Research (JETIR) www.jetir.org f559
© 2022 JETIR November 2022, Volume 9, Issue 11 www.jetir.org (ISSN-2349-5162)
JETIR2211571 Journal of Emerging Technologies and Innovative Research (JETIR) www.jetir.org f560
© 2022 JETIR November 2022, Volume 9, Issue 11 www.jetir.org (ISSN-2349-5162)
behavioral method for navigation that utilized a system capable Extended Kalman Filter (EKF) based object spatial localization
of recognizing visual landmarks combined with a fuzzy-based algorithm, the motion continuity of the UAV can be checked.
obstacle avoidance system was proposed by authors of [48]. In case, the coordinate is not correctly estimated, then it would
Hui et al. (2018)[37] proposed a vanishing point-based be given as an input to the Point-Refine-Net for precise
approach for localization and mapping of autonomous UAVs to prediction of the key points coordinates for the bounding box
be used for safe and robust inspection of power transmission of the target. Furthermore, in order to improve the state
lines. In this proposed method authors regarded the estimation accuracy of the UAV, some researchers have also
transmission tower as an important landmark by which proposed techniques which integrate deep neural networks, and
continuous surveillance and monitoring of the corridor can be reinforcement learning. Following this idea, Afifi and Gadallah
achieved. Similarly, deep reinforcement learning based (2021)[99] proposed a 3-D localization method for a UAV
methods have also been used for estimating the state of the using deep learning and reinforcement learning. In this
UAV and localizing it in the unknown environments without research, authors leverage a 5-G cellular network of 4 ground
the need of a prior map of the environment. Following this base stations, and in order to determine the location of UAV
approach, Huang et al. (2019)[116] proposed a Deep-Q network with minimum error the problem is reduced to optimization
based method for autonomous UAV navigation. In the proposed problem where objective is to minimize the error in the
approach the deep neural networks are integrated with measurement of RSSI (Radio received signal strength index)
reinforcement learning to overcome the limitations of readings of the 4 ground base stations. In order to estimate
reinforcement learning as it cannot be used to solve a complex location of the UAV in real-time a deep learning model is
problem alone. Similarly, Grando et al. (2020)[118] proposed a trained using the correlation between RSSI readings and UAV
mapless navigation strategy for UAVs. The proposed approach localization, further to improve the real-time performance of the
leveraged two deep reinforcement learning based algorithms, proposed model, the authors also introduced reinforcement
Deep deterministic policy gradient (DDPG) [96] and Soft actor learning to their work and compare the localization results with
critic (SAC) [97]. The information about the relative position those obtained from deep learning-based approach. After
of the vehicle from target and vehicle’s velocity are obtained carrying out the analysis of the results, the authors conclude that
through the reading of laser sensor mounted on UAV and for a small region with less dynamic obstacles, deep
localization data. This navigation approach is completely reinforcement learning gives better results but it is more
mapless and doesn’t need any heavy softwares like ORB SLAM computationally expensive, however for larger and crowded
and LSD SLAM to create the map of the environment. For both environments with dynamic obstacles, the deep learning
DDPG and SAC methods, the network consists of 24 inputs and approach performs better. Along with this, there are several
2 outputs. The 24 inputs include 20 laser readings, 2 readings mapless localization techniques which determine the location
of previous linear velocity and yaw angle, and 2 readings are of UAV using feature matching between the UAV-satellite
from the relative position of the UAV and angle to the target. image pairs. Goforth and Lucy (2019)[100] proposed the CNN
And, 2 outputs are values of the linear velocity of the UAV and based method to estimate the position of the UAV in an
the yaw angle which are required to control the UAV. The unknown environment. In this method a convolutional neural
authors of this work have verified the performance of their network extracts feature representations from UAV-satellite
technique by comparing their navigation strategy of UAV with image pairs given as an input to the network. By leveraging
geometry based tracking controller on the Gazebo simulator. In these extracted feature representations, the images of UAV are
the simulation environment without obstacles the geometry aligned with the ground-referenced satellite images and the
based tracking controller performs better whereas, in the location of the UAV is estimated. Similarly, Hou et al.
presence of obstacles the proposed performs better as in that (2020)[101] proposed a deep learning based technique for
case the UAV navigating with the tracking controller collides mapless UAV localization which integrates the digital
with the obstacles. Some researchers have also applied deep evaluation model (DEM) and the deep learning architecture D2-
neural networks to UAV localization problem. Li and Hu Net. The D2-Net architecture is responsible for the extraction
(2021)[98] proposed a mapless, deep learning based of keypoints and matching of satellite-UAV images. The idea
localization technique for the auto landing of unmanned aerial behind this is to combine the obtained geological map of the
vehicles at a specified target. The proposed approach is based area from DEM with the keypoint features to get the 3-D
on ground-vision based methods in which two stereo cameras position coordinates of the key points in the UAV image.
are installed or fixed on two independent Pan-tilt units (PTUs) Subsequently, Bianchi and Barfoot (2021)[8] demonstrated a
symmetrically to capture the images of the vehicle at the time robust and fast method for localization of a UAV using satellite
of landing. Furthermore, this work is based on deep learning, in images which were further utilized for training autoencoders. In
which two deep neural networks are leveraged, Bbox-locate- this work, the authors collected Google Earth (GE) images, and
Net and Point-Refine-Net. The images captured by the ground then an autoencoder model is trained to compact these images
visual system are fed as an input to the Bbox-Net which predicts in order to represent them as a low-dimensional vector without
the coordinates for the bounding box of the target. Then, after distorting important features. The trained autoencoder model is
combining these coordinates with the PTUs’ parameters, the further trained to reduce the size of a real-time image captured
spatial position of the UAV can be determined. On the basis of by UAV which is then compared with the pre-collected
JETIR2211571 Journal of Emerging Technologies and Innovative Research (JETIR) www.jetir.org f561
© 2022 JETIR November 2022, Volume 9, Issue 11 www.jetir.org (ISSN-2349-5162)
autoencoded GE images by using an inner-product kernel. semi-autonomous and fully autonomous disciplines, and this
Hence, localization of a UAV can be achieved by distributing technique for state estimation and building map of the
weights over the relative GE poses. The evaluation results of environment is getting popular rapidly along with the
this work have shown that the proposed approach is able to expeditious growth of visual simultaneous localization and
achieve root mean square error (RMSE) of less than 3m in real- mapping (visual SLAM) methods [4,79]. The size of the UAVs
time experiments. available in the market nowadays is shrinking, which restricts
their competency of lifting payload in some measure. Hence,
B. Sensor based systems researchers and developers have been more focused on the
These UAV systems leverage inertial unit, bioradar, lidar usage of manageable and lightweight visual sensors rather than
sensor, etc to explore the environment with divergence and the conventional heavy sonar and laser radar, etc. Stanford
ability of movement planning after defining the spatial CART Robot [59] carried out the first attempt at single camera-
configuration of the surroundings in a map. Therefore, with the based map-building techniques. Later on, the detection of 3D
aid of these sensors a static map of the environment can be coordinates of images was improved using a modified version
obtained prior to navigation. Typically, environmental maps of an interest operator algorithm. The result that was expected
can be classified into two types: occupancy grids and octree. from this system was the demonstration of 3D coordinates of
Each and every detail of the environment, varying from the 3D obstacles, which were placed on a mesh with each square cell
illustration of the environment to the interdependencies of length 2m. However, this technology is capable of
between environmental elements are included in these maps. In reconstructing the obstacles in the environment but still, it
[22] authors presented a method that used a 3D volumetric cannot model a large-scale world environment. Afterward, in
sensor that can enable an autonomous robot to efficiently order to recover states of cameras and structure of the
explore and map urban environments. In their work, they environment simultaneously, vision-aided SLAM techniques
constructed the 3D model of the environment using a multi- have made a good amount of progress and resulted in four kinds
resolution octree. Later on, for the depiction of the 3D model of of Visual SLAM techniques: feature-based, intensity-based,
the environment, an open-source framework was developed hybrid, and visual-inertial according to the method of image
[33]. Basically, the approach here is to render the 3D processing by visual sensors. Wang et al. (2020) [89] proposed
representation of the environment octree. Although not only and developed an efficient system for autonomous 3D
using the preoccupied area but also unknown and unoccupied exploration with UAV. The authors put forward a systematic
space. In addition to this, by using an octree map compression approach towards robust and efficient autonomous UAV
technique, the represented model is made to occupy less space navigation. For localization of UAV, the authors leveraged a
which makes the system capable of storing and updating the 3D map of the environment which was constructed incrementally
models efficiently. Authors of [27] collected and processed data and preserved during the process of inspection and sensing.
of the surroundings using a stereo camera, which can be This road map is responsible for providing the reward and cost-
leveraged further to produce a 3D map of the environment. The to-go for a candidate region that is to be investigated and these
key of this approach is the extended scan line grouping are the two measures of the next best view (NBV) based
technique that the authors used for precise segmentation of the evaluation. The authors verified their proposed framework in
range data into planar segments. This approach can also various 3D environments and the results obtained exhibit the
effectively confront noise present in the depth estimated by the typical attributes in NBV selection and much better
stereo vision-based algorithm. Dryanovski et al. (2010)[16] performance of their approach in terms of the efficiency of
proposed a method that represents a 3D environment by using aerial analysis as compared to other methods.
a multi-volume occupancy grid which is a cable of explicitly
storing information about both free space and obstacles. In 1) Feature-based SLAM
addition to this, by fusing and filtering in new positive or Feature-based SLAM systems first detect and extract
negative sensor data incrementally, the proposed method can attributes from multi-media data and then use these features as
correct previous sensor readings that were potentially inputs for localization and estimation of motion instead of using
erroneous. the images directly. Generally, the extracted features are
assumed to be uniform to viewpoint changes and rotation, as
C. Visual SLAM based system well as robust to noise and blur. In the past few years, thorough
Many times, it is difficult for a UAV to traverse in the air research on detection and description of features has been
with a previously known map of the surroundings obtained carried out and several types of attribute identifiers and
through UAV sensors due to certain environmental limitations. descriptors have been proposed [47,86]. Therefore, most of the
In addition to this, it would be unfeasible to acquire information recently proposed SLAM algorithms are supposed to operate
of the target location in extreme cases (such as in calamity-hit under this attribute-based framework. In [14] authors proposed
areas). Therefore, in such situations, it would be a more a monocular vision-based localization along with the mapping
efficient and attractive solution to go for simultaneous of a sparse set of natural attributes based on a top-down
localization and mapping at the time of navigation. SLAM Bayesian framework for achieving real-time performance. This
based robot navigation has been tremendously used in both work is a landmark for monocular vision-aided SLAM and has
JETIR2211571 Journal of Emerging Technologies and Innovative Research (JETIR) www.jetir.org f562
© 2022 JETIR November 2022, Volume 9, Issue 11 www.jetir.org (ISSN-2349-5162)
a considerable influence on subsequent research. Klein and better and robust than feature based methods. Authors of [61]
Murray (2007) [41] proposed an algorithm for parallel tracking proposed a monocular visual SLAM algorithm that can perform
and mapping. This is the first algorithm that divides the SLAM in real-time, DTAM, using intensity of the images can also
system into two independent parallel threads: tracking and predict the 6 DOF motion of a camera. At a frame rate derived
mapping which is the standard of current feature-based SLAM from the predicted comprehensive textured depth maps, the
systems. A state-of-the-art approach for large-scale navigation proposed algorithm is capable of generating dense surfaces. To
was proposed by the authors of [53]. Common issues in large- provide the approximation for semi-dense maps authors of [18]
scale environments, such as relocation of trajectories are being proposed and developed an efficacious stochastic intensity
corrected by visual loop-closure detections. Celik et al. (2009) based method, which can be utilized further for the calibration
[12] proposed a visual SLAM-based system for navigation in of images. Rather than optimizing parameters without using a
unknown indoor environments without the aids of GPS. UAVs scale, LSD SLAM [18] uses a different approach that exploits
have to predict the state and range using only a single onboard pose graph optimization, which allows loop closure detection
camera. Representing the environment using energy-based and scale drift correction in real-time by explicitly taking scale
straight architectural lines and feature points in the heart of the factors into account. Krul et al. (2021) [42] presented and
navigation strategy. In [30] researchers presented a vision- developed a visual SLAM-based approach for the localization
based attitude estimation method that leveraged multi-camera of a small UAV with a monocular camera in indoor
parallel tracking and mapping (PTAM) [41]. PTAM first environments for farming and livestock management. The
integrates the estimates of the 3D motion of multiple visual authors compared the performance of two visual SLAM
sensors within an environment and then correlates the mapping algorithms: LSD-SLAM and ORB-SLAM and found that ORB-
modules and state estimation. A state-of-the-art approach is also SLAM based localization suits best at those workplaces.
demonstrated by the authors to calibrate external parameters for Further, this algorithm was tested through several experiments
systems using multiple visual sensors with well-separated fields including waypoint navigation and generation of maps from the
of view. clustered areas in the greenhouse and a dairy farm.
Although, many of the feature-based SLAM methods are able
to reconstruct only a specific set of points as they pull out only 3) Hybrid method
sharp feature points from images. These types of methods are The hybrid method is a combination of both feature based
called sparse feature-based methods. So, researchers have been and intensity based SLAM methods. The first step in the hybrid
anticipating that more enhanced and dense maps of the method includes initialization of feature correspondences by
environment could be reconstructed by the dense feature-based using intensity based SLAM, then in the next step, the camera
methods. A dense energy-based method is exploited by poses are refined continuously using feature descriptor SLAM
Valgaerts et al. (2012) [88] for the estimation of an elementary algorithm. In [20] authors proposed an innovative semi-direct
matrix and for further calibration of similarities by using it. algorithm (SVO), for the estimation of the states of a UAV.
Ranftl et al. (2016) [66] used a segmented optical flow field for Similar to PTAM, this algorithm also implements motion
producing a rich depth map of the surroundings from two estimation and mapping of point clouds in two different threads.
successive frames, which means that adhering to this A more accurate pose estimation can be obtained using SVO by
framework through optimization of a convex program could combining gradient information and pixel brightness with the
result in a dense reconstruction of the scene. calibration of feature points and loss in reprojection error.
Subsequently, for real-time landing spot detection and 3D
2) Intensity based SLAM reconstruction of the surroundings, the authors of [21] proposed
SLAM techniques that leverage the features of captured and developed an algorithm that is computationally efficient
images for building a map of the environment seem to exhibit and capable of providing the best results in real-time
better performance in only ordinary and simple environments. performance. In order to carry out exploration of the real-world
However, they faced difficulties in texture-less environments. environment, a high frame rate visual sensor is required for the
Therefore, the intensity based SLAM came into play. Visual execution of a semi-direct algorithm, as it has limited resources
SLAM algorithms based on this approach optimize geometric to carry out the heavy computation.
parameters by exploiting all the information regarding intensity
present in the images of the surroundings, which can give 4) Visual Inertial SLAM
strength to geometric and photometric deformations present in Bonin-Font et al. (2008) [9] developed a system for the
images. Furthermore, this strategy can find dense resemblances navigation of ground mobile robots in which they utilized laser
so they are capable of reconstructing a dense map at an scanners for accessing 3D point clouds of relatively good
additional price of computation. In [76] researchers presented a quality. Also, nowadays UAVs can be equipped with these laser
state-of-the-art approach for the estimation of scene structure scanners as their size is getting smaller. Though different kinds
and pose of the camera. In this work the authors directly utilized of measurements from different types of sensors can be fused
intensities of the image as observations, thereby, leveraging all together and this can enable a more robust and accurate
data present in images. Hence, for the surroundings with few estimation of the UAV state. A typical SLAM system, extended
feature points, the proposed approach is proved to be much Kalman filter (MSF-EKF) can deal with multiple delayed
JETIR2211571 Journal of Emerging Technologies and Innovative Research (JETIR) www.jetir.org f563
© 2022 JETIR November 2022, Volume 9, Issue 11 www.jetir.org (ISSN-2349-5162)
that leveraged only the speed of light for measuring distance obstacles encountered in the navigation path, then that obtained
between the objects. This approach is simple but efficient as data is transferred to the regular convex bodies, which is further
many insects can detect the obstacles in the surrounding used for generating avoidance trajectories in the shape of
environment using light intensity. At the time of flight, the circular-arc. Then, on the basis of geometric correlation
motion of the image produces a light flow signal on the retina, between UAV and obstacle modelling, the real-time obstacle
this visual flow for the vision-aided navigation of insects avoidance algorithm is developed for UAV navigation. Hence,
provides a variety of information of attributes present in space. the authors of this work formulate the obstacle avoidance
Hence, insects can find out quickly whether they are able to pass problem as a trajectory tracking strategy. Following the idea of
by the obstacles safely or not, based on the intensity of light geometrical approach for collision detection and avoidance,
going through the aperture in the leaf. However, obstacle Aldao et al. (2022)[103] presented and developed a collision
avoidance approaches based on the optical-flow method are not avoidance algorithm for UAV navigation. The authors have
able to acquire precise distance due to which they are not used proposed this method mainly for the purpose of inspection and
in several particular missions. In contrast, SLAM-based monitoring in civil engineering applications. So, by keeping
techniques can come up with a precise estimation of metric this use in mind, the collision avoidance strategy is developed
maps using an advanced SLAM algorithm, thereby making for a UAV navigating between the waypoints following a
UAVs capable of navigating safely with a collision avoidance predefined trajectory. Furthermore, in order to avoid the
feature exploiting more environment information [23]. A novel collision with dynamic obstacles like people, other UAVs, etc,
technique of mapping a prior unknown environment using a the 3-D onboard sensors are also utilized to determine the
SLAM system was proposed in [60]. Furthermore, the authors position of these dynamic objects which are not considered in
leveraged a state-of-the-art artificial potential field technique to the original trajectory. The proposed technique is based on
avoid collision with both static and dynamic obstacles present sense and avoid obstacle avoidance in which first after reaching
in the surrounding environment. Later on, in order to cope up the waypoint of predefined trajectory UAV takes the sample
with the mediocre illumination of indoor environments and pictures and inertial measurements of the room to detect the
holding on the count of feature points, authors of [5] proposed presence of obstacles. If the obstacle is detected then in that case
a procedure for creating adjustable feature points in the map, the position of that obstacle is registered in the point cloud of
which is based on the PTAM (Parallel Tracking and Mapping) the inspection room at that moment of time otherwise the
algorithm. Subsequently, In [19] researchers proposed an scheduled path will be followed. If a new obstacle is detected
approach based on oriented fast and rotated brief SLAM (ORB- after completing the trajectory then its position will be
SLAM) for UAV navigation with obstacle detection and determined using on-board 3-D sensors. Table 2 encapsulates
avoidance aspect. The proposed method first processes data the techniques used for collision detection and avoidance.
from video by figuring out the locations of the aerial vehicle in
a 3D environment and then generates a sparse cloud map. After Table 2: Summary of important methods in collision
this, it enhances the density of the sparse map, and then finally, avoidance.
it generates an obstacle-free map for navigation using potential
fields and rapidly exploring random trees (RRT). Yathirajam et
Collision Methods used References
al. (2020) [90] proposed and developed a real-time system that avoidance
exploits ORB SLAM for building a 3D map of the environment techniques
and for continuously generating a path for UAV to navigate to
the goal position in the shortest time. The main contributions of Vision based Optical flow [77],[70],[28],[26
this approach are: implementation of chain-based path planning ],[1]
with built-in obstacle avoidance feature and generation of the
NMPC [49]
path for UAV with minimum overheads. Subsequently,
Lindqvist et al. (2020) [49] proposed and demonstrated a Potential field ORB-SLAM [19],[90]
Nonlinear Model Predictive Control (NMPC) for obstacle
avoidance and navigation of a UAV. In this work, the authors PTAM [5]
proposed a scheme to identify differences between different
Geometrical Trajectory [102]
kinds of trajectories to predict positions of future obstacles. The
algorithm tracking
authors conduct various laboratory experiments in order to
illustrate the efficacy of the proposed architecture and to prove Sense and avoid Trajectory [103]
that the presented technique delivers computationally stable and tracking; Multi-
fast solutions to obstacle avoidance in dynamic environments. sensor fusion
Subsequently, Guo et al. (2021)[102] proposed geometry based
collision avoidance techniques for real-time UAVs navigating
IV. MOTION PLANNING
autonomously in 3-D dynamic environments. In the proposed
obstacle avoidance technique first, the onboard detection Motion planning refers to the process of navigating safely from
system of the UAV obtains the information about the irregular one place to another in presence of obstacles. It comprises two
JETIR2211571 Journal of Emerging Technologies and Innovative Research (JETIR) www.jetir.org f565
© 2022 JETIR November 2022, Volume 9, Issue 11 www.jetir.org (ISSN-2349-5162)
tasks: path planning and control. Path planning is responsible path planning approach for the generation of a feasible path for
for finding an efficient obstacle free path from source to UAV in the ORB-SLAM framework with dynamic constraints
destination which a UAV can follow, whereas control provides on the length of the path and minimum turn radius. The
necessary commands to UAV for moving through that path presented path planning algorithm enumerates a set of nodes
without colliding with the obstacles. In this section, first we that could move in a force field, thereby permitting the rapid
discuss different types of existing path planning approaches and modifications of the path in real-time as cost function changes.
then we demonstrate the review of the existing literature on Subsequently, Jayaweera and Hanoun (2020) [38]
different strategies of vision-based control for UAV. demonstrated a path planning algorithm for UAVs that enables
them to follow ground moving targets. The proposed technique
A. Path planning
utilized a dynamic artificial potential field (D-APF), the
Planning of a collision-free path for safe navigation of an UAV trajectory generated by this algorithm is smooth and feasible to
is an essential step in the task of autonomous UAV navigation. non-static environments with hindrance and capable of
Path planning is required for finding a best path from the initial handling motion contour for the target moving on the ground
point to the location of the target, on the basis of certain considering the change in their direction and speed. The
performance indicators like the less price of work, the shortest existing path planning techniques, such as graph-based
time of flight, and the shortest flying route. At the time of path algorithms and swarm intelligence algorithms are not capable
planning, the UAV is also required to avoid obstacles. Based on of incorporating UAV dynamic models and flying time into
the environmental information that is to be utilized for evolution. In order to overcome these limitations of existing
computing an optimal path, the problem of path planning can methods Shao et al, (2021) [74] proposed a hierarchical scheme
be classified into two types: global path planning and local path for trajectory optimization with revised particle swarm
planning. The aim of the global path planning algorithm is to optimization (PSO) and Gauss pseudospectral method (GPM).
figure out an optimal path using a geographical map of the The proposed scheme is a two-layered approach. In the first
environment deduced initially. Although, for controlling a real- layer, the authors designed a better version of PSO for path
time UAV the task of global path planning is not enough, planning, then in the second layer after utilizing the waypoints
particularly when there are several other tasks that need to be in path generated by improved PSO, a fitted curve is
carried out forthwith or unexpected hurdles emerging at the constructed and used as the starting values for GPM. After
time of flight. Hence, there is a need for local path planning for comparing these initial values with the ones generated
estimating the path free from collisions in real-time by using randomly, the authors conclude that the designed curve can
data of the sensors taken from surrounding environments. improve the efficiency of GPM significantly. Further, the
1) Global path planning authors validate their presented scheme through plenty of
Global path planner requires a priori geographical map of simulations and the results obtained demonstrate that the
the surrounding environment with locations of starting and proposed technique achieves much better efficiency as
target points to compute a preliminary path, therefore the global compared to existing path planning methods.
map is also known as non-dynamic, that is static map.
Generally, there are two kinds of algorithms that are used for i. Heuristic searching methods
global path planning: Heuristic searching techniques and a The well-known heuristic search method A-star algorithm is
sequence of intelligent algorithms. In [46] authors proposed and the advanced form of the typical Dijkstra algorithm. In the past
developed an algorithm for planning the trajectory of multi- few decades, the A-star has been significantly evolved and
UAV in static environments. The proposed algorithm includes improved, hence lots of enhanced heuristic search methods
three main stages: the generation of initial trajectory, the have been derived from it. A modified A-star algorithm and an
correction of trajectory, and the smooth trajectory planning. orographic database were used in [87] for searching the best
The first phase of the proposed algorithm can be achieved track and building a digital map respectively. The heuristic A-
through MACO which incorporates metropolis measure into the star algorithm was used by the authors of [69], who first
node screening method of ant colony optimization (ACO) that dissected the entire region into square mesh and then leveraged
can efficiently and effectively avoid cascading into the the A-star heuristic algorithm for finding the best path that is
stagnation and local optimal solution. Then in the next phase of based on the value function of multiple points on the grid along
the algorithm the authors of this work proposed three different the estimated path. In [82] authors proposed the sparse A-star
trajectory correction schemes for solving the problem of search (SAS) for computing an optimal path, the presented
collision avoidance. Finally, the discontinuity resulting from heuristic search successfully minimizes the computational
the acute and edged turn in trajectory planning is resolved by complexity by imposing additional impediments and
using the inscribed circle (IC) smooth method. Results obtained limitations to the procedure of searching within space at the
through various laboratory experiments demonstrate the high time of path planning. The D-star algorithm, another name for
effectiveness and feasibility of the proposed solution from the dynamic A-star algorithm, was proposed and developed in
perspectives of obstacle avoidance, optimal solution, and [78] for computing an optimal path in a partially or completely
optimized trajectory in the problem of trajectory planning for unspecified dynamic environment. The algorithm keeps on
UAVs. Yathirajam et al. (2020) [90] proposed a chain-based improving and updating the map obtained from completely new
JETIR2211571 Journal of Emerging Technologies and Innovative Research (JETIR) www.jetir.org f566
© 2022 JETIR November 2022, Volume 9, Issue 11 www.jetir.org (ISSN-2349-5162)
and unknown environments and then amending the path on some unexpected aspects, such as the motion of obstacles in the
detection of new obstacles on its path. Authors of [91] proposed dynamic environments. In order to overcome this problem,
sampling-based path planning similar to rapidly exploring algorithms for local path planning need to be feasible and
random trees (RRT) that can generate an optimal path free from adaptive to the changing attributes of the surrounding
collision when no information of the surrounding environment environment, by using valuable data such as the shape, size, and
is provided initially. In [71] authors proposed and developed a position about different sections of the environment obtained
3D path planning solution for UAVs which makes them capable from multiple sensing devices.
of figuring out a feasible, optimal, and collision-free path in Conventional local path planning techniques include artificial
complicated dynamic environments. In the proposed approach potential field methods, neural network methods, fuzzy logic
authors exploit a probabilistic graph in order to test allowable methods, and spatial search methods, etc. Some general
space without considering the existing obstacles. Whenever methods for local path planning are discussed below. In [81]
planning is required, then the A-star discrete search algorithm authors proposed an artificial field method to navigate a robot
explores the generated probabilistic graph for obtaining an from the local environment into the domain of a metaphysical
optimal collision-free path. Authors validate their proposed artificial potential field. The final point has the “attraction” as
solution in the V-REP simulator and then incorporate it into a well as an object with “repulsion” to the navigating robot, hence
real-time UAV. As a kind of common obstacle in complex 3D these two forces are responsible for moving the robot towards
environments, U-type obstacles might be responsible for the target location. An example of the usage of the artificial
confusing a UAV thus leading to a collision. Therefore, in order gravitational field method for computing the local path through
to overcome this limitation Zhang et al. (2019) [92] proposed the threat area of radar was given in [10]. Genetic algorithms
and developed a state-of-the-art Ant-Colony Optimization can provide a typical framework for solving typical and
(ACO)-based technique called Self Heuristic Ant (SHA) for the complex problems of optimization, especially the ones that are
generation of the optimal, collision-free trajectory in related to the computation of an optimal path. These algorithms
unstructured 3D environments with solid U-type obstacles. In are inspired by the evolution and inheritance concepts of
this approach authors first construct the whole space using a biological phenomena. Problems should be solved by
grid model of workspace and then a novel optimized method leveraging the principle of “survival of the fittest and survival
for path planning of UAV is designed. In order to present ACO competition,” in order to obtain the best solution. A genetic
deadlock, that is, trapping of ants in U-type obstacles in the algorithm consists of five main components: initial population,
absence of an ideal successor node, authors designed two chromosome coding, genetic operation, fitness function, and
different search approaches for selecting the succeeding path control parameters. In [65] authors proposed a path planning
node. Furthermore, Self Heuristic Ant (SHA) is used for solution based on a genetic algorithm for an aircraft. Under the
improving the efficacy of the ACO-based method. Finally, confession of biological functions, neural networks are
results obtained after conducting several deeply investigated established as computational methods. An example of a local
experiments illustrate that the probability of deadlock state can path planning technique implemented using Hopfield networks
be reduced to a great extent with the implementation of was given in [25]. In [64] researchers proposed an ant colony
proposed search strategies. algorithm which is a new type of binocular algorithm inspired
by the ant activity. It is a probabilistic optimization method that
ii. Intelligent algorithms resembles the behavioral attributes of ants. This technique
In the past few decades, researchers tried a lot to work out could accomplish good results by solving a series of onerous
trajectory planning problems using intelligent algorithms and combinatorial optimizations. Table 3 provides an outline of
proposed a variety of intelligent searching techniques. different methods discussed for the task of path planning.
However, the most renowned intelligent algorithms are the
simulated anneal arithmetic (SAA) algorithm and genetic B. Vision-based control
algorithm. In [93] authors used the SAA methods and genetic Shirai and Inoue(1973) [104] introduced the concept of vision-
algorithm in the computation of an optimal path. Crossover and based control in robotics in which the data collected from visual
mutation operations of genetic algorithm and criterion of sensors is used for manipulating and controlling the motion of
Metropolis are used for the evaluation of adaptation function of the autonomous robot. In that era, the performance of
the path, thereby improving the effectiveness of trajectory autonomous robots based on vision-based control was not
planning. Authors of [2] proposed and developed an optimized achieved up to the desire, mainly due to the issue of extracting
global path planning technique using the improved conjugate information from visual sensors. Since 1990, with the
direction method and the simulated annealing algorithm. reasonable improvement in the computing power of personal
computers, there has been a high growth of research and
2) Local path planning development in the field of computer vision-based techniques
Local path planning methods exploit information of the local for robotics applications. After that, the vision-based control for
environment and state estimation of UAV to plan a collision- unmanned aerial vehicles had been considered as an important
free local path dynamically. Path planning in dynamic research problem that needed to be worked upon and till now
environments might become computationally expensive due to this is under rapid development with the objective of attaining
JETIR2211571 Journal of Emerging Technologies and Innovative Research (JETIR) www.jetir.org f567
© 2022 JETIR November 2022, Volume 9, Issue 11 www.jetir.org (ISSN-2349-5162)
complete autonomy in UAVs. The several challenges that still extracted from the image frames captured from a simulated
exist in the vision based control of UAVs are obstacle forest environment. Then, corresponding to the longitudinal
recognition, collision avoidance, and delay in the transfer of strip with minimum distance to the UAV, the real angle is
information to the UAV regarding action that it needs to take. calculated which is then processed further to obtain the
In the past few years, in order to overcome these limitations appropriate yaw/velocity commands for UAV control.
several strategies and solutions based on reinforcement Subsequently, Hu and Wang (2018)[110] proposed the hand-
learning, visual-inertial techniques, and hybrid approaches have gesture based control system for a UAV. The deep learning
been proposed. Few of them are mentioned here. method is trained on the dataset of gesture inputs to predict a
Kendoul et al. (2008)[105] proposed an adaptive controller suitable command for UAV control. First, the random hand
based vision-based control approach in which the UAV is gestures generated by the user are fed as an input input to the
capable of hovering at a certain height and tracking a given leap motion controller which consists of 2 optical sensors and 3
trajectory autonomously. The measurements from the inertial infrared lights and it is responsible for converting the dynamic
measurement unit are merged with the optical flow data for gestures information to the hand skeleton data and provide this
vehicle’s pose estimation and prediction of depth map with skeleton data to the data preprocessing unit which performs the
unknown scale factor use in obstacle detection. Then, finally task of feature selection and scaling on the output of the leap
the presented controller integrates all the measurements motion controller and creates the feature vectors. The composed
according to the control algorithm and provides the necessary feature vectors are delivered to the deep learning module which
commands to the UAV for autonomous navigation. In the recognize the gesture and predict the appropriate command.
progress of their previous work the Kendoul et al. (2009)[117] Finally, the predicted control command is delivered to the UAV
proposed a real time vision-based control for the UAV. The data control module and it processes the prediction to extract the
from the downward looking camera equipped in the UAV was movement command and send it to the UAV over WiFi. In
integrated with IMU measurements through an EKF, and then recent years, some researchers have also introduced
the obtained data was integrated with the designed non-linear reinforcement learning based strategies in the vision-based
controller for achieving the desired performance control of the control problem of aerial vehicles in dynamic environments.
UAV. Similarly, Carillo et al. (2012)[106] has developed a Following this approach, Li et al. (2018)[111] proposed a
quadrotor UAV capable of carrying out autonomous take-off, machine learning based strategy for the autonomous tracking of
navigation, and landing. The stereo camera and IMU sensors a given target. The strategy combines UAV perception and
are installed in the UAV. The data from these sensors is merged control for following a specific target. The proposed approach
using a Kalman filter for improving the accuracy of the leverages the integration of model-free policy gradient method
quadrotor state estimation. The stereo visual odometry and a PID controller for achieving stabilized navigation. The
technique is used for determining the 3D motion of the camera deep convolutional neural network is trained on the set of raw
in the environment. In order to leverage the integration of the images for the classification of target and then model-free
data obtained through depth camera and inertial unit in reinforcement learning technique is used for the generation of
controlling the UAV. Steggano et al. (2013)[107] proposed the high level control actions. The high level actions are then
design of a UAV platform equipped with a RGB-D camera and converted to the low level UAV control commands by a PID
IMU sensors to achieve autonomous vision-based control and controller which is then finally transferred to the quadrotor for
stabilization, especially in the teleoperations. And, the real world navigation and control. Table 4 summarizes the
integration of the information from IMU and RGB-D sensors is above discussed vision-based UAV control methods.
used for the estimation of the UAV velocity, which is then used
to control the UAV. Later on, the technique that utilized only a Table 3: Summary of important methods in path planning
monocular camera for obtaining the control commands for
UAV was proposed by Mannar et al. (2018)[108]. The authors
Types of Path Methods used References
of this work proposed and developed a vision-based control planning
system to avoid aerial obstacles in a forest environment for a
UAV. The proposed algorithm is the enhancement of the Potential field [38],[90]
existing algorithm initially proposed by Michel [109] for
vision-based obstacle avoidance in ground vehicles. The Global Optimization [46],[74]
authors have used this algorithm for UAV control for the first
Intelligent [2],[93]
time. The idea behind this algorithm is that the images captured
by the monocular camera of UAV are first divided into various Heuristic search [87],[82],[78],[91
longitudinal strips then from each strip the texture features are ],[92]
extracted and the weighted combination of these texture
features is used for the estimation of distance to the nearest Hopfield [25]
obstacle. The weights of the prediction network are pre-
Local Artificial [81],[10]
computed at the time of supervised training of the network on potential field
the correspondences of the features to the ground truth distance
JETIR2211571 Journal of Emerging Technologies and Innovative Research (JETIR) www.jetir.org f568
© 2022 JETIR November 2022, Volume 9, Issue 11 www.jetir.org (ISSN-2349-5162)
JETIR2211571 Journal of Emerging Technologies and Innovative Research (JETIR) www.jetir.org f569
© 2022 JETIR November 2022, Volume 9, Issue 11 www.jetir.org (ISSN-2349-5162)
JETIR2211571 Journal of Emerging Technologies and Innovative Research (JETIR) www.jetir.org f571
© 2022 JETIR November 2022, Volume 9, Issue 11 www.jetir.org (ISSN-2349-5162)
[52] Magree, D., and E. N. Johnson. 2014. “Combined Laser and Vision-Aided [72] Santos-Victor, J., G. Sandini, F. Curotto, and S. Garibaldi. 1993.
Inertial Navigation for an Indoor Unmanned Aerial Vehicle.” Paper Presented “Divergent Stereo for Robot Navigation: Learning from Bees.” Paper Presented
at the American Control Conference, IEEE, Portland, USA, June 4–6. at the Computer Vision and Pattern Recognition, IEEE Computer Society
[53] Mahon, I., S. B. Williams, O. Pizarro, and M. Johnson Roberson. 2008. Conference on, New York, USA, June 15–17.
“Efficient View-based SLAM Using Visual Loop Closures.” IEEE [73] Seitz, S. M., B. Curless, J. Diebel, D. Scharstein, and R. Szeliski. 2006. “A
Transactions on Robotics 24 (5): 1002–1014. Comparison and Evaluation of Multi-view Stereo Reconstruction Algorithms.”
[54] Maier, J., and M. Humenberger. 2013. “Movement Detection Based on Paper Presented at the Computer Vision and Pattern Recognition, IEEE
Dense Optical Flow for Unmanned Aerial Vehicles.” International Journal of Computer Society Conference on, New York, USA, June 17-22.
Advanced Robotic Systems 10 (2): 146. [74] Shao, S., He, C., Zhao, Y., & Wu, X. (2021). Efficient Trajectory Planning
[55] Márquez-Gámez, D. 2012. “Towards Visual Navigation in Dynamic and for UAVs Using Hierarchical Optimization. IEEE Access, 9, 60668-60681.
Unknown Environment: Trajectory Learning and following, with Detection and doi:10.1109/access.2021.3073420.
Tracking of Moving Objects.” PhD diss., l’Institut National des Sciences [75] Shen, S., Y. Mulgaonkar, N. Michael, and V. Kumar. 2014. “Multi-Sensor
Appliquées de Toulouse. Matsumoto. Fusion for Robust Autonomous Flight in Indoor and Outdoor Environments
[56] Y., K. Ikeda, M. Inaba, and H. Inoue. 1999. “Visual Navigation Using with a Rotorcraft MAV.” Paper Presented at the Robotics and Automation
Omnidirectional View Sequence.” Paper Presented at the Intelligent Robots and (ICRA), IEEE International Conference on, Hong Kong, China, May 31–June
Systems, IEEE/RSJ International Conference on, Kyongju, South Korea, 7.
October 17 –21. [76] Silveira, G., E. Malis, and P. Rives. 2008. “An Efficient Direct Approach
[57] Maza, I., F. Caballero, J. Capitán, J. R. Martínez-de-Dios, and A. Ollero. to Visual SLAM.” IEEE Transactions on Robotics 24 (5): 969–979.
2011. “Experimental Results in Multi-UAV Coordination for Disaster [77] Srinivasan, M. V., and R. L. Gregory. 1992. “How Bees Exploit Optic
Management and Civil Security Applications.” Journal of Intelligent & Robotic Flow: Behavioural Experiments and Neural Models [and Discussion].”
Systems 61 (1): 563–585. Philosophical Transactions of the Royal Society of London B: Biological
[58] Mirowski, P., R. Pascanu, F. Viola, H. Soyer, A. Ballard, A. Banino, and Sciences 337 (1281): 253–259.
M. Denil. 2017. “Learning to Navigate in Complex Environments.” The 5th [78] Stentz, A. 1994. “Optimal and Efficient Path Planning for Partially-Known
International Conference on Learning Representations, Toulon, France, April Environments.” Paper Presented at the IEEE International Conference on
24–26. Robotics and Automation, San Diego, CA, USA, May 8–13.
[59] Moravec, H. P. 1983. “The Stanford Cart and the CMU Rover.” [79] Strasdat, H., J. M. Montiel, and A. J. Davison. 2012. “Visual SLAM: Why
Proceedings of the IEEE 71 (7): 872–884. Filter?” Image and Vision Computing 30 (2):65–77.
[60] Moreno-Armendáriz, M. A., and H. Calvo. 2014. “Visual SLAM and [80] Strohmeier, M., M. Schäfer, V. Lenders, and I. Martinovic. 2014.
Obstacle Avoidance in Real-Time for Mobile Robots Navigation.” Paper “Realities and Challenges of Nextgen Air Traffic Management: The Case of
Presented at the Mechatronics, Electronics and Automotive Engineering ADS-B.” IEEE Communications Magazine 52 (5): 111–118.
(ICMEAE), IEEE International Conference on, Cuernavaca, Mexico, [81] Sugihara, K., and I. Suzuki. 1996. “Distributed Algorithms for Formation
November 18–21. of Geometric Patterns with Many Mobile Robots.” Journal of Robotic Systems
[61] Newcombe, R. A., S. J. Lovegrove, and A. J. Davison. 2011. “DTAM: 13 (3): 127–139.
Dense Tracking and Mapping in Real-time.” Paper Presented at the Computer [82] Szczerba, R. J., P. Galkowski, I. S. Glicktein, and N. Ternullo. 2000.
Vision (ICCV), IEEE International Conference on, Washington, DC, USA, “Robust Algorithm for Real-time Route Planning.” IEEE Transactions on
November 6–13. Aerospace and Electronic Systems 36 (3): 869–878.
[62] Nourani-Vatani, N., P. Vinicius, K. Borges, J. M. Roberts, and M. V. [83] Szeliski, R. 2010. Computer Vision: Algorithms and Applications.
Srinivasan. 2014. “On the Use of Optical Flow for Scene Change Detection and London: Springer Science & Business Media.
Description.” Journal of Intelligent & Robotic Systems 74 (3–4): 817. [84] Szenher, M. D. 2008. “Visual Homing in Dynamic Indoor Environments.”
[63] Padhy, R.M., Choudhry, S.K., Sa, P.K., & Bakshi, S. (2019). Obstacle Ph.D. diss. The University of Edinburgh.
Avoidance for Unmanned Aerial Vehicles. IEEE Consumer Electronics [85] Trujillo, J., Munguia, R., Urzua, S., Guerra, E., & Grau, A. (2020).
Magazine, 19, 2162-2248. doi: 10.1109/MCE.2019.2892280. Monocular Visual SLAM Based on a Cooperative UAV–Target System.
[64] Parunak, H. V., M. Purcell, and R. O’Connell. 2002. “Digital Pheromones Sensors, 20(12), 3531. doi:10.3390/s20123531.
for Autonomous Coordination of Swarming UAVs.” Paper Presented at the 1st [86] Tuytelaars, T., and K. Mikolajczyk. 2008. “Local Invariant Feature
UAV Conference, Portsmouth, Virginia, May 20–23. Detectors: A Survey.” Foundations and Trends in Computer Graphics and
[65] Pellazar, M. B. 1998. “Vehicle Route Planning with Constraints Using Vision 3 (3): 177–280.
Genetic Algorithms”. Paper Presented at the Aerospace and Electronics [87] Vachtsevanos, G., W. Kim, S. Al-Hasan, F. Rufus, M. Simon, D. Shrage,
Conference, NAECON, 1998, IEEE National, Dayton, USA, July 17. and J. Prasad. 1997. “Autonomous Vehicles: From Flight Control to Mission
[66] Ranftl, R., V. Vineet, Q. Chen, and V. Koltun. 2016. “Dense Monocular Planning Using Fuzzy Logic Techniques.” Paper Presented at the Digital Signal
Depth Estimation in Complex Dynamic Scenes.” Paper Presented at the Processing, 13th International Conference on, IEEE, Santorini, Greece, July 2–
Computer Vision and Pattern Recognition, IEEE Conference on, Las Vegas, 4.
USA, June 27–30. [88] Valgaerts, L., A. Bruhn, M. Mainberger, and J. Weickert. 2012. “Dense
[67] Roberge, V., M. Tarbouchi, and G. Labonté. 2013. “Comparison of Parallel versus Sparse Approaches for Estimating the Fundamental Matrix.”
Genetic Algorithm and Particle Swarm Optimization for Real-time UAV Path International Journal of Computer Vision 96 (2): 212–234.
Planning.” IEEE Transactions on Industrial Informatics 9 (1): 132–141. [89] Wang, C., Ma, H., Chen, W., Liu, L., & Meng, M. Q. (2020). Efficient
[68] Rogers, B., and M. Graham. 1979. “Motion Parallax as an Independent Autonomous Exploration With Incrementally Built Topological Map in 3-D
Cue for Depth Perception.” Perception 8 (2): 125–134. Environments. IEEE Transactions on Instrumentation and Measurement,
[69] Rouse, D. M. 1989. “Route Planning Using Pattern Classification and 69(12), 9853-9865. doi:10.1109/tim.2020.3001816.
Search Techniques.” Aerospace and Electronics Conference, IEEE National, [90] Yathirajam, B., S.M, V. & C.M, A. (2020). Obstacle Avoidance for
Dayton, USA, May 22–26. Unmanned Air Vehicles Using Monocular-SLAM with Chain-Based Path
[70] Ruffier, F., S. Viollet, S. Amic, and N. Franceschini. 2003. “Bio-inspired Planning in GPS Denied Environments. Journal of Aerospace System
Optical Flow Circuits for the Visual Guidance of Micro Air Vehicles.” Paper Engineering, 14(2), 1-11.
Presented at the Circuits and Systems, IEEE International Symposium on, [91] Yershova, A., L. Jaillet, T. Simeon, and S. M. Lavalle. 2005. “Dynamic-
Bangkok, Thailand, May 25–28. Domain RRTs: Efficient Exploration by Controlling the Sampling Domain.”
[71] Sanchez-Lopez, J. L., Wang, M., Olivares-Mendez, M. A., Molina, M., & Paper Presented at the IEEE International Conference on Robotics and
Voos, H. (2018). A Real-Time 3D Path Planning Solution for Collision-Free Automation, Barcelona, Spain, April 18–22.
Navigation of Multirotor Aerial Robots in Dynamic Environments. Journal of [92] Zhang, C., Hu, C., Feng, J., Liu, Z., Zhou, Y., & Zhang, Z. (2019). A Self-
Intelligent & Robotic Systems, 93(1-2), 33-53. doi:10.1007/s10846-018-0809- Heuristic Ant-Based Method for Path Planning of Unmanned Aerial Vehicle in
5. Complex 3-D Space With Dense U-Type Obstacles. IEEE Access, 7, 150775-
150791. doi:10.1109/access.2019.2946448.
JETIR2211571 Journal of Emerging Technologies and Innovative Research (JETIR) www.jetir.org f572
© 2022 JETIR November 2022, Volume 9, Issue 11 www.jetir.org (ISSN-2349-5162)
[93] Zhang, Q., J. Ma, and Q. Liu. 2012. “Path Planning Based Quadtree [115] Lin, S., Jin, L. & Chen, Z., 2021. Real-time monocular vision system for
Representation for Mobile Robot Using Hybrid-Simulated Annealing and Ant UAV autonomous landing in outdoor low-illumination environments. Sensors,
Colony Optimization Algorithm.” Paper Presented at The Intelligent Control 21(18), p.6226.
and Automation (WCICA), 10th World Congress on, IEEE, Beijing, China, [116] Huang, H., 2020. Deep Reinforcement Learning for UAV Navigation
July 6–8. Through Massive MIMO Technique. In IEEE Transactions on Vehicular
[94] González-Jorge, H. et al., 2017. Unmanned aerial systems for Civil Technology , pp. 1117–1121.
Applications: A Review. Drones, 1(1), p.2. [117] Grando, R.B., 2020. Deep Reinforcement Learning for Mapless
[95] Lu, Y. et al., 2018. A survey on vision-based UAV navigation. Geo-spatial Navigation of Unmanned Aerial Vehicles. In 2020 Latin American Robotics
Information Science, 21(1), pp.21–32. Symposium (LARS), 2020 Brazilian Symposium on Robotics (SBR) and 2020
[96] T. P. Lillicrap, J. J. Hunt, A. Pritzel, N. Heess, T. Erez, Y. Tassa, D. Silver, Workshop on Robotics in Education (WRE). Natal, Brazil.
and D. Wierstra, “Continuous control with deep reinforcement learning,” in [118] Kendoul, F., Fantoni, I., Lozano, R.: Adaptive vision-based controller for
ICLR, 2015. small rotorcraft uavs control and guidance. In: Proceedings of the 17th IFAC
[97] T. Haarnoja, A. Zhou, P. Abbeel, and S. Levine, “Soft actor-critic: Off- world congress, pp. 6–11 (2008)
policy maximum entropy deep reinforcement learning with a stochastic actor,”
in ICML, vol. 80, 2018, pp. 1861–1870.
[98] LI, M. & HU, T., 2021. Deep learning enabled localization for UAV
autolanding. Chinese Journal of Aeronautics, 34(5), pp.585–600.
[99] Afifi, G. & Gadallah, Y., 2021. Autonomous 3-D UAV localization using
cellular networks: Deep supervised learning versus reinforcement learning
approaches. IEEE Access, 9, pp.155234–155248.
[100] H. Goforth and S. Lucey, “GPS-denied UAV localization using pre-
existing satellite imagery,” Proceedings - IEEE International Conference
on Robotics and Automation, vol. 2019-May, pp. 2974–2980, 2019.
[101] H. Hou, Q. Xu, C. Lan, W. Lu, Y. Zhang, Z. Cui, and J. Qin, “UAV Pose
Estimation in GNSS-Denied Environment Assisted by Satellite Imagery Deep
Learning Features,” IEEE Access, vol. 9, pp. 6358–6367, 2020.
[102] Guo, J. et al., 2021. Three-dimensional autonomous obstacle avoidance
algorithm for UAV based on circular arc trajectory. International Journal of
Aerospace Engineering, 2021, pp.1–13.
[103] Aldao, E. et al., 2022. UAV obstacle avoidance algorithm to navigate in
Dynamic Building Environments. Drones, 6(1), p.16.
[104] Shirai, Y., Inoue, H., 1973. Guiding a robot by visual feedback in
assembling tasks. Patt. Recogn. 5, 99–108.
[105] Kendoul, F., Fantoni, I., Nonami, K.: Optic flow-based vision system for
autonomous 3d localization and control of small aerial vehicles. Robot.
Autonom. Syst. 57(6), 591–602 (2009)
[106] Carrillo, L.R.G., López, A.E.D., Lozano, R., Pégard, C.: Combining
stereo vision and inertial navigation system for a quad-rotor uav. J. Intelli.
Robot. Syst. 65(1-4), 373–387 (2012).
[107] Stegagno, P.S., 2013. Vision-based Autonomous Control of a Quadrotor
UAV using an Onboard RGB-D Camera and its Application to Haptic
Teleoperation. In the 2nd IFAC Workshop on Research, Education and
Development of Unmanned Aerial Systems, France..
[108] Mannar, S., 2018. Vision-based control for aerial obstacle avoidance in
Forest environments. In the International Federation of Automatic Control.
Bangalore: ELSEVIER, pp. 2405–8963.
[109] Michels, Jeff, Ashutosh Saxena, and Andrew Y. Ng. ”High speed obstacle
avoidance using monocular vision and reinforcement learning.” In Proceedings
of the 22nd international conference on Machine learning, pp. 593-600. ACM,
2005.
[110] Hu, B. & Wang, J., 2018. Deep Learning Based Hand Gesture
Recognition and UAV Flight Controls. In Proceedings of the 24th International
Conference on Automation & Computing. Newcastle University, UK.
[111] Li, S., 2018. Learning Unmanned Aerial Vehicle Control for Autonomous
Target Following. In Proceedings of the Twenty-Seventh International Joint
Conference on Artificial Intelligence (IJCAI-18). HKUST, Hong Kong.
JETIR2211571 Journal of Emerging Technologies and Innovative Research (JETIR) www.jetir.org f573