Professional Documents
Culture Documents
Summary
Summary
Summary
presents dynamic proximal meta policy optimization with covariance matrix adaptation
evolutionary strategies (dynamic-PMPO-CMA) to avoid obstacles and realize autonomous
navigation
Multi-robot coordination (MRC) has higher efficiency than single robot in search and rescue
(SAR) scenario
Elman Fuzzy Adaptive Control for Obstacle Avoidance of Mobile Robots Using
Hybrid
Force/Position Incorporation
addresses a virtual force field between mobile robots and obstacles to keep them away
with a desired distance.
Elman neural network is proposed to compensate the effect of uncertainties between the
dynamic robot model and the obstacles.
uses an Elman fuzzy adaptive controller to adjust the exact distance between the robot
and the obstacles
It is hard to model the dynamics of the environment because of the uncertainty in the
environment, such as observation noise or unknown dynamic obstacles
EFFICIENT META REINFORCEMENT LEARNING VIA META GOAL
GENERATION
Meta reinforcement learning (meta-RL) is able to accelerate the acquisition of new tasks by
learning from past experience.
Given a number of tasks with similar structures, meta-RL methods enable agents learn such
structure from previous experience on many tasks. Thus, when encountering a new task,
agents can quickly adapt to it with only a small amount of experience.
To solve the case of a complex environment with the presence of humans and other
unpredictable moving objects:
where each robot has different options to choose its path to the goal.
Model-free approaches like reinforcement learning has been used to robot motion planning
problem [5], but it requires vast amounts of training data, which can result in low sample
efficiency.
Deep Q Learning combined with CNN (Convolution Neural Network) algorithm, to solve the
problem that robots need to search a comparatively wide area for navigation and move in a
predesigned formation under a given environment in conventional path planning algorithms.
Deep learning combined with reinforcement learning (DRL) has recently been used for a wide
range of tasks, such as games [5], robotics control [6,7] and navigation [8–14]
DRL methods are usually limited to a single scenario and fail to work without additional large amount
of training data if the target environment changes.
Meta-learning [16,17] is similar to human intelligence that can quickly learn a new task from a small
amount of new data.