Degree Type

Thesis

Date of Award

2020

Degree Name

Master of Science

Department

Computer Science

Major

Computer Science

First Advisor

Ali Jannesari

Abstract

Integration of reinforcement learning with unmanned aerial vehicles (UAVs) to achieve autonomous flight has been an active research area in recent years. An important part focuses on collision detection and avoidance as a UAV navigates through an environment. In this thesis, we introduce a new variation of the Deep Q-Network (DQN) algorithm for UAV collision avoidance. Exploration with other variations of DQN for collision avoidance such as D3QN, are typically done through uniform sampling of actions, however, the challenge is environments inherently have sparse rewards resulting in many actions leading to redundant states. We focus on this problem of learning the dynamics of an unseen environment with sparse rewards more efficiently. To this end, we present an algorithm for improved exploration for UAVs. The approach is a guidance based method that uses a Bayesian Gaussian mixture model to compare previously seen states to a predicted next state in order to select the next action. Performance of these approaches was demonstrated in multiple simulation environments using Microsoft AirSim. The proposed algorithm demonstrates a two-fold improvement in average rewards compared to D3QN, after the first 1000 training episodes.

DOI

https://doi.org/10.31274/etd-20200624-163

Copyright Owner

Jeremy Roghair

Language

en

File Format

application/pdf

File Size

36 pages

Available for download on Wednesday, June 16, 2021

Share

COinS