Skip to content

Harnessing reinforcement learning, this repository emulates drone flocking behavior inspired by biological models. This uses a 2D environment with potential field functions, actor-critic models, and boid flocking behavior. Access the dissertation and Matlab simulations for detailed insights.

Notifications You must be signed in to change notification settings

oscell/Biologically-inspired-UAV

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

28 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Biologically Inspired UAV Guidance Using Reinforcement Learning

Overview

This project uses reinforcement learning (RL) with actor-critic model to instill flocking behaviour in drones based on biological models. This involved:

  • Developing a simulation of multi-agent systems in an environment that incorporates potential field functions and boid flocking behavior.
  • Formulating an effective loss function
  • Designing a custom reward function that uses exponential variations to ensure flocking within set steps
  • Drones are directed through gains that boost specific behaviours, with the goal to maximize long-term rewards.

This work highlights RL's promise in UAV systems and suggests transitioning to a more scalable multi-agent environment.

📖 For a comprehensive overview, refer to the dissertation document. 📖

Methods

The methods employed include:

  • Incorporating the boid flocking model with separation, alignment, and cohesion behaviors.
Separation Alignment Cohesion
  • Utilizing the actor-critic RL model to learn gains that control flocking behavior.
  • Layering potential field functions for attraction to destinations.
  • Designing a custom environment and reward shaping to foster flocking and ensure drones reach their destinations.
Rewared Function Potential Field

Results

The project's outcomes are:

  • The RL agent successfully learns the emergent flocking behavior of drones.
  • The significance of reward shaping in promoting desired behaviors is highlighted.
  • Certain challenges, such as some drones not reaching their destinations, underscore the need for further refinement of the reward function.

Trained Agent

Future Work

Potential avenues for future exploration:

  • Transition to a decentralized multi-agent approach to enhance the RL structure.
  • Enrich the environment dynamics and introduce obstacles.
  • Boost simulation speed using techniques like quad trees.
  • Refine the reward function to rectify any undesirable behaviors.

Repository Contents

Matlab Simulation: The custom environment and the RL agent.

Dissertation: The primary document detailing the research, methodology, and findings.

Setup & Installation

git clone --recursive https://github.com/oscell/Biologically-inspired-UAV.git

Reference

Meunier, Oscar. "Biologically inspired UAV guidance: Using reinforcement learning to optimize flocking behaviour." University of Glasgow, 2022.

About

Harnessing reinforcement learning, this repository emulates drone flocking behavior inspired by biological models. This uses a 2D environment with potential field functions, actor-critic models, and boid flocking behavior. Access the dissertation and Matlab simulations for detailed insights.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages