Reinforcement Learning CVRP Save

Dynamic Attention Encoder-Decoder model to learn and design heuristics to solve capacitated vehicle routing problems

Project README

Reinforcement-Learning-CVRP

Introduction

In recent years machine learning is evolving at a phenomenal rate and can tackle tough problems on its own. The recent research work in the field of combinatorial optimization shows that machine learning has the potential to learn and design heuristics better than the traditional heuristics designed by humans. In this project, a reinforcement model with dynamic encoder-decoder architecture is developed that learns to design its heuristics based on the data to solve a large-scale vehicle routing problem with optimality. The trained models produce the near-optimal solution instantly, without the need to retrain the models. When compared, with other heuristic approaches like the Savings Clarke wright algorithm which is implemented in Google’s Operation research tools this reinforcement model outperformed them. This proposed model can be easily extended to solve other variants of VRP problems like multi depot and VRP with Time windows.

alt text

Comparsion of Machine Learning models performance with Savings Algorithm

alt text

Acknowledgements

The Dynamic Attention Model used here for solving VRP is inspired from the paper by Peng on the topic A Deep Reinforcement Learning Algorithm Using Dynamic Attention Model for Vehicle Routing Problems which can be found here

Thanks to Dmitry Eremeev, Alexey Pustynnikov for providing complete code implementation for this paper and making it public.Code implementation for this project can be found in the github repo here.

Open Source Agenda is not affiliated with "Reinforcement Learning CVRP" Project. README Source: JaswanthBadvelu/Reinforcement-Learning-CVRP
Stars
38
Open Issues
1
Last Commit
3 years ago

Open Source Agenda Badge

Open Source Agenda Rating