Update README.md
Browse files
README.md
CHANGED
|
@@ -8,6 +8,16 @@ pipeline_tag: reinforcement-learning
|
|
| 8 |
Model used for solving Capacitated Vehicle Routing Problem (CVRP). The CVRP is a variant of the vehicle routing problem (VRP) in which vehicles have a limited carrying capacity and must visit a set of customer locations to deliver or collect items.
|
| 9 |
Model is based on GitHub repo [HERE](https://github.com/d-eremeev/ADM-VRP), and was used for medium.com article **"Vaccine Supply Chain Optimization with AI-Powered Capacitated Vehicle Routing Problem(CVRP)"**.
|
| 10 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 11 |
|
| 12 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 13 |
|
|
|
|
| 8 |
Model used for solving Capacitated Vehicle Routing Problem (CVRP). The CVRP is a variant of the vehicle routing problem (VRP) in which vehicles have a limited carrying capacity and must visit a set of customer locations to deliver or collect items.
|
| 9 |
Model is based on GitHub repo [HERE](https://github.com/d-eremeev/ADM-VRP), and was used for medium.com article **"Vaccine Supply Chain Optimization with AI-Powered Capacitated Vehicle Routing Problem(CVRP)"**.
|
| 10 |
|
| 11 |
+
**Dynamic Attention Model (AM-D) Approach**:
|
| 12 |
+
After vehicle returns to depot, the remaining nodes could be considered as a new (smaller) instance (graph) to be solved.
|
| 13 |
+
Idea: update embedding of the remaining nodes using encoder after agent arrives back to depot.
|
| 14 |
+
**Implementation**:
|
| 15 |
+
- Force RL agent to wait for others once it arrives to .
|
| 16 |
+
- When every agent is in depot, apply encoder with mask to the whole batch.
|
| 17 |
|
| 18 |
|
| 19 |
+
If you want to train your own model with AM-D approach:
|
| 20 |
+
1. Prepare data (depo location Lat/Long, nodes location Lat/Long and capacity of the vehicles)
|
| 21 |
+
2. Transform data with TensorFlow tranform_to_tensor [Here is Gist](https://gist.github.com/PiotrKrosniak/f488eea5b31a2d61e21554041a1ee59b) example with transforming from Pandas Data Frame
|
| 22 |
+
3. Train the model using
|
| 23 |
|