philippds commited on
Commit
1a0384e
1 Parent(s): c59dfd3

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -1
README.md CHANGED
@@ -58,4 +58,17 @@ model-index:
58
  value: 419.7281740188599 +/- 230.5992870123718
59
  name: Cumulative Reward
60
  verified: true
61
- ---This model serves as the baseline for the **Aerial Wildfire Suppression** environment, trained and tested on task <code>0</code> with difficulty <code>1</code> using the Proximal Policy Optimization (PPO) algorithm.<br><br>Environment: **Aerial Wildfire Suppression**<br>Task: <code>0</code><br>Difficulty: <code>1</code><br>Algorithm: <code>PPO</code><br>Episode Length: <code>3000</code><br>Training <code>max_steps</code>: <code>1800000</code><br>Testing <code>max_steps</code>: <code>180000</code><br><br>Train & Test [Scripts](https://github.com/hivex-research/hivex)<br>Download the [Environment](https://github.com/hivex-research/hivex-environments)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
58
  value: 419.7281740188599 +/- 230.5992870123718
59
  name: Cumulative Reward
60
  verified: true
61
+ ---
62
+
63
+ This model serves as the baseline for the **Aerial Wildfire Suppression** environment, trained and tested on task <code>0</code> with difficulty <code>1</code> using the Proximal Policy Optimization (PPO) algorithm.<br><br>
64
+
65
+ Environment: **Aerial Wildfire Suppression**<br>
66
+ Task: <code>0</code><br>
67
+ Difficulty: <code>1</code><br>
68
+ Algorithm: <code>PPO</code><br>
69
+ Episode Length: <code>3000</code><br>
70
+ Training <code>max_steps</code>: <code>1800000</code><br>
71
+ Testing <code>max_steps</code>: <code>180000</code><br><br>
72
+
73
+ Train & Test [Scripts](https://github.com/hivex-research/hivex)<br>
74
+ Download the [Environment](https://github.com/hivex-research/hivex-environments)