Update README.md
Browse files
README.md
CHANGED
@@ -29,4 +29,16 @@ model-index:
|
|
29 |
value: 142.19907608032227 +/- 19.368785745326573
|
30 |
name: "Local Reward"
|
31 |
verified: true
|
32 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
29 |
value: 142.19907608032227 +/- 19.368785745326573
|
30 |
name: "Local Reward"
|
31 |
verified: true
|
32 |
+
---
|
33 |
+
|
34 |
+
This model serves as the baseline for the **Ocean Plastic Collection** environment, trained and tested on task <code>1</code> using the Proximal Policy Optimization (PPO) algorithm.<br>
|
35 |
+
<br>
|
36 |
+
Environment: **Ocean Plastic Collection**<br>
|
37 |
+
Task: <code>1</code><br>
|
38 |
+
Algorithm: <code>PPO</code><br>
|
39 |
+
Episode Length: <code>5000</code><br>
|
40 |
+
Training <code>max_steps</code>: <code>3000000</code><br>
|
41 |
+
Testing <code>max_steps</code>: <code>150000</code><br>
|
42 |
+
<br>
|
43 |
+
Train & Test [Scripts](https://github.com/hivex-research/hivex)<br>
|
44 |
+
Download the [Environment](https://github.com/hivex-research/hivex-environments)
|