philippds commited on
Commit
c6d0442
1 Parent(s): 0b6555c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -1
README.md CHANGED
@@ -29,4 +29,16 @@ model-index:
29
  value: 142.19907608032227 +/- 19.368785745326573
30
  name: "Local Reward"
31
  verified: true
32
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
29
  value: 142.19907608032227 +/- 19.368785745326573
30
  name: "Local Reward"
31
  verified: true
32
+ ---
33
+
34
+ This model serves as the baseline for the **Ocean Plastic Collection** environment, trained and tested on task <code>1</code> using the Proximal Policy Optimization (PPO) algorithm.<br>
35
+ <br>
36
+ Environment: **Ocean Plastic Collection**<br>
37
+ Task: <code>1</code><br>
38
+ Algorithm: <code>PPO</code><br>
39
+ Episode Length: <code>5000</code><br>
40
+ Training <code>max_steps</code>: <code>3000000</code><br>
41
+ Testing <code>max_steps</code>: <code>150000</code><br>
42
+ <br>
43
+ Train & Test [Scripts](https://github.com/hivex-research/hivex)<br>
44
+ Download the [Environment](https://github.com/hivex-research/hivex-environments)