hishamcse commited on
Commit
61672cf
โ€ข
1 Parent(s): 6ef4984

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +35 -27
README.md CHANGED
@@ -7,31 +7,39 @@ tags:
7
  - ML-Agents-SoccerTwos
8
  ---
9
 
10
- # **poca** Agent playing **SoccerTwos**
11
- This is a trained model of a **poca** agent playing **SoccerTwos**
12
- using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
13
-
14
- To see full code, visit: https://www.kaggle.com/code/syedjarullahhisham/drl-huggingface-unit-7-marl-soccer-2vs2
15
-
16
- ## Usage (with ML-Agents)
17
- The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/
18
-
19
- We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
20
- - A *short tutorial* where you teach Huggy the Dog ๐Ÿถ to fetch the stick and then play with him directly in your
21
- browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction
22
- - A *longer tutorial* to understand how works ML-Agents:
23
- https://huggingface.co/learn/deep-rl-course/unit5/introduction
24
-
25
- ### Resume the training
26
- ```bash
27
- mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
28
- ```
29
-
30
- ### Watch your Agent play
31
- You can watch your agent **playing directly in your browser**
32
-
33
- 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity
34
- 2. Step 1: Find your model_id: hishamcse/poca-SoccerTwos
35
- 3. Step 2: Select your *.nn /*.onnx file
36
- 4. Click on Watch the agent play ๐Ÿ‘€
 
 
 
 
 
 
 
 
37
 
 
7
  - ML-Agents-SoccerTwos
8
  ---
9
 
10
+ # **poca** Agent playing **SoccerTwos**
11
+ This is a trained model of a **poca** agent playing **SoccerTwos**
12
+ using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
13
+
14
+ ## Codes
15
+
16
+ Github repos(Give a star if found useful):
17
+ * https://github.com/hishamcse/Advanced-DRL-Renegades-Game-Bots
18
+ * https://github.com/hishamcse/DRL-Renegades-Game-Bots
19
+
20
+ Kaggle Notebook:
21
+ * https://www.kaggle.com/code/syedjarullahhisham/drl-huggingface-unit-7-marl-soccer-2vs2
22
+
23
+
24
+ ## Usage (with ML-Agents)
25
+ The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/
26
+
27
+ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
28
+ - A *short tutorial* where you teach Huggy the Dog ๐Ÿถ to fetch the stick and then play with him directly in your
29
+ browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction
30
+ - A *longer tutorial* to understand how works ML-Agents:
31
+ https://huggingface.co/learn/deep-rl-course/unit5/introduction
32
+
33
+ ### Resume the training
34
+ ```bash
35
+ mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
36
+ ```
37
+
38
+ ### Watch your Agent play
39
+ You can watch your agent **playing directly in your browser**
40
+
41
+ 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity
42
+ 2. Step 1: Find your model_id: hishamcse/poca-SoccerTwos
43
+ 3. Step 2: Select your *.nn /*.onnx file
44
+ 4. Click on Watch the agent play ๐Ÿ‘€
45