hishamcse commited on
Commit
6ef4984
โ€ข
1 Parent(s): 480abfc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +36 -34
README.md CHANGED
@@ -1,35 +1,37 @@
1
- ---
2
- library_name: ml-agents
3
- tags:
4
- - SoccerTwos
5
- - deep-reinforcement-learning
6
- - reinforcement-learning
7
- - ML-Agents-SoccerTwos
8
- ---
9
-
10
- # **poca** Agent playing **SoccerTwos**
11
- This is a trained model of a **poca** agent playing **SoccerTwos**
12
- using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
13
-
14
- ## Usage (with ML-Agents)
15
- The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/
16
-
17
- We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
18
- - A *short tutorial* where you teach Huggy the Dog ๐Ÿถ to fetch the stick and then play with him directly in your
19
- browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction
20
- - A *longer tutorial* to understand how works ML-Agents:
21
- https://huggingface.co/learn/deep-rl-course/unit5/introduction
22
-
23
- ### Resume the training
24
- ```bash
25
- mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
26
- ```
27
-
28
- ### Watch your Agent play
29
- You can watch your agent **playing directly in your browser**
30
-
31
- 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity
32
- 2. Step 1: Find your model_id: hishamcse/poca-SoccerTwos
33
- 3. Step 2: Select your *.nn /*.onnx file
34
- 4. Click on Watch the agent play ๐Ÿ‘€
 
 
35
 
 
1
+ ---
2
+ library_name: ml-agents
3
+ tags:
4
+ - SoccerTwos
5
+ - deep-reinforcement-learning
6
+ - reinforcement-learning
7
+ - ML-Agents-SoccerTwos
8
+ ---
9
+
10
+ # **poca** Agent playing **SoccerTwos**
11
+ This is a trained model of a **poca** agent playing **SoccerTwos**
12
+ using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
13
+
14
+ To see full code, visit: https://www.kaggle.com/code/syedjarullahhisham/drl-huggingface-unit-7-marl-soccer-2vs2
15
+
16
+ ## Usage (with ML-Agents)
17
+ The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/
18
+
19
+ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
20
+ - A *short tutorial* where you teach Huggy the Dog ๐Ÿถ to fetch the stick and then play with him directly in your
21
+ browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction
22
+ - A *longer tutorial* to understand how works ML-Agents:
23
+ https://huggingface.co/learn/deep-rl-course/unit5/introduction
24
+
25
+ ### Resume the training
26
+ ```bash
27
+ mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
28
+ ```
29
+
30
+ ### Watch your Agent play
31
+ You can watch your agent **playing directly in your browser**
32
+
33
+ 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity
34
+ 2. Step 1: Find your model_id: hishamcse/poca-SoccerTwos
35
+ 3. Step 2: Select your *.nn /*.onnx file
36
+ 4. Click on Watch the agent play ๐Ÿ‘€
37