micheljperez commited on
Commit
4658da4
1 Parent(s): 12e2a0a
Huggy.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2b92309003666901febd79a3ee11fc28af357e454f96b38edaefc073ceb3ca20
3
- size 2270373
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44d133ce7c4f9f017a38c5302950f3a14f0c682e4e85bb86b95f3b2d2044f504
3
+ size 2273983
Huggy/Huggy-1199991.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5768dea29d3b84817158429d315fb0cddcff1f800fdbae4a06bb3752b117729
3
+ size 2273983
Huggy/Huggy-1199991.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a418157b2e5c5954787ce1761669c35d8b0509c05cc281ac25106ad2d2a2e632
3
+ size 13509028
Huggy/Huggy-1399993.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8693dc08d59fc769fa25053495b8dd3f5681f6d3c19bae1450a3bce9feaea6bd
3
+ size 2273983
Huggy/Huggy-1399993.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31da4a6577c5e5b4b5cd8243f01f85ebd4003e2268fa38b91fc3ebdb30a74ed1
3
+ size 13509028
Huggy/Huggy-1599947.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1181c93bc7313fc05219df54b85510eb2201be736558d6d15fa3c02960b1ff25
3
+ size 2273983
Huggy/Huggy-1599947.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f417bde7e613677d798fcd09f307cb9f94d0ada16debbb31a5a3980ee54957c
3
+ size 13509028
Huggy/Huggy-1799952.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9975c62ba1fd424fa92ce3a3e79e98363c3f3e0b051da5f39d88797a2a67c78e
3
+ size 2273983
Huggy/Huggy-1799952.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4b0e0f7e6912eeb3d03cb5bcbacd1f76e5ac2861b6167eda893fb0c4fd7e907
3
+ size 13509028
Huggy/Huggy-199642.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3316d26a703c540de84581a7d6b98f3052d501cacb3c22592f4903cf24e4ca25
3
+ size 2273983
Huggy/Huggy-199642.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3534bb96536d6616fcfbcd27aaa8496e34e6359bf2c03b933791400bdad28a3d
3
+ size 13508945
Huggy/Huggy-1999979.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50740e75a60166a8f63325cb78137cd29ce3adbbe444f02a4c050c17473a876d
3
+ size 2273983
Huggy/Huggy-1999979.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55e9214787f8e68cd8b1c4ec88b95858db9e7a2c0fae7bc5e8a9d39edce573ac
3
+ size 13509028
Huggy/Huggy-2000044.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44d133ce7c4f9f017a38c5302950f3a14f0c682e4e85bb86b95f3b2d2044f504
3
+ size 2273983
Huggy/Huggy-2000044.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a1fd8c07ea96154c92adbc9b22cfb5828a97ff72a6ea5b5570e32291fc36516
3
+ size 13509028
Huggy/Huggy-399995.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f8d6e7dc9630259bdc14893945598ff71abd7561fa79b8d305282890f82f239
3
+ size 2273983
Huggy/Huggy-399995.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:448a23eddcb4a9858f1b6833624d64fa286aa11ae41b849165af0ac84d2e621b
3
+ size 13508945
Huggy/Huggy-599953.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5152242934991ea4ddee10976eb8d240e7dba844c96835fba83fc3576b08e894
3
+ size 2273983
Huggy/Huggy-599953.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bc8cfb8bbe6696456e299df22a8c6f8a11ac9fcaa66d17c29e5062d81c00be4
3
+ size 13508945
Huggy/Huggy-799938.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5231ced0b9246ed3b7b5df70836006a61f741e04ccc69e09f451f8acbbbf6f0
3
+ size 2273983
Huggy/Huggy-799938.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dcad4664ae88a6fbbd22a42aa735db0de6dabfacff2e95793fab33c369ed3584
3
+ size 13508945
Huggy/Huggy-999969.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:289eda62ecf90d8b95d5792ffe326ad9365cf9d5a590d9652ca12732983f5b11
3
+ size 2273983
Huggy/Huggy-999969.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6f7a8baa25ee0195f3044af959e8442c57e5e02e757ac435fa7e4d48714c8ae
3
+ size 13508945
Huggy/checkpoint.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4c9021d2a2cbab60aa966dbaab73eb57481610fb41f0efb0633c4e334aff40be
3
- size 13505250
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:654ef81d97c0282175a2426e4fe22c55fef3b66dc5b598a539b6c8101e2632ab
3
+ size 13508779
Huggy/events.out.tfevents.1696951108.8742e47e82d7.5733.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57395899053b0df18a8fa210a7f0f5531962349890b4d2b90901c3d8557d189a
3
+ size 246433
README.md CHANGED
@@ -1,31 +1,35 @@
1
-
2
  ---
3
- tags:
4
- - unity-ml-agents
5
- - ml-agents
6
- - deep-reinforcement-learning
7
- - reinforcement-learning
8
- - ML-Agents-Huggy
9
- library_name: ml-agents
10
  ---
11
-
12
  # **ppo** Agent playing **Huggy**
13
- This is a trained model of a **ppo** agent playing **Huggy** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
14
-
 
15
  ## Usage (with ML-Agents)
16
- The Documentation: https://github.com/huggingface/ml-agents#get-started
17
- We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
18
 
 
 
 
 
 
19
 
20
  ### Resume the training
21
- ```
22
  mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
23
  ```
 
24
  ### Watch your Agent play
25
- You can watch your agent **playing directly in your browser:**.
26
-
27
- 1. Go to https://huggingface.co/spaces/unity/ML-Agents-Huggy
28
- 2. Step 1: Write your model_id: micheljperez/ppo-Huggy
29
  3. Step 2: Select your *.nn /*.onnx file
30
  4. Click on Watch the agent play 👀
31
 
 
 
1
  ---
2
+ library_name: ml-agents
3
+ tags:
4
+ - Huggy
5
+ - deep-reinforcement-learning
6
+ - reinforcement-learning
7
+ - ML-Agents-Huggy
 
8
  ---
9
+
10
  # **ppo** Agent playing **Huggy**
11
+ This is a trained model of a **ppo** agent playing **Huggy**
12
+ using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
13
+
14
  ## Usage (with ML-Agents)
15
+ The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/
 
16
 
17
+ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
18
+ - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your
19
+ browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction
20
+ - A *longer tutorial* to understand how works ML-Agents:
21
+ https://huggingface.co/learn/deep-rl-course/unit5/introduction
22
 
23
  ### Resume the training
24
+ ```bash
25
  mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
26
  ```
27
+
28
  ### Watch your Agent play
29
+ You can watch your agent **playing directly in your browser**
30
+
31
+ 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity
32
+ 2. Step 1: Find your model_id: micheljperez/ppo-Huggy
33
  3. Step 2: Select your *.nn /*.onnx file
34
  4. Click on Watch the agent play 👀
35
 
config.json CHANGED
@@ -1 +1 @@
1
- {"default_settings": null, "behaviors": {"Huggy": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 2048, "buffer_size": 20480, "learning_rate": 0.0003, "beta": 0.005, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "network_settings": {"normalize": true, "hidden_units": 512, "num_layers": 3, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.995, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 15, "checkpoint_interval": 200000, "max_steps": 2000000, "time_horizon": 1000, "summary_freq": 50000, "threaded": false, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./trained-envs-executables/linux/Huggy/Huggy", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true}, "environment_parameters": null, "checkpoint_settings": {"run_id": "Huggy", "initialize_from": null, "load_model": false, "resume": false, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
 
1
+ {"default_settings": null, "behaviors": {"Huggy": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 2048, "buffer_size": 20480, "learning_rate": 0.0003, "beta": 0.005, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "shared_critic": false, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "checkpoint_interval": 200000, "network_settings": {"normalize": true, "hidden_units": 512, "num_layers": 3, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.995, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 15, "even_checkpoints": false, "max_steps": 2000000, "time_horizon": 1000, "summary_freq": 50000, "threaded": false, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./trained-envs-executables/linux/Huggy/Huggy", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "timeout_wait": 60, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true}, "environment_parameters": null, "checkpoint_settings": {"run_id": "Huggy", "initialize_from": null, "load_model": false, "resume": false, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
configuration.yaml CHANGED
@@ -10,9 +10,11 @@ behaviors:
10
  epsilon: 0.2
11
  lambd: 0.95
12
  num_epoch: 3
 
13
  learning_rate_schedule: linear
14
  beta_schedule: linear
15
  epsilon_schedule: linear
 
16
  network_settings:
17
  normalize: true
18
  hidden_units: 512
@@ -35,7 +37,7 @@ behaviors:
35
  deterministic: false
36
  init_path: null
37
  keep_checkpoints: 15
38
- checkpoint_interval: 200000
39
  max_steps: 2000000
40
  time_horizon: 1000
41
  summary_freq: 50000
@@ -48,6 +50,7 @@ env_settings:
48
  base_port: 5005
49
  num_envs: 1
50
  num_areas: 1
 
51
  seed: -1
52
  max_lifetime_restarts: 10
53
  restarts_rate_limit_n: 1
 
10
  epsilon: 0.2
11
  lambd: 0.95
12
  num_epoch: 3
13
+ shared_critic: false
14
  learning_rate_schedule: linear
15
  beta_schedule: linear
16
  epsilon_schedule: linear
17
+ checkpoint_interval: 200000
18
  network_settings:
19
  normalize: true
20
  hidden_units: 512
 
37
  deterministic: false
38
  init_path: null
39
  keep_checkpoints: 15
40
+ even_checkpoints: false
41
  max_steps: 2000000
42
  time_horizon: 1000
43
  summary_freq: 50000
 
50
  base_port: 5005
51
  num_envs: 1
52
  num_areas: 1
53
+ timeout_wait: 60
54
  seed: -1
55
  max_lifetime_restarts: 10
56
  restarts_rate_limit_n: 1
run_logs/Player-0.log CHANGED
@@ -13,28 +13,28 @@ NullGfxDevice:
13
  Version: NULL 1.0 [1.0]
14
  Renderer: Null Device
15
  Vendor: Unity Technologies
16
- ALSA lib confmisc.c:767:(parse_card) cannot find card '0'
17
- ALSA lib conf.c:4732:(_snd_config_evaluate) function snd_func_card_driver returned error: No such file or directory
18
- ALSA lib confmisc.c:392:(snd_func_concat) error evaluating strings
19
- ALSA lib conf.c:4732:(_snd_config_evaluate) function snd_func_concat returned error: No such file or directory
20
- ALSA lib confmisc.c:1246:(snd_func_refer) error evaluating name
21
- ALSA lib conf.c:4732:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
22
- ALSA lib conf.c:5220:(snd_config_expand) Evaluate error: No such file or directory
23
- ALSA lib pcm.c:2642:(snd_pcm_open_noupdate) Unknown PCM default
24
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
25
  Forced to initialize FMOD to to the device driver's system output rate 48000, this may impact performance and/or give inconsistent experiences compared to selected sample rate 48000
26
- ALSA lib confmisc.c:767:(parse_card) cannot find card '0'
27
- ALSA lib conf.c:4732:(_snd_config_evaluate) function snd_func_card_driver returned error: No such file or directory
28
- ALSA lib confmisc.c:392:(snd_func_concat) error evaluating strings
29
- ALSA lib conf.c:4732:(_snd_config_evaluate) function snd_func_concat returned error: No such file or directory
30
- ALSA lib confmisc.c:1246:(snd_func_refer) error evaluating name
31
- ALSA lib conf.c:4732:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
32
- ALSA lib conf.c:5220:(snd_config_expand) Evaluate error: No such file or directory
33
- ALSA lib pcm.c:2642:(snd_pcm_open_noupdate) Unknown PCM default
34
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
35
  FMOD initialized on nosound output
36
  Begin MonoManager ReloadAssembly
37
- - Completed reload, in 0.090 seconds
38
  ERROR: Shader Hidden/Universal Render Pipeline/Blit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
39
  ERROR: Shader Hidden/Universal Render Pipeline/CopyDepth shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
40
  ERROR: Shader Hidden/Universal Render Pipeline/ScreenSpaceShadows shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
@@ -78,14 +78,14 @@ ERROR: Shader Universal Render Pipeline/Lit shader is not supported on this GPU
78
  WARNING: Shader Unsupported: 'Universal Render Pipeline/Lit' - All subshaders removed
79
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
80
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
81
- UnloadTime: 0.798761 ms
82
  requesting resize 84 x 84
83
  Setting up 1 worker threads for Enlighten.
84
  Memory Statistics:
85
  [ALLOC_TEMP_TLS] TLS Allocator
86
  StackAllocators :
87
  [ALLOC_TEMP_MAIN]
88
- Peak usage frame count: [4.0 KB-8.0 KB]: 26752 frames, [2.0 MB-4.0 MB]: 1 frames
89
  Initial Block Size 4.0 MB
90
  Current Block Size 4.0 MB
91
  Peak Allocated Bytes 3.6 MB
@@ -93,7 +93,7 @@ Memory Statistics:
93
  [ALLOC_TEMP_Loading.AsyncRead]
94
  Initial Block Size 64.0 KB
95
  Current Block Size 64.0 KB
96
- Peak Allocated Bytes 192 B
97
  Overflow Count 0
98
  [ALLOC_TEMP_Loading.PreloadManager]
99
  Initial Block Size 256.0 KB
@@ -201,22 +201,22 @@ Memory Statistics:
201
  Peak Allocated Bytes 0 B
202
  Overflow Count 0
203
  [ALLOC_DEFAULT] Dual Thread Allocator
204
- Peak main deferred allocation count 108
205
  [ALLOC_BUCKET]
206
  Large Block size 4.0 MB
207
  Used Block count 1
208
  Peak Allocated bytes 1.4 MB
209
  [ALLOC_DEFAULT_MAIN]
210
- Peak usage frame count: [16.0 MB-32.0 MB]: 26753 frames
211
  Requested Block Size 16.0 MB
212
  Peak Block count 1
213
- Peak Allocated memory 23.7 MB
214
  Peak Large allocation bytes 16.0 MB
215
  [ALLOC_DEFAULT_THREAD]
216
- Peak usage frame count: [2.0 MB-4.0 MB]: 26753 frames
217
  Requested Block Size 16.0 MB
218
  Peak Block count 1
219
- Peak Allocated memory 2.7 MB
220
  Peak Large allocation bytes 0 B
221
  [ALLOC_TEMP_JOB_1_FRAME]
222
  Initial Block Size 2.0 MB
@@ -245,13 +245,13 @@ Memory Statistics:
245
  Used Block count 1
246
  Peak Allocated bytes 1.4 MB
247
  [ALLOC_GFX_MAIN]
248
- Peak usage frame count: [32.0 KB-64.0 KB]: 26752 frames, [64.0 KB-128.0 KB]: 1 frames
249
  Requested Block Size 16.0 MB
250
  Peak Block count 1
251
  Peak Allocated memory 65.6 KB
252
  Peak Large allocation bytes 0 B
253
  [ALLOC_GFX_THREAD]
254
- Peak usage frame count: [64.0 KB-128.0 KB]: 26753 frames
255
  Requested Block Size 16.0 MB
256
  Peak Block count 1
257
  Peak Allocated memory 81.8 KB
@@ -263,13 +263,13 @@ Memory Statistics:
263
  Used Block count 1
264
  Peak Allocated bytes 1.4 MB
265
  [ALLOC_CACHEOBJECTS_MAIN]
266
- Peak usage frame count: [1.0 MB-2.0 MB]: 26752 frames, [16.0 MB-32.0 MB]: 1 frames
267
  Requested Block Size 4.0 MB
268
  Peak Block count 2
269
  Peak Allocated memory 30.6 MB
270
  Peak Large allocation bytes 24.9 MB
271
  [ALLOC_CACHEOBJECTS_THREAD]
272
- Peak usage frame count: [0.5 MB-1.0 MB]: 26752 frames, [2.0 MB-4.0 MB]: 1 frames
273
  Requested Block Size 4.0 MB
274
  Peak Block count 1
275
  Peak Allocated memory 2.6 MB
@@ -281,13 +281,13 @@ Memory Statistics:
281
  Used Block count 1
282
  Peak Allocated bytes 1.4 MB
283
  [ALLOC_TYPETREE_MAIN]
284
- Peak usage frame count: [0-1.0 KB]: 26753 frames
285
  Requested Block Size 2.0 MB
286
  Peak Block count 1
287
  Peak Allocated memory 1.0 KB
288
  Peak Large allocation bytes 0 B
289
  [ALLOC_TYPETREE_THREAD]
290
- Peak usage frame count: [4.0 KB-8.0 KB]: 26753 frames
291
  Requested Block Size 2.0 MB
292
  Peak Block count 1
293
  Peak Allocated memory 7.3 KB
 
13
  Version: NULL 1.0 [1.0]
14
  Renderer: Null Device
15
  Vendor: Unity Technologies
16
+ ALSA lib confmisc.c:855:(parse_card) cannot find card '0'
17
+ ALSA lib conf.c:5178:(_snd_config_evaluate) function snd_func_card_inum returned error: No such file or directory
18
+ ALSA lib confmisc.c:422:(snd_func_concat) error evaluating strings
19
+ ALSA lib conf.c:5178:(_snd_config_evaluate) function snd_func_concat returned error: No such file or directory
20
+ ALSA lib confmisc.c:1334:(snd_func_refer) error evaluating name
21
+ ALSA lib conf.c:5178:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
22
+ ALSA lib conf.c:5701:(snd_config_expand) Evaluate error: No such file or directory
23
+ ALSA lib pcm.c:2664:(snd_pcm_open_noupdate) Unknown PCM default
24
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
25
  Forced to initialize FMOD to to the device driver's system output rate 48000, this may impact performance and/or give inconsistent experiences compared to selected sample rate 48000
26
+ ALSA lib confmisc.c:855:(parse_card) cannot find card '0'
27
+ ALSA lib conf.c:5178:(_snd_config_evaluate) function snd_func_card_inum returned error: No such file or directory
28
+ ALSA lib confmisc.c:422:(snd_func_concat) error evaluating strings
29
+ ALSA lib conf.c:5178:(_snd_config_evaluate) function snd_func_concat returned error: No such file or directory
30
+ ALSA lib confmisc.c:1334:(snd_func_refer) error evaluating name
31
+ ALSA lib conf.c:5178:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
32
+ ALSA lib conf.c:5701:(snd_config_expand) Evaluate error: No such file or directory
33
+ ALSA lib pcm.c:2664:(snd_pcm_open_noupdate) Unknown PCM default
34
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
35
  FMOD initialized on nosound output
36
  Begin MonoManager ReloadAssembly
37
+ - Completed reload, in 0.087 seconds
38
  ERROR: Shader Hidden/Universal Render Pipeline/Blit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
39
  ERROR: Shader Hidden/Universal Render Pipeline/CopyDepth shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
40
  ERROR: Shader Hidden/Universal Render Pipeline/ScreenSpaceShadows shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
 
78
  WARNING: Shader Unsupported: 'Universal Render Pipeline/Lit' - All subshaders removed
79
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
80
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
81
+ UnloadTime: 0.697067 ms
82
  requesting resize 84 x 84
83
  Setting up 1 worker threads for Enlighten.
84
  Memory Statistics:
85
  [ALLOC_TEMP_TLS] TLS Allocator
86
  StackAllocators :
87
  [ALLOC_TEMP_MAIN]
88
+ Peak usage frame count: [4.0 KB-8.0 KB]: 26753 frames, [2.0 MB-4.0 MB]: 1 frames
89
  Initial Block Size 4.0 MB
90
  Current Block Size 4.0 MB
91
  Peak Allocated Bytes 3.6 MB
 
93
  [ALLOC_TEMP_Loading.AsyncRead]
94
  Initial Block Size 64.0 KB
95
  Current Block Size 64.0 KB
96
+ Peak Allocated Bytes 144 B
97
  Overflow Count 0
98
  [ALLOC_TEMP_Loading.PreloadManager]
99
  Initial Block Size 256.0 KB
 
201
  Peak Allocated Bytes 0 B
202
  Overflow Count 0
203
  [ALLOC_DEFAULT] Dual Thread Allocator
204
+ Peak main deferred allocation count 313
205
  [ALLOC_BUCKET]
206
  Large Block size 4.0 MB
207
  Used Block count 1
208
  Peak Allocated bytes 1.4 MB
209
  [ALLOC_DEFAULT_MAIN]
210
+ Peak usage frame count: [16.0 MB-32.0 MB]: 26754 frames
211
  Requested Block Size 16.0 MB
212
  Peak Block count 1
213
+ Peak Allocated memory 24.2 MB
214
  Peak Large allocation bytes 16.0 MB
215
  [ALLOC_DEFAULT_THREAD]
216
+ Peak usage frame count: [1.0 MB-2.0 MB]: 26753 frames, [2.0 MB-4.0 MB]: 1 frames
217
  Requested Block Size 16.0 MB
218
  Peak Block count 1
219
+ Peak Allocated memory 2.3 MB
220
  Peak Large allocation bytes 0 B
221
  [ALLOC_TEMP_JOB_1_FRAME]
222
  Initial Block Size 2.0 MB
 
245
  Used Block count 1
246
  Peak Allocated bytes 1.4 MB
247
  [ALLOC_GFX_MAIN]
248
+ Peak usage frame count: [32.0 KB-64.0 KB]: 26753 frames, [64.0 KB-128.0 KB]: 1 frames
249
  Requested Block Size 16.0 MB
250
  Peak Block count 1
251
  Peak Allocated memory 65.6 KB
252
  Peak Large allocation bytes 0 B
253
  [ALLOC_GFX_THREAD]
254
+ Peak usage frame count: [64.0 KB-128.0 KB]: 26754 frames
255
  Requested Block Size 16.0 MB
256
  Peak Block count 1
257
  Peak Allocated memory 81.8 KB
 
263
  Used Block count 1
264
  Peak Allocated bytes 1.4 MB
265
  [ALLOC_CACHEOBJECTS_MAIN]
266
+ Peak usage frame count: [1.0 MB-2.0 MB]: 26753 frames, [16.0 MB-32.0 MB]: 1 frames
267
  Requested Block Size 4.0 MB
268
  Peak Block count 2
269
  Peak Allocated memory 30.6 MB
270
  Peak Large allocation bytes 24.9 MB
271
  [ALLOC_CACHEOBJECTS_THREAD]
272
+ Peak usage frame count: [0.5 MB-1.0 MB]: 26753 frames, [2.0 MB-4.0 MB]: 1 frames
273
  Requested Block Size 4.0 MB
274
  Peak Block count 1
275
  Peak Allocated memory 2.6 MB
 
281
  Used Block count 1
282
  Peak Allocated bytes 1.4 MB
283
  [ALLOC_TYPETREE_MAIN]
284
+ Peak usage frame count: [0-1.0 KB]: 26754 frames
285
  Requested Block Size 2.0 MB
286
  Peak Block count 1
287
  Peak Allocated memory 1.0 KB
288
  Peak Large allocation bytes 0 B
289
  [ALLOC_TYPETREE_THREAD]
290
+ Peak usage frame count: [4.0 KB-8.0 KB]: 26754 frames
291
  Requested Block Size 2.0 MB
292
  Peak Block count 1
293
  Peak Allocated memory 7.3 KB
run_logs/timers.json CHANGED
@@ -2,135 +2,135 @@
2
  "name": "root",
3
  "gauges": {
4
  "Huggy.Policy.Entropy.mean": {
5
- "value": 1.4014652967453003,
6
- "min": 1.4014652967453003,
7
- "max": 1.427881121635437,
8
  "count": 40
9
  },
10
  "Huggy.Policy.Entropy.sum": {
11
- "value": 70493.703125,
12
- "min": 69124.515625,
13
- "max": 77285.28125,
14
  "count": 40
15
  },
16
  "Huggy.Environment.EpisodeLength.mean": {
17
- "value": 76.02153846153846,
18
- "min": 71.00287769784173,
19
- "max": 412.26446280991735,
20
  "count": 40
21
  },
22
  "Huggy.Environment.EpisodeLength.sum": {
23
- "value": 49414.0,
24
- "min": 49283.0,
25
- "max": 50157.0,
26
  "count": 40
27
  },
28
  "Huggy.Step.mean": {
29
- "value": 1999982.0,
30
- "min": 49663.0,
31
- "max": 1999982.0,
32
  "count": 40
33
  },
34
  "Huggy.Step.sum": {
35
- "value": 1999982.0,
36
- "min": 49663.0,
37
- "max": 1999982.0,
38
  "count": 40
39
  },
40
  "Huggy.Policy.ExtrinsicValueEstimate.mean": {
41
- "value": 2.491381883621216,
42
- "min": 0.05321761593222618,
43
- "max": 2.535512924194336,
44
  "count": 40
45
  },
46
  "Huggy.Policy.ExtrinsicValueEstimate.sum": {
47
- "value": 1619.398193359375,
48
- "min": 6.386114120483398,
49
- "max": 1732.599365234375,
50
  "count": 40
51
  },
52
  "Huggy.Environment.CumulativeReward.mean": {
53
- "value": 3.859221896850146,
54
- "min": 2.0314500451087953,
55
- "max": 4.025593738311277,
56
  "count": 40
57
  },
58
  "Huggy.Environment.CumulativeReward.sum": {
59
- "value": 2508.4942329525948,
60
- "min": 243.77400541305542,
61
- "max": 2689.615422785282,
62
  "count": 40
63
  },
64
  "Huggy.Policy.ExtrinsicReward.mean": {
65
- "value": 3.859221896850146,
66
- "min": 2.0314500451087953,
67
- "max": 4.025593738311277,
68
  "count": 40
69
  },
70
  "Huggy.Policy.ExtrinsicReward.sum": {
71
- "value": 2508.4942329525948,
72
- "min": 243.77400541305542,
73
- "max": 2689.615422785282,
74
  "count": 40
75
  },
76
  "Huggy.Losses.PolicyLoss.mean": {
77
- "value": 0.015274040847240636,
78
- "min": 0.014337154147647703,
79
- "max": 0.022562845268597205,
80
  "count": 40
81
  },
82
  "Huggy.Losses.PolicyLoss.sum": {
83
- "value": 0.04582212254172191,
84
- "min": 0.030435713407496224,
85
- "max": 0.05985438954157871,
86
  "count": 40
87
  },
88
  "Huggy.Losses.ValueLoss.mean": {
89
- "value": 0.05871611125767231,
90
- "min": 0.02263389518484473,
91
- "max": 0.06464658478895824,
92
  "count": 40
93
  },
94
  "Huggy.Losses.ValueLoss.sum": {
95
- "value": 0.17614833377301692,
96
- "min": 0.04526779036968946,
97
- "max": 0.17614833377301692,
98
  "count": 40
99
  },
100
  "Huggy.Policy.LearningRate.mean": {
101
- "value": 3.9006486998166695e-06,
102
- "min": 3.9006486998166695e-06,
103
- "max": 0.00029530425156524994,
104
  "count": 40
105
  },
106
  "Huggy.Policy.LearningRate.sum": {
107
- "value": 1.1701946099450008e-05,
108
- "min": 1.1701946099450008e-05,
109
- "max": 0.0008438560687146499,
110
  "count": 40
111
  },
112
  "Huggy.Policy.Epsilon.mean": {
113
- "value": 0.10130018333333335,
114
- "min": 0.10130018333333335,
115
- "max": 0.19843475000000005,
116
  "count": 40
117
  },
118
  "Huggy.Policy.Epsilon.sum": {
119
- "value": 0.30390055000000005,
120
- "min": 0.20773025,
121
- "max": 0.5812853500000001,
122
  "count": 40
123
  },
124
  "Huggy.Policy.Beta.mean": {
125
- "value": 7.487914833333338e-05,
126
- "min": 7.487914833333338e-05,
127
- "max": 0.004921894025,
128
  "count": 40
129
  },
130
  "Huggy.Policy.Beta.sum": {
131
- "value": 0.00022463744500000014,
132
- "min": 0.00022463744500000014,
133
- "max": 0.014066138965000002,
134
  "count": 40
135
  },
136
  "Huggy.IsTraining.mean": {
@@ -148,74 +148,67 @@
148
  },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
- "start_time_seconds": "1677608027",
152
- "python_version": "3.8.10 (default, Nov 14 2022, 12:59:47) \n[GCC 9.4.0]",
153
  "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics",
154
- "mlagents_version": "0.29.0.dev0",
155
- "mlagents_envs_version": "0.29.0.dev0",
156
  "communication_protocol_version": "1.5.0",
157
- "pytorch_version": "1.8.1+cu102",
158
- "numpy_version": "1.22.4",
159
- "end_time_seconds": "1677610434"
160
  },
161
- "total": 2406.9226794270003,
162
  "count": 1,
163
- "self": 0.4445138000000952,
164
  "children": {
165
  "run_training.setup": {
166
- "total": 0.12011708199997884,
167
  "count": 1,
168
- "self": 0.12011708199997884
169
  },
170
  "TrainerController.start_learning": {
171
- "total": 2406.358048545,
172
  "count": 1,
173
- "self": 4.258204464015307,
174
  "children": {
175
  "TrainerController._reset_env": {
176
- "total": 10.087536569999997,
177
  "count": 1,
178
- "self": 10.087536569999997
179
  },
180
  "TrainerController.advance": {
181
- "total": 2391.8980703499847,
182
- "count": 233484,
183
- "self": 4.460104079784287,
184
  "children": {
185
  "env_step": {
186
- "total": 1858.03990899415,
187
- "count": 233484,
188
- "self": 1560.4666971691663,
189
  "children": {
190
  "SubprocessEnvManager._take_step": {
191
- "total": 294.76814985495935,
192
- "count": 233484,
193
- "self": 16.276135231838225,
194
  "children": {
195
  "TorchPolicy.evaluate": {
196
- "total": 278.4920146231211,
197
- "count": 222942,
198
- "self": 70.52789814602488,
199
- "children": {
200
- "TorchPolicy.sample_actions": {
201
- "total": 207.96411647709624,
202
- "count": 222942,
203
- "self": 207.96411647709624
204
- }
205
- }
206
  }
207
  }
208
  },
209
  "workers": {
210
- "total": 2.805061970024326,
211
- "count": 233484,
212
  "self": 0.0,
213
  "children": {
214
  "worker_root": {
215
- "total": 2397.999344846901,
216
- "count": 233484,
217
  "is_parallel": true,
218
- "self": 1129.1211434049176,
219
  "children": {
220
  "run_training.setup": {
221
  "total": 0.0,
@@ -224,48 +217,48 @@
224
  "self": 0.0,
225
  "children": {
226
  "steps_from_proto": {
227
- "total": 0.0009612540000034642,
228
  "count": 1,
229
  "is_parallel": true,
230
- "self": 0.0003112280001005274,
231
  "children": {
232
  "_process_rank_one_or_two_observation": {
233
- "total": 0.0006500259999029367,
234
  "count": 2,
235
  "is_parallel": true,
236
- "self": 0.0006500259999029367
237
  }
238
  }
239
  },
240
  "UnityEnvironment.step": {
241
- "total": 0.028854474999889135,
242
  "count": 1,
243
  "is_parallel": true,
244
- "self": 0.000299982999877102,
245
  "children": {
246
  "UnityEnvironment._generate_step_input": {
247
- "total": 0.00021390400002019305,
248
  "count": 1,
249
  "is_parallel": true,
250
- "self": 0.00021390400002019305
251
  },
252
  "communicator.exchange": {
253
- "total": 0.027621409999937896,
254
  "count": 1,
255
  "is_parallel": true,
256
- "self": 0.027621409999937896
257
  },
258
  "steps_from_proto": {
259
- "total": 0.0007191780000539438,
260
  "count": 1,
261
  "is_parallel": true,
262
- "self": 0.0002631540000948007,
263
  "children": {
264
  "_process_rank_one_or_two_observation": {
265
- "total": 0.0004560239999591431,
266
  "count": 2,
267
  "is_parallel": true,
268
- "self": 0.0004560239999591431
269
  }
270
  }
271
  }
@@ -274,34 +267,34 @@
274
  }
275
  },
276
  "UnityEnvironment.step": {
277
- "total": 1268.8782014419833,
278
- "count": 233483,
279
  "is_parallel": true,
280
- "self": 39.23732533685984,
281
  "children": {
282
  "UnityEnvironment._generate_step_input": {
283
- "total": 80.49768388703205,
284
- "count": 233483,
285
  "is_parallel": true,
286
- "self": 80.49768388703205
287
  },
288
  "communicator.exchange": {
289
- "total": 1056.9094856630145,
290
- "count": 233483,
291
  "is_parallel": true,
292
- "self": 1056.9094856630145
293
  },
294
  "steps_from_proto": {
295
- "total": 92.23370655507688,
296
- "count": 233483,
297
  "is_parallel": true,
298
- "self": 37.41109621402484,
299
  "children": {
300
  "_process_rank_one_or_two_observation": {
301
- "total": 54.82261034105204,
302
- "count": 466966,
303
  "is_parallel": true,
304
- "self": 54.82261034105204
305
  }
306
  }
307
  }
@@ -314,31 +307,31 @@
314
  }
315
  },
316
  "trainer_advance": {
317
- "total": 529.3980572760504,
318
- "count": 233484,
319
- "self": 6.3370501321692245,
320
  "children": {
321
  "process_trajectory": {
322
- "total": 167.7143627108842,
323
- "count": 233484,
324
- "self": 166.50670340888382,
325
  "children": {
326
  "RLTrainer._checkpoint": {
327
- "total": 1.2076593020003656,
328
  "count": 10,
329
- "self": 1.2076593020003656
330
  }
331
  }
332
  },
333
  "_update_policy": {
334
- "total": 355.346644432997,
335
  "count": 97,
336
- "self": 297.7006297049936,
337
  "children": {
338
  "TorchPPOOptimizer.update": {
339
- "total": 57.646014728003365,
340
  "count": 2910,
341
- "self": 57.646014728003365
342
  }
343
  }
344
  }
@@ -347,19 +340,19 @@
347
  }
348
  },
349
  "trainer_threads": {
350
- "total": 8.340002750628628e-07,
351
  "count": 1,
352
- "self": 8.340002750628628e-07
353
  },
354
  "TrainerController._save_models": {
355
- "total": 0.11423632700007147,
356
  "count": 1,
357
- "self": 0.002008240000122896,
358
  "children": {
359
  "RLTrainer._checkpoint": {
360
- "total": 0.11222808699994857,
361
  "count": 1,
362
- "self": 0.11222808699994857
363
  }
364
  }
365
  }
 
2
  "name": "root",
3
  "gauges": {
4
  "Huggy.Policy.Entropy.mean": {
5
+ "value": 1.403228998184204,
6
+ "min": 1.403228998184204,
7
+ "max": 1.4289727210998535,
8
  "count": 40
9
  },
10
  "Huggy.Policy.Entropy.sum": {
11
+ "value": 70002.8828125,
12
+ "min": 69019.6796875,
13
+ "max": 77445.3203125,
14
  "count": 40
15
  },
16
  "Huggy.Environment.EpisodeLength.mean": {
17
+ "value": 69.04341736694678,
18
+ "min": 69.04341736694678,
19
+ "max": 391.3828125,
20
  "count": 40
21
  },
22
  "Huggy.Environment.EpisodeLength.sum": {
23
+ "value": 49297.0,
24
+ "min": 48641.0,
25
+ "max": 50097.0,
26
  "count": 40
27
  },
28
  "Huggy.Step.mean": {
29
+ "value": 1999979.0,
30
+ "min": 49617.0,
31
+ "max": 1999979.0,
32
  "count": 40
33
  },
34
  "Huggy.Step.sum": {
35
+ "value": 1999979.0,
36
+ "min": 49617.0,
37
+ "max": 1999979.0,
38
  "count": 40
39
  },
40
  "Huggy.Policy.ExtrinsicValueEstimate.mean": {
41
+ "value": 2.529850721359253,
42
+ "min": 0.09101362526416779,
43
+ "max": 2.562394857406616,
44
  "count": 40
45
  },
46
  "Huggy.Policy.ExtrinsicValueEstimate.sum": {
47
+ "value": 1806.3133544921875,
48
+ "min": 11.558730125427246,
49
+ "max": 1806.3133544921875,
50
  "count": 40
51
  },
52
  "Huggy.Environment.CumulativeReward.mean": {
53
+ "value": 3.9115888817804536,
54
+ "min": 1.9256142039937296,
55
+ "max": 4.080791214987032,
56
  "count": 40
57
  },
58
  "Huggy.Environment.CumulativeReward.sum": {
59
+ "value": 2792.8744615912437,
60
+ "min": 244.55300390720367,
61
+ "max": 2823.9075207710266,
62
  "count": 40
63
  },
64
  "Huggy.Policy.ExtrinsicReward.mean": {
65
+ "value": 3.9115888817804536,
66
+ "min": 1.9256142039937296,
67
+ "max": 4.080791214987032,
68
  "count": 40
69
  },
70
  "Huggy.Policy.ExtrinsicReward.sum": {
71
+ "value": 2792.8744615912437,
72
+ "min": 244.55300390720367,
73
+ "max": 2823.9075207710266,
74
  "count": 40
75
  },
76
  "Huggy.Losses.PolicyLoss.mean": {
77
+ "value": 0.017578730093858515,
78
+ "min": 0.013388852636368635,
79
+ "max": 0.020930114932222445,
80
  "count": 40
81
  },
82
  "Huggy.Losses.PolicyLoss.sum": {
83
+ "value": 0.05273619028157554,
84
+ "min": 0.02677770527273727,
85
+ "max": 0.059735836560139435,
86
  "count": 40
87
  },
88
  "Huggy.Losses.ValueLoss.mean": {
89
+ "value": 0.0627057037419743,
90
+ "min": 0.023507859713087477,
91
+ "max": 0.0627057037419743,
92
  "count": 40
93
  },
94
  "Huggy.Losses.ValueLoss.sum": {
95
+ "value": 0.1881171112259229,
96
+ "min": 0.047015719426174954,
97
+ "max": 0.1881171112259229,
98
  "count": 40
99
  },
100
  "Huggy.Policy.LearningRate.mean": {
101
+ "value": 3.958048680683332e-06,
102
+ "min": 3.958048680683332e-06,
103
+ "max": 0.000295338826553725,
104
  "count": 40
105
  },
106
  "Huggy.Policy.LearningRate.sum": {
107
+ "value": 1.1874146042049995e-05,
108
+ "min": 1.1874146042049995e-05,
109
+ "max": 0.0008438790187069998,
110
  "count": 40
111
  },
112
  "Huggy.Policy.Epsilon.mean": {
113
+ "value": 0.10131931666666667,
114
+ "min": 0.10131931666666667,
115
+ "max": 0.19844627499999995,
116
  "count": 40
117
  },
118
  "Huggy.Policy.Epsilon.sum": {
119
+ "value": 0.30395795000000003,
120
+ "min": 0.20777125000000002,
121
+ "max": 0.581293,
122
  "count": 40
123
  },
124
  "Huggy.Policy.Beta.mean": {
125
+ "value": 7.583390166666665e-05,
126
+ "min": 7.583390166666665e-05,
127
+ "max": 0.0049224691224999985,
128
  "count": 40
129
  },
130
  "Huggy.Policy.Beta.sum": {
131
+ "value": 0.00022750170499999996,
132
+ "min": 0.00022750170499999996,
133
+ "max": 0.014066520700000001,
134
  "count": 40
135
  },
136
  "Huggy.IsTraining.mean": {
 
148
  },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
+ "start_time_seconds": "1696951107",
152
+ "python_version": "3.10.12 (main, Jun 11 2023, 05:26:28) [GCC 11.4.0]",
153
  "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics",
154
+ "mlagents_version": "1.1.0.dev0",
155
+ "mlagents_envs_version": "1.1.0.dev0",
156
  "communication_protocol_version": "1.5.0",
157
+ "pytorch_version": "2.0.1+cu118",
158
+ "numpy_version": "1.21.2",
159
+ "end_time_seconds": "1696953531"
160
  },
161
+ "total": 2423.802595407,
162
  "count": 1,
163
+ "self": 0.43636329399987517,
164
  "children": {
165
  "run_training.setup": {
166
+ "total": 0.04781813099998544,
167
  "count": 1,
168
+ "self": 0.04781813099998544
169
  },
170
  "TrainerController.start_learning": {
171
+ "total": 2423.318413982,
172
  "count": 1,
173
+ "self": 4.413656039996113,
174
  "children": {
175
  "TrainerController._reset_env": {
176
+ "total": 8.021537040000112,
177
  "count": 1,
178
+ "self": 8.021537040000112
179
  },
180
  "TrainerController.advance": {
181
+ "total": 2410.784739200004,
182
+ "count": 233890,
183
+ "self": 4.560497000141822,
184
  "children": {
185
  "env_step": {
186
+ "total": 1846.5977701539052,
187
+ "count": 233890,
188
+ "self": 1531.5861030200733,
189
  "children": {
190
  "SubprocessEnvManager._take_step": {
191
+ "total": 312.20134225886727,
192
+ "count": 233890,
193
+ "self": 16.499976662820472,
194
  "children": {
195
  "TorchPolicy.evaluate": {
196
+ "total": 295.7013655960468,
197
+ "count": 222945,
198
+ "self": 295.7013655960468
 
 
 
 
 
 
 
199
  }
200
  }
201
  },
202
  "workers": {
203
+ "total": 2.8103248749646355,
204
+ "count": 233890,
205
  "self": 0.0,
206
  "children": {
207
  "worker_root": {
208
+ "total": 2415.992196399164,
209
+ "count": 233890,
210
  "is_parallel": true,
211
+ "self": 1171.2062316571828,
212
  "children": {
213
  "run_training.setup": {
214
  "total": 0.0,
 
217
  "self": 0.0,
218
  "children": {
219
  "steps_from_proto": {
220
+ "total": 0.000846455000100832,
221
  "count": 1,
222
  "is_parallel": true,
223
+ "self": 0.00024190499993892445,
224
  "children": {
225
  "_process_rank_one_or_two_observation": {
226
+ "total": 0.0006045500001619075,
227
  "count": 2,
228
  "is_parallel": true,
229
+ "self": 0.0006045500001619075
230
  }
231
  }
232
  },
233
  "UnityEnvironment.step": {
234
+ "total": 0.02694922000000588,
235
  "count": 1,
236
  "is_parallel": true,
237
+ "self": 0.0002893549999498646,
238
  "children": {
239
  "UnityEnvironment._generate_step_input": {
240
+ "total": 0.00023497400002270297,
241
  "count": 1,
242
  "is_parallel": true,
243
+ "self": 0.00023497400002270297
244
  },
245
  "communicator.exchange": {
246
+ "total": 0.025645232999977452,
247
  "count": 1,
248
  "is_parallel": true,
249
+ "self": 0.025645232999977452
250
  },
251
  "steps_from_proto": {
252
+ "total": 0.0007796580000558606,
253
  "count": 1,
254
  "is_parallel": true,
255
+ "self": 0.00023580899983244308,
256
  "children": {
257
  "_process_rank_one_or_two_observation": {
258
+ "total": 0.0005438490002234175,
259
  "count": 2,
260
  "is_parallel": true,
261
+ "self": 0.0005438490002234175
262
  }
263
  }
264
  }
 
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
+ "total": 1244.7859647419814,
271
+ "count": 233889,
272
  "is_parallel": true,
273
+ "self": 40.70974441420867,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
+ "total": 82.11496812792711,
277
+ "count": 233889,
278
  "is_parallel": true,
279
+ "self": 82.11496812792711
280
  },
281
  "communicator.exchange": {
282
+ "total": 1022.8206100708637,
283
+ "count": 233889,
284
  "is_parallel": true,
285
+ "self": 1022.8206100708637
286
  },
287
  "steps_from_proto": {
288
+ "total": 99.14064212898188,
289
+ "count": 233889,
290
  "is_parallel": true,
291
+ "self": 35.78881004893856,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
+ "total": 63.35183208004332,
295
+ "count": 467778,
296
  "is_parallel": true,
297
+ "self": 63.35183208004332
298
  }
299
  }
300
  }
 
307
  }
308
  },
309
  "trainer_advance": {
310
+ "total": 559.6264720459569,
311
+ "count": 233890,
312
+ "self": 6.173125561966344,
313
  "children": {
314
  "process_trajectory": {
315
+ "total": 152.44454011298876,
316
+ "count": 233890,
317
+ "self": 151.1038633239889,
318
  "children": {
319
  "RLTrainer._checkpoint": {
320
+ "total": 1.3406767889998719,
321
  "count": 10,
322
+ "self": 1.3406767889998719
323
  }
324
  }
325
  },
326
  "_update_policy": {
327
+ "total": 401.0088063710018,
328
  "count": 97,
329
+ "self": 339.93564413300624,
330
  "children": {
331
  "TorchPPOOptimizer.update": {
332
+ "total": 61.073162237995575,
333
  "count": 2910,
334
+ "self": 61.073162237995575
335
  }
336
  }
337
  }
 
340
  }
341
  },
342
  "trainer_threads": {
343
+ "total": 9.7599968285067e-07,
344
  "count": 1,
345
+ "self": 9.7599968285067e-07
346
  },
347
  "TrainerController._save_models": {
348
+ "total": 0.09848072600016167,
349
  "count": 1,
350
+ "self": 0.001958457000000635,
351
  "children": {
352
  "RLTrainer._checkpoint": {
353
+ "total": 0.09652226900016103,
354
  "count": 1,
355
+ "self": 0.09652226900016103
356
  }
357
  }
358
  }
run_logs/training_status.json CHANGED
@@ -2,118 +2,118 @@
2
  "Huggy": {
3
  "checkpoints": [
4
  {
5
- "steps": 199839,
6
- "file_path": "results/Huggy/Huggy/Huggy-199839.onnx",
7
- "reward": 3.267228267648641,
8
- "creation_time": 1677608273.439319,
9
  "auxillary_file_paths": [
10
- "results/Huggy/Huggy/Huggy-199839.pt"
11
  ]
12
  },
13
  {
14
- "steps": 399856,
15
- "file_path": "results/Huggy/Huggy/Huggy-399856.onnx",
16
- "reward": 4.03699004650116,
17
- "creation_time": 1677608514.0762005,
18
  "auxillary_file_paths": [
19
- "results/Huggy/Huggy/Huggy-399856.pt"
20
  ]
21
  },
22
  {
23
- "steps": 599866,
24
- "file_path": "results/Huggy/Huggy/Huggy-599866.onnx",
25
- "reward": 3.9803427024321123,
26
- "creation_time": 1677608755.036398,
27
  "auxillary_file_paths": [
28
- "results/Huggy/Huggy/Huggy-599866.pt"
29
  ]
30
  },
31
  {
32
- "steps": 799958,
33
- "file_path": "results/Huggy/Huggy/Huggy-799958.onnx",
34
- "reward": 3.97494601169411,
35
- "creation_time": 1677608993.1662903,
36
  "auxillary_file_paths": [
37
- "results/Huggy/Huggy/Huggy-799958.pt"
38
  ]
39
  },
40
  {
41
- "steps": 999993,
42
- "file_path": "results/Huggy/Huggy/Huggy-999993.onnx",
43
- "reward": 4.005736847718556,
44
- "creation_time": 1677609233.2743168,
45
  "auxillary_file_paths": [
46
- "results/Huggy/Huggy/Huggy-999993.pt"
47
  ]
48
  },
49
  {
50
- "steps": 1199965,
51
- "file_path": "results/Huggy/Huggy/Huggy-1199965.onnx",
52
- "reward": 3.940881059613339,
53
- "creation_time": 1677609475.2061155,
54
  "auxillary_file_paths": [
55
- "results/Huggy/Huggy/Huggy-1199965.pt"
56
  ]
57
  },
58
  {
59
- "steps": 1399969,
60
- "file_path": "results/Huggy/Huggy/Huggy-1399969.onnx",
61
- "reward": 3.925844192504883,
62
- "creation_time": 1677609715.3567898,
63
  "auxillary_file_paths": [
64
- "results/Huggy/Huggy/Huggy-1399969.pt"
65
  ]
66
  },
67
  {
68
- "steps": 1599902,
69
- "file_path": "results/Huggy/Huggy/Huggy-1599902.onnx",
70
- "reward": 3.940251687027159,
71
- "creation_time": 1677609952.1858134,
72
  "auxillary_file_paths": [
73
- "results/Huggy/Huggy/Huggy-1599902.pt"
74
  ]
75
  },
76
  {
77
- "steps": 1799949,
78
- "file_path": "results/Huggy/Huggy/Huggy-1799949.onnx",
79
- "reward": 3.8176064431667327,
80
- "creation_time": 1677610194.7683113,
81
  "auxillary_file_paths": [
82
- "results/Huggy/Huggy/Huggy-1799949.pt"
83
  ]
84
  },
85
  {
86
- "steps": 1999982,
87
- "file_path": "results/Huggy/Huggy/Huggy-1999982.onnx",
88
- "reward": 3.5897830616344106,
89
- "creation_time": 1677610433.8496847,
90
  "auxillary_file_paths": [
91
- "results/Huggy/Huggy/Huggy-1999982.pt"
92
  ]
93
  },
94
  {
95
- "steps": 2000065,
96
- "file_path": "results/Huggy/Huggy/Huggy-2000065.onnx",
97
- "reward": 3.5975924760867386,
98
- "creation_time": 1677610433.9677625,
99
  "auxillary_file_paths": [
100
- "results/Huggy/Huggy/Huggy-2000065.pt"
101
  ]
102
  }
103
  ],
104
  "final_checkpoint": {
105
- "steps": 2000065,
106
  "file_path": "results/Huggy/Huggy.onnx",
107
- "reward": 3.5975924760867386,
108
- "creation_time": 1677610433.9677625,
109
  "auxillary_file_paths": [
110
- "results/Huggy/Huggy/Huggy-2000065.pt"
111
  ]
112
  }
113
  },
114
  "metadata": {
115
  "stats_format_version": "0.3.0",
116
- "mlagents_version": "0.29.0.dev0",
117
- "torch_version": "1.8.1+cu102"
118
  }
119
  }
 
2
  "Huggy": {
3
  "checkpoints": [
4
  {
5
+ "steps": 199642,
6
+ "file_path": "results/Huggy/Huggy/Huggy-199642.onnx",
7
+ "reward": 3.3408852869814094,
8
+ "creation_time": 1696951351.066657,
9
  "auxillary_file_paths": [
10
+ "results/Huggy/Huggy/Huggy-199642.pt"
11
  ]
12
  },
13
  {
14
+ "steps": 399995,
15
+ "file_path": "results/Huggy/Huggy/Huggy-399995.onnx",
16
+ "reward": 3.8581132473129975,
17
+ "creation_time": 1696951588.3978288,
18
  "auxillary_file_paths": [
19
+ "results/Huggy/Huggy/Huggy-399995.pt"
20
  ]
21
  },
22
  {
23
+ "steps": 599953,
24
+ "file_path": "results/Huggy/Huggy/Huggy-599953.onnx",
25
+ "reward": 3.8165545654296875,
26
+ "creation_time": 1696951829.5304844,
27
  "auxillary_file_paths": [
28
+ "results/Huggy/Huggy/Huggy-599953.pt"
29
  ]
30
  },
31
  {
32
+ "steps": 799938,
33
+ "file_path": "results/Huggy/Huggy/Huggy-799938.onnx",
34
+ "reward": 3.9907866439548147,
35
+ "creation_time": 1696952068.3049042,
36
  "auxillary_file_paths": [
37
+ "results/Huggy/Huggy/Huggy-799938.pt"
38
  ]
39
  },
40
  {
41
+ "steps": 999969,
42
+ "file_path": "results/Huggy/Huggy/Huggy-999969.onnx",
43
+ "reward": 3.9267869674798215,
44
+ "creation_time": 1696952309.4688065,
45
  "auxillary_file_paths": [
46
+ "results/Huggy/Huggy/Huggy-999969.pt"
47
  ]
48
  },
49
  {
50
+ "steps": 1199991,
51
+ "file_path": "results/Huggy/Huggy/Huggy-1199991.onnx",
52
+ "reward": 4.183989282264266,
53
+ "creation_time": 1696952553.7304053,
54
  "auxillary_file_paths": [
55
+ "results/Huggy/Huggy/Huggy-1199991.pt"
56
  ]
57
  },
58
  {
59
+ "steps": 1399993,
60
+ "file_path": "results/Huggy/Huggy/Huggy-1399993.onnx",
61
+ "reward": 3.392833543320497,
62
+ "creation_time": 1696952800.2592373,
63
  "auxillary_file_paths": [
64
+ "results/Huggy/Huggy/Huggy-1399993.pt"
65
  ]
66
  },
67
  {
68
+ "steps": 1599947,
69
+ "file_path": "results/Huggy/Huggy/Huggy-1599947.onnx",
70
+ "reward": 3.9308800111340068,
71
+ "creation_time": 1696953041.6891658,
72
  "auxillary_file_paths": [
73
+ "results/Huggy/Huggy/Huggy-1599947.pt"
74
  ]
75
  },
76
  {
77
+ "steps": 1799952,
78
+ "file_path": "results/Huggy/Huggy/Huggy-1799952.onnx",
79
+ "reward": 3.958530896546825,
80
+ "creation_time": 1696953286.6476877,
81
  "auxillary_file_paths": [
82
+ "results/Huggy/Huggy/Huggy-1799952.pt"
83
  ]
84
  },
85
  {
86
+ "steps": 1999979,
87
+ "file_path": "results/Huggy/Huggy/Huggy-1999979.onnx",
88
+ "reward": 4.117876688171836,
89
+ "creation_time": 1696953531.0502884,
90
  "auxillary_file_paths": [
91
+ "results/Huggy/Huggy/Huggy-1999979.pt"
92
  ]
93
  },
94
  {
95
+ "steps": 2000044,
96
+ "file_path": "results/Huggy/Huggy/Huggy-2000044.onnx",
97
+ "reward": 4.121083632458088,
98
+ "creation_time": 1696953531.1514134,
99
  "auxillary_file_paths": [
100
+ "results/Huggy/Huggy/Huggy-2000044.pt"
101
  ]
102
  }
103
  ],
104
  "final_checkpoint": {
105
+ "steps": 2000044,
106
  "file_path": "results/Huggy/Huggy.onnx",
107
+ "reward": 4.121083632458088,
108
+ "creation_time": 1696953531.1514134,
109
  "auxillary_file_paths": [
110
+ "results/Huggy/Huggy/Huggy-2000044.pt"
111
  ]
112
  }
113
  },
114
  "metadata": {
115
  "stats_format_version": "0.3.0",
116
+ "mlagents_version": "1.1.0.dev0",
117
+ "torch_version": "2.0.1+cu118"
118
  }
119
  }