{ "name": "root", "gauges": { "Agent.Policy.Entropy.mean": { "value": 1.3696762323379517, "min": 1.3696684837341309, "max": 1.4317792654037476, "count": 200 }, "Agent.Policy.Entropy.sum": { "value": 8213.9482421875, "min": 6821.7373046875, "max": 9953.8876953125, "count": 200 }, "Agent.DroneBasedReforestation.TreeDropCount.mean": { "value": 0.9420289855072463, "min": 0.4444444444444444, "max": 0.9871794871794872, "count": 200 }, "Agent.DroneBasedReforestation.TreeDropCount.sum": { "value": 65.0, "min": 8.0, "max": 124.0, "count": 200 }, "Agent.DroneBasedReforestation.RechargeEnergyCount.mean": { "value": 11.695652173913043, "min": 6.0, "max": 44.8, "count": 200 }, "Agent.DroneBasedReforestation.RechargeEnergyCount.sum": { "value": 807.0, "min": 114.0, "max": 1460.0, "count": 200 }, "Agent.DroneBasedReforestation.SaveLocationCount.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.SaveLocationCount.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.OutofEnergyCount.mean": { "value": 0.057971014492753624, "min": 0.007936507936507936, "max": 0.4444444444444444, "count": 200 }, "Agent.DroneBasedReforestation.OutofEnergyCount.sum": { "value": 4.0, "min": 1.0, "max": 10.0, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceUntilTreeDrop.mean": { "value": 54.99568212205085, "min": 14.928998240718135, "max": 64.69170888264973, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceUntilTreeDrop.sum": { "value": 3794.702066421509, "min": 359.63262462615967, "max": 6308.236831665039, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeTreeDropReward.mean": { "value": 3.6662469488447127, "min": 0.9191970278819402, "max": 6.476189487705044, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeTreeDropReward.sum": { "value": 252.97103947028518, "min": 21.192142456769943, "max": 538.1379623115063, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceReward.mean": { "value": 1.6787470760552778, "min": 0.26963984171549477, "max": 2.0073165723255704, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceReward.sum": { "value": 115.83354824781418, "min": 7.631449937820435, "max": 189.3757051229477, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeNormalizedDistanceUntilTreeDrop.mean": { "value": 0.16787470779989078, "min": 0.026963984469572704, "max": 0.20073165957416808, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeNormalizedDistanceUntilTreeDrop.sum": { "value": 11.583354838192463, "min": 0.7631450071930885, "max": 18.937570460140705, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistancetoExistingTrees.mean": { "value": 62.13030809077664, "min": 27.802483320236206, "max": 80.10559981664022, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistancetoExistingTrees.sum": { "value": 4286.991258263588, "min": 500.4446997642517, "max": 9092.85719871521, "count": 200 }, "Agent.Environment.LessonNumber.difficulty.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.LessonNumber.difficulty.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.LessonNumber.task.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.LessonNumber.task.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.EpisodeLength.mean": { "value": 85.8695652173913, "min": 46.19047619047619, "max": 299.8333333333333, "count": 200 }, "Agent.Environment.EpisodeLength.sum": { "value": 5925.0, "min": 4839.0, "max": 7086.0, "count": 200 }, "Agent.Step.mean": { "value": 1199642.0, "min": 5673.0, "max": 1199642.0, "count": 200 }, "Agent.Step.sum": { "value": 1199642.0, "min": 5673.0, "max": 1199642.0, "count": 200 }, "Agent.Policy.CuriosityValueEstimate.mean": { "value": 0.19184796512126923, "min": 0.013155708089470863, "max": 1.0088117122650146, "count": 200 }, "Agent.Policy.CuriosityValueEstimate.sum": { "value": 13.237509727478027, "min": 0.394671231508255, "max": 36.34642791748047, "count": 200 }, "Agent.Policy.ExtrinsicValueEstimate.mean": { "value": 50.770591735839844, "min": 0.05967150628566742, "max": 54.52288055419922, "count": 200 }, "Agent.Policy.ExtrinsicValueEstimate.sum": { "value": 3503.1708984375, "min": 1.4917876720428467, "max": 6622.75146484375, "count": 200 }, "Agent.Environment.CumulativeReward.mean": { "value": 100.76390773317088, "min": 47.78406707950247, "max": 104.8796310018301, "count": 200 }, "Agent.Environment.CumulativeReward.sum": { "value": 6952.709633588791, "min": 956.0870161354542, "max": 13109.953875228763, "count": 200 }, "Agent.Policy.CuriosityReward.mean": { "value": 0.23652483907568714, "min": 0.0, "max": 5.984732621056693, "count": 200 }, "Agent.Policy.CuriosityReward.sum": { "value": 16.320213896222413, "min": 0.0, "max": 209.46564173698425, "count": 200 }, "Agent.Policy.ExtrinsicReward.mean": { "value": 90.68751952268076, "min": 43.00566218708021, "max": 94.3916706160307, "count": 200 }, "Agent.Policy.ExtrinsicReward.sum": { "value": 6257.438847064972, "min": 860.4783912152052, "max": 11798.958827003837, "count": 200 }, "Agent.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 200 }, "Agent.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 200 }, "Agent.Losses.PolicyLoss.mean": { "value": 0.0248545864985014, "min": 0.01224414041886727, "max": 0.03292681433958933, "count": 138 }, "Agent.Losses.PolicyLoss.sum": { "value": 0.0248545864985014, "min": 0.01224414041886727, "max": 0.03292681433958933, "count": 138 }, "Agent.Losses.ValueLoss.mean": { "value": 101.52218373616536, "min": 29.132872978846233, "max": 143.23273150126138, "count": 138 }, "Agent.Losses.ValueLoss.sum": { "value": 101.52218373616536, "min": 29.132872978846233, "max": 143.23273150126138, "count": 138 }, "Agent.Policy.LearningRate.mean": { "value": 2.1390992869999907e-06, "min": 2.1390992869999907e-06, "max": 0.00029776350074549996, "count": 138 }, "Agent.Policy.LearningRate.sum": { "value": 2.1390992869999907e-06, "min": 2.1390992869999907e-06, "max": 0.00029776350074549996, "count": 138 }, "Agent.Policy.Epsilon.mean": { "value": 0.100713, "min": 0.100713, "max": 0.1992545, "count": 138 }, "Agent.Policy.Epsilon.sum": { "value": 0.100713, "min": 0.100713, "max": 0.1992545, "count": 138 }, "Agent.Policy.Beta.mean": { "value": 4.557869999999984e-05, "min": 4.557869999999984e-05, "max": 0.00496279955, "count": 138 }, "Agent.Policy.Beta.sum": { "value": 4.557869999999984e-05, "min": 4.557869999999984e-05, "max": 0.00496279955, "count": 138 }, "Agent.Losses.CuriosityForwardLoss.mean": { "value": 0.023440313059836626, "min": 0.01844055384087066, "max": 0.6002316027879715, "count": 138 }, "Agent.Losses.CuriosityForwardLoss.sum": { "value": 0.023440313059836626, "min": 0.01844055384087066, "max": 0.6002316027879715, "count": 138 }, "Agent.Losses.CuriosityInverseLoss.mean": { "value": 2.2776452600955963, "min": 2.24655544757843, "max": 3.353024572134018, "count": 138 }, "Agent.Losses.CuriosityInverseLoss.sum": { "value": 2.2776452600955963, "min": 2.24655544757843, "max": 3.353024572134018, "count": 138 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1717367557", "python_version": "3.9.18 (main, Sep 11 2023, 14:09:26) [MSC v.1916 64 bit (AMD64)]", "command_line_arguments": "C:\\Users\\pdsie\\anaconda3\\envs\\mlagents20\\Scripts\\mlagents-learn c:/users/pdsie/documents/hivex/src/hivex/training/baseline/ml_agents/configs/mlagents/tmp/train/DroneBasedReforestation_difficulty_8_task_3_run_id_1_train.yaml --run-id=DroneBasedReforestation/train/DroneBasedReforestation_difficulty_8_task_3_run_id_1_train --base-port 5007", "mlagents_version": "0.30.0", "mlagents_envs_version": "0.30.0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.7.1+cu110", "numpy_version": "1.21.0", "end_time_seconds": "1717371457" }, "total": 3899.8439110000004, "count": 1, "self": 0.6335378000003402, "children": { "run_training.setup": { "total": 0.055460000000000065, "count": 1, "self": 0.055460000000000065 }, "TrainerController.start_learning": { "total": 3899.1549132, "count": 1, "self": 6.664154499924734, "children": { "TrainerController._reset_env": { "total": 2.0787866, "count": 1, "self": 2.0787866 }, "TrainerController.advance": { "total": 3890.247656600075, "count": 403276, "self": 6.144649200065487, "children": { "env_step": { "total": 3884.1030074000096, "count": 403276, "self": 2103.623677900101, "children": { "SubprocessEnvManager._take_step": { "total": 1776.412495199966, "count": 403276, "self": 11.736166999943407, "children": { "TorchPolicy.evaluate": { "total": 1764.6763282000227, "count": 400021, "self": 1764.6763282000227 } } }, "workers": { "total": 4.066834299942425, "count": 403276, "self": 0.0, "children": { "worker_root": { "total": 3891.0592357000355, "count": 403276, "is_parallel": true, "self": 2010.1671302000898, "children": { "steps_from_proto": { "total": 0.006613300000000155, "count": 1, "is_parallel": true, "self": 0.00010339999999997573, "children": { "_process_maybe_compressed_observation": { "total": 0.006461600000000178, "count": 2, "is_parallel": true, "self": 3.1600000000242545e-05, "children": { "_observation_to_np_array": { "total": 0.006429999999999936, "count": 3, "is_parallel": true, "self": 2.9399999999846216e-05, "children": { "process_pixels": { "total": 0.0064006000000000896, "count": 3, "is_parallel": true, "self": 0.00024199999999985344, "children": { "image_decompress": { "total": 0.006158600000000236, "count": 3, "is_parallel": true, "self": 0.006158600000000236 } } } } } } }, "_process_rank_one_or_two_observation": { "total": 4.830000000000112e-05, "count": 2, "is_parallel": true, "self": 4.830000000000112e-05 } } }, "UnityEnvironment.step": { "total": 1880.8854921999457, "count": 403276, "is_parallel": true, "self": 17.952159799998753, "children": { "UnityEnvironment._generate_step_input": { "total": 17.947403399987053, "count": 403276, "is_parallel": true, "self": 17.947403399987053 }, "communicator.exchange": { "total": 1697.2936717999964, "count": 403276, "is_parallel": true, "self": 1697.2936717999964 }, "steps_from_proto": { "total": 147.6922571999635, "count": 403276, "is_parallel": true, "self": 29.59189109999437, "children": { "_process_maybe_compressed_observation": { "total": 105.86420470010437, "count": 806552, "is_parallel": true, "self": 8.060683799925513, "children": { "_observation_to_np_array": { "total": 97.80352090017885, "count": 1210485, "is_parallel": true, "self": 7.620016500243906, "children": { "process_pixels": { "total": 90.18350439993495, "count": 1210485, "is_parallel": true, "self": 42.0744073999156, "children": { "image_decompress": { "total": 48.10909700001935, "count": 1210485, "is_parallel": true, "self": 48.10909700001935 } } } } } } }, "_process_rank_one_or_two_observation": { "total": 12.236161399864773, "count": 806552, "is_parallel": true, "self": 12.236161399864773 } } } } } } } } } } } } }, "trainer_threads": { "total": 2.6400000024295878e-05, "count": 1, "self": 2.6400000024295878e-05, "children": { "thread_root": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "trainer_advance": { "total": 3895.108201099968, "count": 196344, "is_parallel": true, "self": 5.70519819991614, "children": { "process_trajectory": { "total": 3227.662570100049, "count": 196344, "is_parallel": true, "self": 3227.240592800049, "children": { "RLTrainer._checkpoint": { "total": 0.42197729999998046, "count": 2, "is_parallel": true, "self": 0.42197729999998046 } } }, "_update_policy": { "total": 661.7404328000028, "count": 138, "is_parallel": true, "self": 447.0390076000049, "children": { "TorchPPOOptimizer.update": { "total": 214.70142519999794, "count": 3342, "is_parallel": true, "self": 214.70142519999794 } } } } } } } } }, "TrainerController._save_models": { "total": 0.16428910000013275, "count": 1, "self": 0.00740429999996195, "children": { "RLTrainer._checkpoint": { "total": 0.1568848000001708, "count": 1, "self": 0.1568848000001708 } } } } } } }