philippds's picture
Upload 12 files
f37e4b7 verified
{
"name": "root",
"gauges": {
"Agent.Policy.Entropy.mean": {
"value": 1.461394190788269,
"min": 1.4189385175704956,
"max": 1.4627084732055664,
"count": 200
},
"Agent.Policy.Entropy.sum": {
"value": 8768.365234375,
"min": 7108.82421875,
"max": 10005.6728515625,
"count": 200
},
"Agent.DroneBasedReforestation.TreeDropCount.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.TreeDropCount.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.RechargeEnergyCount.mean": {
"value": 85.73333333333333,
"min": 0.0,
"max": 522.4,
"count": 200
},
"Agent.DroneBasedReforestation.RechargeEnergyCount.sum": {
"value": 1286.0,
"min": 0.0,
"max": 7836.0,
"count": 200
},
"Agent.DroneBasedReforestation.SaveLocationCount.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.SaveLocationCount.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.OutofEnergyCount.mean": {
"value": 0.6666666666666666,
"min": 0.375,
"max": 0.7333333333333333,
"count": 200
},
"Agent.DroneBasedReforestation.OutofEnergyCount.sum": {
"value": 10.0,
"min": 7.0,
"max": 14.0,
"count": 200
},
"Agent.DroneBasedReforestation.CumulativeDistanceUntilTreeDrop.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.CumulativeDistanceUntilTreeDrop.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.CumulativeTreeDropReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.CumulativeTreeDropReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.CumulativeDistanceReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.CumulativeDistanceReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.CumulativeNormalizedDistanceUntilTreeDrop.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.CumulativeNormalizedDistanceUntilTreeDrop.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.CumulativeDistancetoExistingTrees.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.CumulativeDistancetoExistingTrees.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.DroneBasedReforestation.FurthestDistanceExplored.mean": {
"value": 146.21323369344074,
"min": 92.7341601451238,
"max": 176.34202880859374,
"count": 200
},
"Agent.DroneBasedReforestation.FurthestDistanceExplored.sum": {
"value": 2193.1985054016113,
"min": 1417.522566318512,
"max": 2846.2834649086,
"count": 200
},
"Agent.Environment.LessonNumber.difficulty.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.Environment.LessonNumber.difficulty.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.Environment.LessonNumber.task.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.Environment.LessonNumber.task.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 200
},
"Agent.Environment.EpisodeLength.mean": {
"value": 399.0,
"min": 273.375,
"max": 399.0,
"count": 200
},
"Agent.Environment.EpisodeLength.sum": {
"value": 5985.0,
"min": 4833.0,
"max": 6855.0,
"count": 200
},
"Agent.Step.mean": {
"value": 1199889.0,
"min": 5600.0,
"max": 1199889.0,
"count": 200
},
"Agent.Step.sum": {
"value": 1199889.0,
"min": 5600.0,
"max": 1199889.0,
"count": 200
},
"Agent.Policy.CuriosityValueEstimate.mean": {
"value": 0.1302863210439682,
"min": 0.031902581453323364,
"max": 1.057988166809082,
"count": 200
},
"Agent.Policy.CuriosityValueEstimate.sum": {
"value": 1.9542948007583618,
"min": 0.4466361403465271,
"max": 18.928512573242188,
"count": 200
},
"Agent.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.41766637563705444,
"min": -0.24882428348064423,
"max": 1.1357226371765137,
"count": 200
},
"Agent.Policy.ExtrinsicValueEstimate.sum": {
"value": 6.264995574951172,
"min": -4.230012893676758,
"max": 20.25455665588379,
"count": 200
},
"Agent.Environment.CumulativeReward.mean": {
"value": 8.835175029436748,
"min": -1.1326333165168763,
"max": 15.139918671920896,
"count": 200
},
"Agent.Environment.CumulativeReward.sum": {
"value": 132.5276254415512,
"min": -16.989499747753143,
"max": 277.330514757894,
"count": 200
},
"Agent.Policy.CuriosityReward.mean": {
"value": 0.493207577864329,
"min": 0.0,
"max": 18.167254384358724,
"count": 200
},
"Agent.Policy.CuriosityReward.sum": {
"value": 7.398113667964935,
"min": 0.0,
"max": 272.50881576538086,
"count": 200
},
"Agent.Policy.ExtrinsicReward.mean": {
"value": 7.95165540377299,
"min": -1.0193702220916747,
"max": 13.625924177467823,
"count": 200
},
"Agent.Policy.ExtrinsicReward.sum": {
"value": 119.27483105659485,
"min": -15.290553331375122,
"max": 249.59742173738778,
"count": 200
},
"Agent.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 200
},
"Agent.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 200
},
"Agent.Losses.PolicyLoss.mean": {
"value": 0.028155309342158336,
"min": 0.013494075605800996,
"max": 0.034375812800135463,
"count": 140
},
"Agent.Losses.PolicyLoss.sum": {
"value": 0.028155309342158336,
"min": 0.013494075605800996,
"max": 0.034375812800135463,
"count": 140
},
"Agent.Losses.ValueLoss.mean": {
"value": 8.478612542152405,
"min": 0.0005081328054075129,
"max": 8.902225154417533,
"count": 140
},
"Agent.Losses.ValueLoss.sum": {
"value": 8.478612542152405,
"min": 0.0005081328054075129,
"max": 8.902225154417533,
"count": 140
},
"Agent.Policy.LearningRate.mean": {
"value": 2.7849990750015264e-08,
"min": 2.7849990750015264e-08,
"max": 0.0002979000007,
"count": 140
},
"Agent.Policy.LearningRate.sum": {
"value": 2.7849990750015264e-08,
"min": 2.7849990750015264e-08,
"max": 0.0002979000007,
"count": 140
},
"Agent.Policy.Epsilon.mean": {
"value": 0.10000925000000001,
"min": 0.10000925000000001,
"max": 0.1993,
"count": 140
},
"Agent.Policy.Epsilon.sum": {
"value": 0.10000925000000001,
"min": 0.10000925000000001,
"max": 0.1993,
"count": 140
},
"Agent.Policy.Beta.mean": {
"value": 1.0461575000000252e-05,
"min": 1.0461575000000252e-05,
"max": 0.00496507,
"count": 140
},
"Agent.Policy.Beta.sum": {
"value": 1.0461575000000252e-05,
"min": 1.0461575000000252e-05,
"max": 0.00496507,
"count": 140
},
"Agent.Losses.CuriosityForwardLoss.mean": {
"value": 0.011916863693234822,
"min": 0.010664537665434182,
"max": 0.5835290277997652,
"count": 140
},
"Agent.Losses.CuriosityForwardLoss.sum": {
"value": 0.011916863693234822,
"min": 0.010664537665434182,
"max": 0.5835290277997652,
"count": 140
},
"Agent.Losses.CuriosityInverseLoss.mean": {
"value": 2.4481616616249084,
"min": 2.3409460683663688,
"max": 3.310828596353531,
"count": 140
},
"Agent.Losses.CuriosityInverseLoss.sum": {
"value": 2.4481616616249084,
"min": 2.3409460683663688,
"max": 3.310828596353531,
"count": 140
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1717292782",
"python_version": "3.9.18 (main, Sep 11 2023, 14:09:26) [MSC v.1916 64 bit (AMD64)]",
"command_line_arguments": "C:\\Users\\pdsie\\anaconda3\\envs\\mlagents20\\Scripts\\mlagents-learn c:/users/pdsie/documents/hivex/src/hivex/training/baseline/ml_agents/configs/mlagents/tmp/train/DroneBasedReforestation_difficulty_5_task_6_run_id_1_train.yaml --run-id=DroneBasedReforestation/train/DroneBasedReforestation_difficulty_5_task_6_run_id_1_train --base-port 5007",
"mlagents_version": "0.30.0",
"mlagents_envs_version": "0.30.0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.7.1+cu110",
"numpy_version": "1.21.0",
"end_time_seconds": "1717296310"
},
"total": 3527.6842628,
"count": 1,
"self": 0.271194800000103,
"children": {
"run_training.setup": {
"total": 0.0522089,
"count": 1,
"self": 0.0522089
},
"TrainerController.start_learning": {
"total": 3527.3608590999997,
"count": 1,
"self": 4.724687999997059,
"children": {
"TrainerController._reset_env": {
"total": 2.0738813,
"count": 1,
"self": 2.0738813
},
"TrainerController.advance": {
"total": 3520.388816000003,
"count": 401225,
"self": 4.241707900048368,
"children": {
"env_step": {
"total": 3516.1471080999545,
"count": 401225,
"self": 1496.8510475001858,
"children": {
"SubprocessEnvManager._take_step": {
"total": 2016.2752550000253,
"count": 401225,
"self": 10.122549799896206,
"children": {
"TorchPolicy.evaluate": {
"total": 2006.152705200129,
"count": 400363,
"self": 2006.152705200129
}
}
},
"workers": {
"total": 3.0208055997434866,
"count": 401225,
"self": 0.0,
"children": {
"worker_root": {
"total": 3520.4959868000665,
"count": 401225,
"is_parallel": true,
"self": 2203.313819800009,
"children": {
"steps_from_proto": {
"total": 0.006579399999999902,
"count": 1,
"is_parallel": true,
"self": 0.00010719999999975194,
"children": {
"_process_maybe_compressed_observation": {
"total": 0.006426100000000101,
"count": 2,
"is_parallel": true,
"self": 3.2999999999949736e-05,
"children": {
"_observation_to_np_array": {
"total": 0.0063931000000001514,
"count": 3,
"is_parallel": true,
"self": 3.219999999992673e-05,
"children": {
"process_pixels": {
"total": 0.006360900000000225,
"count": 3,
"is_parallel": true,
"self": 0.000252800000000164,
"children": {
"image_decompress": {
"total": 0.006108100000000061,
"count": 3,
"is_parallel": true,
"self": 0.006108100000000061
}
}
}
}
}
}
},
"_process_rank_one_or_two_observation": {
"total": 4.610000000004888e-05,
"count": 2,
"is_parallel": true,
"self": 4.610000000004888e-05
}
}
},
"UnityEnvironment.step": {
"total": 1317.175587600058,
"count": 401225,
"is_parallel": true,
"self": 16.29968470018207,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 19.04841950003623,
"count": 401225,
"is_parallel": true,
"self": 19.04841950003623
},
"communicator.exchange": {
"total": 1147.2209986000285,
"count": 401225,
"is_parallel": true,
"self": 1147.2209986000285
},
"steps_from_proto": {
"total": 134.6064847998111,
"count": 401225,
"is_parallel": true,
"self": 26.638156099796177,
"children": {
"_process_maybe_compressed_observation": {
"total": 96.88266109995595,
"count": 802450,
"is_parallel": true,
"self": 7.688365399733755,
"children": {
"_observation_to_np_array": {
"total": 89.1942957002222,
"count": 1204305,
"is_parallel": true,
"self": 7.565408300169878,
"children": {
"process_pixels": {
"total": 81.62888740005232,
"count": 1204305,
"is_parallel": true,
"self": 38.62657160014098,
"children": {
"image_decompress": {
"total": 43.002315799911344,
"count": 1204305,
"is_parallel": true,
"self": 43.002315799911344
}
}
}
}
}
}
},
"_process_rank_one_or_two_observation": {
"total": 11.085667600058988,
"count": 802450,
"is_parallel": true,
"self": 11.085667600058988
}
}
}
}
}
}
}
}
}
}
}
}
},
"trainer_threads": {
"total": 3.879999985656468e-05,
"count": 1,
"self": 3.879999985656468e-05,
"children": {
"thread_root": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"trainer_advance": {
"total": 3523.306334700067,
"count": 168218,
"is_parallel": true,
"self": 5.020779300031336,
"children": {
"process_trajectory": {
"total": 2763.2327570000357,
"count": 168218,
"is_parallel": true,
"self": 2762.7916066000353,
"children": {
"RLTrainer._checkpoint": {
"total": 0.44115040000042427,
"count": 2,
"is_parallel": true,
"self": 0.44115040000042427
}
}
},
"_update_policy": {
"total": 755.0527984000001,
"count": 140,
"is_parallel": true,
"self": 502.3291928999985,
"children": {
"TorchPPOOptimizer.update": {
"total": 252.72360550000167,
"count": 3372,
"is_parallel": true,
"self": 252.72360550000167
}
}
}
}
}
}
}
}
},
"TrainerController._save_models": {
"total": 0.1734350000001541,
"count": 1,
"self": 0.012856700000156707,
"children": {
"RLTrainer._checkpoint": {
"total": 0.1605782999999974,
"count": 1,
"self": 0.1605782999999974
}
}
}
}
}
}
}