philippds's picture
Upload 9 files
200cb7b verified
{
"name": "root",
"gauges": {
"Agent.Policy.Entropy.mean": {
"value": 3.927981376647949,
"min": 3.8751437664031982,
"max": 4.380889415740967,
"count": 100
},
"Agent.Policy.Entropy.sum": {
"value": 21741.376953125,
"min": 6804.640625,
"max": 28357.529296875,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualResourceCount.mean": {
"value": 0.8925924428083278,
"min": 0.5888889440231853,
"max": 0.8999999215205511,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualResourceCount.sum": {
"value": 24.099995955824852,
"min": 7.500000648200512,
"max": 31.899996653199196,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoNeighbours.mean": {
"value": 121.29669867621527,
"min": 10.817269179556105,
"max": 175.8681265513102,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoNeighbours.sum": {
"value": 3275.0108642578125,
"min": 251.57147407531738,
"max": 3275.0108642578125,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoSelf.mean": {
"value": 0.3946781733409605,
"min": 0.078398619717215,
"max": 11.990447444220385,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoSelf.sum": {
"value": 10.656310680205934,
"min": 0.7055875774549349,
"max": 247.520980425179,
"count": 100
},
"Agent.WildfireResourceManagement.CollectivePerformance.mean": {
"value": 82.87572337962963,
"min": 19.81304868062337,
"max": 141.02953423394098,
"count": 100
},
"Agent.WildfireResourceManagement.CollectivePerformance.sum": {
"value": 2237.64453125,
"min": 440.5219039916992,
"max": 2280.235191345215,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualPerformance.mean": {
"value": 44.38486165056626,
"min": 10.05215057068401,
"max": 69.64021114508311,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualPerformance.sum": {
"value": 1198.391264565289,
"min": 240.03722366690636,
"max": 1198.391264565289,
"count": 100
},
"Agent.Environment.LessonNumber.difficulty.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.LessonNumber.difficulty.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.LessonNumber.task.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.LessonNumber.task.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.EpisodeLength.mean": {
"value": 204.66666666666666,
"min": 93.75,
"max": 497.0,
"count": 100
},
"Agent.Environment.EpisodeLength.sum": {
"value": 5526.0,
"min": 1719.0,
"max": 6831.0,
"count": 100
},
"Agent.Step.mean": {
"value": 449812.0,
"min": 4398.0,
"max": 449812.0,
"count": 100
},
"Agent.Step.sum": {
"value": 449812.0,
"min": 4398.0,
"max": 449812.0,
"count": 100
},
"Agent.Policy.CuriosityValueEstimate.mean": {
"value": 0.006913419347256422,
"min": -0.0063050780445337296,
"max": 0.12888027727603912,
"count": 100
},
"Agent.Policy.CuriosityValueEstimate.sum": {
"value": 0.145181804895401,
"min": -0.1460437923669815,
"max": 3.479767322540283,
"count": 100
},
"Agent.Policy.ExtrinsicValueEstimate.mean": {
"value": 42.51063919067383,
"min": 1.8513860702514648,
"max": 50.33930206298828,
"count": 100
},
"Agent.Policy.ExtrinsicValueEstimate.sum": {
"value": 892.7234497070312,
"min": 46.28465270996094,
"max": 1408.047119140625,
"count": 100
},
"Agent.Environment.CumulativeReward.mean": {
"value": 228.74893406459265,
"min": 46.219611057421055,
"max": 290.4283050537109,
"count": 100
},
"Agent.Environment.CumulativeReward.sum": {
"value": 4803.727615356445,
"min": 1038.7959022521973,
"max": 4803.727615356445,
"count": 100
},
"Agent.Policy.CuriosityReward.mean": {
"value": 0.013974522018716448,
"min": 0.011297034891322255,
"max": 0.18975708599794994,
"count": 100
},
"Agent.Policy.CuriosityReward.sum": {
"value": 0.2934649623930454,
"min": 0.22433854267001152,
"max": 4.174655891954899,
"count": 100
},
"Agent.Policy.ExtrinsicReward.mean": {
"value": 228.74893406459265,
"min": 46.219611057421055,
"max": 290.4283050537109,
"count": 100
},
"Agent.Policy.ExtrinsicReward.sum": {
"value": 4803.727615356445,
"min": 1038.7959022521973,
"max": 4803.727615356445,
"count": 100
},
"Agent.Losses.PolicyLoss.mean": {
"value": 0.07224695668662981,
"min": 0.055277328552870916,
"max": 0.08372609338543394,
"count": 99
},
"Agent.Losses.PolicyLoss.sum": {
"value": 0.14449391337325962,
"min": 0.055277328552870916,
"max": 0.15900131116321137,
"count": 99
},
"Agent.Losses.ValueLoss.mean": {
"value": 225.9630463348935,
"min": 5.860518129248368,
"max": 225.9630463348935,
"count": 99
},
"Agent.Losses.ValueLoss.sum": {
"value": 451.926092669787,
"min": 5.860518129248368,
"max": 451.926092669787,
"count": 99
},
"Agent.Policy.LearningRate.mean": {
"value": 1.7790994069999974e-06,
"min": 1.7790994069999974e-06,
"max": 0.00029834400055199994,
"count": 99
},
"Agent.Policy.LearningRate.sum": {
"value": 3.5581988139999947e-06,
"min": 3.5581988139999947e-06,
"max": 0.0005912520029159999,
"count": 99
},
"Agent.Policy.Epsilon.mean": {
"value": 0.100593,
"min": 0.100593,
"max": 0.19944800000000004,
"count": 99
},
"Agent.Policy.Epsilon.sum": {
"value": 0.201186,
"min": 0.10142999999999998,
"max": 0.3970840000000001,
"count": 99
},
"Agent.Policy.Beta.mean": {
"value": 6.924069999999992e-05,
"min": 6.924069999999992e-05,
"max": 0.009944855199999998,
"count": 99
},
"Agent.Policy.Beta.sum": {
"value": 0.00013848139999999984,
"min": 0.00013848139999999984,
"max": 0.019708691599999995,
"count": 99
},
"Agent.Losses.CuriosityForwardLoss.mean": {
"value": 0.0028177442157939587,
"min": 0.0024900277077004014,
"max": 0.3886077385722545,
"count": 99
},
"Agent.Losses.CuriosityForwardLoss.sum": {
"value": 0.0056354884315879174,
"min": 0.0024900277077004014,
"max": 0.3886077385722545,
"count": 99
},
"Agent.Losses.CuriosityInverseLoss.mean": {
"value": 3.9829481481387896,
"min": 3.857355904957605,
"max": 4.39014532691554,
"count": 99
},
"Agent.Losses.CuriosityInverseLoss.sum": {
"value": 7.965896296277579,
"min": 3.857355904957605,
"max": 8.717200434508444,
"count": 99
},
"Agent.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 100
},
"Agent.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 100
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1716671234",
"python_version": "3.9.18 (main, Sep 11 2023, 14:09:26) [MSC v.1916 64 bit (AMD64)]",
"command_line_arguments": "C:\\Users\\pdsie\\anaconda3\\envs\\mlagents20\\Scripts\\mlagents-learn c:/users/pdsie/documents/hivex/src/hivex/training/baseline/ml_agents/configs/mlagents/tmp/train/WildfireResourceManagement_difficulty_8_task_0_run_id_0_train.yaml --run-id=WildfireResourceManagement/train/WildfireResourceManagement_difficulty_8_task_0_run_id_0_train --base-port 5006",
"mlagents_version": "0.30.0",
"mlagents_envs_version": "0.30.0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.7.1+cu110",
"numpy_version": "1.21.0",
"end_time_seconds": "1716673111"
},
"total": 1877.4795203,
"count": 1,
"self": 0.2012331000003087,
"children": {
"run_training.setup": {
"total": 0.057823100000000016,
"count": 1,
"self": 0.057823100000000016
},
"TrainerController.start_learning": {
"total": 1877.2204640999998,
"count": 1,
"self": 1.2166694000222833,
"children": {
"TrainerController._reset_env": {
"total": 2.1442506999999997,
"count": 1,
"self": 2.1442506999999997
},
"TrainerController.advance": {
"total": 1873.7536908999778,
"count": 50157,
"self": 1.0207458000056704,
"children": {
"env_step": {
"total": 1872.732945099972,
"count": 50157,
"self": 1584.8345274999333,
"children": {
"SubprocessEnvManager._take_step": {
"total": 287.39775570001245,
"count": 50157,
"self": 1.9877286999894181,
"children": {
"TorchPolicy.evaluate": {
"total": 285.41002700002304,
"count": 50157,
"self": 285.41002700002304
}
}
},
"workers": {
"total": 0.5006619000262265,
"count": 50157,
"self": 0.0,
"children": {
"worker_root": {
"total": 1874.5358007000282,
"count": 50157,
"is_parallel": true,
"self": 352.3841908000252,
"children": {
"steps_from_proto": {
"total": 0.000249600000000072,
"count": 1,
"is_parallel": true,
"self": 0.00011890000000014389,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0001306999999999281,
"count": 2,
"is_parallel": true,
"self": 0.0001306999999999281
}
}
},
"UnityEnvironment.step": {
"total": 1522.151360300003,
"count": 50157,
"is_parallel": true,
"self": 3.7453855999726784,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 4.608074200023985,
"count": 50157,
"is_parallel": true,
"self": 4.608074200023985
},
"communicator.exchange": {
"total": 1503.4560867999946,
"count": 50157,
"is_parallel": true,
"self": 1503.4560867999946
},
"steps_from_proto": {
"total": 10.34181370001179,
"count": 50157,
"is_parallel": true,
"self": 5.356283000016858,
"children": {
"_process_rank_one_or_two_observation": {
"total": 4.985530699994931,
"count": 100314,
"is_parallel": true,
"self": 4.985530699994931
}
}
}
}
}
}
}
}
}
}
}
}
},
"trainer_threads": {
"total": 2.2599999965677853e-05,
"count": 1,
"self": 2.2599999965677853e-05,
"children": {
"thread_root": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"trainer_advance": {
"total": 1873.6321939000304,
"count": 93646,
"is_parallel": true,
"self": 3.4664417000290086,
"children": {
"process_trajectory": {
"total": 1485.2527727000006,
"count": 93646,
"is_parallel": true,
"self": 1485.2527727000006
},
"_update_policy": {
"total": 384.9129795000009,
"count": 148,
"is_parallel": true,
"self": 188.9460652999935,
"children": {
"TorchPPOOptimizer.update": {
"total": 195.9669142000074,
"count": 10284,
"is_parallel": true,
"self": 195.9669142000074
}
}
}
}
}
}
}
}
},
"TrainerController._save_models": {
"total": 0.1058304999999109,
"count": 1,
"self": 0.014507299999877432,
"children": {
"RLTrainer._checkpoint": {
"total": 0.09132320000003347,
"count": 1,
"self": 0.09132320000003347
}
}
}
}
}
}
}