philippds's picture
Upload 9 files
148a5e8 verified
raw
history blame
19.6 kB
{
"name": "root",
"gauges": {
"Agent.Policy.Entropy.mean": {
"value": 3.9017300605773926,
"min": 3.8140387535095215,
"max": 4.3714494705200195,
"count": 100
},
"Agent.Policy.Entropy.sum": {
"value": 21244.919921875,
"min": 8572.8515625,
"max": 32024.3671875,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualResourceCount.mean": {
"value": 0.8933332022693422,
"min": 0.6500000520436852,
"max": 0.8999998793005943,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualResourceCount.sum": {
"value": 40.1999941021204,
"min": 7.4999992325901985,
"max": 62.59999257326126,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoNeighbours.mean": {
"value": 455.88364140722484,
"min": 53.61029998461405,
"max": 2424.1639472113716,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoNeighbours.sum": {
"value": 20514.76386332512,
"min": 2894.956199169159,
"max": 41510.122690439224,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoSelf.mean": {
"value": 0.30548179356548943,
"min": 0.05765622271535954,
"max": 4.903017125965562,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoSelf.sum": {
"value": 13.746680710447023,
"min": 1.0378120088764717,
"max": 110.67634117603302,
"count": 100
},
"Agent.WildfireResourceManagement.CollectivePerformance.mean": {
"value": 31.89151145087348,
"min": 14.07080864906311,
"max": 141.05380249023438,
"count": 100
},
"Agent.WildfireResourceManagement.CollectivePerformance.sum": {
"value": 1435.1180152893066,
"min": 498.6379585266113,
"max": 2363.7911472320557,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualPerformance.mean": {
"value": 16.32038382192239,
"min": 7.486613300111559,
"max": 88.22554196251764,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualPerformance.sum": {
"value": 734.4172719865076,
"min": 224.51533195376396,
"max": 1192.3509817123413,
"count": 100
},
"Agent.Environment.LessonNumber.difficulty.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.LessonNumber.difficulty.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.LessonNumber.task.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.LessonNumber.task.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.EpisodeLength.mean": {
"value": 120.2,
"min": 65.16666666666667,
"max": 436.0,
"count": 100
},
"Agent.Environment.EpisodeLength.sum": {
"value": 5409.0,
"min": 2052.0,
"max": 7722.0,
"count": 100
},
"Agent.Step.mean": {
"value": 449904.0,
"min": 4441.0,
"max": 449904.0,
"count": 100
},
"Agent.Step.sum": {
"value": 449904.0,
"min": 4441.0,
"max": 449904.0,
"count": 100
},
"Agent.Policy.CuriosityValueEstimate.mean": {
"value": -0.009130329824984074,
"min": -0.08197839558124542,
"max": 0.1488860845565796,
"count": 100
},
"Agent.Policy.CuriosityValueEstimate.sum": {
"value": -0.38347384333610535,
"min": -2.705286979675293,
"max": 5.070252895355225,
"count": 100
},
"Agent.Policy.ExtrinsicValueEstimate.mean": {
"value": 267.82708740234375,
"min": 0.8720908761024475,
"max": 337.3840637207031,
"count": 100
},
"Agent.Policy.ExtrinsicValueEstimate.sum": {
"value": 11248.7373046875,
"min": 35.7557258605957,
"max": 14407.11328125,
"count": 100
},
"Agent.Environment.CumulativeReward.mean": {
"value": 302.4706774666196,
"min": 76.75898319337426,
"max": 2716.065643310547,
"count": 100
},
"Agent.Environment.CumulativeReward.sum": {
"value": 12703.768453598022,
"min": 3147.1183109283447,
"max": 35308.85336303711,
"count": 100
},
"Agent.Policy.CuriosityReward.mean": {
"value": 0.006048896419261361,
"min": 0.005580994091799547,
"max": 0.18199072705536354,
"count": 100
},
"Agent.Policy.CuriosityReward.sum": {
"value": 0.25405364960897714,
"min": 0.16184882866218686,
"max": 7.461619809269905,
"count": 100
},
"Agent.Policy.ExtrinsicReward.mean": {
"value": 302.4706774666196,
"min": 76.75898319337426,
"max": 2716.065643310547,
"count": 100
},
"Agent.Policy.ExtrinsicReward.sum": {
"value": 12703.768453598022,
"min": 3147.1183109283447,
"max": 35308.85336303711,
"count": 100
},
"Agent.Losses.PolicyLoss.mean": {
"value": 0.06425880937903025,
"min": 0.05538995831233004,
"max": 0.0795730839303711,
"count": 99
},
"Agent.Losses.PolicyLoss.sum": {
"value": 0.1285176187580605,
"min": 0.05538995831233004,
"max": 0.2090949601998242,
"count": 99
},
"Agent.Losses.ValueLoss.mean": {
"value": 5232.771832818718,
"min": 9.863332196518227,
"max": 21489.855577256945,
"count": 99
},
"Agent.Losses.ValueLoss.sum": {
"value": 10465.543665637437,
"min": 9.863332196518227,
"max": 29943.240201822915,
"count": 99
},
"Agent.Policy.LearningRate.mean": {
"value": 1.6080994640000096e-06,
"min": 1.6080994640000096e-06,
"max": 0.00029842200052599994,
"count": 99
},
"Agent.Policy.LearningRate.sum": {
"value": 3.216198928000019e-06,
"min": 3.216198928000019e-06,
"max": 0.0005854140048619999,
"count": 99
},
"Agent.Policy.Epsilon.mean": {
"value": 0.10053600000000001,
"min": 0.10053600000000001,
"max": 0.19947400000000004,
"count": 99
},
"Agent.Policy.Epsilon.sum": {
"value": 0.20107200000000003,
"min": 0.101584,
"max": 0.490534,
"count": 99
},
"Agent.Policy.Beta.mean": {
"value": 6.354640000000032e-05,
"min": 6.354640000000032e-05,
"max": 0.009947452599999998,
"count": 99
},
"Agent.Policy.Beta.sum": {
"value": 0.00012709280000000063,
"min": 0.00012709280000000063,
"max": 0.0195142862,
"count": 99
},
"Agent.Losses.CuriosityForwardLoss.mean": {
"value": 0.0025058726663171734,
"min": 0.0014629171960804537,
"max": 0.37563059092671786,
"count": 99
},
"Agent.Losses.CuriosityForwardLoss.sum": {
"value": 0.005011745332634347,
"min": 0.001611489856396943,
"max": 0.37563059092671786,
"count": 99
},
"Agent.Losses.CuriosityInverseLoss.mean": {
"value": 3.8885767201156565,
"min": 3.79713204795239,
"max": 4.390431465925993,
"count": 99
},
"Agent.Losses.CuriosityInverseLoss.sum": {
"value": 7.777153440231313,
"min": 3.79713204795239,
"max": 12.45364493658303,
"count": 99
},
"Agent.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 100
},
"Agent.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 100
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1716691861",
"python_version": "3.9.18 (main, Sep 11 2023, 14:09:26) [MSC v.1916 64 bit (AMD64)]",
"command_line_arguments": "C:\\Users\\pdsie\\anaconda3\\envs\\mlagents20\\Scripts\\mlagents-learn c:/users/pdsie/documents/hivex/src/hivex/training/baseline/ml_agents/configs/mlagents/tmp/train/WildfireResourceManagement_difficulty_1_task_2_run_id_1_train.yaml --run-id=WildfireResourceManagement/train/WildfireResourceManagement_difficulty_1_task_2_run_id_1_train --base-port 5006",
"mlagents_version": "0.30.0",
"mlagents_envs_version": "0.30.0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.7.1+cu110",
"numpy_version": "1.21.0",
"end_time_seconds": "1716693456"
},
"total": 1595.0740624,
"count": 1,
"self": 0.1920430000000124,
"children": {
"run_training.setup": {
"total": 0.05646689999999999,
"count": 1,
"self": 0.05646689999999999
},
"TrainerController.start_learning": {
"total": 1594.8255525,
"count": 1,
"self": 1.1509333000137758,
"children": {
"TrainerController._reset_env": {
"total": 2.0875927,
"count": 1,
"self": 2.0875927
},
"TrainerController.advance": {
"total": 1591.484710599986,
"count": 50201,
"self": 1.0372009999775855,
"children": {
"env_step": {
"total": 1590.4475096000085,
"count": 50201,
"self": 1316.8220555999865,
"children": {
"SubprocessEnvManager._take_step": {
"total": 273.16257040000954,
"count": 50201,
"self": 1.735983200026169,
"children": {
"TorchPolicy.evaluate": {
"total": 271.4265871999834,
"count": 50201,
"self": 271.4265871999834
}
}
},
"workers": {
"total": 0.4628836000123995,
"count": 50201,
"self": 0.0,
"children": {
"worker_root": {
"total": 1592.2150314999903,
"count": 50201,
"is_parallel": true,
"self": 333.5507993999897,
"children": {
"steps_from_proto": {
"total": 0.0002470999999999446,
"count": 1,
"is_parallel": true,
"self": 0.00012069999999986258,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.000126400000000082,
"count": 2,
"is_parallel": true,
"self": 0.000126400000000082
}
}
},
"UnityEnvironment.step": {
"total": 1258.6639850000006,
"count": 50201,
"is_parallel": true,
"self": 3.258150600005365,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 4.005038999999481,
"count": 50201,
"is_parallel": true,
"self": 4.005038999999481
},
"communicator.exchange": {
"total": 1242.0187904000136,
"count": 50201,
"is_parallel": true,
"self": 1242.0187904000136
},
"steps_from_proto": {
"total": 9.382004999982113,
"count": 50201,
"is_parallel": true,
"self": 4.845075099989843,
"children": {
"_process_rank_one_or_two_observation": {
"total": 4.536929899992271,
"count": 100402,
"is_parallel": true,
"self": 4.536929899992271
}
}
}
}
}
}
}
}
}
}
}
}
},
"trainer_threads": {
"total": 2.8299999939918052e-05,
"count": 1,
"self": 2.8299999939918052e-05,
"children": {
"thread_root": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"trainer_advance": {
"total": 1591.7403541000006,
"count": 76587,
"is_parallel": true,
"self": 2.4294420000276205,
"children": {
"process_trajectory": {
"total": 1220.6428112999731,
"count": 76587,
"is_parallel": true,
"self": 1220.6428112999731
},
"_update_policy": {
"total": 368.6681007999998,
"count": 150,
"is_parallel": true,
"self": 179.46155949999405,
"children": {
"TorchPPOOptimizer.update": {
"total": 189.20654130000577,
"count": 10314,
"is_parallel": true,
"self": 189.20654130000577
}
}
}
}
}
}
}
}
},
"TrainerController._save_models": {
"total": 0.10228760000018156,
"count": 1,
"self": 0.014001800000187359,
"children": {
"RLTrainer._checkpoint": {
"total": 0.0882857999999942,
"count": 1,
"self": 0.0882857999999942
}
}
}
}
}
}
}