philippds's picture
Upload 9 files
2859f8c verified
raw
history blame
19.6 kB
{
"name": "root",
"gauges": {
"Agent.Policy.Entropy.mean": {
"value": 3.840954303741455,
"min": 3.824573040008545,
"max": 4.376417636871338,
"count": 100
},
"Agent.Policy.Entropy.sum": {
"value": 20913.99609375,
"min": 8704.244140625,
"max": 32441.447265625,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualResourceCount.mean": {
"value": 0.8844443354341719,
"min": 0.6462963503543977,
"max": 0.8999998793005943,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualResourceCount.sum": {
"value": 39.799995094537735,
"min": 7.299999713897705,
"max": 61.39999267458916,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoNeighbours.mean": {
"value": 44.65506867253118,
"min": 5.183370861742231,
"max": 204.7496075100369,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoNeighbours.sum": {
"value": 2009.4780902639031,
"min": 279.9020265340805,
"max": 3452.556253552437,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoSelf.mean": {
"value": 0.2664910261799399,
"min": 0.1047856067177943,
"max": 10.791231377671162,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoSelf.sum": {
"value": 11.992096178097295,
"min": 2.3146310513839126,
"max": 299.0446553565562,
"count": 100
},
"Agent.WildfireResourceManagement.CollectivePerformance.mean": {
"value": 32.170304446750215,
"min": 14.268592225180733,
"max": 150.80484178331164,
"count": 100
},
"Agent.WildfireResourceManagement.CollectivePerformance.sum": {
"value": 1447.6637001037598,
"min": 443.10023975372314,
"max": 2396.3430638313293,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualPerformance.mean": {
"value": 16.353615532673622,
"min": 7.39306378364563,
"max": 80.06239223480225,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualPerformance.sum": {
"value": 735.9126989703129,
"min": 194.66514086723328,
"max": 1118.2563444972038,
"count": 100
},
"Agent.Environment.LessonNumber.difficulty.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.LessonNumber.difficulty.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.LessonNumber.task.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.LessonNumber.task.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.EpisodeLength.mean": {
"value": 120.2,
"min": 65.16666666666667,
"max": 436.0,
"count": 100
},
"Agent.Environment.EpisodeLength.sum": {
"value": 5409.0,
"min": 2052.0,
"max": 7722.0,
"count": 100
},
"Agent.Step.mean": {
"value": 449904.0,
"min": 4441.0,
"max": 449904.0,
"count": 100
},
"Agent.Step.sum": {
"value": 449904.0,
"min": 4441.0,
"max": 449904.0,
"count": 100
},
"Agent.Policy.CuriosityValueEstimate.mean": {
"value": 0.0020331591367721558,
"min": -0.01999807544052601,
"max": 0.11944062262773514,
"count": 100
},
"Agent.Policy.CuriosityValueEstimate.sum": {
"value": 0.08539268374443054,
"min": -0.9399095177650452,
"max": 4.180421829223633,
"count": 100
},
"Agent.Policy.ExtrinsicValueEstimate.mean": {
"value": 40.62111282348633,
"min": 0.2851587235927582,
"max": 45.50162887573242,
"count": 100
},
"Agent.Policy.ExtrinsicValueEstimate.sum": {
"value": 1706.086669921875,
"min": 11.691507339477539,
"max": 2222.79443359375,
"count": 100
},
"Agent.Environment.CumulativeReward.mean": {
"value": 52.5391713096982,
"min": 17.885152323032493,
"max": 361.33335172213043,
"count": 100
},
"Agent.Environment.CumulativeReward.sum": {
"value": 2206.645195007324,
"min": 968.859085559845,
"max": 4737.388134002686,
"count": 100
},
"Agent.Policy.CuriosityReward.mean": {
"value": 0.006350158051299374,
"min": 0.005583618249712062,
"max": 0.18180645166373835,
"count": 100
},
"Agent.Policy.CuriosityReward.sum": {
"value": 0.26670663815457374,
"min": 0.1619249292416498,
"max": 7.454064518213272,
"count": 100
},
"Agent.Policy.ExtrinsicReward.mean": {
"value": 52.5391713096982,
"min": 17.885152323032493,
"max": 361.33335172213043,
"count": 100
},
"Agent.Policy.ExtrinsicReward.sum": {
"value": 2206.645195007324,
"min": 968.859085559845,
"max": 4737.388134002686,
"count": 100
},
"Agent.Losses.PolicyLoss.mean": {
"value": 0.063536120154438,
"min": 0.05455417974271636,
"max": 0.07857325897849579,
"count": 99
},
"Agent.Losses.PolicyLoss.sum": {
"value": 0.127072240308876,
"min": 0.05455417974271636,
"max": 0.20946030842661412,
"count": 99
},
"Agent.Losses.ValueLoss.mean": {
"value": 104.72049866689343,
"min": 0.962940000825458,
"max": 255.13963254292807,
"count": 99
},
"Agent.Losses.ValueLoss.sum": {
"value": 209.44099733378687,
"min": 0.962940000825458,
"max": 391.84998753865557,
"count": 99
},
"Agent.Policy.LearningRate.mean": {
"value": 1.6080994640000096e-06,
"min": 1.6080994640000096e-06,
"max": 0.00029842200052599994,
"count": 99
},
"Agent.Policy.LearningRate.sum": {
"value": 3.216198928000019e-06,
"min": 3.216198928000019e-06,
"max": 0.0005854140048619999,
"count": 99
},
"Agent.Policy.Epsilon.mean": {
"value": 0.10053600000000001,
"min": 0.10053600000000001,
"max": 0.19947400000000004,
"count": 99
},
"Agent.Policy.Epsilon.sum": {
"value": 0.20107200000000003,
"min": 0.101584,
"max": 0.490534,
"count": 99
},
"Agent.Policy.Beta.mean": {
"value": 6.354640000000032e-05,
"min": 6.354640000000032e-05,
"max": 0.009947452599999998,
"count": 99
},
"Agent.Policy.Beta.sum": {
"value": 0.00012709280000000063,
"min": 0.00012709280000000063,
"max": 0.0195142862,
"count": 99
},
"Agent.Losses.CuriosityForwardLoss.mean": {
"value": 0.002479477593393257,
"min": 0.001664362004136338,
"max": 0.37563059092671786,
"count": 99
},
"Agent.Losses.CuriosityForwardLoss.sum": {
"value": 0.004958955186786514,
"min": 0.0018040557356832165,
"max": 0.37563059092671786,
"count": 99
},
"Agent.Losses.CuriosityInverseLoss.mean": {
"value": 3.85361607520901,
"min": 3.781625284868128,
"max": 4.390431465925993,
"count": 99
},
"Agent.Losses.CuriosityInverseLoss.sum": {
"value": 7.70723215041802,
"min": 3.781625284868128,
"max": 12.324647450573229,
"count": 99
},
"Agent.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 100
},
"Agent.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 100
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1716633990",
"python_version": "3.9.18 (main, Sep 11 2023, 14:09:26) [MSC v.1916 64 bit (AMD64)]",
"command_line_arguments": "C:\\Users\\pdsie\\anaconda3\\envs\\mlagents20\\Scripts\\mlagents-learn c:/users/pdsie/documents/hivex/src/hivex/training/baseline/ml_agents/configs/mlagents/tmp/train/WildfireResourceManagement_difficulty_1_task_0_run_id_0_train.yaml --run-id=WildfireResourceManagement/train/WildfireResourceManagement_difficulty_1_task_0_run_id_0_train --base-port 5006",
"mlagents_version": "0.30.0",
"mlagents_envs_version": "0.30.0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.7.1+cu110",
"numpy_version": "1.21.0",
"end_time_seconds": "1716635594"
},
"total": 1604.2192071,
"count": 1,
"self": 0.2187648999999965,
"children": {
"run_training.setup": {
"total": 0.05890530000000005,
"count": 1,
"self": 0.05890530000000005
},
"TrainerController.start_learning": {
"total": 1603.9415368999998,
"count": 1,
"self": 1.1650278999932198,
"children": {
"TrainerController._reset_env": {
"total": 2.0822798999999996,
"count": 1,
"self": 2.0822798999999996
},
"TrainerController.advance": {
"total": 1600.6075365000065,
"count": 50201,
"self": 1.0454456000204573,
"children": {
"env_step": {
"total": 1599.562090899986,
"count": 50201,
"self": 1321.5207513999762,
"children": {
"SubprocessEnvManager._take_step": {
"total": 277.5657273999984,
"count": 50201,
"self": 1.8167531000220833,
"children": {
"TorchPolicy.evaluate": {
"total": 275.74897429997634,
"count": 50201,
"self": 275.74897429997634
}
}
},
"workers": {
"total": 0.47561210001138976,
"count": 50201,
"self": 0.0,
"children": {
"worker_root": {
"total": 1601.3577724999916,
"count": 50201,
"is_parallel": true,
"self": 343.3875831000139,
"children": {
"steps_from_proto": {
"total": 0.00026009999999998534,
"count": 1,
"is_parallel": true,
"self": 0.00011960000000010851,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.00014049999999987683,
"count": 2,
"is_parallel": true,
"self": 0.00014049999999987683
}
}
},
"UnityEnvironment.step": {
"total": 1257.9699292999776,
"count": 50201,
"is_parallel": true,
"self": 3.6019024999789053,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 4.785271999992973,
"count": 50201,
"is_parallel": true,
"self": 4.785271999992973
},
"communicator.exchange": {
"total": 1239.082793400009,
"count": 50201,
"is_parallel": true,
"self": 1239.082793400009
},
"steps_from_proto": {
"total": 10.499961399996645,
"count": 50201,
"is_parallel": true,
"self": 5.412605099983018,
"children": {
"_process_rank_one_or_two_observation": {
"total": 5.087356300013626,
"count": 100402,
"is_parallel": true,
"self": 5.087356300013626
}
}
}
}
}
}
}
}
}
}
}
}
},
"trainer_threads": {
"total": 2.220000010311196e-05,
"count": 1,
"self": 2.220000010311196e-05,
"children": {
"thread_root": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"trainer_advance": {
"total": 1600.8696196000012,
"count": 76131,
"is_parallel": true,
"self": 2.347892499994032,
"children": {
"process_trajectory": {
"total": 1216.8828683000079,
"count": 76131,
"is_parallel": true,
"self": 1216.8828683000079
},
"_update_policy": {
"total": 381.63885879999924,
"count": 150,
"is_parallel": true,
"self": 188.25779600000348,
"children": {
"TorchPPOOptimizer.update": {
"total": 193.38106279999576,
"count": 10314,
"is_parallel": true,
"self": 193.38106279999576
}
}
}
}
}
}
}
}
},
"TrainerController._save_models": {
"total": 0.0866704000000027,
"count": 1,
"self": 0.005369399999835878,
"children": {
"RLTrainer._checkpoint": {
"total": 0.08130100000016682,
"count": 1,
"self": 0.08130100000016682
}
}
}
}
}
}
}