philippds's picture
Upload 9 files
28e2ceb verified
{
"name": "root",
"gauges": {
"Agent.Policy.Entropy.mean": {
"value": 3.9668312072753906,
"min": 3.7804253101348877,
"max": 4.380436420440674,
"count": 100
},
"Agent.Policy.Entropy.sum": {
"value": 19457.306640625,
"min": 5902.4345703125,
"max": 31717.58203125,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualResourceCount.mean": {
"value": 0.8925926861939607,
"min": 0.5805555743475755,
"max": 0.9000000953674316,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualResourceCount.sum": {
"value": 24.10000252723694,
"min": 7.600000739097595,
"max": 40.000004172325134,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoNeighbours.mean": {
"value": 2.1979813239088766,
"min": 0.3762214953016433,
"max": 12.163262406984964,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoNeighbours.sum": {
"value": 59.345495745539665,
"min": 3.3859934577147897,
"max": 366.9930374722462,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoSelf.mean": {
"value": 195.94763296621818,
"min": 45.09299181732867,
"max": 500.01685587565106,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoSelf.sum": {
"value": 5290.586090087891,
"min": 1623.347705423832,
"max": 8641.529499053955,
"count": 100
},
"Agent.WildfireResourceManagement.CollectivePerformance.mean": {
"value": 46.293033599853516,
"min": 25.176106135050457,
"max": 117.2500237358941,
"count": 100
},
"Agent.WildfireResourceManagement.CollectivePerformance.sum": {
"value": 1249.911907196045,
"min": 448.02629470825195,
"max": 2048.968963623047,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualPerformance.mean": {
"value": 23.532668590545654,
"min": 13.759139292769962,
"max": 58.621494717068146,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualPerformance.sum": {
"value": 635.3820519447327,
"min": 211.78832626342773,
"max": 1010.5693788528442,
"count": 100
},
"Agent.Environment.LessonNumber.difficulty.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.LessonNumber.difficulty.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.LessonNumber.task.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.LessonNumber.task.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.EpisodeLength.mean": {
"value": 181.66666666666666,
"min": 96.33333333333333,
"max": 473.0,
"count": 100
},
"Agent.Environment.EpisodeLength.sum": {
"value": 4905.0,
"min": 1512.0,
"max": 7803.0,
"count": 100
},
"Agent.Step.mean": {
"value": 449977.0,
"min": 4405.0,
"max": 449977.0,
"count": 100
},
"Agent.Step.sum": {
"value": 449977.0,
"min": 4405.0,
"max": 449977.0,
"count": 100
},
"Agent.Policy.CuriosityValueEstimate.mean": {
"value": 0.00578606640920043,
"min": -0.045637745410203934,
"max": 0.09990525245666504,
"count": 100
},
"Agent.Policy.CuriosityValueEstimate.sum": {
"value": 0.11572133004665375,
"min": -0.8686326742172241,
"max": 2.7380776405334473,
"count": 100
},
"Agent.Policy.ExtrinsicValueEstimate.mean": {
"value": 89.14933776855469,
"min": 3.151017189025879,
"max": 131.51524353027344,
"count": 100
},
"Agent.Policy.ExtrinsicValueEstimate.sum": {
"value": 1782.98681640625,
"min": 81.92644500732422,
"max": 4076.97265625,
"count": 100
},
"Agent.Environment.CumulativeReward.mean": {
"value": 321.5479022026062,
"min": 96.69844296039679,
"max": 614.6264218417081,
"count": 100
},
"Agent.Environment.CumulativeReward.sum": {
"value": 6430.958044052124,
"min": 2918.829081296921,
"max": 8660.427490234375,
"count": 100
},
"Agent.Policy.CuriosityReward.mean": {
"value": 0.03222924910951406,
"min": 0.014380116129102997,
"max": 0.26607194998198086,
"count": 100
},
"Agent.Policy.CuriosityReward.sum": {
"value": 0.6445849821902812,
"min": 0.48400826193392277,
"max": 5.135645776987076,
"count": 100
},
"Agent.Policy.ExtrinsicReward.mean": {
"value": 321.5479022026062,
"min": 96.69844296039679,
"max": 614.6264218417081,
"count": 100
},
"Agent.Policy.ExtrinsicReward.sum": {
"value": 6430.958044052124,
"min": 2918.829081296921,
"max": 8660.427490234375,
"count": 100
},
"Agent.Losses.PolicyLoss.mean": {
"value": 0.06731047691286993,
"min": 0.053174213441606194,
"max": 0.08211484194571297,
"count": 99
},
"Agent.Losses.PolicyLoss.sum": {
"value": 0.13462095382573985,
"min": 0.053174213441606194,
"max": 0.16422968389142595,
"count": 99
},
"Agent.Losses.ValueLoss.mean": {
"value": 787.9369762449553,
"min": 100.27956397860657,
"max": 1110.4575428903838,
"count": 99
},
"Agent.Losses.ValueLoss.sum": {
"value": 1575.8739524899106,
"min": 100.27956397860657,
"max": 2220.9150857807676,
"count": 99
},
"Agent.Policy.LearningRate.mean": {
"value": 1.4610995129999955e-06,
"min": 1.4610995129999955e-06,
"max": 0.00029848200050600006,
"count": 99
},
"Agent.Policy.LearningRate.sum": {
"value": 2.922199025999991e-06,
"min": 2.922199025999991e-06,
"max": 0.0005858280047239997,
"count": 99
},
"Agent.Policy.Epsilon.mean": {
"value": 0.10048699999999999,
"min": 0.10048699999999999,
"max": 0.199494,
"count": 99
},
"Agent.Policy.Epsilon.sum": {
"value": 0.20097399999999999,
"min": 0.101576,
"max": 0.3952760000000001,
"count": 99
},
"Agent.Policy.Beta.mean": {
"value": 5.865129999999986e-05,
"min": 5.865129999999986e-05,
"max": 0.009949450600000001,
"count": 99
},
"Agent.Policy.Beta.sum": {
"value": 0.00011730259999999973,
"min": 0.00011730259999999973,
"max": 0.0195280724,
"count": 99
},
"Agent.Losses.CuriosityForwardLoss.mean": {
"value": 0.0059364226382802535,
"min": 0.004110809447916431,
"max": 0.4862046896242628,
"count": 99
},
"Agent.Losses.CuriosityForwardLoss.sum": {
"value": 0.011872845276560507,
"min": 0.004110809447916431,
"max": 0.4862046896242628,
"count": 99
},
"Agent.Losses.CuriosityInverseLoss.mean": {
"value": 3.808633321162426,
"min": 3.682709720399645,
"max": 4.390673973981072,
"count": 99
},
"Agent.Losses.CuriosityInverseLoss.sum": {
"value": 7.617266642324852,
"min": 3.682709720399645,
"max": 8.669611559973823,
"count": 99
},
"Agent.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 100
},
"Agent.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 100
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1716645566",
"python_version": "3.9.18 (main, Sep 11 2023, 14:09:26) [MSC v.1916 64 bit (AMD64)]",
"command_line_arguments": "C:\\Users\\pdsie\\anaconda3\\envs\\mlagents20\\Scripts\\mlagents-learn c:/users/pdsie/documents/hivex/src/hivex/training/baseline/ml_agents/configs/mlagents/tmp/train/WildfireResourceManagement_difficulty_3_task_1_run_id_0_train.yaml --run-id=WildfireResourceManagement/train/WildfireResourceManagement_difficulty_3_task_1_run_id_0_train --base-port 5006",
"mlagents_version": "0.30.0",
"mlagents_envs_version": "0.30.0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.7.1+cu110",
"numpy_version": "1.21.0",
"end_time_seconds": "1716647330"
},
"total": 1763.6787115,
"count": 1,
"self": 0.18825479999986783,
"children": {
"run_training.setup": {
"total": 0.049356999999999984,
"count": 1,
"self": 0.049356999999999984
},
"TrainerController.start_learning": {
"total": 1763.4410997,
"count": 1,
"self": 1.1662981999820659,
"children": {
"TrainerController._reset_env": {
"total": 2.1075382,
"count": 1,
"self": 2.1075382
},
"TrainerController.advance": {
"total": 1760.0768772000179,
"count": 50074,
"self": 1.137395200033552,
"children": {
"env_step": {
"total": 1758.9394819999843,
"count": 50074,
"self": 1476.5500912999805,
"children": {
"SubprocessEnvManager._take_step": {
"total": 281.8926474000154,
"count": 50074,
"self": 1.9207899000046496,
"children": {
"TorchPolicy.evaluate": {
"total": 279.97185750001074,
"count": 50074,
"self": 279.97185750001074
}
}
},
"workers": {
"total": 0.49674329998837274,
"count": 50074,
"self": 0.0,
"children": {
"worker_root": {
"total": 1760.8244052999887,
"count": 50074,
"is_parallel": true,
"self": 346.75208069998916,
"children": {
"steps_from_proto": {
"total": 0.00026729999999997034,
"count": 1,
"is_parallel": true,
"self": 0.00012269999999969805,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0001446000000002723,
"count": 2,
"is_parallel": true,
"self": 0.0001446000000002723
}
}
},
"UnityEnvironment.step": {
"total": 1414.0720572999996,
"count": 50074,
"is_parallel": true,
"self": 3.771616100005076,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 4.718567200012059,
"count": 50074,
"is_parallel": true,
"self": 4.718567200012059
},
"communicator.exchange": {
"total": 1394.9378351999794,
"count": 50074,
"is_parallel": true,
"self": 1394.9378351999794
},
"steps_from_proto": {
"total": 10.644038800003129,
"count": 50074,
"is_parallel": true,
"self": 5.410248600030676,
"children": {
"_process_rank_one_or_two_observation": {
"total": 5.233790199972453,
"count": 100148,
"is_parallel": true,
"self": 5.233790199972453
}
}
}
}
}
}
}
}
}
}
}
}
},
"trainer_threads": {
"total": 2.130000007127819e-05,
"count": 1,
"self": 2.130000007127819e-05,
"children": {
"thread_root": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"trainer_advance": {
"total": 1760.1467308999916,
"count": 87516,
"is_parallel": true,
"self": 3.1565607999880285,
"children": {
"process_trajectory": {
"total": 1389.8533329000027,
"count": 87516,
"is_parallel": true,
"self": 1389.8533329000027
},
"_update_policy": {
"total": 367.13683720000097,
"count": 141,
"is_parallel": true,
"self": 180.21210090000932,
"children": {
"TorchPPOOptimizer.update": {
"total": 186.92473629999165,
"count": 10335,
"is_parallel": true,
"self": 186.92473629999165
}
}
}
}
}
}
}
}
},
"TrainerController._save_models": {
"total": 0.09036480000008851,
"count": 1,
"self": 0.006183100000043851,
"children": {
"RLTrainer._checkpoint": {
"total": 0.08418170000004466,
"count": 1,
"self": 0.08418170000004466
}
}
}
}
}
}
}