philippds's picture
Upload 9 files
64e22b5 verified
{
"name": "root",
"gauges": {
"Agent.Policy.Entropy.mean": {
"value": 3.9218742847442627,
"min": 3.863060474395752,
"max": 4.382823467254639,
"count": 100
},
"Agent.Policy.Entropy.sum": {
"value": 21354.60546875,
"min": 8286.625,
"max": 31645.068359375,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualResourceCount.mean": {
"value": 0.8733334170447455,
"min": 0.650000056044923,
"max": 0.8888889882299635,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualResourceCount.sum": {
"value": 39.30000376701355,
"min": 7.100000649690628,
"max": 61.100006096065044,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoNeighbours.mean": {
"value": 1.7044366454084714,
"min": 0.43675118446764016,
"max": 14.50437198082606,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoNeighbours.sum": {
"value": 76.69964904338121,
"min": 13.791587144136429,
"max": 404.41131377220154,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoSelf.mean": {
"value": 138.11908293565114,
"min": 15.732518797671354,
"max": 485.51227484809027,
"count": 100
},
"Agent.WildfireResourceManagement.RewardforMovingResourcestoSelf.sum": {
"value": 6215.358732104301,
"min": 849.5560150742531,
"max": 8050.53844833374,
"count": 100
},
"Agent.WildfireResourceManagement.CollectivePerformance.mean": {
"value": 31.67512978447808,
"min": 14.841589812879208,
"max": 114.47105407714844,
"count": 100
},
"Agent.WildfireResourceManagement.CollectivePerformance.sum": {
"value": 1425.3808403015137,
"min": 529.9125142097473,
"max": 2131.192500114441,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualPerformance.mean": {
"value": 16.451286547879377,
"min": 6.683963431252374,
"max": 57.487370596991646,
"count": 100
},
"Agent.WildfireResourceManagement.IndividualPerformance.sum": {
"value": 740.307894654572,
"min": 230.80473151803017,
"max": 1005.2699508666992,
"count": 100
},
"Agent.Environment.LessonNumber.difficulty.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.LessonNumber.difficulty.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.LessonNumber.task.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.LessonNumber.task.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"Agent.Environment.EpisodeLength.mean": {
"value": 120.2,
"min": 65.16666666666667,
"max": 436.0,
"count": 100
},
"Agent.Environment.EpisodeLength.sum": {
"value": 5409.0,
"min": 2052.0,
"max": 7722.0,
"count": 100
},
"Agent.Step.mean": {
"value": 449904.0,
"min": 4441.0,
"max": 449904.0,
"count": 100
},
"Agent.Step.sum": {
"value": 449904.0,
"min": 4441.0,
"max": 449904.0,
"count": 100
},
"Agent.Policy.CuriosityValueEstimate.mean": {
"value": 0.01566745899617672,
"min": -0.04759064316749573,
"max": 0.1646767556667328,
"count": 100
},
"Agent.Policy.CuriosityValueEstimate.sum": {
"value": 0.6580332517623901,
"min": -1.878260850906372,
"max": 4.296153545379639,
"count": 100
},
"Agent.Policy.ExtrinsicValueEstimate.mean": {
"value": 85.21717834472656,
"min": 0.376181960105896,
"max": 99.24287414550781,
"count": 100
},
"Agent.Policy.ExtrinsicValueEstimate.sum": {
"value": 3579.121337890625,
"min": 15.423460960388184,
"max": 3766.849853515625,
"count": 100
},
"Agent.Environment.CumulativeReward.mean": {
"value": 165.82417479015533,
"min": 34.71144700631863,
"max": 534.9340178171793,
"count": 100
},
"Agent.Environment.CumulativeReward.sum": {
"value": 6964.615341186523,
"min": 1423.1693272590637,
"max": 8177.103809833527,
"count": 100
},
"Agent.Policy.CuriosityReward.mean": {
"value": 0.018203601307634796,
"min": 0.014499624579477666,
"max": 0.18180511673776115,
"count": 100
},
"Agent.Policy.CuriosityReward.sum": {
"value": 0.7645512549206614,
"min": 0.4937439579516649,
"max": 7.454009786248207,
"count": 100
},
"Agent.Policy.ExtrinsicReward.mean": {
"value": 165.82417479015533,
"min": 34.71144700631863,
"max": 534.9340178171793,
"count": 100
},
"Agent.Policy.ExtrinsicReward.sum": {
"value": 6964.615341186523,
"min": 1423.1693272590637,
"max": 8177.103809833527,
"count": 100
},
"Agent.Losses.PolicyLoss.mean": {
"value": 0.06870737995969312,
"min": 0.055507686713503465,
"max": 0.08021375311013215,
"count": 99
},
"Agent.Losses.PolicyLoss.sum": {
"value": 0.13741475991938623,
"min": 0.055507686713503465,
"max": 0.2052263682775899,
"count": 99
},
"Agent.Losses.ValueLoss.mean": {
"value": 934.7500864268123,
"min": 12.23641684320238,
"max": 934.7500864268123,
"count": 99
},
"Agent.Losses.ValueLoss.sum": {
"value": 1869.5001728536247,
"min": 12.23641684320238,
"max": 1869.5001728536247,
"count": 99
},
"Agent.Policy.LearningRate.mean": {
"value": 1.6080994640000096e-06,
"min": 1.6080994640000096e-06,
"max": 0.00029842200052599994,
"count": 99
},
"Agent.Policy.LearningRate.sum": {
"value": 3.216198928000019e-06,
"min": 3.216198928000019e-06,
"max": 0.0005854140048619999,
"count": 99
},
"Agent.Policy.Epsilon.mean": {
"value": 0.10053600000000001,
"min": 0.10053600000000001,
"max": 0.19947400000000004,
"count": 99
},
"Agent.Policy.Epsilon.sum": {
"value": 0.20107200000000003,
"min": 0.101584,
"max": 0.490534,
"count": 99
},
"Agent.Policy.Beta.mean": {
"value": 6.354640000000032e-05,
"min": 6.354640000000032e-05,
"max": 0.009947452599999998,
"count": 99
},
"Agent.Policy.Beta.sum": {
"value": 0.00012709280000000063,
"min": 0.00012709280000000063,
"max": 0.0195142862,
"count": 99
},
"Agent.Losses.CuriosityForwardLoss.mean": {
"value": 0.007479374189700755,
"min": 0.004942750531262246,
"max": 0.37563059092671786,
"count": 99
},
"Agent.Losses.CuriosityForwardLoss.sum": {
"value": 0.01495874837940151,
"min": 0.004942750531262246,
"max": 0.37563059092671786,
"count": 99
},
"Agent.Losses.CuriosityInverseLoss.mean": {
"value": 3.7495019280190123,
"min": 3.7219206725849823,
"max": 4.390431465925993,
"count": 99
},
"Agent.Losses.CuriosityInverseLoss.sum": {
"value": 7.4990038560380246,
"min": 3.7219206725849823,
"max": 12.085247370206488,
"count": 99
},
"Agent.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 100
},
"Agent.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 100
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1716635595",
"python_version": "3.9.18 (main, Sep 11 2023, 14:09:26) [MSC v.1916 64 bit (AMD64)]",
"command_line_arguments": "C:\\Users\\pdsie\\anaconda3\\envs\\mlagents20\\Scripts\\mlagents-learn c:/users/pdsie/documents/hivex/src/hivex/training/baseline/ml_agents/configs/mlagents/tmp/train/WildfireResourceManagement_difficulty_1_task_1_run_id_0_train.yaml --run-id=WildfireResourceManagement/train/WildfireResourceManagement_difficulty_1_task_1_run_id_0_train --base-port 5006",
"mlagents_version": "0.30.0",
"mlagents_envs_version": "0.30.0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.7.1+cu110",
"numpy_version": "1.21.0",
"end_time_seconds": "1716637228"
},
"total": 1632.6059326,
"count": 1,
"self": 0.21384130000001278,
"children": {
"run_training.setup": {
"total": 0.05671159999999997,
"count": 1,
"self": 0.05671159999999997
},
"TrainerController.start_learning": {
"total": 1632.3353797,
"count": 1,
"self": 1.204931100003705,
"children": {
"TrainerController._reset_env": {
"total": 2.1452037,
"count": 1,
"self": 2.1452037
},
"TrainerController.advance": {
"total": 1628.8838285999961,
"count": 50201,
"self": 0.9999493999921469,
"children": {
"env_step": {
"total": 1627.883879200004,
"count": 50201,
"self": 1356.462256299996,
"children": {
"SubprocessEnvManager._take_step": {
"total": 270.95744470000267,
"count": 50201,
"self": 1.7378636999908963,
"children": {
"TorchPolicy.evaluate": {
"total": 269.2195810000118,
"count": 50201,
"self": 269.2195810000118
}
}
},
"workers": {
"total": 0.4641782000053478,
"count": 50201,
"self": 0.0,
"children": {
"worker_root": {
"total": 1629.7476933999687,
"count": 50201,
"is_parallel": true,
"self": 332.937152899965,
"children": {
"steps_from_proto": {
"total": 0.00024489999999999235,
"count": 1,
"is_parallel": true,
"self": 0.0001184999999996883,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.00012640000000030405,
"count": 2,
"is_parallel": true,
"self": 0.00012640000000030405
}
}
},
"UnityEnvironment.step": {
"total": 1296.8102956000037,
"count": 50201,
"is_parallel": true,
"self": 3.2137171999620477,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 4.057409000002163,
"count": 50201,
"is_parallel": true,
"self": 4.057409000002163
},
"communicator.exchange": {
"total": 1279.9483216000249,
"count": 50201,
"is_parallel": true,
"self": 1279.9483216000249
},
"steps_from_proto": {
"total": 9.590847800014732,
"count": 50201,
"is_parallel": true,
"self": 4.8921236000355,
"children": {
"_process_rank_one_or_two_observation": {
"total": 4.698724199979232,
"count": 100402,
"is_parallel": true,
"self": 4.698724199979232
}
}
}
}
}
}
}
}
}
}
}
}
},
"trainer_threads": {
"total": 2.0500000118772732e-05,
"count": 1,
"self": 2.0500000118772732e-05,
"children": {
"thread_root": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"trainer_advance": {
"total": 1629.2049924999953,
"count": 78999,
"is_parallel": true,
"self": 2.4634891000175685,
"children": {
"process_trajectory": {
"total": 1259.2101674999785,
"count": 78999,
"is_parallel": true,
"self": 1259.2101674999785
},
"_update_policy": {
"total": 367.53133589999936,
"count": 150,
"is_parallel": true,
"self": 179.1598848999912,
"children": {
"TorchPPOOptimizer.update": {
"total": 188.37145100000816,
"count": 10314,
"is_parallel": true,
"self": 188.37145100000816
}
}
}
}
}
}
}
}
},
"TrainerController._save_models": {
"total": 0.10139580000009119,
"count": 1,
"self": 0.015529600000036226,
"children": {
"RLTrainer._checkpoint": {
"total": 0.08586620000005496,
"count": 1,
"self": 0.08586620000005496
}
}
}
}
}
}
}