yumingyi's picture
First Push
b361c0c
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 1.6854889392852783,
"min": 1.6446720361709595,
"max": 3.295754909515381,
"count": 1284
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 33655.84375,
"min": 2636.5986328125,
"max": 108795.6640625,
"count": 1284
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 67.13513513513513,
"min": 50.903225806451616,
"max": 999.0,
"count": 1284
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 19872.0,
"min": 3588.0,
"max": 28476.0,
"count": 1284
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1621.2359015459922,
"min": 1198.8400592488654,
"max": 1685.056753026278,
"count": 1279
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 239942.91342880685,
"min": 2397.6801184977307,
"max": 311853.19746006106,
"count": 1279
},
"SoccerTwos.Step.mean": {
"value": 12839958.0,
"min": 9254.0,
"max": 12839958.0,
"count": 1284
},
"SoccerTwos.Step.sum": {
"value": 12839958.0,
"min": 9254.0,
"max": 12839958.0,
"count": 1284
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": -0.023442860692739487,
"min": -0.13118010759353638,
"max": 0.22188475728034973,
"count": 1284
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": -3.4461004734039307,
"min": -17.523700714111328,
"max": 23.360881805419922,
"count": 1284
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": -0.025601858273148537,
"min": -0.13272199034690857,
"max": 0.22091476619243622,
"count": 1284
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": -3.7634730339050293,
"min": -17.610153198242188,
"max": 22.95449447631836,
"count": 1284
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 1284
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 1284
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": 0.07953741355818145,
"min": -0.6037677430337475,
"max": 0.7360842096178155,
"count": 1284
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": 11.691999793052673,
"min": -56.56719982624054,
"max": 69.92799991369247,
"count": 1284
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": 0.07953741355818145,
"min": -0.6037677430337475,
"max": 0.7360842096178155,
"count": 1284
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": 11.691999793052673,
"min": -56.56719982624054,
"max": 69.92799991369247,
"count": 1284
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 1284
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 1284
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.017124803007269897,
"min": 0.011296885545016266,
"max": 0.01993974419310689,
"count": 311
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.017124803007269897,
"min": 0.011296885545016266,
"max": 0.01993974419310689,
"count": 311
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.08500608255465826,
"min": 0.000957246491452679,
"max": 0.11027985538045565,
"count": 311
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.08500608255465826,
"min": 0.000957246491452679,
"max": 0.11027985538045565,
"count": 311
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.08708574796716372,
"min": 0.0009718410981198152,
"max": 0.11275544961293539,
"count": 311
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.08708574796716372,
"min": 0.0009718410981198152,
"max": 0.11275544961293539,
"count": 311
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0004359855728028881,
"min": 0.0004359855728028881,
"max": 0.0004997833100433381,
"count": 311
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0004359855728028881,
"min": 0.0004359855728028881,
"max": 0.0004997833100433381,
"count": 311
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.18719711199999994,
"min": 0.18719711199999994,
"max": 0.199956662,
"count": 311
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.18719711199999994,
"min": 0.18719711199999994,
"max": 0.199956662,
"count": 311
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.004361135888799998,
"min": 0.004361135888799998,
"max": 0.0049978374338,
"count": 311
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.004361135888799998,
"min": 0.004361135888799998,
"max": 0.0049978374338,
"count": 311
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1680263325",
"python_version": "3.9.16 (main, Mar 8 2023, 10:39:24) [MSC v.1916 64 bit (AMD64)]",
"command_line_arguments": "C:\\Users\\zhino\\anaconda3\\envs\\rl\\Scripts\\mlagents-learn ./config/poca/SoccerTwos2.yaml --env=./training-envs-executables/SoccerTwos.exe --run-id=SoccerTwos2 --no-graphics",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "2.0.0+cpu",
"numpy_version": "1.21.2",
"end_time_seconds": "1680315752"
},
"total": 52428.5448929,
"count": 1,
"self": 0.614798199996585,
"children": {
"run_training.setup": {
"total": 0.2207328999999998,
"count": 1,
"self": 0.2207328999999998
},
"TrainerController.start_learning": {
"total": 52427.7093618,
"count": 1,
"self": 31.030055797578825,
"children": {
"TrainerController._reset_env": {
"total": 7.169895799999432,
"count": 65,
"self": 7.169895799999432
},
"TrainerController.advance": {
"total": 52389.29691510243,
"count": 870702,
"self": 33.048831299718586,
"children": {
"env_step": {
"total": 20574.75001490272,
"count": 870702,
"self": 16087.847553507745,
"children": {
"SubprocessEnvManager._take_step": {
"total": 4467.564729396231,
"count": 870702,
"self": 176.7991257948488,
"children": {
"TorchPolicy.evaluate": {
"total": 4290.765603601382,
"count": 1616590,
"self": 4290.765603601382
}
}
},
"workers": {
"total": 19.33773199874161,
"count": 870702,
"self": 0.0,
"children": {
"worker_root": {
"total": 52316.70364719866,
"count": 870702,
"is_parallel": true,
"self": 39472.28898839814,
"children": {
"steps_from_proto": {
"total": 0.14765100000977327,
"count": 130,
"is_parallel": true,
"self": 0.030188600070926164,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.1174623999388471,
"count": 520,
"is_parallel": true,
"self": 0.1174623999388471
}
}
},
"UnityEnvironment.step": {
"total": 12844.26700780051,
"count": 870702,
"is_parallel": true,
"self": 667.9905243010526,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 525.8878633989822,
"count": 870702,
"is_parallel": true,
"self": 525.8878633989822
},
"communicator.exchange": {
"total": 9606.97778750024,
"count": 870702,
"is_parallel": true,
"self": 9606.97778750024
},
"steps_from_proto": {
"total": 2043.4108326002352,
"count": 1741404,
"is_parallel": true,
"self": 421.6623869006894,
"children": {
"_process_rank_one_or_two_observation": {
"total": 1621.7484456995458,
"count": 6965616,
"is_parallel": true,
"self": 1621.7484456995458
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 31781.49806889999,
"count": 870702,
"self": 210.89493119909457,
"children": {
"process_trajectory": {
"total": 4605.6541179009155,
"count": 870702,
"self": 4599.036565900917,
"children": {
"RLTrainer._checkpoint": {
"total": 6.6175519999987955,
"count": 25,
"self": 6.6175519999987955
}
}
},
"_update_policy": {
"total": 26964.94901979998,
"count": 312,
"self": 2240.8157296002682,
"children": {
"TorchPOCAOptimizer.update": {
"total": 24724.133290199712,
"count": 9347,
"self": 24724.133290199712
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.5999976312741637e-06,
"count": 1,
"self": 1.5999976312741637e-06
},
"TrainerController._save_models": {
"total": 0.21249349999561673,
"count": 1,
"self": 0.021035699995991308,
"children": {
"RLTrainer._checkpoint": {
"total": 0.19145779999962542,
"count": 1,
"self": 0.19145779999962542
}
}
}
}
}
}
}