This commit is contained in:
2024-05-03 13:41:30 +02:00
parent 1e53774e99
commit 34adad0d55
159 changed files with 5919 additions and 163 deletions

View File

@@ -0,0 +1,320 @@
{
"name": "root",
"gauges": {
"HotrodTrack2.Policy.Entropy.mean": {
"value": 1.4688674211502075,
"min": 1.4606980085372925,
"max": 1.8379020690917969,
"count": 226
},
"HotrodTrack2.Policy.Entropy.sum": {
"value": 14547.6630859375,
"min": 11550.4892578125,
"max": 18437.833984375,
"count": 226
},
"HotrodTrack2.Step.mean": {
"value": 2779972.0,
"min": 529965.0,
"max": 2779972.0,
"count": 226
},
"HotrodTrack2.Step.sum": {
"value": 2779972.0,
"min": 529965.0,
"max": 2779972.0,
"count": 226
},
"HotrodTrack2.Policy.ExtrinsicValueEstimate.mean": {
"value": 4.824958324432373,
"min": 3.2828564643859863,
"max": 5.060272216796875,
"count": 226
},
"HotrodTrack2.Policy.ExtrinsicValueEstimate.sum": {
"value": 805.76806640625,
"min": 376.0504455566406,
"max": 845.0654296875,
"count": 226
},
"HotrodTrack2.Environment.EpisodeLength.mean": {
"value": 437.0869565217391,
"min": 414.2083333333333,
"max": 692.4285714285714,
"count": 226
},
"HotrodTrack2.Environment.EpisodeLength.sum": {
"value": 10053.0,
"min": 4260.0,
"max": 11493.0,
"count": 226
},
"HotrodTrack2.Environment.CumulativeReward.mean": {
"value": 23.611490470215518,
"min": 13.981027029367397,
"max": 24.549402566183183,
"count": 226
},
"HotrodTrack2.Environment.CumulativeReward.sum": {
"value": 543.0642808149569,
"min": 111.84821623493917,
"max": 590.2947730850428,
"count": 226
},
"HotrodTrack2.Policy.ExtrinsicReward.mean": {
"value": 23.611490470215518,
"min": 13.981027029367397,
"max": 24.549402566183183,
"count": 226
},
"HotrodTrack2.Policy.ExtrinsicReward.sum": {
"value": 543.0642808149569,
"min": 111.84821623493917,
"max": 590.2947730850428,
"count": 226
},
"HotrodTrack2.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 226
},
"HotrodTrack2.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 226
},
"HotrodTrack2.Losses.PolicyLoss.mean": {
"value": 0.10946844393568211,
"min": 0.10277282706882078,
"max": 0.12013487166725573,
"count": 219
},
"HotrodTrack2.Losses.PolicyLoss.sum": {
"value": 0.10946844393568211,
"min": 0.10277282706882078,
"max": 0.12013487166725573,
"count": 219
},
"HotrodTrack2.Losses.ValueLoss.mean": {
"value": 0.34673709232148114,
"min": 0.3059479927023252,
"max": 0.47748435271707007,
"count": 219
},
"HotrodTrack2.Losses.ValueLoss.sum": {
"value": 0.34673709232148114,
"min": 0.3059479927023252,
"max": 0.47748435271707007,
"count": 219
},
"HotrodTrack2.Policy.LearningRate.mean": {
"value": 0.0001335651554783,
"min": 0.0001335651554783,
"max": 0.00026795113068295995,
"count": 219
},
"HotrodTrack2.Policy.LearningRate.sum": {
"value": 0.0001335651554783,
"min": 0.0001335651554783,
"max": 0.00026795113068295995,
"count": 219
},
"HotrodTrack2.Policy.Epsilon.mean": {
"value": 0.14452170000000003,
"min": 0.14452170000000003,
"max": 0.18931704,
"count": 219
},
"HotrodTrack2.Policy.Epsilon.sum": {
"value": 0.14452170000000003,
"min": 0.14452170000000003,
"max": 0.18931704,
"count": 219
},
"HotrodTrack2.Policy.Beta.mean": {
"value": 0.022266397830000003,
"min": 0.022266397830000003,
"max": 0.04465958829600001,
"count": 219
},
"HotrodTrack2.Policy.Beta.sum": {
"value": 0.022266397830000003,
"min": 0.022266397830000003,
"max": 0.04465958829600001,
"count": 219
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1714037967",
"python_version": "3.9.13 (tags/v3.9.13:6de2ca5, May 17 2022, 16:36:42) [MSC v.1929 64 bit (AMD64)]",
"command_line_arguments": "C:\\Users\\SukkertoppenDDU\\AppData\\Local\\Programs\\Python\\Python39\\Scripts\\mlagents-learn config/LLL.yaml --run-id=HotrodTrack2 --resume --env=Builds\\HotrodTrack2 part2\\Racesm_L.exe",
"mlagents_version": "0.30.0",
"mlagents_envs_version": "0.30.0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "2.3.0+cpu",
"numpy_version": "1.21.2",
"end_time_seconds": "1714041202"
},
"total": 3234.6716136,
"count": 1,
"self": 0.18517130000009274,
"children": {
"run_training.setup": {
"total": 0.06581130000000002,
"count": 1,
"self": 0.06581130000000002
},
"TrainerController.start_learning": {
"total": 3234.420631,
"count": 1,
"self": 5.433118900041791,
"children": {
"TrainerController._reset_env": {
"total": 5.5146695,
"count": 1,
"self": 5.5146695
},
"TrainerController.advance": {
"total": 3223.422524499958,
"count": 286047,
"self": 4.9960255999667424,
"children": {
"env_step": {
"total": 1793.434575099959,
"count": 286047,
"self": 1334.5602916998732,
"children": {
"SubprocessEnvManager._take_step": {
"total": 455.704912300043,
"count": 286047,
"self": 17.3271994000836,
"children": {
"TorchPolicy.evaluate": {
"total": 438.3777128999594,
"count": 282350,
"self": 438.3777128999594
}
}
},
"workers": {
"total": 3.169371100042804,
"count": 286046,
"self": 0.0,
"children": {
"worker_root": {
"total": 3224.0514847999452,
"count": 286046,
"is_parallel": true,
"self": 2178.365734600079,
"children": {
"steps_from_proto": {
"total": 0.0008873999999998716,
"count": 1,
"is_parallel": true,
"self": 0.00021310000000074325,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0006742999999991284,
"count": 6,
"is_parallel": true,
"self": 0.0006742999999991284
}
}
},
"UnityEnvironment.step": {
"total": 1045.6848627998659,
"count": 286046,
"is_parallel": true,
"self": 33.65867239990712,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 61.004303999989204,
"count": 286046,
"is_parallel": true,
"self": 61.004303999989204
},
"communicator.exchange": {
"total": 848.6281209999536,
"count": 286046,
"is_parallel": true,
"self": 848.6281209999536
},
"steps_from_proto": {
"total": 102.3937654000159,
"count": 286046,
"is_parallel": true,
"self": 39.71735850005038,
"children": {
"_process_rank_one_or_two_observation": {
"total": 62.676406899965514,
"count": 1716276,
"is_parallel": true,
"self": 62.676406899965514
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 1424.991923800032,
"count": 286046,
"self": 7.689493400025185,
"children": {
"process_trajectory": {
"total": 226.76586240000722,
"count": 286046,
"self": 226.49428270000703,
"children": {
"RLTrainer._checkpoint": {
"total": 0.2715797000001885,
"count": 4,
"self": 0.2715797000001885
}
}
},
"_update_policy": {
"total": 1190.5365679999998,
"count": 219,
"self": 426.9760664000719,
"children": {
"TorchPPOOptimizer.update": {
"total": 763.5605015999279,
"count": 134649,
"self": 763.5605015999279
}
}
}
}
}
}
},
"trainer_threads": {
"total": 8.000001798791345e-07,
"count": 1,
"self": 8.000001798791345e-07
},
"TrainerController._save_models": {
"total": 0.05031730000018797,
"count": 1,
"self": 0.007384699999875011,
"children": {
"RLTrainer._checkpoint": {
"total": 0.04293260000031296,
"count": 1,
"self": 0.04293260000031296
}
}
}
}
}
}
}