{ "name": "root", "gauges": { "HotrodTrack2.Policy.Entropy.mean": { "value": 1.4688674211502075, "min": 1.4606980085372925, "max": 1.8379020690917969, "count": 226 }, "HotrodTrack2.Policy.Entropy.sum": { "value": 14547.6630859375, "min": 11550.4892578125, "max": 18437.833984375, "count": 226 }, "HotrodTrack2.Step.mean": { "value": 2779972.0, "min": 529965.0, "max": 2779972.0, "count": 226 }, "HotrodTrack2.Step.sum": { "value": 2779972.0, "min": 529965.0, "max": 2779972.0, "count": 226 }, "HotrodTrack2.Policy.ExtrinsicValueEstimate.mean": { "value": 4.824958324432373, "min": 3.2828564643859863, "max": 5.060272216796875, "count": 226 }, "HotrodTrack2.Policy.ExtrinsicValueEstimate.sum": { "value": 805.76806640625, "min": 376.0504455566406, "max": 845.0654296875, "count": 226 }, "HotrodTrack2.Environment.EpisodeLength.mean": { "value": 437.0869565217391, "min": 414.2083333333333, "max": 692.4285714285714, "count": 226 }, "HotrodTrack2.Environment.EpisodeLength.sum": { "value": 10053.0, "min": 4260.0, "max": 11493.0, "count": 226 }, "HotrodTrack2.Environment.CumulativeReward.mean": { "value": 23.611490470215518, "min": 13.981027029367397, "max": 24.549402566183183, "count": 226 }, "HotrodTrack2.Environment.CumulativeReward.sum": { "value": 543.0642808149569, "min": 111.84821623493917, "max": 590.2947730850428, "count": 226 }, "HotrodTrack2.Policy.ExtrinsicReward.mean": { "value": 23.611490470215518, "min": 13.981027029367397, "max": 24.549402566183183, "count": 226 }, "HotrodTrack2.Policy.ExtrinsicReward.sum": { "value": 543.0642808149569, "min": 111.84821623493917, "max": 590.2947730850428, "count": 226 }, "HotrodTrack2.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 226 }, "HotrodTrack2.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 226 }, "HotrodTrack2.Losses.PolicyLoss.mean": { "value": 0.10946844393568211, "min": 0.10277282706882078, "max": 0.12013487166725573, "count": 219 }, "HotrodTrack2.Losses.PolicyLoss.sum": { "value": 0.10946844393568211, "min": 0.10277282706882078, "max": 0.12013487166725573, "count": 219 }, "HotrodTrack2.Losses.ValueLoss.mean": { "value": 0.34673709232148114, "min": 0.3059479927023252, "max": 0.47748435271707007, "count": 219 }, "HotrodTrack2.Losses.ValueLoss.sum": { "value": 0.34673709232148114, "min": 0.3059479927023252, "max": 0.47748435271707007, "count": 219 }, "HotrodTrack2.Policy.LearningRate.mean": { "value": 0.0001335651554783, "min": 0.0001335651554783, "max": 0.00026795113068295995, "count": 219 }, "HotrodTrack2.Policy.LearningRate.sum": { "value": 0.0001335651554783, "min": 0.0001335651554783, "max": 0.00026795113068295995, "count": 219 }, "HotrodTrack2.Policy.Epsilon.mean": { "value": 0.14452170000000003, "min": 0.14452170000000003, "max": 0.18931704, "count": 219 }, "HotrodTrack2.Policy.Epsilon.sum": { "value": 0.14452170000000003, "min": 0.14452170000000003, "max": 0.18931704, "count": 219 }, "HotrodTrack2.Policy.Beta.mean": { "value": 0.022266397830000003, "min": 0.022266397830000003, "max": 0.04465958829600001, "count": 219 }, "HotrodTrack2.Policy.Beta.sum": { "value": 0.022266397830000003, "min": 0.022266397830000003, "max": 0.04465958829600001, "count": 219 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1714037967", "python_version": "3.9.13 (tags/v3.9.13:6de2ca5, May 17 2022, 16:36:42) [MSC v.1929 64 bit (AMD64)]", "command_line_arguments": "C:\\Users\\SukkertoppenDDU\\AppData\\Local\\Programs\\Python\\Python39\\Scripts\\mlagents-learn config/LLL.yaml --run-id=HotrodTrack2 --resume --env=Builds\\HotrodTrack2 part2\\Racesm_L.exe", "mlagents_version": "0.30.0", "mlagents_envs_version": "0.30.0", "communication_protocol_version": "1.5.0", "pytorch_version": "2.3.0+cpu", "numpy_version": "1.21.2", "end_time_seconds": "1714041202" }, "total": 3234.6716136, "count": 1, "self": 0.18517130000009274, "children": { "run_training.setup": { "total": 0.06581130000000002, "count": 1, "self": 0.06581130000000002 }, "TrainerController.start_learning": { "total": 3234.420631, "count": 1, "self": 5.433118900041791, "children": { "TrainerController._reset_env": { "total": 5.5146695, "count": 1, "self": 5.5146695 }, "TrainerController.advance": { "total": 3223.422524499958, "count": 286047, "self": 4.9960255999667424, "children": { "env_step": { "total": 1793.434575099959, "count": 286047, "self": 1334.5602916998732, "children": { "SubprocessEnvManager._take_step": { "total": 455.704912300043, "count": 286047, "self": 17.3271994000836, "children": { "TorchPolicy.evaluate": { "total": 438.3777128999594, "count": 282350, "self": 438.3777128999594 } } }, "workers": { "total": 3.169371100042804, "count": 286046, "self": 0.0, "children": { "worker_root": { "total": 3224.0514847999452, "count": 286046, "is_parallel": true, "self": 2178.365734600079, "children": { "steps_from_proto": { "total": 0.0008873999999998716, "count": 1, "is_parallel": true, "self": 0.00021310000000074325, "children": { "_process_rank_one_or_two_observation": { "total": 0.0006742999999991284, "count": 6, "is_parallel": true, "self": 0.0006742999999991284 } } }, "UnityEnvironment.step": { "total": 1045.6848627998659, "count": 286046, "is_parallel": true, "self": 33.65867239990712, "children": { "UnityEnvironment._generate_step_input": { "total": 61.004303999989204, "count": 286046, "is_parallel": true, "self": 61.004303999989204 }, "communicator.exchange": { "total": 848.6281209999536, "count": 286046, "is_parallel": true, "self": 848.6281209999536 }, "steps_from_proto": { "total": 102.3937654000159, "count": 286046, "is_parallel": true, "self": 39.71735850005038, "children": { "_process_rank_one_or_two_observation": { "total": 62.676406899965514, "count": 1716276, "is_parallel": true, "self": 62.676406899965514 } } } } } } } } } } }, "trainer_advance": { "total": 1424.991923800032, "count": 286046, "self": 7.689493400025185, "children": { "process_trajectory": { "total": 226.76586240000722, "count": 286046, "self": 226.49428270000703, "children": { "RLTrainer._checkpoint": { "total": 0.2715797000001885, "count": 4, "self": 0.2715797000001885 } } }, "_update_policy": { "total": 1190.5365679999998, "count": 219, "self": 426.9760664000719, "children": { "TorchPPOOptimizer.update": { "total": 763.5605015999279, "count": 134649, "self": 763.5605015999279 } } } } } } }, "trainer_threads": { "total": 8.000001798791345e-07, "count": 1, "self": 8.000001798791345e-07 }, "TrainerController._save_models": { "total": 0.05031730000018797, "count": 1, "self": 0.007384699999875011, "children": { "RLTrainer._checkpoint": { "total": 0.04293260000031296, "count": 1, "self": 0.04293260000031296 } } } } } } }