{ "name": "root", "gauges": { "BusTrack2.Policy.Entropy.mean": { "value": 1.1329643726348877, "min": 0.8422980904579163, "max": 2.189603328704834, "count": 260 }, "BusTrack2.Policy.Entropy.sum": { "value": 11302.4521484375, "min": 8396.02734375, "max": 22141.26953125, "count": 260 }, "BusTrack2.Step.mean": { "value": 2599948.0, "min": 9984.0, "max": 2599948.0, "count": 260 }, "BusTrack2.Step.sum": { "value": 2599948.0, "min": 9984.0, "max": 2599948.0, "count": 260 }, "BusTrack2.Policy.ExtrinsicValueEstimate.mean": { "value": 5.365256309509277, "min": -1.2024420499801636, "max": 5.365256309509277, "count": 260 }, "BusTrack2.Policy.ExtrinsicValueEstimate.sum": { "value": 928.1893310546875, "min": -238.08352661132812, "max": 946.985107421875, "count": 260 }, "BusTrack2.Environment.EpisodeLength.mean": { "value": 400.2307692307692, "min": 199.0, "max": 687.4, "count": 260 }, "BusTrack2.Environment.EpisodeLength.sum": { "value": 10406.0, "min": 7857.0, "max": 11564.0, "count": 260 }, "BusTrack2.Environment.CumulativeReward.mean": { "value": 24.59055496064516, "min": -0.12888464100554606, "max": 25.026868030428886, "count": 260 }, "BusTrack2.Environment.CumulativeReward.sum": { "value": 639.3544289767742, "min": -6.18646276826621, "max": 664.3267043232918, "count": 260 }, "BusTrack2.Policy.ExtrinsicReward.mean": { "value": 24.59055496064516, "min": -0.12888464100554606, "max": 25.026868030428886, "count": 260 }, "BusTrack2.Policy.ExtrinsicReward.sum": { "value": 639.3544289767742, "min": -6.18646276826621, "max": 664.3267043232918, "count": 260 }, "BusTrack2.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 260 }, "BusTrack2.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 260 }, "BusTrack2.Losses.PolicyLoss.mean": { "value": 0.11128140402090834, "min": 0.10049134346719678, "max": 0.12087385670970027, "count": 252 }, "BusTrack2.Losses.PolicyLoss.sum": { "value": 0.11128140402090834, "min": 0.10049134346719678, "max": 0.12087385670970027, "count": 252 }, "BusTrack2.Losses.ValueLoss.mean": { "value": 0.2692062053738571, "min": 0.002542790192565585, "max": 0.442403101896852, "count": 252 }, "BusTrack2.Losses.ValueLoss.sum": { "value": 0.2692062053738571, "min": 0.002542790192565585, "max": 0.442403101896852, "count": 252 }, "BusTrack2.Policy.LearningRate.mean": { "value": 0.00014448515183829996, "min": 0.00014448515183829996, "max": 0.00029936256021247993, "count": 252 }, "BusTrack2.Policy.LearningRate.sum": { "value": 0.00014448515183829996, "min": 0.00014448515183829996, "max": 0.00029936256021247993, "count": 252 }, "BusTrack2.Policy.Epsilon.mean": { "value": 0.14816169999999998, "min": 0.14816169999999998, "max": 0.19978752000000002, "count": 252 }, "BusTrack2.Policy.Epsilon.sum": { "value": 0.14816169999999998, "min": 0.14816169999999998, "max": 0.19978752000000002, "count": 252 }, "BusTrack2.Policy.Beta.mean": { "value": 0.024086033829999996, "min": 0.024086033829999996, "max": 0.04989378124799999, "count": 252 }, "BusTrack2.Policy.Beta.sum": { "value": 0.024086033829999996, "min": 0.024086033829999996, "max": 0.04989378124799999, "count": 252 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1714041537", "python_version": "3.9.13 (tags/v3.9.13:6de2ca5, May 17 2022, 16:36:42) [MSC v.1929 64 bit (AMD64)]", "command_line_arguments": "C:\\Users\\SukkertoppenDDU\\AppData\\Local\\Programs\\Python\\Python39\\Scripts\\mlagents-learn config/LLL.yaml --force --run-id=BusTrack2 --env=Builds\\BusTrack2 part1\\Racesm_L.exe", "mlagents_version": "0.30.0", "mlagents_envs_version": "0.30.0", "communication_protocol_version": "1.5.0", "pytorch_version": "2.3.0+cpu", "numpy_version": "1.21.2", "end_time_seconds": "1714045697" }, "total": 4160.6230477, "count": 1, "self": 1.1064352000003055, "children": { "run_training.setup": { "total": 0.06439529999999993, "count": 1, "self": 0.06439529999999993 }, "TrainerController.start_learning": { "total": 4159.4522172, "count": 1, "self": 6.86582220004675, "children": { "TrainerController._reset_env": { "total": 8.1418034, "count": 1, "self": 8.1418034 }, "TrainerController.advance": { "total": 4144.3929840999535, "count": 329660, "self": 6.501555599788844, "children": { "env_step": { "total": 2443.7344946999947, "count": 329660, "self": 1881.5832031000455, "children": { "SubprocessEnvManager._take_step": { "total": 558.1467443999725, "count": 329660, "self": 20.998963999995112, "children": { "TorchPolicy.evaluate": { "total": 537.1477803999774, "count": 325349, "self": 537.1477803999774 } } }, "workers": { "total": 4.004547199976585, "count": 329659, "self": 0.0, "children": { "worker_root": { "total": 4145.043805499976, "count": 329659, "is_parallel": true, "self": 2624.752221099995, "children": { "steps_from_proto": { "total": 0.0003552999999998363, "count": 1, "is_parallel": true, "self": 0.00012589999999956802, "children": { "_process_rank_one_or_two_observation": { "total": 0.00022940000000026828, "count": 6, "is_parallel": true, "self": 0.00022940000000026828 } } }, "UnityEnvironment.step": { "total": 1520.2912290999807, "count": 329659, "is_parallel": true, "self": 41.948900199665786, "children": { "UnityEnvironment._generate_step_input": { "total": 74.12176200007869, "count": 329659, "is_parallel": true, "self": 74.12176200007869 }, "communicator.exchange": { "total": 1279.3069929000733, "count": 329659, "is_parallel": true, "self": 1279.3069929000733 }, "steps_from_proto": { "total": 124.91357400016305, "count": 329659, "is_parallel": true, "self": 48.08493170013452, "children": { "_process_rank_one_or_two_observation": { "total": 76.82864230002853, "count": 1977954, "is_parallel": true, "self": 76.82864230002853 } } } } } } } } } } }, "trainer_advance": { "total": 1694.1569338001696, "count": 329659, "self": 9.622891000080472, "children": { "process_trajectory": { "total": 267.8110550000872, "count": 329659, "self": 267.3326875000871, "children": { "RLTrainer._checkpoint": { "total": 0.478367500000104, "count": 5, "self": 0.478367500000104 } } }, "_update_policy": { "total": 1416.7229878000019, "count": 253, "self": 563.0196431000599, "children": { "TorchPPOOptimizer.update": { "total": 853.703344699942, "count": 155727, "self": 853.703344699942 } } } } } } }, "trainer_threads": { "total": 6.999998731771484e-07, "count": 1, "self": 6.999998731771484e-07 }, "TrainerController._save_models": { "total": 0.05160679999971762, "count": 1, "self": 0.006841899999017187, "children": { "RLTrainer._checkpoint": { "total": 0.044764900000700436, "count": 1, "self": 0.044764900000700436 } } } } } } }