{ "name": "root", "gauges": { "RacecarTrack2.Policy.Entropy.mean": { "value": 1.3993903398513794, "min": 1.3993903398513794, "max": 1.8428030014038086, "count": 132 }, "RacecarTrack2.Policy.Entropy.sum": { "value": 13993.9033203125, "min": 8835.998046875, "max": 18604.939453125, "count": 132 }, "RacecarTrack2.Step.mean": { "value": 2029969.0, "min": 719975.0, "max": 2029969.0, "count": 132 }, "RacecarTrack2.Step.sum": { "value": 2029969.0, "min": 719975.0, "max": 2029969.0, "count": 132 }, "RacecarTrack2.Policy.ExtrinsicValueEstimate.mean": { "value": 7.826251983642578, "min": 4.873648643493652, "max": 7.84055233001709, "count": 132 }, "RacecarTrack2.Policy.ExtrinsicValueEstimate.sum": { "value": 1361.767822265625, "min": 360.6499938964844, "max": 1361.767822265625, "count": 132 }, "RacecarTrack2.Environment.EpisodeLength.mean": { "value": 236.46666666666667, "min": 235.62790697674419, "max": 421.25, "count": 132 }, "RacecarTrack2.Environment.EpisodeLength.sum": { "value": 10641.0, "min": 3370.0, "max": 11103.0, "count": 132 }, "RacecarTrack2.Environment.CumulativeReward.mean": { "value": 24.35301135579745, "min": 16.950463268905878, "max": 24.41715809371736, "count": 132 }, "RacecarTrack2.Environment.CumulativeReward.sum": { "value": 1095.8855110108852, "min": 135.60370615124702, "max": 1095.8855110108852, "count": 132 }, "RacecarTrack2.Policy.ExtrinsicReward.mean": { "value": 24.35301135579745, "min": 16.950463268905878, "max": 24.41715809371736, "count": 132 }, "RacecarTrack2.Policy.ExtrinsicReward.sum": { "value": 1095.8855110108852, "min": 135.60370615124702, "max": 1095.8855110108852, "count": 132 }, "RacecarTrack2.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 132 }, "RacecarTrack2.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 132 }, "RacecarTrack2.Losses.PolicyLoss.mean": { "value": 0.11060398251198732, "min": 0.10318752745090941, "max": 0.1207956799904142, "count": 127 }, "RacecarTrack2.Losses.PolicyLoss.sum": { "value": 0.11060398251198732, "min": 0.10318752745090941, "max": 0.1207956799904142, "count": 127 }, "RacecarTrack2.Losses.ValueLoss.mean": { "value": 0.1292799282001286, "min": 0.09924579720671584, "max": 0.5470766848180352, "count": 127 }, "RacecarTrack2.Losses.ValueLoss.sum": { "value": 0.1292799282001286, "min": 0.09924579720671584, "max": 0.5470766848180352, "count": 127 }, "RacecarTrack2.Policy.LearningRate.mean": { "value": 0.00017878480040508003, "min": 0.00017878480040508003, "max": 0.00025645633451456005, "count": 127 }, "RacecarTrack2.Policy.LearningRate.sum": { "value": 0.00017878480040508003, "min": 0.00017878480040508003, "max": 0.00025645633451456005, "count": 127 }, "RacecarTrack2.Policy.Epsilon.mean": { "value": 0.15959492, "min": 0.15959492, "max": 0.18548543999999997, "count": 127 }, "RacecarTrack2.Policy.Epsilon.sum": { "value": 0.15959492, "min": 0.15959492, "max": 0.18548543999999997, "count": 127 }, "RacecarTrack2.Policy.Beta.mean": { "value": 0.029801500508000005, "min": 0.029801500508000005, "max": 0.042744171456000006, "count": 127 }, "RacecarTrack2.Policy.Beta.sum": { "value": 0.029801500508000005, "min": 0.029801500508000005, "max": 0.042744171456000006, "count": 127 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1714033897", "python_version": "3.9.13 (tags/v3.9.13:6de2ca5, May 17 2022, 16:36:42) [MSC v.1929 64 bit (AMD64)]", "command_line_arguments": "C:\\Users\\SukkertoppenDDU\\AppData\\Local\\Programs\\Python\\Python39\\Scripts\\mlagents-learn config/LLL.yaml --run-id=RacecarTrack2 --resume --env=Builds\\RacecarTrack2 part2\\Racesm_L.exe", "mlagents_version": "0.30.0", "mlagents_envs_version": "0.30.0", "communication_protocol_version": "1.5.0", "pytorch_version": "2.3.0+cpu", "numpy_version": "1.21.2", "end_time_seconds": "1714035849" }, "total": 1951.6701458999999, "count": 1, "self": 0.2424239999998008, "children": { "run_training.setup": { "total": 0.06349230000000006, "count": 1, "self": 0.06349230000000006 }, "TrainerController.start_learning": { "total": 1951.3642296, "count": 1, "self": 3.1981367999824215, "children": { "TrainerController._reset_env": { "total": 5.4445986, "count": 1, "self": 5.4445986 }, "TrainerController.advance": { "total": 1942.6599118000177, "count": 168232, "self": 3.1001032000438045, "children": { "env_step": { "total": 1085.9256885000104, "count": 168232, "self": 834.5117967000376, "children": { "SubprocessEnvManager._take_step": { "total": 249.48164799997926, "count": 168232, "self": 9.625592899954654, "children": { "TorchPolicy.evaluate": { "total": 239.8560551000246, "count": 164419, "self": 239.8560551000246 } } }, "workers": { "total": 1.9322437999935467, "count": 168232, "self": 0.0, "children": { "worker_root": { "total": 1940.8134554999986, "count": 168232, "is_parallel": true, "self": 1271.5039737000238, "children": { "steps_from_proto": { "total": 0.00046560000000006596, "count": 1, "is_parallel": true, "self": 0.00016540000000198063, "children": { "_process_rank_one_or_two_observation": { "total": 0.00030019999999808533, "count": 6, "is_parallel": true, "self": 0.00030019999999808533 } } }, "UnityEnvironment.step": { "total": 669.3090161999748, "count": 168232, "is_parallel": true, "self": 20.44959419988686, "children": { "UnityEnvironment._generate_step_input": { "total": 32.56759710003403, "count": 168232, "is_parallel": true, "self": 32.56759710003403 }, "communicator.exchange": { "total": 556.4037336000775, "count": 168232, "is_parallel": true, "self": 556.4037336000775 }, "steps_from_proto": { "total": 59.888091299976395, "count": 168232, "is_parallel": true, "self": 22.70735709995128, "children": { "_process_rank_one_or_two_observation": { "total": 37.180734200025114, "count": 1009392, "is_parallel": true, "self": 37.180734200025114 } } } } } } } } } } }, "trainer_advance": { "total": 853.6341200999634, "count": 168232, "self": 4.475069199966924, "children": { "process_trajectory": { "total": 121.44999709999652, "count": 168232, "self": 121.23218689999649, "children": { "RLTrainer._checkpoint": { "total": 0.21781020000003082, "count": 3, "self": 0.21781020000003082 } } }, "_update_policy": { "total": 727.7090538, "count": 128, "self": 272.1004473000106, "children": { "TorchPPOOptimizer.update": { "total": 455.6086064999894, "count": 78350, "self": 455.6086064999894 } } } } } } }, "trainer_threads": { "total": 9.000000318337698e-07, "count": 1, "self": 9.000000318337698e-07 }, "TrainerController._save_models": { "total": 0.06158149999987472, "count": 1, "self": 0.009829899999886038, "children": { "RLTrainer._checkpoint": { "total": 0.051751599999988684, "count": 1, "self": 0.051751599999988684 } } } } } } }