{ "name": "root", "gauges": { "BensonV6.Policy.Entropy.mean": { "value": 1.5862776041030884, "min": 1.365457534790039, "max": 2.193558931350708, "count": 33 }, "BensonV6.Policy.Entropy.sum": { "value": 15957.9521484375, "min": 13736.5029296875, "max": 22374.30078125, "count": 33 }, "BensonV6.Environment.EpisodeLength.mean": { "value": 230.2325581395349, "min": 59.56547619047619, "max": 345.2, "count": 33 }, "BensonV6.Environment.EpisodeLength.sum": { "value": 9900.0, "min": 8960.0, "max": 11034.0, "count": 33 }, "BensonV6.Step.mean": { "value": 329973.0, "min": 9995.0, "max": 329973.0, "count": 33 }, "BensonV6.Step.sum": { "value": 329973.0, "min": 9995.0, "max": 329973.0, "count": 33 }, "BensonV6.Policy.ExtrinsicValueEstimate.mean": { "value": 5.722393989562988, "min": 0.2746765613555908, "max": 5.809690475463867, "count": 33 }, "BensonV6.Policy.ExtrinsicValueEstimate.sum": { "value": 989.97412109375, "min": 46.69501495361328, "max": 1022.5054931640625, "count": 33 }, "BensonV6.Environment.CumulativeReward.mean": { "value": 19.246166612901508, "min": -0.019117560070522267, "max": 23.05193606035318, "count": 33 }, "BensonV6.Environment.CumulativeReward.sum": { "value": 827.5851643547649, "min": -3.1161622914951295, "max": 922.0774424141273, "count": 33 }, "BensonV6.Policy.ExtrinsicReward.mean": { "value": 19.246166612901508, "min": -0.019117560070522267, "max": 23.05193606035318, "count": 33 }, "BensonV6.Policy.ExtrinsicReward.sum": { "value": 827.5851643547649, "min": -3.1161622914951295, "max": 922.0774424141273, "count": 33 }, "BensonV6.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 33 }, "BensonV6.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 33 }, "BensonV6.Losses.PolicyLoss.mean": { "value": 0.11866796751207544, "min": 0.0957306937479255, "max": 0.12068313519878321, "count": 32 }, "BensonV6.Losses.PolicyLoss.sum": { "value": 0.11866796751207544, "min": 0.0957306937479255, "max": 0.12068313519878321, "count": 32 }, "BensonV6.Losses.ValueLoss.mean": { "value": 0.6118914237836512, "min": 0.021924442883459776, "max": 0.7137479831288064, "count": 32 }, "BensonV6.Losses.ValueLoss.sum": { "value": 0.6118914237836512, "min": 0.021924442883459776, "max": 0.7137479831288064, "count": 32 }, "BensonV6.Policy.LearningRate.mean": { "value": 0.00026712281095906667, "min": 0.00026712281095906667, "max": 0.00029897050034316655, "count": 32 }, "BensonV6.Policy.LearningRate.sum": { "value": 0.00026712281095906667, "min": 0.00026712281095906667, "max": 0.00029897050034316655, "count": 32 }, "BensonV6.Policy.Epsilon.mean": { "value": 0.18904093333333336, "min": 0.18904093333333336, "max": 0.19965683333333328, "count": 32 }, "BensonV6.Policy.Epsilon.sum": { "value": 0.18904093333333336, "min": 0.18904093333333336, "max": 0.19965683333333328, "count": 32 }, "BensonV6.Policy.Beta.mean": { "value": 0.04452156257333334, "min": 0.04452156257333334, "max": 0.049828450983333335, "count": 32 }, "BensonV6.Policy.Beta.sum": { "value": 0.04452156257333334, "min": 0.04452156257333334, "max": 0.049828450983333335, "count": 32 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1713623140", "python_version": "3.9.13 (tags/v3.9.13:6de2ca5, May 17 2022, 16:36:42) [MSC v.1929 64 bit (AMD64)]", "command_line_arguments": "C:\\Users\\noahk\\Documents\\Unity projects\\Racesm\\.venv\\Scripts\\mlagents-learn config/BensonV6.yaml --run-id=BensonV6 --force", "mlagents_version": "0.30.0", "mlagents_envs_version": "0.30.0", "communication_protocol_version": "1.5.0", "pytorch_version": "2.2.2+cu118", "numpy_version": "1.21.2", "end_time_seconds": "1713623853" }, "total": 712.8463069000001, "count": 1, "self": 0.00745920000008482, "children": { "run_training.setup": { "total": 0.07786059999999995, "count": 1, "self": 0.07786059999999995 }, "TrainerController.start_learning": { "total": 712.7609871, "count": 1, "self": 0.7619565000039756, "children": { "TrainerController._reset_env": { "total": 5.778238099999999, "count": 1, "self": 5.778238099999999 }, "TrainerController.advance": { "total": 706.039172599996, "count": 34801, "self": 0.612005799983308, "children": { "env_step": { "total": 415.03360060000045, "count": 34801, "self": 265.1050452999982, "children": { "SubprocessEnvManager._take_step": { "total": 149.45768900000039, "count": 34801, "self": 2.116340399996801, "children": { "TorchPolicy.evaluate": { "total": 147.34134860000358, "count": 33434, "self": 147.34134860000358 } } }, "workers": { "total": 0.47086630000189267, "count": 34800, "self": 0.0, "children": { "worker_root": { "total": 706.6658678000041, "count": 34800, "is_parallel": true, "self": 482.1807981999955, "children": { "steps_from_proto": { "total": 0.0005944000000006611, "count": 1, "is_parallel": true, "self": 0.00020790000000125985, "children": { "_process_rank_one_or_two_observation": { "total": 0.0003864999999994012, "count": 6, "is_parallel": true, "self": 0.0003864999999994012 } } }, "UnityEnvironment.step": { "total": 224.4844752000086, "count": 34800, "is_parallel": true, "self": 5.25407010000518, "children": { "UnityEnvironment._generate_step_input": { "total": 5.428810699995278, "count": 34800, "is_parallel": true, "self": 5.428810699995278 }, "communicator.exchange": { "total": 198.2789811000104, "count": 34800, "is_parallel": true, "self": 198.2789811000104 }, "steps_from_proto": { "total": 15.522613299997753, "count": 34800, "is_parallel": true, "self": 5.879353500006703, "children": { "_process_rank_one_or_two_observation": { "total": 9.64325979999105, "count": 208800, "is_parallel": true, "self": 9.64325979999105 } } } } } } } } } } }, "trainer_advance": { "total": 290.3935662000123, "count": 34800, "self": 1.3219589000069618, "children": { "process_trajectory": { "total": 40.38120300000543, "count": 34800, "self": 40.38120300000543 }, "_update_policy": { "total": 248.69040429999993, "count": 32, "self": 46.19149409999946, "children": { "TorchPPOOptimizer.update": { "total": 202.49891020000047, "count": 19674, "self": 202.49891020000047 } } } } } } }, "trainer_threads": { "total": 1.6999999843392288e-06, "count": 1, "self": 1.6999999843392288e-06 }, "TrainerController._save_models": { "total": 0.18161820000000262, "count": 1, "self": 0.011474000000021078, "children": { "RLTrainer._checkpoint": { "total": 0.17014419999998154, "count": 1, "self": 0.17014419999998154 } } } } } } }