{ "name": "root", "gauges": { "BusTrack2.Policy.Entropy.mean": { "value": 2.1863579750061035, "min": 2.1749253273010254, "max": 2.1947851181030273, "count": 5 }, "BusTrack2.Policy.Entropy.sum": { "value": 21968.525390625, "min": 21644.857421875, "max": 22053.201171875, "count": 5 }, "BusTrack2.Step.mean": { "value": 49984.0, "min": 9984.0, "max": 49984.0, "count": 5 }, "BusTrack2.Step.sum": { "value": 49984.0, "min": 9984.0, "max": 49984.0, "count": 5 }, "BusTrack2.Policy.ExtrinsicValueEstimate.mean": { "value": 0.0749274343252182, "min": 0.0749274343252182, "max": 0.2529940903186798, "count": 5 }, "BusTrack2.Policy.ExtrinsicValueEstimate.sum": { "value": 11.98838996887207, "min": 11.98838996887207, "max": 40.22605895996094, "count": 5 }, "BusTrack2.Environment.EpisodeLength.mean": { "value": 499.0, "min": 499.0, "max": 499.0, "count": 5 }, "BusTrack2.Environment.EpisodeLength.sum": { "value": 7984.0, "min": 7984.0, "max": 11976.0, "count": 5 }, "BusTrack2.Environment.CumulativeReward.mean": { "value": -0.30098484913126894, "min": -0.7619615873187761, "max": -0.19303640429825464, "count": 5 }, "BusTrack2.Environment.CumulativeReward.sum": { "value": -5.116742435231572, "min": -17.52511650833185, "max": -3.088582468772074, "count": 5 }, "BusTrack2.Policy.ExtrinsicReward.mean": { "value": -0.30098484913126894, "min": -0.7619615873187761, "max": -0.19303640429825464, "count": 5 }, "BusTrack2.Policy.ExtrinsicReward.sum": { "value": -5.116742435231572, "min": -17.52511650833185, "max": -3.088582468772074, "count": 5 }, "BusTrack2.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 5 }, "BusTrack2.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 5 }, "BusTrack2.Losses.PolicyLoss.mean": { "value": 0.11586630168607447, "min": 0.11050240805861049, "max": 0.11743067037271877, "count": 4 }, "BusTrack2.Losses.PolicyLoss.sum": { "value": 0.11586630168607447, "min": 0.11050240805861049, "max": 0.11743067037271877, "count": 4 }, "BusTrack2.Losses.ValueLoss.mean": { "value": 0.0026205912706508275, "min": 0.0026205912706508275, "max": 0.01049612028478976, "count": 4 }, "BusTrack2.Losses.ValueLoss.sum": { "value": 0.0026205912706508275, "min": 0.0026205912706508275, "max": 0.01049612028478976, "count": 4 }, "BusTrack2.Policy.LearningRate.mean": { "value": 0.00029747712084095994, "min": 0.00029747712084095994, "max": 0.0002993664002111999, "count": 4 }, "BusTrack2.Policy.LearningRate.sum": { "value": 0.00029747712084095994, "min": 0.00029747712084095994, "max": 0.0002993664002111999, "count": 4 }, "BusTrack2.Policy.Epsilon.mean": { "value": 0.19915904000000004, "min": 0.19915904000000004, "max": 0.19978880000000004, "count": 4 }, "BusTrack2.Policy.Epsilon.sum": { "value": 0.19915904000000004, "min": 0.19915904000000004, "max": 0.19978880000000004, "count": 4 }, "BusTrack2.Policy.Beta.mean": { "value": 0.04957960409599999, "min": 0.04957960409599999, "max": 0.049894421120000004, "count": 4 }, "BusTrack2.Policy.Beta.sum": { "value": 0.04957960409599999, "min": 0.04957960409599999, "max": 0.049894421120000004, "count": 4 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1713956580", "python_version": "3.9.13 (tags/v3.9.13:6de2ca5, May 17 2022, 16:36:42) [MSC v.1929 64 bit (AMD64)]", "command_line_arguments": "C:\\Users\\noahk\\Documents\\Unity projects\\Racesm\\.venv\\Scripts\\mlagents-learn config/BusTrack2.yaml --run-id=BusTrack2 --force", "mlagents_version": "0.30.0", "mlagents_envs_version": "0.30.0", "communication_protocol_version": "1.5.0", "pytorch_version": "2.2.2+cu118", "numpy_version": "1.21.2", "end_time_seconds": "1713956677" }, "total": 96.8911192, "count": 1, "self": 0.004357900000016457, "children": { "run_training.setup": { "total": 0.08156290000000022, "count": 1, "self": 0.08156290000000022 }, "TrainerController.start_learning": { "total": 96.8051984, "count": 1, "self": 0.10506819999997674, "children": { "TrainerController._reset_env": { "total": 9.1141883, "count": 1, "self": 9.1141883 }, "TrainerController.advance": { "total": 87.46949260000002, "count": 6585, "self": 0.09444109999992634, "children": { "env_step": { "total": 49.186684400000495, "count": 6585, "self": 21.425331300000444, "children": { "SubprocessEnvManager._take_step": { "total": 27.69659450000002, "count": 6585, "self": 0.28653870000042403, "children": { "TorchPolicy.evaluate": { "total": 27.410055799999597, "count": 6585, "self": 27.410055799999597 } } }, "workers": { "total": 0.06475860000003131, "count": 6584, "self": 0.0, "children": { "worker_root": { "total": 88.33605969999992, "count": 6584, "is_parallel": true, "self": 72.29476539999987, "children": { "steps_from_proto": { "total": 0.0005465000000004494, "count": 1, "is_parallel": true, "self": 0.00022170000000087953, "children": { "_process_rank_one_or_two_observation": { "total": 0.0003247999999995699, "count": 6, "is_parallel": true, "self": 0.0003247999999995699 } } }, "UnityEnvironment.step": { "total": 16.04074780000004, "count": 6584, "is_parallel": true, "self": 0.6872718000007936, "children": { "UnityEnvironment._generate_step_input": { "total": 0.7578845999999206, "count": 6584, "is_parallel": true, "self": 0.7578845999999206 }, "communicator.exchange": { "total": 12.44123759999931, "count": 6584, "is_parallel": true, "self": 12.44123759999931 }, "steps_from_proto": { "total": 2.154353800000017, "count": 6584, "is_parallel": true, "self": 0.8423023000004495, "children": { "_process_rank_one_or_two_observation": { "total": 1.3120514999995674, "count": 39504, "is_parallel": true, "self": 1.3120514999995674 } } } } } } } } } } }, "trainer_advance": { "total": 38.1883670999996, "count": 6584, "self": 0.13078589999910406, "children": { "process_trajectory": { "total": 4.465832300000498, "count": 6584, "self": 4.465832300000498 }, "_update_policy": { "total": 33.5917489, "count": 5, "self": 6.835913400000042, "children": { "TorchPPOOptimizer.update": { "total": 26.755835499999957, "count": 3147, "self": 26.755835499999957 } } } } } } }, "trainer_threads": { "total": 1.1999999998124622e-06, "count": 1, "self": 1.1999999998124622e-06 }, "TrainerController._save_models": { "total": 0.1164480999999995, "count": 1, "self": 0.0068753000000043585, "children": { "RLTrainer._checkpoint": { "total": 0.10957279999999514, "count": 1, "self": 0.10957279999999514 } } } } } } }