benson V5 har cooket

This commit is contained in:
2024-04-20 14:45:47 +02:00
parent 2d017b7a59
commit 65607aebc4
322 changed files with 127875 additions and 143 deletions

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

View File

@@ -0,0 +1,76 @@
default_settings: null
behaviors:
BensonV5:
trainer_type: ppo
hyperparameters:
batch_size: 50
buffer_size: 10240
learning_rate: 0.0003
beta: 0.05
epsilon: 0.2
lambd: 0.95
num_epoch: 3
shared_critic: false
learning_rate_schedule: linear
beta_schedule: linear
epsilon_schedule: linear
network_settings:
normalize: false
hidden_units: 128
num_layers: 2
vis_encode_type: simple
memory: null
goal_conditioning_type: hyper
deterministic: false
reward_signals:
extrinsic:
gamma: 0.99
strength: 1.0
network_settings:
normalize: false
hidden_units: 128
num_layers: 2
vis_encode_type: simple
memory: null
goal_conditioning_type: hyper
deterministic: false
init_path: null
keep_checkpoints: 5
checkpoint_interval: 500000
max_steps: 30000000
time_horizon: 64
summary_freq: 10000
threaded: false
self_play: null
behavioral_cloning: null
env_settings:
env_path: null
env_args: null
base_port: 5005
num_envs: 1
num_areas: 1
seed: -1
max_lifetime_restarts: 10
restarts_rate_limit_n: 1
restarts_rate_limit_period_s: 60
engine_settings:
width: 84
height: 84
quality_level: 5
time_scale: 20
target_frame_rate: -1
capture_frame_rate: 60
no_graphics: false
environment_parameters: null
checkpoint_settings:
run_id: BensonV5
initialize_from: null
load_model: false
resume: true
force: false
train_model: false
inference: false
results_dir: results
torch_settings:
device: null
debug: false

View File

@@ -0,0 +1,320 @@
{
"name": "root",
"gauges": {
"BensonV5.Policy.Entropy.mean": {
"value": 1.748111367225647,
"min": 1.6609280109405518,
"max": 1.7695776224136353,
"count": 21
},
"BensonV5.Policy.Entropy.sum": {
"value": 17306.302734375,
"min": 1062.993896484375,
"max": 17801.951171875,
"count": 21
},
"BensonV5.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 21
},
"BensonV5.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 21
},
"BensonV5.Step.mean": {
"value": 699957.0,
"min": 509973.0,
"max": 699957.0,
"count": 20
},
"BensonV5.Step.sum": {
"value": 699957.0,
"min": 509973.0,
"max": 699957.0,
"count": 20
},
"BensonV5.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.51871919631958,
"min": 1.69606351852417,
"max": 2.540252685546875,
"count": 20
},
"BensonV5.Policy.ExtrinsicValueEstimate.sum": {
"value": 438.25714111328125,
"min": 298.5071716308594,
"max": 452.16497802734375,
"count": 20
},
"BensonV5.Environment.EpisodeLength.mean": {
"value": 265.8611111111111,
"min": 215.65217391304347,
"max": 275.2647058823529,
"count": 20
},
"BensonV5.Environment.EpisodeLength.sum": {
"value": 9571.0,
"min": 8653.0,
"max": 10639.0,
"count": 20
},
"BensonV5.Environment.CumulativeReward.mean": {
"value": 11.074494913614924,
"min": 8.316233371874612,
"max": 11.346780761749157,
"count": 20
},
"BensonV5.Environment.CumulativeReward.sum": {
"value": 398.68181689013727,
"min": 316.01686813123524,
"max": 448.9062953479588,
"count": 20
},
"BensonV5.Policy.ExtrinsicReward.mean": {
"value": 11.074494913614924,
"min": 8.316233371874612,
"max": 11.346780761749157,
"count": 20
},
"BensonV5.Policy.ExtrinsicReward.sum": {
"value": 398.68181689013727,
"min": 316.01686813123524,
"max": 448.9062953479588,
"count": 20
},
"BensonV5.Losses.PolicyLoss.mean": {
"value": 0.11478592688132577,
"min": 0.10645271842309384,
"max": 0.12377680898482028,
"count": 19
},
"BensonV5.Losses.PolicyLoss.sum": {
"value": 0.11478592688132577,
"min": 0.10645271842309384,
"max": 0.12377680898482028,
"count": 19
},
"BensonV5.Losses.ValueLoss.mean": {
"value": 0.09043678960846208,
"min": 0.03595422747870529,
"max": 0.18858122416749235,
"count": 19
},
"BensonV5.Losses.ValueLoss.sum": {
"value": 0.09043678960846208,
"min": 0.03595422747870529,
"max": 0.18858122416749235,
"count": 19
},
"BensonV5.Policy.LearningRate.mean": {
"value": 0.00029305035231655006,
"min": 0.00029305035231655006,
"max": 0.00029489777170074337,
"count": 19
},
"BensonV5.Policy.LearningRate.sum": {
"value": 0.00029305035231655006,
"min": 0.00029305035231655006,
"max": 0.00029489777170074337,
"count": 19
},
"BensonV5.Policy.Epsilon.mean": {
"value": 0.19768345,
"min": 0.19768345,
"max": 0.19829925666666665,
"count": 19
},
"BensonV5.Policy.Epsilon.sum": {
"value": 0.19768345,
"min": 0.19768345,
"max": 0.19829925666666665,
"count": 19
},
"BensonV5.Policy.Beta.mean": {
"value": 0.04884195665500001,
"min": 0.04884195665500001,
"max": 0.04914979840766667,
"count": 19
},
"BensonV5.Policy.Beta.sum": {
"value": 0.04884195665500001,
"min": 0.04884195665500001,
"max": 0.04914979840766667,
"count": 19
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1713616468",
"python_version": "3.9.13 (tags/v3.9.13:6de2ca5, May 17 2022, 16:36:42) [MSC v.1929 64 bit (AMD64)]",
"command_line_arguments": "C:\\Users\\noahk\\Documents\\Unity projects\\Racesm\\.venv\\Scripts\\mlagents-learn config/BensonV5.yaml --run-id=BensonV5 --resume",
"mlagents_version": "0.30.0",
"mlagents_envs_version": "0.30.0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "2.2.2+cu118",
"numpy_version": "1.21.2",
"end_time_seconds": "1713617033"
},
"total": 565.8408033999999,
"count": 1,
"self": 0.0047095999999555715,
"children": {
"run_training.setup": {
"total": 0.10145319999999991,
"count": 1,
"self": 0.10145319999999991
},
"TrainerController.start_learning": {
"total": 565.7346405999999,
"count": 1,
"self": 0.4873183999975481,
"children": {
"TrainerController._reset_env": {
"total": 7.1156246,
"count": 1,
"self": 7.1156246
},
"TrainerController.advance": {
"total": 558.0471050000024,
"count": 21588,
"self": 0.41232360000719837,
"children": {
"env_step": {
"total": 385.2072878999983,
"count": 21588,
"self": 281.618686800001,
"children": {
"SubprocessEnvManager._take_step": {
"total": 103.27503330000239,
"count": 21588,
"self": 1.2959513999997796,
"children": {
"TorchPolicy.evaluate": {
"total": 101.97908190000261,
"count": 20945,
"self": 101.97908190000261
}
}
},
"workers": {
"total": 0.3135677999948836,
"count": 21587,
"self": 0.0,
"children": {
"worker_root": {
"total": 558.988520400006,
"count": 21587,
"is_parallel": true,
"self": 302.9514653000143,
"children": {
"steps_from_proto": {
"total": 0.0006380999999997528,
"count": 1,
"is_parallel": true,
"self": 0.00020650000000088653,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0004315999999988662,
"count": 6,
"is_parallel": true,
"self": 0.0004315999999988662
}
}
},
"UnityEnvironment.step": {
"total": 256.0364169999917,
"count": 21587,
"is_parallel": true,
"self": 3.5628550999919923,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 3.2092912000003135,
"count": 21587,
"is_parallel": true,
"self": 3.2092912000003135
},
"communicator.exchange": {
"total": 239.03681639999718,
"count": 21587,
"is_parallel": true,
"self": 239.03681639999718
},
"steps_from_proto": {
"total": 10.227454300002195,
"count": 21587,
"is_parallel": true,
"self": 3.78097849998377,
"children": {
"_process_rank_one_or_two_observation": {
"total": 6.446475800018425,
"count": 129522,
"is_parallel": true,
"self": 6.446475800018425
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 172.4274934999969,
"count": 21587,
"self": 0.8109245999987706,
"children": {
"process_trajectory": {
"total": 25.179432899998258,
"count": 21587,
"self": 25.036942499998258,
"children": {
"RLTrainer._checkpoint": {
"total": 0.1424903999999998,
"count": 1,
"self": 0.1424903999999998
}
}
},
"_update_policy": {
"total": 146.43713599999987,
"count": 20,
"self": 28.81576090000226,
"children": {
"TorchPPOOptimizer.update": {
"total": 117.6213750999976,
"count": 12297,
"self": 117.6213750999976
}
}
}
}
}
}
},
"trainer_threads": {
"total": 7.999999525054591e-07,
"count": 1,
"self": 7.999999525054591e-07
},
"TrainerController._save_models": {
"total": 0.08459179999999833,
"count": 1,
"self": 0.00801210000008723,
"children": {
"RLTrainer._checkpoint": {
"total": 0.0765796999999111,
"count": 1,
"self": 0.0765796999999111
}
}
}
}
}
}
}

View File

@@ -0,0 +1,47 @@
{
"BensonV5": {
"checkpoints": [
{
"steps": 372750,
"file_path": "results\\BensonV5\\BensonV5\\BensonV5-372750.onnx",
"reward": 4.475778912527944,
"creation_time": 1713612293.8306386,
"auxillary_file_paths": [
"results\\BensonV5\\BensonV5\\BensonV5-372750.pt"
]
},
{
"steps": 499938,
"file_path": "results\\BensonV5\\BensonV5\\BensonV5-499938.onnx",
"reward": null,
"creation_time": 1713616476.991812,
"auxillary_file_paths": [
"results\\BensonV5\\BensonV5\\BensonV5-499938.pt"
]
},
{
"steps": 709053,
"file_path": "results\\BensonV5\\BensonV5\\BensonV5-709053.onnx",
"reward": 9.909191635747751,
"creation_time": 1713617033.95885,
"auxillary_file_paths": [
"results\\BensonV5\\BensonV5\\BensonV5-709053.pt"
]
}
],
"final_checkpoint": {
"steps": 709053,
"file_path": "results\\BensonV5\\BensonV5.onnx",
"reward": 9.909191635747751,
"creation_time": 1713617033.95885,
"auxillary_file_paths": [
"results\\BensonV5\\BensonV5\\BensonV5-709053.pt"
]
}
},
"metadata": {
"stats_format_version": "0.3.0",
"mlagents_version": "0.30.0",
"torch_version": "2.2.2+cu118"
}
}