First Push
Browse files- SnowballTarget.onnx +1 -1
- SnowballTarget/SnowballTarget-149984.onnx +1 -1
- SnowballTarget/SnowballTarget-149984.pt +1 -1
- SnowballTarget/SnowballTarget-199984.onnx +1 -1
- SnowballTarget/SnowballTarget-199984.pt +1 -1
- SnowballTarget/SnowballTarget-200112.onnx +1 -1
- SnowballTarget/SnowballTarget-200112.pt +1 -1
- SnowballTarget/SnowballTarget-49936.onnx +1 -1
- SnowballTarget/SnowballTarget-49936.pt +1 -1
- SnowballTarget/SnowballTarget-99960.onnx +1 -1
- SnowballTarget/SnowballTarget-99960.pt +1 -1
- SnowballTarget/checkpoint.pt +1 -1
- SnowballTarget/events.out.tfevents.1694619472.84664ca06e8d.8721.0 +3 -0
- config.json +1 -1
- configuration.yaml +2 -2
- run_logs/Player-0.log +7 -10
- run_logs/timers.json +141 -141
- run_logs/training_status.json +12 -12
SnowballTarget.onnx
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 646532
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:76dd774faf42882541dc8350c37ce847692c609a81b78416794bdd0f03947d52
|
| 3 |
size 646532
|
SnowballTarget/SnowballTarget-149984.onnx
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 646532
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:13bbc7d278f99c86c2183fd94d2b0e9fa453d1c01181999bb013f3b3fe82e752
|
| 3 |
size 646532
|
SnowballTarget/SnowballTarget-149984.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4648906
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d1187cdbc7212b505af4b386a8c30f635806b19d90e5894f2aca97291f669586
|
| 3 |
size 4648906
|
SnowballTarget/SnowballTarget-199984.onnx
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 646532
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:76dd774faf42882541dc8350c37ce847692c609a81b78416794bdd0f03947d52
|
| 3 |
size 646532
|
SnowballTarget/SnowballTarget-199984.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4648906
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e71e2c39a7e96bab1ba1a956d9e427391c06f298f4b4d424094d91936dd399ce
|
| 3 |
size 4648906
|
SnowballTarget/SnowballTarget-200112.onnx
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 646532
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:76dd774faf42882541dc8350c37ce847692c609a81b78416794bdd0f03947d52
|
| 3 |
size 646532
|
SnowballTarget/SnowballTarget-200112.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4648906
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9a81b457a8849a8a53fd446cfc539482a5a39fb9bb56ee13679f5808634d97af
|
| 3 |
size 4648906
|
SnowballTarget/SnowballTarget-49936.onnx
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 646532
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:77b02d0efa9c527262bdd2dbbf2b294fdac66e8c0f8a85c8aeae2ee78a37fc54
|
| 3 |
size 646532
|
SnowballTarget/SnowballTarget-49936.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4648906
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:006070d1b26d5ad7b61159cb7d3ac28e4a67822bdd8e571c7803aff16178253c
|
| 3 |
size 4648906
|
SnowballTarget/SnowballTarget-99960.onnx
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 646532
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9de75d96d27c340e3fb6eddf78ac4493d7dd85221b5f33a2c0c50830151a29f8
|
| 3 |
size 646532
|
SnowballTarget/SnowballTarget-99960.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4648906
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:02143d42cbbbb81293fcbd8598c6cd54d7bcfea07ca4e2e5af2fc8bb5b64a658
|
| 3 |
size 4648906
|
SnowballTarget/checkpoint.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4648906
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9a81b457a8849a8a53fd446cfc539482a5a39fb9bb56ee13679f5808634d97af
|
| 3 |
size 4648906
|
SnowballTarget/events.out.tfevents.1694619472.84664ca06e8d.8721.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:137643d64ecb39fe91e649edf5592e28dd9aff6afa2fbf0ecceb4c9f10968524
|
| 3 |
+
size 29468
|
config.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"default_settings": null, "behaviors": {"SnowballTarget": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate":
|
|
|
|
| 1 |
+
{"default_settings": null, "behaviors": {"SnowballTarget": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate": 0.0003, "beta": 0.005, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "shared_critic": false, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "checkpoint_interval": 50000, "network_settings": {"normalize": false, "hidden_units": 256, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}, "curiosity": {"gamma": 0.99, "strength": 0.05, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "learning_rate": 0.0003, "encoding_size": null}}, "init_path": null, "keep_checkpoints": 10, "even_checkpoints": false, "max_steps": 200000, "time_horizon": 64, "summary_freq": 10000, "threaded": true, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./training-envs-executables/linux/SnowballTarget/SnowballTarget", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true}, "environment_parameters": null, "checkpoint_settings": {"run_id": "SnowballTarget1", "initialize_from": null, "load_model": false, "resume": false, "force": true, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
|
configuration.yaml
CHANGED
|
@@ -5,7 +5,7 @@ behaviors:
|
|
| 5 |
hyperparameters:
|
| 6 |
batch_size: 128
|
| 7 |
buffer_size: 2048
|
| 8 |
-
learning_rate:
|
| 9 |
beta: 0.005
|
| 10 |
epsilon: 0.2
|
| 11 |
lambd: 0.95
|
|
@@ -81,7 +81,7 @@ checkpoint_settings:
|
|
| 81 |
initialize_from: null
|
| 82 |
load_model: false
|
| 83 |
resume: false
|
| 84 |
-
force:
|
| 85 |
train_model: false
|
| 86 |
inference: false
|
| 87 |
results_dir: results
|
|
|
|
| 5 |
hyperparameters:
|
| 6 |
batch_size: 128
|
| 7 |
buffer_size: 2048
|
| 8 |
+
learning_rate: 0.0003
|
| 9 |
beta: 0.005
|
| 10 |
epsilon: 0.2
|
| 11 |
lambd: 0.95
|
|
|
|
| 81 |
initialize_from: null
|
| 82 |
load_model: false
|
| 83 |
resume: false
|
| 84 |
+
force: true
|
| 85 |
train_model: false
|
| 86 |
inference: false
|
| 87 |
results_dir: results
|
run_logs/Player-0.log
CHANGED
|
@@ -2,9 +2,6 @@ Mono path[0] = '/content/ml-agents/training-envs-executables/linux/SnowballTarge
|
|
| 2 |
Mono config path = '/content/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/MonoBleedingEdge/etc'
|
| 3 |
Preloaded 'lib_burst_generated.so'
|
| 4 |
Preloaded 'libgrpc_csharp_ext.x64.so'
|
| 5 |
-
PlayerPrefs - Creating folder: /root/.config/unity3d/Hugging Face
|
| 6 |
-
PlayerPrefs - Creating folder: /root/.config/unity3d/Hugging Face/SnowballTarget
|
| 7 |
-
Unable to load player prefs
|
| 8 |
Initialize engine version: 2021.3.14f1 (eee1884e7226)
|
| 9 |
[Subsystems] Discovering subsystems at path /content/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/UnitySubsystems
|
| 10 |
Forcing GfxDevice: Null
|
|
@@ -34,7 +31,7 @@ ALSA lib pcm.c:2664:(snd_pcm_open_noupdate) Unknown PCM default
|
|
| 34 |
FMOD failed to initialize the output device.: "Error initializing output device. " (60)
|
| 35 |
FMOD initialized on nosound output
|
| 36 |
Begin MonoManager ReloadAssembly
|
| 37 |
-
- Completed reload, in 0.
|
| 38 |
ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
|
| 39 |
ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
|
| 40 |
ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
|
|
@@ -45,7 +42,7 @@ ERROR: Shader Standard shader is not supported on this GPU (none of subshaders/f
|
|
| 45 |
WARNING: Shader Unsupported: 'Standard' - All subshaders removed
|
| 46 |
WARNING: Shader Did you use #pragma only_renderers and omit this platform?
|
| 47 |
WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
|
| 48 |
-
UnloadTime: 0.
|
| 49 |
ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
|
| 50 |
requesting resize 84 x 84
|
| 51 |
Setting up 1 worker threads for Enlighten.
|
|
@@ -169,7 +166,7 @@ Memory Statistics:
|
|
| 169 |
Peak Allocated Bytes 0 B
|
| 170 |
Overflow Count 0
|
| 171 |
[ALLOC_DEFAULT] Dual Thread Allocator
|
| 172 |
-
Peak main deferred allocation count
|
| 173 |
[ALLOC_BUCKET]
|
| 174 |
Large Block size 4.0 MB
|
| 175 |
Used Block count 1
|
|
@@ -178,13 +175,13 @@ Memory Statistics:
|
|
| 178 |
Peak usage frame count: [4.0 MB-8.0 MB]: 5463 frames
|
| 179 |
Requested Block Size 16.0 MB
|
| 180 |
Peak Block count 1
|
| 181 |
-
Peak Allocated memory 6.
|
| 182 |
Peak Large allocation bytes 0 B
|
| 183 |
[ALLOC_DEFAULT_THREAD]
|
| 184 |
Peak usage frame count: [16.0 MB-32.0 MB]: 5463 frames
|
| 185 |
Requested Block Size 16.0 MB
|
| 186 |
Peak Block count 1
|
| 187 |
-
Peak Allocated memory 17.
|
| 188 |
Peak Large allocation bytes 16.0 MB
|
| 189 |
[ALLOC_TEMP_JOB_1_FRAME]
|
| 190 |
Initial Block Size 2.0 MB
|
|
@@ -213,10 +210,10 @@ Memory Statistics:
|
|
| 213 |
Used Block count 1
|
| 214 |
Peak Allocated bytes 1.0 MB
|
| 215 |
[ALLOC_GFX_MAIN]
|
| 216 |
-
Peak usage frame count: [32.0 KB-64.0 KB]:
|
| 217 |
Requested Block Size 16.0 MB
|
| 218 |
Peak Block count 1
|
| 219 |
-
Peak Allocated memory
|
| 220 |
Peak Large allocation bytes 0 B
|
| 221 |
[ALLOC_GFX_THREAD]
|
| 222 |
Peak usage frame count: [32.0 KB-64.0 KB]: 5463 frames
|
|
|
|
| 2 |
Mono config path = '/content/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/MonoBleedingEdge/etc'
|
| 3 |
Preloaded 'lib_burst_generated.so'
|
| 4 |
Preloaded 'libgrpc_csharp_ext.x64.so'
|
|
|
|
|
|
|
|
|
|
| 5 |
Initialize engine version: 2021.3.14f1 (eee1884e7226)
|
| 6 |
[Subsystems] Discovering subsystems at path /content/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/UnitySubsystems
|
| 7 |
Forcing GfxDevice: Null
|
|
|
|
| 31 |
FMOD failed to initialize the output device.: "Error initializing output device. " (60)
|
| 32 |
FMOD initialized on nosound output
|
| 33 |
Begin MonoManager ReloadAssembly
|
| 34 |
+
- Completed reload, in 0.124 seconds
|
| 35 |
ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
|
| 36 |
ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
|
| 37 |
ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
|
|
|
|
| 42 |
WARNING: Shader Unsupported: 'Standard' - All subshaders removed
|
| 43 |
WARNING: Shader Did you use #pragma only_renderers and omit this platform?
|
| 44 |
WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
|
| 45 |
+
UnloadTime: 0.893295 ms
|
| 46 |
ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
|
| 47 |
requesting resize 84 x 84
|
| 48 |
Setting up 1 worker threads for Enlighten.
|
|
|
|
| 166 |
Peak Allocated Bytes 0 B
|
| 167 |
Overflow Count 0
|
| 168 |
[ALLOC_DEFAULT] Dual Thread Allocator
|
| 169 |
+
Peak main deferred allocation count 36
|
| 170 |
[ALLOC_BUCKET]
|
| 171 |
Large Block size 4.0 MB
|
| 172 |
Used Block count 1
|
|
|
|
| 175 |
Peak usage frame count: [4.0 MB-8.0 MB]: 5463 frames
|
| 176 |
Requested Block Size 16.0 MB
|
| 177 |
Peak Block count 1
|
| 178 |
+
Peak Allocated memory 6.6 MB
|
| 179 |
Peak Large allocation bytes 0 B
|
| 180 |
[ALLOC_DEFAULT_THREAD]
|
| 181 |
Peak usage frame count: [16.0 MB-32.0 MB]: 5463 frames
|
| 182 |
Requested Block Size 16.0 MB
|
| 183 |
Peak Block count 1
|
| 184 |
+
Peak Allocated memory 17.8 MB
|
| 185 |
Peak Large allocation bytes 16.0 MB
|
| 186 |
[ALLOC_TEMP_JOB_1_FRAME]
|
| 187 |
Initial Block Size 2.0 MB
|
|
|
|
| 210 |
Used Block count 1
|
| 211 |
Peak Allocated bytes 1.0 MB
|
| 212 |
[ALLOC_GFX_MAIN]
|
| 213 |
+
Peak usage frame count: [32.0 KB-64.0 KB]: 5121 frames, [64.0 KB-128.0 KB]: 342 frames
|
| 214 |
Requested Block Size 16.0 MB
|
| 215 |
Peak Block count 1
|
| 216 |
+
Peak Allocated memory 66.4 KB
|
| 217 |
Peak Large allocation bytes 0 B
|
| 218 |
[ALLOC_GFX_THREAD]
|
| 219 |
Peak usage frame count: [32.0 KB-64.0 KB]: 5463 frames
|
run_logs/timers.json
CHANGED
|
@@ -2,15 +2,15 @@
|
|
| 2 |
"name": "root",
|
| 3 |
"gauges": {
|
| 4 |
"SnowballTarget.Policy.Entropy.mean": {
|
| 5 |
-
"value":
|
| 6 |
-
"min":
|
| 7 |
-
"max": 2.
|
| 8 |
"count": 20
|
| 9 |
},
|
| 10 |
"SnowballTarget.Policy.Entropy.sum": {
|
| 11 |
-
"value":
|
| 12 |
-
"min":
|
| 13 |
-
"max":
|
| 14 |
"count": 20
|
| 15 |
},
|
| 16 |
"SnowballTarget.Step.mean": {
|
|
@@ -26,27 +26,27 @@
|
|
| 26 |
"count": 20
|
| 27 |
},
|
| 28 |
"SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
|
| 29 |
-
"value":
|
| 30 |
-
"min": 0.
|
| 31 |
-
"max":
|
| 32 |
"count": 20
|
| 33 |
},
|
| 34 |
"SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
|
| 35 |
-
"value":
|
| 36 |
-
"min":
|
| 37 |
-
"max":
|
| 38 |
"count": 20
|
| 39 |
},
|
| 40 |
"SnowballTarget.Policy.CuriosityValueEstimate.mean": {
|
| 41 |
-
"value": 0.
|
| 42 |
-
"min": 0.
|
| 43 |
-
"max": 0.
|
| 44 |
"count": 20
|
| 45 |
},
|
| 46 |
"SnowballTarget.Policy.CuriosityValueEstimate.sum": {
|
| 47 |
-
"value":
|
| 48 |
-
"min":
|
| 49 |
-
"max":
|
| 50 |
"count": 20
|
| 51 |
},
|
| 52 |
"SnowballTarget.Environment.EpisodeLength.mean": {
|
|
@@ -62,39 +62,39 @@
|
|
| 62 |
"count": 20
|
| 63 |
},
|
| 64 |
"SnowballTarget.Losses.PolicyLoss.mean": {
|
| 65 |
-
"value": 0.
|
| 66 |
-
"min": 0.
|
| 67 |
-
"max": 0.
|
| 68 |
"count": 20
|
| 69 |
},
|
| 70 |
"SnowballTarget.Losses.PolicyLoss.sum": {
|
| 71 |
-
"value": 0.
|
| 72 |
-
"min": 0.
|
| 73 |
-
"max": 0.
|
| 74 |
"count": 20
|
| 75 |
},
|
| 76 |
"SnowballTarget.Losses.ValueLoss.mean": {
|
| 77 |
-
"value": 0.
|
| 78 |
-
"min": 0.
|
| 79 |
-
"max": 0.
|
| 80 |
"count": 20
|
| 81 |
},
|
| 82 |
"SnowballTarget.Losses.ValueLoss.sum": {
|
| 83 |
-
"value": 0.
|
| 84 |
-
"min": 0.
|
| 85 |
-
"max": 0.
|
| 86 |
"count": 20
|
| 87 |
},
|
| 88 |
"SnowballTarget.Policy.LearningRate.mean": {
|
| 89 |
-
"value":
|
| 90 |
-
"min":
|
| 91 |
-
"max":
|
| 92 |
"count": 20
|
| 93 |
},
|
| 94 |
"SnowballTarget.Policy.LearningRate.sum": {
|
| 95 |
-
"value":
|
| 96 |
-
"min":
|
| 97 |
-
"max": 0.
|
| 98 |
"count": 20
|
| 99 |
},
|
| 100 |
"SnowballTarget.Policy.Epsilon.mean": {
|
|
@@ -122,63 +122,63 @@
|
|
| 122 |
"count": 20
|
| 123 |
},
|
| 124 |
"SnowballTarget.Losses.CuriosityForwardLoss.mean": {
|
| 125 |
-
"value": 0.
|
| 126 |
-
"min": 0.
|
| 127 |
-
"max": 0.
|
| 128 |
"count": 20
|
| 129 |
},
|
| 130 |
"SnowballTarget.Losses.CuriosityForwardLoss.sum": {
|
| 131 |
-
"value": 0.
|
| 132 |
-
"min": 0.
|
| 133 |
-
"max": 0.
|
| 134 |
"count": 20
|
| 135 |
},
|
| 136 |
"SnowballTarget.Losses.CuriosityInverseLoss.mean": {
|
| 137 |
-
"value":
|
| 138 |
-
"min":
|
| 139 |
-
"max": 2.
|
| 140 |
"count": 20
|
| 141 |
},
|
| 142 |
"SnowballTarget.Losses.CuriosityInverseLoss.sum": {
|
| 143 |
-
"value":
|
| 144 |
-
"min":
|
| 145 |
-
"max": 13.
|
| 146 |
"count": 20
|
| 147 |
},
|
| 148 |
"SnowballTarget.Environment.CumulativeReward.mean": {
|
| 149 |
-
"value":
|
| 150 |
-
"min": 3.
|
| 151 |
-
"max":
|
| 152 |
"count": 20
|
| 153 |
},
|
| 154 |
"SnowballTarget.Environment.CumulativeReward.sum": {
|
| 155 |
-
"value":
|
| 156 |
-
"min":
|
| 157 |
-
"max":
|
| 158 |
"count": 20
|
| 159 |
},
|
| 160 |
"SnowballTarget.Policy.ExtrinsicReward.mean": {
|
| 161 |
-
"value":
|
| 162 |
-
"min": 3.
|
| 163 |
-
"max":
|
| 164 |
"count": 20
|
| 165 |
},
|
| 166 |
"SnowballTarget.Policy.ExtrinsicReward.sum": {
|
| 167 |
-
"value":
|
| 168 |
-
"min":
|
| 169 |
-
"max":
|
| 170 |
"count": 20
|
| 171 |
},
|
| 172 |
"SnowballTarget.Policy.CuriosityReward.mean": {
|
| 173 |
-
"value": 0.
|
| 174 |
-
"min": 0.
|
| 175 |
-
"max": 0.
|
| 176 |
"count": 20
|
| 177 |
},
|
| 178 |
"SnowballTarget.Policy.CuriosityReward.sum": {
|
| 179 |
-
"value":
|
| 180 |
-
"min":
|
| 181 |
-
"max": 36.
|
| 182 |
"count": 20
|
| 183 |
},
|
| 184 |
"SnowballTarget.IsTraining.mean": {
|
|
@@ -196,67 +196,67 @@
|
|
| 196 |
},
|
| 197 |
"metadata": {
|
| 198 |
"timer_format_version": "0.1.0",
|
| 199 |
-
"start_time_seconds": "
|
| 200 |
"python_version": "3.10.12 (main, Jun 11 2023, 05:26:28) [GCC 11.4.0]",
|
| 201 |
-
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget1 --no-graphics",
|
| 202 |
"mlagents_version": "0.31.0.dev0",
|
| 203 |
"mlagents_envs_version": "0.31.0.dev0",
|
| 204 |
"communication_protocol_version": "1.5.0",
|
| 205 |
"pytorch_version": "1.11.0+cu102",
|
| 206 |
"numpy_version": "1.21.2",
|
| 207 |
-
"end_time_seconds": "
|
| 208 |
},
|
| 209 |
-
"total":
|
| 210 |
"count": 1,
|
| 211 |
-
"self": 0.
|
| 212 |
"children": {
|
| 213 |
"run_training.setup": {
|
| 214 |
-
"total": 0.
|
| 215 |
"count": 1,
|
| 216 |
-
"self": 0.
|
| 217 |
},
|
| 218 |
"TrainerController.start_learning": {
|
| 219 |
-
"total":
|
| 220 |
"count": 1,
|
| 221 |
-
"self": 0.
|
| 222 |
"children": {
|
| 223 |
"TrainerController._reset_env": {
|
| 224 |
-
"total": 4.
|
| 225 |
"count": 1,
|
| 226 |
-
"self": 4.
|
| 227 |
},
|
| 228 |
"TrainerController.advance": {
|
| 229 |
-
"total":
|
| 230 |
-
"count":
|
| 231 |
-
"self": 0.
|
| 232 |
"children": {
|
| 233 |
"env_step": {
|
| 234 |
-
"total":
|
| 235 |
-
"count":
|
| 236 |
-
"self":
|
| 237 |
"children": {
|
| 238 |
"SubprocessEnvManager._take_step": {
|
| 239 |
-
"total":
|
| 240 |
-
"count":
|
| 241 |
-
"self": 1.
|
| 242 |
"children": {
|
| 243 |
"TorchPolicy.evaluate": {
|
| 244 |
-
"total":
|
| 245 |
-
"count":
|
| 246 |
-
"self":
|
| 247 |
}
|
| 248 |
}
|
| 249 |
},
|
| 250 |
"workers": {
|
| 251 |
-
"total": 0.
|
| 252 |
-
"count":
|
| 253 |
"self": 0.0,
|
| 254 |
"children": {
|
| 255 |
"worker_root": {
|
| 256 |
-
"total":
|
| 257 |
-
"count":
|
| 258 |
"is_parallel": true,
|
| 259 |
-
"self":
|
| 260 |
"children": {
|
| 261 |
"run_training.setup": {
|
| 262 |
"total": 0.0,
|
|
@@ -265,48 +265,48 @@
|
|
| 265 |
"self": 0.0,
|
| 266 |
"children": {
|
| 267 |
"steps_from_proto": {
|
| 268 |
-
"total": 0.
|
| 269 |
"count": 1,
|
| 270 |
"is_parallel": true,
|
| 271 |
-
"self": 0.
|
| 272 |
"children": {
|
| 273 |
"_process_rank_one_or_two_observation": {
|
| 274 |
-
"total": 0.
|
| 275 |
"count": 10,
|
| 276 |
"is_parallel": true,
|
| 277 |
-
"self": 0.
|
| 278 |
}
|
| 279 |
}
|
| 280 |
},
|
| 281 |
"UnityEnvironment.step": {
|
| 282 |
-
"total": 0.
|
| 283 |
"count": 1,
|
| 284 |
"is_parallel": true,
|
| 285 |
-
"self": 0.
|
| 286 |
"children": {
|
| 287 |
"UnityEnvironment._generate_step_input": {
|
| 288 |
-
"total": 0.
|
| 289 |
"count": 1,
|
| 290 |
"is_parallel": true,
|
| 291 |
-
"self": 0.
|
| 292 |
},
|
| 293 |
"communicator.exchange": {
|
| 294 |
-
"total": 0.
|
| 295 |
"count": 1,
|
| 296 |
"is_parallel": true,
|
| 297 |
-
"self": 0.
|
| 298 |
},
|
| 299 |
"steps_from_proto": {
|
| 300 |
-
"total": 0.
|
| 301 |
"count": 1,
|
| 302 |
"is_parallel": true,
|
| 303 |
-
"self": 0.
|
| 304 |
"children": {
|
| 305 |
"_process_rank_one_or_two_observation": {
|
| 306 |
-
"total": 0.
|
| 307 |
"count": 10,
|
| 308 |
"is_parallel": true,
|
| 309 |
-
"self": 0.
|
| 310 |
}
|
| 311 |
}
|
| 312 |
}
|
|
@@ -315,34 +315,34 @@
|
|
| 315 |
}
|
| 316 |
},
|
| 317 |
"UnityEnvironment.step": {
|
| 318 |
-
"total":
|
| 319 |
-
"count":
|
| 320 |
"is_parallel": true,
|
| 321 |
-
"self": 10.
|
| 322 |
"children": {
|
| 323 |
"UnityEnvironment._generate_step_input": {
|
| 324 |
-
"total": 5.
|
| 325 |
-
"count":
|
| 326 |
"is_parallel": true,
|
| 327 |
-
"self": 5.
|
| 328 |
},
|
| 329 |
"communicator.exchange": {
|
| 330 |
-
"total":
|
| 331 |
-
"count":
|
| 332 |
"is_parallel": true,
|
| 333 |
-
"self":
|
| 334 |
},
|
| 335 |
"steps_from_proto": {
|
| 336 |
-
"total":
|
| 337 |
-
"count":
|
| 338 |
"is_parallel": true,
|
| 339 |
-
"self":
|
| 340 |
"children": {
|
| 341 |
"_process_rank_one_or_two_observation": {
|
| 342 |
-
"total":
|
| 343 |
-
"count":
|
| 344 |
"is_parallel": true,
|
| 345 |
-
"self":
|
| 346 |
}
|
| 347 |
}
|
| 348 |
}
|
|
@@ -357,9 +357,9 @@
|
|
| 357 |
}
|
| 358 |
},
|
| 359 |
"trainer_threads": {
|
| 360 |
-
"total": 0.
|
| 361 |
"count": 1,
|
| 362 |
-
"self": 0.
|
| 363 |
"children": {
|
| 364 |
"thread_root": {
|
| 365 |
"total": 0.0,
|
|
@@ -368,36 +368,36 @@
|
|
| 368 |
"self": 0.0,
|
| 369 |
"children": {
|
| 370 |
"trainer_advance": {
|
| 371 |
-
"total":
|
| 372 |
-
"count":
|
| 373 |
"is_parallel": true,
|
| 374 |
-
"self": 8.
|
| 375 |
"children": {
|
| 376 |
"process_trajectory": {
|
| 377 |
-
"total":
|
| 378 |
-
"count":
|
| 379 |
"is_parallel": true,
|
| 380 |
-
"self":
|
| 381 |
"children": {
|
| 382 |
"RLTrainer._checkpoint": {
|
| 383 |
-
"total":
|
| 384 |
"count": 4,
|
| 385 |
"is_parallel": true,
|
| 386 |
-
"self":
|
| 387 |
}
|
| 388 |
}
|
| 389 |
},
|
| 390 |
"_update_policy": {
|
| 391 |
-
"total":
|
| 392 |
"count": 90,
|
| 393 |
"is_parallel": true,
|
| 394 |
-
"self":
|
| 395 |
"children": {
|
| 396 |
"TorchPPOOptimizer.update": {
|
| 397 |
-
"total": 93.
|
| 398 |
"count": 4584,
|
| 399 |
"is_parallel": true,
|
| 400 |
-
"self": 93.
|
| 401 |
}
|
| 402 |
}
|
| 403 |
}
|
|
@@ -408,14 +408,14 @@
|
|
| 408 |
}
|
| 409 |
},
|
| 410 |
"TrainerController._save_models": {
|
| 411 |
-
"total": 0.
|
| 412 |
"count": 1,
|
| 413 |
-
"self": 0.
|
| 414 |
"children": {
|
| 415 |
"RLTrainer._checkpoint": {
|
| 416 |
-
"total": 0.
|
| 417 |
"count": 1,
|
| 418 |
-
"self": 0.
|
| 419 |
}
|
| 420 |
}
|
| 421 |
}
|
|
|
|
| 2 |
"name": "root",
|
| 3 |
"gauges": {
|
| 4 |
"SnowballTarget.Policy.Entropy.mean": {
|
| 5 |
+
"value": 0.9894982576370239,
|
| 6 |
+
"min": 0.9754210710525513,
|
| 7 |
+
"max": 2.8669533729553223,
|
| 8 |
"count": 20
|
| 9 |
},
|
| 10 |
"SnowballTarget.Policy.Entropy.sum": {
|
| 11 |
+
"value": 9480.3828125,
|
| 12 |
+
"min": 9480.3828125,
|
| 13 |
+
"max": 29549.6875,
|
| 14 |
"count": 20
|
| 15 |
},
|
| 16 |
"SnowballTarget.Step.mean": {
|
|
|
|
| 26 |
"count": 20
|
| 27 |
},
|
| 28 |
"SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
|
| 29 |
+
"value": 12.786215782165527,
|
| 30 |
+
"min": 0.4517129361629486,
|
| 31 |
+
"max": 12.806447982788086,
|
| 32 |
"count": 20
|
| 33 |
},
|
| 34 |
"SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
|
| 35 |
+
"value": 2493.31201171875,
|
| 36 |
+
"min": 87.63230895996094,
|
| 37 |
+
"max": 2617.53857421875,
|
| 38 |
"count": 20
|
| 39 |
},
|
| 40 |
"SnowballTarget.Policy.CuriosityValueEstimate.mean": {
|
| 41 |
+
"value": 0.17987486720085144,
|
| 42 |
+
"min": 0.10761666297912598,
|
| 43 |
+
"max": 0.26051655411720276,
|
| 44 |
"count": 20
|
| 45 |
},
|
| 46 |
"SnowballTarget.Policy.CuriosityValueEstimate.sum": {
|
| 47 |
+
"value": 35.075599670410156,
|
| 48 |
+
"min": 20.87763214111328,
|
| 49 |
+
"max": 53.40589141845703,
|
| 50 |
"count": 20
|
| 51 |
},
|
| 52 |
"SnowballTarget.Environment.EpisodeLength.mean": {
|
|
|
|
| 62 |
"count": 20
|
| 63 |
},
|
| 64 |
"SnowballTarget.Losses.PolicyLoss.mean": {
|
| 65 |
+
"value": 0.06864661709208604,
|
| 66 |
+
"min": 0.06325657534191623,
|
| 67 |
+
"max": 0.07433176024967586,
|
| 68 |
"count": 20
|
| 69 |
},
|
| 70 |
"SnowballTarget.Losses.PolicyLoss.sum": {
|
| 71 |
+
"value": 0.27458646836834416,
|
| 72 |
+
"min": 0.2530263013676649,
|
| 73 |
+
"max": 0.3716588012483793,
|
| 74 |
"count": 20
|
| 75 |
},
|
| 76 |
"SnowballTarget.Losses.ValueLoss.mean": {
|
| 77 |
+
"value": 0.10565935024151614,
|
| 78 |
+
"min": 0.05770217951537822,
|
| 79 |
+
"max": 0.16127740534972038,
|
| 80 |
"count": 20
|
| 81 |
},
|
| 82 |
"SnowballTarget.Losses.ValueLoss.sum": {
|
| 83 |
+
"value": 0.42263740096606456,
|
| 84 |
+
"min": 0.23080871806151287,
|
| 85 |
+
"max": 0.8063870267486019,
|
| 86 |
"count": 20
|
| 87 |
},
|
| 88 |
"SnowballTarget.Policy.LearningRate.mean": {
|
| 89 |
+
"value": 8.082097306000005e-06,
|
| 90 |
+
"min": 8.082097306000005e-06,
|
| 91 |
+
"max": 0.000291882002706,
|
| 92 |
"count": 20
|
| 93 |
},
|
| 94 |
"SnowballTarget.Policy.LearningRate.sum": {
|
| 95 |
+
"value": 3.232838922400002e-05,
|
| 96 |
+
"min": 3.232838922400002e-05,
|
| 97 |
+
"max": 0.00138516003828,
|
| 98 |
"count": 20
|
| 99 |
},
|
| 100 |
"SnowballTarget.Policy.Epsilon.mean": {
|
|
|
|
| 122 |
"count": 20
|
| 123 |
},
|
| 124 |
"SnowballTarget.Losses.CuriosityForwardLoss.mean": {
|
| 125 |
+
"value": 0.03605778130027009,
|
| 126 |
+
"min": 0.035578419041692046,
|
| 127 |
+
"max": 0.1175546848680824,
|
| 128 |
"count": 20
|
| 129 |
},
|
| 130 |
"SnowballTarget.Losses.CuriosityForwardLoss.sum": {
|
| 131 |
+
"value": 0.14423112520108036,
|
| 132 |
+
"min": 0.14423112520108036,
|
| 133 |
+
"max": 0.4702187394723296,
|
| 134 |
"count": 20
|
| 135 |
},
|
| 136 |
"SnowballTarget.Losses.CuriosityInverseLoss.mean": {
|
| 137 |
+
"value": 0.8827689254985136,
|
| 138 |
+
"min": 0.8652849678899728,
|
| 139 |
+
"max": 2.7858384264012175,
|
| 140 |
"count": 20
|
| 141 |
},
|
| 142 |
"SnowballTarget.Losses.CuriosityInverseLoss.sum": {
|
| 143 |
+
"value": 3.5310757019940544,
|
| 144 |
+
"min": 3.4890115354575366,
|
| 145 |
+
"max": 13.03318906765358,
|
| 146 |
"count": 20
|
| 147 |
},
|
| 148 |
"SnowballTarget.Environment.CumulativeReward.mean": {
|
| 149 |
+
"value": 25.022727272727273,
|
| 150 |
+
"min": 3.2954545454545454,
|
| 151 |
+
"max": 25.022727272727273,
|
| 152 |
"count": 20
|
| 153 |
},
|
| 154 |
"SnowballTarget.Environment.CumulativeReward.sum": {
|
| 155 |
+
"value": 1101.0,
|
| 156 |
+
"min": 145.0,
|
| 157 |
+
"max": 1375.0,
|
| 158 |
"count": 20
|
| 159 |
},
|
| 160 |
"SnowballTarget.Policy.ExtrinsicReward.mean": {
|
| 161 |
+
"value": 25.022727272727273,
|
| 162 |
+
"min": 3.2954545454545454,
|
| 163 |
+
"max": 25.022727272727273,
|
| 164 |
"count": 20
|
| 165 |
},
|
| 166 |
"SnowballTarget.Policy.ExtrinsicReward.sum": {
|
| 167 |
+
"value": 1101.0,
|
| 168 |
+
"min": 145.0,
|
| 169 |
+
"max": 1375.0,
|
| 170 |
"count": 20
|
| 171 |
},
|
| 172 |
"SnowballTarget.Policy.CuriosityReward.mean": {
|
| 173 |
+
"value": 0.37121628858784045,
|
| 174 |
+
"min": 0.36311526602049443,
|
| 175 |
+
"max": 0.6572515787048774,
|
| 176 |
"count": 20
|
| 177 |
},
|
| 178 |
"SnowballTarget.Policy.CuriosityReward.sum": {
|
| 179 |
+
"value": 16.33351669786498,
|
| 180 |
+
"min": 16.193131546955556,
|
| 181 |
+
"max": 36.14883682876825,
|
| 182 |
"count": 20
|
| 183 |
},
|
| 184 |
"SnowballTarget.IsTraining.mean": {
|
|
|
|
| 196 |
},
|
| 197 |
"metadata": {
|
| 198 |
"timer_format_version": "0.1.0",
|
| 199 |
+
"start_time_seconds": "1694619470",
|
| 200 |
"python_version": "3.10.12 (main, Jun 11 2023, 05:26:28) [GCC 11.4.0]",
|
| 201 |
+
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget1 --no-graphics --force",
|
| 202 |
"mlagents_version": "0.31.0.dev0",
|
| 203 |
"mlagents_envs_version": "0.31.0.dev0",
|
| 204 |
"communication_protocol_version": "1.5.0",
|
| 205 |
"pytorch_version": "1.11.0+cu102",
|
| 206 |
"numpy_version": "1.21.2",
|
| 207 |
+
"end_time_seconds": "1694620008"
|
| 208 |
},
|
| 209 |
+
"total": 537.871770708,
|
| 210 |
"count": 1,
|
| 211 |
+
"self": 0.42486783499998637,
|
| 212 |
"children": {
|
| 213 |
"run_training.setup": {
|
| 214 |
+
"total": 0.07099536200007606,
|
| 215 |
"count": 1,
|
| 216 |
+
"self": 0.07099536200007606
|
| 217 |
},
|
| 218 |
"TrainerController.start_learning": {
|
| 219 |
+
"total": 537.3759075109999,
|
| 220 |
"count": 1,
|
| 221 |
+
"self": 0.5942737139978362,
|
| 222 |
"children": {
|
| 223 |
"TrainerController._reset_env": {
|
| 224 |
+
"total": 4.729345418999856,
|
| 225 |
"count": 1,
|
| 226 |
+
"self": 4.729345418999856
|
| 227 |
},
|
| 228 |
"TrainerController.advance": {
|
| 229 |
+
"total": 531.9056004700019,
|
| 230 |
+
"count": 18207,
|
| 231 |
+
"self": 0.28875258900757217,
|
| 232 |
"children": {
|
| 233 |
"env_step": {
|
| 234 |
+
"total": 531.6168478809943,
|
| 235 |
+
"count": 18207,
|
| 236 |
+
"self": 410.4874470380071,
|
| 237 |
"children": {
|
| 238 |
"SubprocessEnvManager._take_step": {
|
| 239 |
+
"total": 120.82789380598388,
|
| 240 |
+
"count": 18207,
|
| 241 |
+
"self": 1.8920219630301744,
|
| 242 |
"children": {
|
| 243 |
"TorchPolicy.evaluate": {
|
| 244 |
+
"total": 118.9358718429537,
|
| 245 |
+
"count": 18207,
|
| 246 |
+
"self": 118.9358718429537
|
| 247 |
}
|
| 248 |
}
|
| 249 |
},
|
| 250 |
"workers": {
|
| 251 |
+
"total": 0.30150703700337544,
|
| 252 |
+
"count": 18207,
|
| 253 |
"self": 0.0,
|
| 254 |
"children": {
|
| 255 |
"worker_root": {
|
| 256 |
+
"total": 535.6665731939938,
|
| 257 |
+
"count": 18207,
|
| 258 |
"is_parallel": true,
|
| 259 |
+
"self": 284.15536370497443,
|
| 260 |
"children": {
|
| 261 |
"run_training.setup": {
|
| 262 |
"total": 0.0,
|
|
|
|
| 265 |
"self": 0.0,
|
| 266 |
"children": {
|
| 267 |
"steps_from_proto": {
|
| 268 |
+
"total": 0.003331277999905069,
|
| 269 |
"count": 1,
|
| 270 |
"is_parallel": true,
|
| 271 |
+
"self": 0.0008317099998293997,
|
| 272 |
"children": {
|
| 273 |
"_process_rank_one_or_two_observation": {
|
| 274 |
+
"total": 0.0024995680000756693,
|
| 275 |
"count": 10,
|
| 276 |
"is_parallel": true,
|
| 277 |
+
"self": 0.0024995680000756693
|
| 278 |
}
|
| 279 |
}
|
| 280 |
},
|
| 281 |
"UnityEnvironment.step": {
|
| 282 |
+
"total": 0.04511771600004977,
|
| 283 |
"count": 1,
|
| 284 |
"is_parallel": true,
|
| 285 |
+
"self": 0.0007030890001260559,
|
| 286 |
"children": {
|
| 287 |
"UnityEnvironment._generate_step_input": {
|
| 288 |
+
"total": 0.00042302299993934867,
|
| 289 |
"count": 1,
|
| 290 |
"is_parallel": true,
|
| 291 |
+
"self": 0.00042302299993934867
|
| 292 |
},
|
| 293 |
"communicator.exchange": {
|
| 294 |
+
"total": 0.04046062499992331,
|
| 295 |
"count": 1,
|
| 296 |
"is_parallel": true,
|
| 297 |
+
"self": 0.04046062499992331
|
| 298 |
},
|
| 299 |
"steps_from_proto": {
|
| 300 |
+
"total": 0.003530979000061052,
|
| 301 |
"count": 1,
|
| 302 |
"is_parallel": true,
|
| 303 |
+
"self": 0.0015479049998248229,
|
| 304 |
"children": {
|
| 305 |
"_process_rank_one_or_two_observation": {
|
| 306 |
+
"total": 0.001983074000236229,
|
| 307 |
"count": 10,
|
| 308 |
"is_parallel": true,
|
| 309 |
+
"self": 0.001983074000236229
|
| 310 |
}
|
| 311 |
}
|
| 312 |
}
|
|
|
|
| 315 |
}
|
| 316 |
},
|
| 317 |
"UnityEnvironment.step": {
|
| 318 |
+
"total": 251.5112094890194,
|
| 319 |
+
"count": 18206,
|
| 320 |
"is_parallel": true,
|
| 321 |
+
"self": 10.637280265021445,
|
| 322 |
"children": {
|
| 323 |
"UnityEnvironment._generate_step_input": {
|
| 324 |
+
"total": 5.376301449994344,
|
| 325 |
+
"count": 18206,
|
| 326 |
"is_parallel": true,
|
| 327 |
+
"self": 5.376301449994344
|
| 328 |
},
|
| 329 |
"communicator.exchange": {
|
| 330 |
+
"total": 198.88422318497965,
|
| 331 |
+
"count": 18206,
|
| 332 |
"is_parallel": true,
|
| 333 |
+
"self": 198.88422318497965
|
| 334 |
},
|
| 335 |
"steps_from_proto": {
|
| 336 |
+
"total": 36.61340458902396,
|
| 337 |
+
"count": 18206,
|
| 338 |
"is_parallel": true,
|
| 339 |
+
"self": 6.762891233860273,
|
| 340 |
"children": {
|
| 341 |
"_process_rank_one_or_two_observation": {
|
| 342 |
+
"total": 29.850513355163685,
|
| 343 |
+
"count": 182060,
|
| 344 |
"is_parallel": true,
|
| 345 |
+
"self": 29.850513355163685
|
| 346 |
}
|
| 347 |
}
|
| 348 |
}
|
|
|
|
| 357 |
}
|
| 358 |
},
|
| 359 |
"trainer_threads": {
|
| 360 |
+
"total": 0.00011929000038435333,
|
| 361 |
"count": 1,
|
| 362 |
+
"self": 0.00011929000038435333,
|
| 363 |
"children": {
|
| 364 |
"thread_root": {
|
| 365 |
"total": 0.0,
|
|
|
|
| 368 |
"self": 0.0,
|
| 369 |
"children": {
|
| 370 |
"trainer_advance": {
|
| 371 |
+
"total": 528.7277012818795,
|
| 372 |
+
"count": 395869,
|
| 373 |
"is_parallel": true,
|
| 374 |
+
"self": 8.865601085034768,
|
| 375 |
"children": {
|
| 376 |
"process_trajectory": {
|
| 377 |
+
"total": 270.5945444238446,
|
| 378 |
+
"count": 395869,
|
| 379 |
"is_parallel": true,
|
| 380 |
+
"self": 269.1041468868448,
|
| 381 |
"children": {
|
| 382 |
"RLTrainer._checkpoint": {
|
| 383 |
+
"total": 1.490397536999808,
|
| 384 |
"count": 4,
|
| 385 |
"is_parallel": true,
|
| 386 |
+
"self": 1.490397536999808
|
| 387 |
}
|
| 388 |
}
|
| 389 |
},
|
| 390 |
"_update_policy": {
|
| 391 |
+
"total": 249.26755577300014,
|
| 392 |
"count": 90,
|
| 393 |
"is_parallel": true,
|
| 394 |
+
"self": 155.99453570301625,
|
| 395 |
"children": {
|
| 396 |
"TorchPPOOptimizer.update": {
|
| 397 |
+
"total": 93.27302006998389,
|
| 398 |
"count": 4584,
|
| 399 |
"is_parallel": true,
|
| 400 |
+
"self": 93.27302006998389
|
| 401 |
}
|
| 402 |
}
|
| 403 |
}
|
|
|
|
| 408 |
}
|
| 409 |
},
|
| 410 |
"TrainerController._save_models": {
|
| 411 |
+
"total": 0.14656861799994658,
|
| 412 |
"count": 1,
|
| 413 |
+
"self": 0.0010552509997978632,
|
| 414 |
"children": {
|
| 415 |
"RLTrainer._checkpoint": {
|
| 416 |
+
"total": 0.14551336700014872,
|
| 417 |
"count": 1,
|
| 418 |
+
"self": 0.14551336700014872
|
| 419 |
}
|
| 420 |
}
|
| 421 |
}
|
run_logs/training_status.json
CHANGED
|
@@ -4,8 +4,8 @@
|
|
| 4 |
{
|
| 5 |
"steps": 49936,
|
| 6 |
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.onnx",
|
| 7 |
-
"reward":
|
| 8 |
-
"creation_time":
|
| 9 |
"auxillary_file_paths": [
|
| 10 |
"results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.pt"
|
| 11 |
]
|
|
@@ -13,8 +13,8 @@
|
|
| 13 |
{
|
| 14 |
"steps": 99960,
|
| 15 |
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.onnx",
|
| 16 |
-
"reward":
|
| 17 |
-
"creation_time":
|
| 18 |
"auxillary_file_paths": [
|
| 19 |
"results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.pt"
|
| 20 |
]
|
|
@@ -22,8 +22,8 @@
|
|
| 22 |
{
|
| 23 |
"steps": 149984,
|
| 24 |
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.onnx",
|
| 25 |
-
"reward":
|
| 26 |
-
"creation_time":
|
| 27 |
"auxillary_file_paths": [
|
| 28 |
"results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.pt"
|
| 29 |
]
|
|
@@ -31,8 +31,8 @@
|
|
| 31 |
{
|
| 32 |
"steps": 199984,
|
| 33 |
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.onnx",
|
| 34 |
-
"reward":
|
| 35 |
-
"creation_time":
|
| 36 |
"auxillary_file_paths": [
|
| 37 |
"results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.pt"
|
| 38 |
]
|
|
@@ -40,8 +40,8 @@
|
|
| 40 |
{
|
| 41 |
"steps": 200112,
|
| 42 |
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.onnx",
|
| 43 |
-
"reward":
|
| 44 |
-
"creation_time":
|
| 45 |
"auxillary_file_paths": [
|
| 46 |
"results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
|
| 47 |
]
|
|
@@ -50,8 +50,8 @@
|
|
| 50 |
"final_checkpoint": {
|
| 51 |
"steps": 200112,
|
| 52 |
"file_path": "results/SnowballTarget1/SnowballTarget.onnx",
|
| 53 |
-
"reward":
|
| 54 |
-
"creation_time":
|
| 55 |
"auxillary_file_paths": [
|
| 56 |
"results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
|
| 57 |
]
|
|
|
|
| 4 |
{
|
| 5 |
"steps": 49936,
|
| 6 |
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.onnx",
|
| 7 |
+
"reward": 13.545454545454545,
|
| 8 |
+
"creation_time": 1694619608.0852468,
|
| 9 |
"auxillary_file_paths": [
|
| 10 |
"results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.pt"
|
| 11 |
]
|
|
|
|
| 13 |
{
|
| 14 |
"steps": 99960,
|
| 15 |
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.onnx",
|
| 16 |
+
"reward": 24.0,
|
| 17 |
+
"creation_time": 1694619738.5600145,
|
| 18 |
"auxillary_file_paths": [
|
| 19 |
"results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.pt"
|
| 20 |
]
|
|
|
|
| 22 |
{
|
| 23 |
"steps": 149984,
|
| 24 |
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.onnx",
|
| 25 |
+
"reward": 25.272727272727273,
|
| 26 |
+
"creation_time": 1694619876.8527107,
|
| 27 |
"auxillary_file_paths": [
|
| 28 |
"results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.pt"
|
| 29 |
]
|
|
|
|
| 31 |
{
|
| 32 |
"steps": 199984,
|
| 33 |
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.onnx",
|
| 34 |
+
"reward": 24.09090909090909,
|
| 35 |
+
"creation_time": 1694620008.1859448,
|
| 36 |
"auxillary_file_paths": [
|
| 37 |
"results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.pt"
|
| 38 |
]
|
|
|
|
| 40 |
{
|
| 41 |
"steps": 200112,
|
| 42 |
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.onnx",
|
| 43 |
+
"reward": 24.09090909090909,
|
| 44 |
+
"creation_time": 1694620008.3675275,
|
| 45 |
"auxillary_file_paths": [
|
| 46 |
"results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
|
| 47 |
]
|
|
|
|
| 50 |
"final_checkpoint": {
|
| 51 |
"steps": 200112,
|
| 52 |
"file_path": "results/SnowballTarget1/SnowballTarget.onnx",
|
| 53 |
+
"reward": 24.09090909090909,
|
| 54 |
+
"creation_time": 1694620008.3675275,
|
| 55 |
"auxillary_file_paths": [
|
| 56 |
"results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
|
| 57 |
]
|