rohit23 commited on
Commit
275d4be
·
verified ·
1 Parent(s): f72b14d

Pushing Unit 5 Pyramids1 model

Browse files
Pyramids.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8119c613b2b885b2b46dca785a07376a22c7b832bfcc45f1908167d254f44c4
3
+ size 1420238
Pyramids/Pyramids-0.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0658db6f81b9f303ea4f86e0970bd6ca133c892a043a4982ba3e942deb5ed705
3
+ size 1418249
Pyramids/Pyramids-0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:781cc01a75034a1980268b4cc367117d362b59d25daefa0d645dd7604b253289
3
+ size 2995554
Pyramids/Pyramids-1000012.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8119c613b2b885b2b46dca785a07376a22c7b832bfcc45f1908167d254f44c4
3
+ size 1420238
Pyramids/Pyramids-1000012.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d61670d10b37b90efac3fa936c40464bfbb0742d79ae575689233a4b06b9d891
3
+ size 8656252
Pyramids/Pyramids-499981.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:950c9d9a8741da24696dee281a070ad04914117688f26f0b9d299bca5634e84a
3
+ size 1420238
Pyramids/Pyramids-499981.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e6dcd56533f76892721b2ff3d1ca3212e7825139aad0a01bc8f65de3d9d867d
3
+ size 8656173
Pyramids/Pyramids-999884.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8119c613b2b885b2b46dca785a07376a22c7b832bfcc45f1908167d254f44c4
3
+ size 1420238
Pyramids/Pyramids-999884.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44b560a5f54e77001c5c29fd501c31713c9ce21147c8ed4504c00e9f46999c43
3
+ size 8656173
Pyramids/checkpoint.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e315e3e9d4878dfc818de50b2fbd3f39f106b339e3b64d0d05ec0c9c8f52259
3
+ size 8655778
Pyramids/events.out.tfevents.1774334075.paul23.6488.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2854b0a97743dd6ce2135502ee302783712678d7467c8c79bf13d094550fee4
3
+ size 313961
Pyramids/events.out.tfevents.1774351333.paul23.63896.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c76ca5249a5d96c21b8984b586dce625ffc33f4a2b1e538c7121f0b5fea12ee
3
+ size 1423
README.md ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: ml-agents
3
+ tags:
4
+ - Pyramids
5
+ - deep-reinforcement-learning
6
+ - reinforcement-learning
7
+ - ML-Agents-Pyramids
8
+ ---
9
+
10
+ # **ppo** Agent playing **Pyramids**
11
+ This is a trained model of a **ppo** agent playing **Pyramids**
12
+ using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
13
+
14
+ ## Usage (with ML-Agents)
15
+ The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/
16
+
17
+ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
18
+ - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your
19
+ browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction
20
+ - A *longer tutorial* to understand how works ML-Agents:
21
+ https://huggingface.co/learn/deep-rl-course/unit5/introduction
22
+
23
+ ### Resume the training
24
+ ```bash
25
+ mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
26
+ ```
27
+
28
+ ### Watch your Agent play
29
+ You can watch your agent **playing directly in your browser**
30
+
31
+ 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity
32
+ 2. Step 1: Find your model_id: rohit23/ppo-Pyramids-version001
33
+ 3. Step 2: Select your *.nn /*.onnx file
34
+ 4. Click on Watch the agent play 👀
35
+
config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"default_settings": null, "behaviors": {"Pyramids": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate": 0.0003, "beta": 0.01, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "shared_critic": false, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "checkpoint_interval": 500000, "network_settings": {"normalize": false, "hidden_units": 512, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}, "rnd": {"gamma": 0.99, "strength": 0.01, "network_settings": {"normalize": false, "hidden_units": 64, "num_layers": 3, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "learning_rate": 0.0001, "encoding_size": null}}, "init_path": null, "keep_checkpoints": 5, "even_checkpoints": false, "max_steps": 1000000, "time_horizon": 128, "summary_freq": 30000, "threaded": false, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "C:\\Users\\rohit\\Downloads\\RL-learn\\training-envs-executables\\windows\\Pyramids\\UnityEnvironment.exe", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "timeout_wait": 60, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": false}, "environment_parameters": null, "checkpoint_settings": {"run_id": "Pyramids1", "initialize_from": null, "load_model": false, "resume": true, "force": false, "train_model": false, "inference": true, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
configuration.yaml ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ default_settings: null
2
+ behaviors:
3
+ Pyramids:
4
+ trainer_type: ppo
5
+ hyperparameters:
6
+ batch_size: 128
7
+ buffer_size: 2048
8
+ learning_rate: 0.0003
9
+ beta: 0.01
10
+ epsilon: 0.2
11
+ lambd: 0.95
12
+ num_epoch: 3
13
+ shared_critic: false
14
+ learning_rate_schedule: linear
15
+ beta_schedule: linear
16
+ epsilon_schedule: linear
17
+ checkpoint_interval: 500000
18
+ network_settings:
19
+ normalize: false
20
+ hidden_units: 512
21
+ num_layers: 2
22
+ vis_encode_type: simple
23
+ memory: null
24
+ goal_conditioning_type: hyper
25
+ deterministic: false
26
+ reward_signals:
27
+ extrinsic:
28
+ gamma: 0.99
29
+ strength: 1.0
30
+ network_settings:
31
+ normalize: false
32
+ hidden_units: 128
33
+ num_layers: 2
34
+ vis_encode_type: simple
35
+ memory: null
36
+ goal_conditioning_type: hyper
37
+ deterministic: false
38
+ rnd:
39
+ gamma: 0.99
40
+ strength: 0.01
41
+ network_settings:
42
+ normalize: false
43
+ hidden_units: 64
44
+ num_layers: 3
45
+ vis_encode_type: simple
46
+ memory: null
47
+ goal_conditioning_type: hyper
48
+ deterministic: false
49
+ learning_rate: 0.0001
50
+ encoding_size: null
51
+ init_path: null
52
+ keep_checkpoints: 5
53
+ even_checkpoints: false
54
+ max_steps: 1000000
55
+ time_horizon: 128
56
+ summary_freq: 30000
57
+ threaded: false
58
+ self_play: null
59
+ behavioral_cloning: null
60
+ env_settings:
61
+ env_path: C:\Users\rohit\Downloads\RL-learn\training-envs-executables\windows\Pyramids\UnityEnvironment.exe
62
+ env_args: null
63
+ base_port: 5005
64
+ num_envs: 1
65
+ num_areas: 1
66
+ timeout_wait: 60
67
+ seed: -1
68
+ max_lifetime_restarts: 10
69
+ restarts_rate_limit_n: 1
70
+ restarts_rate_limit_period_s: 60
71
+ engine_settings:
72
+ width: 84
73
+ height: 84
74
+ quality_level: 5
75
+ time_scale: 20
76
+ target_frame_rate: -1
77
+ capture_frame_rate: 60
78
+ no_graphics: false
79
+ environment_parameters: null
80
+ checkpoint_settings:
81
+ run_id: Pyramids1
82
+ initialize_from: null
83
+ load_model: false
84
+ resume: true
85
+ force: false
86
+ train_model: false
87
+ inference: true
88
+ results_dir: results
89
+ torch_settings:
90
+ device: null
91
+ debug: false
run_logs/Player-0.log ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Mono path[0] = 'C:/Users/rohit/Downloads/RL-learn/training-envs-executables/windows/Pyramids/UnityEnvironment_Data/Managed'
2
+ Mono config path = 'C:/Users/rohit/Downloads/RL-learn/training-envs-executables/windows/Pyramids/MonoBleedingEdge/etc'
3
+ CodeReloadManager initialized
4
+ Input System module state changed to: Initialized.
5
+ [Physics::Module] Initialized fallback backend.
6
+ [Physics::Module] Id: 0xdecafbad
7
+ Initialize engine version: 6000.4.0f1 (8cf496087c8f)
8
+ [Subsystems] Discovering subsystems at path C:/Users/rohit/Downloads/RL-learn/training-envs-executables/windows/Pyramids/UnityEnvironment_Data/UnitySubsystems
9
+ [D3D12 Device Filter] Vendor Name: NVIDIA
10
+ [D3D12 Device Filter] Device Name: NVIDIA GeForce RTX 3050 6GB Laptop GPU
11
+ [D3D12 Device Filter] Driver Version: 32.0.15.5585
12
+ [D3D12 Device Filter] Feature Level: 12.2
13
+ [D3D12 Device Filter] Graphics Memory: 6002 MB
14
+ [D3D12 Device Filter] Processor Count: 12
15
+ [D3D12 Device Filter] Device Type: Discrete
16
+ GfxDevice: creating device client; kGfxThreadingModeThreaded
17
+ Direct3D:
18
+ Version: Direct3D 11.0 [level 11.1]
19
+ Renderer: NVIDIA GeForce RTX 3050 6GB Laptop GPU (ID=0x25ac)
20
+ Vendor: NVIDIA
21
+ VRAM: 6002 MB
22
+ Driver: 32.0.15.5585
23
+ Begin MonoManager ReloadAssembly
24
+ - Finished resetting the current domain, in 0.002 seconds
25
+ [Physics::Module] Selected backend.
26
+ [Physics::Module] Name: PhysX
27
+ [Physics::Module] Id: 0xf2b8ea05
28
+ [Physics::Module] SDK Version: 4.1.2
29
+ [Physics::Module] Integration Version: 1.0.0
30
+ [Physics::Module] Threading Mode: Multi-Threaded
31
+ <RI> Initializing input.
32
+ Using Windows.Gaming.Input
33
+ XInput1_3.dll not found. Trying XInput9_1_0.dll instead...
34
+ <RI> Input initialized.
35
+ UnloadTime: 0.645500 ms
36
+ Registered Communicator in Agent.
37
+ [Physics::Module] Cleanup current backend.
38
+ [Physics::Module] Id: 0xf2b8ea05
39
+ Input System module state changed to: ShutdownInProgress.
40
+ Input Polling Thread exited.
41
+ Input System module state changed to: Shutdown.
42
+ CodeReloadManager destroyed
run_logs/timers.json ADDED
@@ -0,0 +1,143 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "root",
3
+ "metadata": {
4
+ "timer_format_version": "0.1.0",
5
+ "start_time_seconds": "1774351324",
6
+ "python_version": "3.10.11 | packaged by Anaconda, Inc. | (main, May 16 2023, 00:55:32) [MSC v.1916 64 bit (AMD64)]",
7
+ "command_line_arguments": "C:\\Users\\rohit\\miniconda3\\envs\\rl-learn\\Scripts\\mlagents-learn C:\\Users\\rohit\\Downloads\\RL-learn\\config\\ppo\\PyramidsRND.yaml --env=C:\\Users\\rohit\\Downloads\\RL-learn\\training-envs-executables\\windows\\Pyramids\\UnityEnvironment.exe --run-id=Pyramids1 --inference --resume",
8
+ "mlagents_version": "1.0.0",
9
+ "mlagents_envs_version": "1.1.0",
10
+ "communication_protocol_version": "1.5.0",
11
+ "pytorch_version": "2.6.0+cu124",
12
+ "numpy_version": "1.23.5",
13
+ "end_time_seconds": "1774351372"
14
+ },
15
+ "total": 48.87714009999763,
16
+ "count": 1,
17
+ "self": 10.006929499912076,
18
+ "children": {
19
+ "run_training.setup": {
20
+ "total": 0.1279296000720933,
21
+ "count": 1,
22
+ "self": 0.1279296000720933
23
+ },
24
+ "TrainerController.start_learning": {
25
+ "total": 38.74228100001346,
26
+ "count": 1,
27
+ "self": 0.029113499796949327,
28
+ "children": {
29
+ "TrainerController._reset_env": {
30
+ "total": 11.584887400036678,
31
+ "count": 1,
32
+ "self": 11.584887400036678
33
+ },
34
+ "TrainerController.advance": {
35
+ "total": 27.128279200172983,
36
+ "count": 1205,
37
+ "self": 0.030219002277590334,
38
+ "children": {
39
+ "env_step": {
40
+ "total": 24.78010079928208,
41
+ "count": 1205,
42
+ "self": 19.980615599779412,
43
+ "children": {
44
+ "SubprocessEnvManager._take_step": {
45
+ "total": 4.782365499762818,
46
+ "count": 1206,
47
+ "self": 0.1007295026211068,
48
+ "children": {
49
+ "TorchPolicy.evaluate": {
50
+ "total": 4.681635997141711,
51
+ "count": 1175,
52
+ "self": 4.681635997141711
53
+ }
54
+ }
55
+ },
56
+ "workers": {
57
+ "total": 0.017119699739851058,
58
+ "count": 1204,
59
+ "self": 0.0,
60
+ "children": {
61
+ "worker_root": {
62
+ "total": 23.39705480169505,
63
+ "count": 1204,
64
+ "is_parallel": true,
65
+ "self": 11.783425701782107,
66
+ "children": {
67
+ "steps_from_proto": {
68
+ "total": 0.0033790001180022955,
69
+ "count": 2,
70
+ "is_parallel": true,
71
+ "self": 0.0006664003012701869,
72
+ "children": {
73
+ "_process_rank_one_or_two_observation": {
74
+ "total": 0.0027125998167321086,
75
+ "count": 16,
76
+ "is_parallel": true,
77
+ "self": 0.0027125998167321086
78
+ }
79
+ }
80
+ },
81
+ "UnityEnvironment.step": {
82
+ "total": 11.61025009979494,
83
+ "count": 1204,
84
+ "is_parallel": true,
85
+ "self": 0.3885261978721246,
86
+ "children": {
87
+ "UnityEnvironment._generate_step_input": {
88
+ "total": 0.316357901901938,
89
+ "count": 1204,
90
+ "is_parallel": true,
91
+ "self": 0.316357901901938
92
+ },
93
+ "communicator.exchange": {
94
+ "total": 9.808723400346935,
95
+ "count": 1204,
96
+ "is_parallel": true,
97
+ "self": 9.808723400346935
98
+ },
99
+ "steps_from_proto": {
100
+ "total": 1.0966425996739417,
101
+ "count": 1204,
102
+ "is_parallel": true,
103
+ "self": 0.2563022974645719,
104
+ "children": {
105
+ "_process_rank_one_or_two_observation": {
106
+ "total": 0.8403403022093698,
107
+ "count": 9632,
108
+ "is_parallel": true,
109
+ "self": 0.8403403022093698
110
+ }
111
+ }
112
+ }
113
+ }
114
+ }
115
+ }
116
+ }
117
+ }
118
+ }
119
+ }
120
+ },
121
+ "trainer_advance": {
122
+ "total": 2.3179593986133114,
123
+ "count": 1204,
124
+ "self": 0.03145549667533487,
125
+ "children": {
126
+ "process_trajectory": {
127
+ "total": 2.2865039019379765,
128
+ "count": 1204,
129
+ "self": 2.2865039019379765
130
+ }
131
+ }
132
+ }
133
+ }
134
+ },
135
+ "trainer_threads": {
136
+ "total": 9.00006853044033e-07,
137
+ "count": 1,
138
+ "self": 9.00006853044033e-07
139
+ }
140
+ }
141
+ }
142
+ }
143
+ }
run_logs/training_status.json ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "Pyramids": {
3
+ "checkpoints": [
4
+ {
5
+ "steps": 499981,
6
+ "file_path": "results\\Pyramids1\\Pyramids\\Pyramids-499981.onnx",
7
+ "reward": 1.364499993622303,
8
+ "creation_time": 1774335910.8615959,
9
+ "auxillary_file_paths": [
10
+ "results\\Pyramids1\\Pyramids\\Pyramids-499981.pt"
11
+ ]
12
+ },
13
+ {
14
+ "steps": 999884,
15
+ "file_path": "results\\Pyramids1\\Pyramids\\Pyramids-999884.onnx",
16
+ "reward": 1.488333319624265,
17
+ "creation_time": 1774337884.5285158,
18
+ "auxillary_file_paths": [
19
+ "results\\Pyramids1\\Pyramids\\Pyramids-999884.pt"
20
+ ]
21
+ },
22
+ {
23
+ "steps": 1000012,
24
+ "file_path": "results\\Pyramids1\\Pyramids\\Pyramids-1000012.onnx",
25
+ "reward": 1.488333319624265,
26
+ "creation_time": 1774337884.713831,
27
+ "auxillary_file_paths": [
28
+ "results\\Pyramids1\\Pyramids\\Pyramids-1000012.pt"
29
+ ]
30
+ }
31
+ ],
32
+ "final_checkpoint": {
33
+ "steps": 1000012,
34
+ "file_path": "results\\Pyramids1\\Pyramids.onnx",
35
+ "reward": 1.488333319624265,
36
+ "creation_time": 1774337884.713831,
37
+ "auxillary_file_paths": [
38
+ "results\\Pyramids1\\Pyramids\\Pyramids-1000012.pt"
39
+ ]
40
+ }
41
+ },
42
+ "metadata": {
43
+ "stats_format_version": "0.3.0",
44
+ "mlagents_version": "1.0.0",
45
+ "torch_version": "2.6.0+cu124"
46
+ }
47
+ }