VisionaryKunal commited on
Commit
bd8fb2e
·
verified ·
1 Parent(s): f521289

Add trained model and demo video

Browse files
Files changed (3) hide show
  1. .gitattributes +1 -0
  2. 3DBall_Demo_(1).mp4 +3 -0
  3. README.md +35 -54
.gitattributes CHANGED
@@ -35,3 +35,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  3DBall_Demo.mp4 filter=lfs diff=lfs merge=lfs -text
37
  3DBall_Demo[[:space:]](online-video-cutter.com).mp4 filter=lfs diff=lfs merge=lfs -text
 
 
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  3DBall_Demo.mp4 filter=lfs diff=lfs merge=lfs -text
37
  3DBall_Demo[[:space:]](online-video-cutter.com).mp4 filter=lfs diff=lfs merge=lfs -text
38
+ 3DBall_Demo_(1).mp4 filter=lfs diff=lfs merge=lfs -text
3DBall_Demo_(1).mp4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58a928222520572fa05bd7b2c8315ea4f481f36a5432b4cbf5fe98b7491a61fa
3
+ size 1611615
README.md CHANGED
@@ -1,54 +1,35 @@
1
- ---
2
- library_name: ml-agents
3
- tags:
4
- - 3d-ball
5
- - deep-reinforcement-learning
6
- - reinforcement-learning
7
- - ppo
8
- - unity-ml-agents
9
- ---
10
-
11
- # 3DBall Trained Agent
12
-
13
- This is a trained model of a PPO agent playing the 3DBall environment, created using the Unity ML-Agents library. The agent learns to balance a ball on a moving platform for as long as possible.
14
-
15
- ### Training Hyperparameters
16
-
17
- The agent was trained using the following configuration from the `3DBall.yaml` file:
18
-
19
- ```yaml
20
- behaviors:
21
- 3DBall:
22
- trainer_type: ppo
23
- hyperparameters:
24
- learning_rate: 0.0003
25
- learning_rate_schedule: linear
26
- beta: 0.0005
27
- epsilon: 0.2
28
- lambd: 0.95
29
- num_epoch: 3
30
- buffer_size: 2048
31
- batch_size: 256
32
- time_horizon: 1024
33
- network_settings:
34
- normalize: false
35
- hidden_units: 128
36
- num_layers: 2
37
- vis_encode_type: simple
38
- reward_signals:
39
- extrinsic:
40
- gamma: 0.99
41
- strength: 1.0
42
- checkpoint_interval: 500000
43
- threaded: true
44
- ```
45
-
46
- ### Video Demo
47
-
48
- Here is a video of the trained agent in action, demonstrating the learned behavior.
49
-
50
- <video controls width="100%">
51
- <source src="[results/3DBallTraining/`3DBall_Demo (online-video-cutter.com).mp4`]("C:/Users/KUNAL/OneDrive/Desktop/Pirate_AP/DeepRL/results/3DBallTraining/`3DBall_Demo (online-video-cutter.com).mp4`")" type="video/mp4">
52
- Your browser does not support the video tag.
53
- </video>
54
-
 
1
+ ---
2
+ library_name: ml-agents
3
+ tags:
4
+ - 3DBall
5
+ - deep-reinforcement-learning
6
+ - reinforcement-learning
7
+ - ML-Agents-3DBall
8
+ ---
9
+
10
+ # **ppo** Agent playing **3DBall**
11
+ This is a trained model of a **ppo** agent playing **3DBall**
12
+ using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
13
+
14
+ ## Usage (with ML-Agents)
15
+ The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/
16
+
17
+ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
18
+ - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your
19
+ browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction
20
+ - A *longer tutorial* to understand how works ML-Agents:
21
+ https://huggingface.co/learn/deep-rl-course/unit5/introduction
22
+
23
+ ### Resume the training
24
+ ```bash
25
+ mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
26
+ ```
27
+
28
+ ### Watch your Agent play
29
+ You can watch your agent **playing directly in your browser**
30
+
31
+ 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity
32
+ 2. Step 1: Find your model_id: VisionaryKunal/3DBall-MLAgents
33
+ 3. Step 2: Select your *.nn /*.onnx file
34
+ 4. Click on Watch the agent play 👀
35
+