--- license: mit task_categories: - reinforcement-learning - robotics language: - en tags: - drone-navigation - rl-dataset - threejs - ppo - telemetry - pathfinding --- [![Website](https://img.shields.io/badge/webXOS.netlify.app-Explore_Apps-00d4aa?style=for-the-badge&logo=netlify&logoColor=white)](https://webxos.netlify.app) [![GitHub](https://img.shields.io/badge/GitHub-webxos/webxos-181717?style=for-the-badge&logo=github&logoColor=white)](https://github.com/webxos/webxos) [![Hugging Face](https://img.shields.io/badge/Hugging_Face-🤗_webxos-FFD21E?style=for-the-badge&logo=huggingface&logoColor=white)](https://huggingface.co/webxos) [![Follow on X](https://img.shields.io/badge/Follow_@webxos-1DA1F2?style=for-the-badge&logo=x&logoColor=white)](https://x.com/webxos) # DRONE FSD DATASET Single training run (1 epoch, 4 iterations, 198 steps) of a drone navigating a 60×60 room with 15 static + 12 floating obstacles. This dataset was generated with the MIRROR IDE by webXOS. Download the app in the /mirror/ folder to train your own similar datasets. **Final performance (after 2456 frames):** - Best time: **43.821 s** - Success rate: **0.0%** (reached SE corner in best run but did not complete full pattern) - Collisions: **0** in final recorded path - Avg reward: **0.0732** - Cumulative reward: **49.24** - Final exploration rate: **0.784** - Final learning rate: **5.40e-4** ## Network - Architecture: `[256 → 128 → 64 → 32]` (MLP policy/value heads) - Exported: 2026-01-17 03:32 UTC ## Files | File | Description | Size | |-----------------------------|------------------------------------------|----------| | `enhanced_network.json` | Final policy weights + shapes + LR | ~small | | `metadata.json` | Training summary & config | ~small | | `successful_paths.json` | Best 3 partial successes (times, paths) | ~small | | `enhanced_telemetry.jsonl` | Full per-frame telemetry (2456 lines) | ~2.4 MB | | `enhanced_telemetry.csv` | Same data in CSV format | ~1.8 MB | | `training_experiences.jsonl`| PPO-style transitions (state, action, reward, next) | ~1.2 MB | ## Environment - Room: 60 units - Difficulty: 1 - Obstacles: 15 static + 12 floating (0.2–0.5 speed, bounce energy 0.8) - Pattern targets: NW → SE → NE → SW → CENTER - Reward: mostly distance-based + small shaping ## Intended Use - Analyze early-stage PPO behavior on 3D continuous control - Study exploration vs exploitation trade-off (ε still ~78% at end) - Visualize drone trajectories in Three.js / Unity / similar - Baseline for future drone racing / obstacle avoidance models