| --- |
| library_name: transformers |
| tags: |
| - generated_from_trainer |
| model-index: |
| - name: dragon_interceptor |
| results: [] |
| --- |
| # π Dragon Interceptor (267k) |
|
|
| **Dragon Interceptor** is an ultra-compact, high-speed generative model designed for aviation-inspired structural synthesis. Optimized for **35W mobile hardware**, it bridges the gap between neural text generation and 2D pixel blueprints. |
|
|
| []() |
| []() |
| []() |
|
|
| --- |
|
|
| ## π Performance HUD |
| Benchmarks recorded on **Intel Core i5-10210U** (Surface Pro setup): |
|
|
| * **Weight Loading:** 2055.47 it/s |
| * **Inference Speed:** ~286 iterations/sec |
| * **Full Image Gen (28x28):** ~0.88s |
| * **Brute Force Throughput:** 1.2 - 2.5 images/sec (Turbo Mode) |
|
|
| ## π οΈ Architecture |
| The model utilizes a **GPT-2 Causal LM** backbone, repurposed for spatial data: |
| - **Vocab Size:** 256 (Mapped to 8-bit grayscale intensity) |
| - **Sequence Length:** 784 (Fixed $28 \times 28$ positional embeddings) |
| - **Parameters:** 267,000 (Fits entirely within L2/L3 CPU cache) |
|
|
| ## π°οΈ Key Features |
|
|
| ### 1. Mosaic Synthesis |
| By utilizing a sliding-window context bridge, the model can bypass its 784-position limit to generate seamless, multi-tile blueprints. |
| - **Tile Resolution:** 28x28 |
| - **Global Resolution:** 112x112 (4x4 Mosaic) or custom strips. |
|
|
| ### 2. Seed Brute-Forcing |
| The "Titan DNA" protocol allows for mass-generation of 1,000+ seeds to map the latent space for specific aircraft parts (wings, fuselages, tail fins). |
|
|
| ### 3. Thermal-Aware Inference |
| Optimized for the Surface Pro's 35W power envelope. Uses raw Torch inference with **KV-Caching** to maintain stable frame rates even under thermal pressure. |
|
|
| ## πΌοΈ Sample Inference & Visualization |
| Use the following code to generate a high-fidelity "Radar Scan" from a specific seed. This snippet is optimized for 4K displays and technical clarity. |
|
|
| ```python |
| import torch |
| from transformers import AutoModelForCausalLM |
| import matplotlib.pyplot as plt |
| import os |
| # Load Dragon Interceptor |
| model = AutoModelForCausalLM.from_pretrained("MightyDragon-Dev/dragon_interceptor") |
| |
| # Generate a 28x28 Blueprint (Random Seed) |
| seed_id = os.urandom(1)[0] % 10**6 # Random seed for variability |
| print(f"π Generating Dragon Blueprint with Seed {seed_id}...") |
| input_ids = torch.tensor([[seed_id]]) |
| output = model.generate(input_ids, max_length=784, min_length=784, do_sample=True, temperature=0.7) |
| |
| # Reshape and Render |
| blueprint = output[0].view(28, 28).detach().numpy() |
| |
| plt.figure(figsize=(8, 8), dpi=120) |
| plt.imshow(blueprint, cmap='magma', interpolation='lanczos') |
| plt.title(f"Dragon Interceptor: Sector Scan (Seed {seed_id})", color='white') |
| plt.style.use('dark_background') |
| plt.axis('off') |
| plt.show() |
| |