--- license: apache-2.0 tags: - tiny-llm - dgpo - distillation --- # DGPO Tiny Pure Tiny 0.43M parameter language model trained with DGPO (Distillation-Guided Policy Optimization). - - Python pickle model (1.8MB) - - GGUF format Usage: ```python import pickle with open("dgpo_tiny.pkl", "rb") as f: model = pickle.load(f) ```