--- license: agpl-3.0 tags: - text-generation - education - curriculum - course-design pipeline_tag: text-generation language: - en datasets: - BAAI/Infinity-Instruct base_model: - mistralai/Mistral-7B-Instruct-v0.3 --- # Antlia-Learn-7B ## Summary Antlia-Learn-7B is a 7B-parameter instruction-tuned model for instructional design. It drafts concise short micro-courses and complex long-form course outlines, including learning outcomes, section structure, short practice activities, and lightweight checkpoints. This model is released under the GNU Affero General Public License v3.0 (AGPL-3.0). ## Intended use * Draft micro-courses with one goal, a 3–4 point outline (minutes in parentheses), and 1–2 short practice tasks. * Draft long-form course outlines with section objectives, activities, checkpoints, and assessment ideas. * Rewrite learning objectives in plain language and propose short Socratic prompts. **Out of scope** Safety-critical or regulated domains (medical, legal, financial), processing of personal/regulated data, and unsupervised deployment without human review. ## Inputs and outputs **Input:** English prompts describing topic and scope (e.g., “Create a micro-course on threading a needle” or “Design a 3–4 hour course on building a car engine for advanced beginners”). **Output:** English text: for micro-courses, a goal + outline + practice tasks; for longer courses, multi-section plans with objectives, activities, checkpoints, and assessment ideas. ## Recommended decoding The repository includes `generation_config.json` tuned for concise, low-repetition drafting: * `do_sample: true` * `temperature: 0.35` * `top_p: 0.85` * `top_k: 40` * `repetition_penalty: 1.3` * `no_repeat_ngram_size: 5` * `max_new_tokens: 120` * `min_new_tokens: 16` ## Model details * **Architecture:** Decoder-only transformer, \~7B parameters. * **Precision:** Standard HF weights; quantized inference supported downstream. * **Alignment:** Supervised instruction-tuning focused on clear, compact educational scaffolds; no RLHF in this release. ## Training Supervised instruction-tuning across staged passes to encourage concise goals, minute-budgeted outlines, checkpoints, and practice prompts. Exact mixtures, volumes, and hyperparameters are proprietary. ## Limitations and risks * Not a knowledge base; facts may be incomplete or outdated. * Tends toward brevity; for richer syllabi, include explicit section counts, activities, and assessment style. * Can produce generic phrasing for very broad prompts—be specific. ## Safety All outputs require human review for accuracy and appropriateness. Do not use to generate harmful, discriminatory, or unsafe content. ## License * **Weights:** GNU Affero General Public License v3.0 * **Use:** Your use must comply with the license and any applicable upstream licenses. ## Third-party notice (required) This model’s training process made use of content from **BAAI/Infinity-Instruct**, licensed under **Creative Commons Attribution-ShareAlike 4.0 International**. License: [https://creativecommons.org/licenses/by-sa/4.0/](https://creativecommons.org/licenses/by-sa/4.0/) Dataset page: [https://huggingface.co/datasets/BAAI/Infinity-Instruct](https://huggingface.co/datasets/BAAI/Infinity-Instruct) ## Contact Maintainer: **OOMU** — **[antlia@oomu.ai](mailto:antlia@oomu.ai)**