Antlia-Learn-7B / README.md
oomu-ai's picture
Update README.md
bea3784 verified
---
license: agpl-3.0
tags:
- text-generation
- education
- curriculum
- course-design
pipeline_tag: text-generation
language:
- en
datasets:
- BAAI/Infinity-Instruct
base_model:
- mistralai/Mistral-7B-Instruct-v0.3
---
# Antlia-Learn-7B
## Summary
Antlia-Learn-7B is a 7B-parameter instruction-tuned model for instructional design. It drafts concise short micro-courses and complex long-form course outlines, including learning outcomes, section structure, short practice activities, and lightweight checkpoints. This model is released under the GNU Affero General Public License v3.0 (AGPL-3.0).
## Intended use
* Draft micro-courses with one goal, a 3–4 point outline (minutes in parentheses), and 1–2 short practice tasks.
* Draft long-form course outlines with section objectives, activities, checkpoints, and assessment ideas.
* Rewrite learning objectives in plain language and propose short Socratic prompts.
**Out of scope**
Safety-critical or regulated domains (medical, legal, financial), processing of personal/regulated data, and unsupervised deployment without human review.
## Inputs and outputs
**Input:** English prompts describing topic and scope (e.g., “Create a micro-course on threading a needle” or “Design a 3–4 hour course on building a car engine for advanced beginners”).
**Output:** English text: for micro-courses, a goal + outline + practice tasks; for longer courses, multi-section plans with objectives, activities, checkpoints, and assessment ideas.
## Recommended decoding
The repository includes `generation_config.json` tuned for concise, low-repetition drafting:
* `do_sample: true`
* `temperature: 0.35`
* `top_p: 0.85`
* `top_k: 40`
* `repetition_penalty: 1.3`
* `no_repeat_ngram_size: 5`
* `max_new_tokens: 120`
* `min_new_tokens: 16`
## Model details
* **Architecture:** Decoder-only transformer, \~7B parameters.
* **Precision:** Standard HF weights; quantized inference supported downstream.
* **Alignment:** Supervised instruction-tuning focused on clear, compact educational scaffolds; no RLHF in this release.
## Training
Supervised instruction-tuning across staged passes to encourage concise goals, minute-budgeted outlines, checkpoints, and practice prompts. Exact mixtures, volumes, and hyperparameters are proprietary.
## Limitations and risks
* Not a knowledge base; facts may be incomplete or outdated.
* Tends toward brevity; for richer syllabi, include explicit section counts, activities, and assessment style.
* Can produce generic phrasing for very broad prompts—be specific.
## Safety
All outputs require human review for accuracy and appropriateness. Do not use to generate harmful, discriminatory, or unsafe content.
## License
* **Weights:** GNU Affero General Public License v3.0
* **Use:** Your use must comply with the license and any applicable upstream licenses.
## Third-party notice (required)
This model’s training process made use of content from **BAAI/Infinity-Instruct**, licensed under **Creative Commons Attribution-ShareAlike 4.0 International**.
License: [https://creativecommons.org/licenses/by-sa/4.0/](https://creativecommons.org/licenses/by-sa/4.0/)
Dataset page: [https://huggingface.co/datasets/BAAI/Infinity-Instruct](https://huggingface.co/datasets/BAAI/Infinity-Instruct)
## Contact
Maintainer: **OOMU****[antlia@oomu.ai](mailto:antlia@oomu.ai)**