๐ง Model Overview
| Field | Details |
|---|---|
| Model Name | mjpsm/checkin-generator-distilgpt2 |
| Base Model | distilgpt2 |
| Task | Text Generation (Causal Language Modeling) |
| Training Data | ~20,000 cleaned student check-ins |
| Framework | Hugging Face Transformers |
| Use Case | Generate CIC-style check-ins from prompts |
โจ Example
Input:
Today i worked on
Output:
Today i worked on making some progress on getting the authentication set up. It's been a bit of a struggle, but I think i'm finally starting to get the hang of it
โ๏ธ How to Use
๐น Install Dependencies
pip install transformers torch
๐น Load Model & Tokenizer
from transformers import AutoTokenizer, AutoModelForCausalLM
model_name = "mjpsm/checkin-generator-distilgpt2"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)
๐น Generate Text
import torch
def generate(prompt, max_length=50):
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(
inputs["input_ids"],
max_length=max_length,
do_sample=True,
top_k=50,
top_p=0.95,
temperature=0.8,
pad_token_id=tokenizer.eos_token_id
)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
print(generate("today i worked on"))
๐ฏ Use Cases
- ๐งโ๐ป Student check-in generator
- ๐ค Slack / Discord bots for daily reflections
- ๐ Writing assistance tools
- ๐ง AI coaching and feedback systems
- ๐ Auto-completion for journaling platforms
โ ๏ธ Limitations
- May occasionally repeat phrases or generate generic responses
- Performance depends on prompt quality
- Not designed for factual accuracy or external knowledge retrieval
๐ฎ Future Improvements
- Add topic classification (e.g., debugging, frontend, ML)
- Improve dataset diversity for richer outputs
- Deploy as an API or integrate into CIC workflows
- Add reinforcement learning or prompt tuning
๐ค Author
Mazamesso Meba (Mazzy)
๐ค Hugging Face: https://huggingface.co/mjpsm
๐ Notes
This project demonstrates the power of fine-tuning pretrained language models for domain-specific text generation. Instead of training from scratch, leveraging existing models allows for faster development and significantly better results.
- Downloads last month
- 74
Model tree for mjpsm/checkin-generator-distilgpt2
Base model
distilbert/distilgpt2