๐Ÿง  Model Overview

Field Details
Model Name mjpsm/checkin-generator-distilgpt2
Base Model distilgpt2
Task Text Generation (Causal Language Modeling)
Training Data ~20,000 cleaned student check-ins
Framework Hugging Face Transformers
Use Case Generate CIC-style check-ins from prompts

โœจ Example

Input:

Today i worked on

Output:

Today i worked on making some progress on getting the authentication set up. It's been a bit of a struggle, but I think i'm finally starting to get the hang of it

โš™๏ธ How to Use

๐Ÿ”น Install Dependencies

pip install transformers torch

๐Ÿ”น Load Model & Tokenizer

from transformers import AutoTokenizer, AutoModelForCausalLM

model_name = "mjpsm/checkin-generator-distilgpt2"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

๐Ÿ”น Generate Text

import torch

def generate(prompt, max_length=50):
    inputs = tokenizer(prompt, return_tensors="pt")
    outputs = model.generate(
        inputs["input_ids"],
        max_length=max_length,
        do_sample=True,
        top_k=50,
        top_p=0.95,
        temperature=0.8,
        pad_token_id=tokenizer.eos_token_id
    )
    return tokenizer.decode(outputs[0], skip_special_tokens=True)

print(generate("today i worked on"))

๐ŸŽฏ Use Cases

  • ๐Ÿง‘โ€๐Ÿ’ป Student check-in generator
  • ๐Ÿค– Slack / Discord bots for daily reflections
  • ๐Ÿ“š Writing assistance tools
  • ๐Ÿง  AI coaching and feedback systems
  • ๐Ÿ“ Auto-completion for journaling platforms

โš ๏ธ Limitations

  • May occasionally repeat phrases or generate generic responses
  • Performance depends on prompt quality
  • Not designed for factual accuracy or external knowledge retrieval

๐Ÿ”ฎ Future Improvements

  • Add topic classification (e.g., debugging, frontend, ML)
  • Improve dataset diversity for richer outputs
  • Deploy as an API or integrate into CIC workflows
  • Add reinforcement learning or prompt tuning

๐Ÿ‘ค Author

Mazamesso Meba (Mazzy)
๐Ÿค— Hugging Face: https://huggingface.co/mjpsm


๐Ÿ“Œ Notes

This project demonstrates the power of fine-tuning pretrained language models for domain-specific text generation. Instead of training from scratch, leveraging existing models allows for faster development and significantly better results.

Downloads last month
74
Safetensors
Model size
81.9M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for mjpsm/checkin-generator-distilgpt2

Finetuned
(1420)
this model

Space using mjpsm/checkin-generator-distilgpt2 1