QuestCrafter / README.md
Gkop's picture
Upload folder using huggingface_hub
62896ef verified

QuestCrafter-AI_Project

QuestCrafter is a lightweight generative AI project that builds a "Dungeon Master" for RPG quests. It fine-tunes a small language model on curated prompt-to-quest data, compares baseline vs fine-tuned outputs, evaluates quality with automatic metrics and a human rubric, and delivers a simple interactive demo (Streamlit or Gradio).

Setup

  • Python 3.10+
  • Create a virtual env and install dependencies:
    • pip install -r requirements.txt

Repository structure (W1)

  • data/ raw and processed datasets
  • scripts/ data and training scripts (W1 starts with download_data.py)
  • models/ model checkpoints (later)
  • docs/ project docs (board, roles)

Data pipeline (W1)

We use the Reddit Jokes dataset (CSV). The script cleans, filters, splits, and exports JSONL files with a consistent schema.

Download + preprocess

Example: python download_data.py --dataset redditjokes --local_csv path/to/reddit_jokes.csv

Output: data/raw/redditjokes/train.jsonl, val.jsonl, test.jsonl

JSONL schema

Each line has:

  • prompt
  • response
  • source
  • optional metadata (e.g., score, author)

Filters (defaults)

  • Prompt: 5 to 300 characters (if prompt exists)
  • Response: 20 to 800 characters
  • Drops [deleted] / [removed]

Override with: --min_prompt_chars, --max_prompt_chars, --min_response_chars, --max_response_chars, --keep_deleted

Team docs

  • GitHub board and issues: docs/github_board.md
  • Roles, branches, and tasks: docs/team_roles.md

Upload dataset to Hugging Face

Use the script below to upload archive.zip to your dataset repo.

  1. Install deps: pip install -r requirements.txt

  2. Upload (choose one):

  • With token env: set HF_TOKEN=your_token python upload_to_hf.py --repo_id GemimaOndele/questcrafter-dataset --file "C:\Users\gemim\OneDrive\Bureau\M1-cours-Data engineer\MSC 1 AI\Semestre 2\Foundations of machine learning and datascience\Project\archive.zip"

  • With token argument: python upload_to_hf.py --token your_token --repo_id GemimaOndele/questcrafter-dataset --file "C:\Users\gemim\OneDrive\Bureau\M1-cours-Data engineer\MSC 1 AI\Semestre 2\Foundations of machine learning and datascience\Project\archive.zip"

If you already logged in with huggingface-cli login, the script will use that cached token.