qwen3-32b Fine-tuned for Story Creativity Evaluation

Fine-tuned qwen3-32b model for evaluating story creativity (1-5 scale).

Model Details

  • Base Model: Qwen/qwen3-32b
  • Method: LoRA (r=8, alpha=16)
  • Training Data: 5000 story creativity evaluations
  • Epochs: 3
  • Final Loss: ~1.26

Usage

from transformers import AutoTokenizer, AutoModelForCausalLM
from peft import PeftModel
import torch

# Load models
base_model = AutoModelForCausalLM.from_pretrained(
    "Qwen/qwen3-32b",
    torch_dtype=torch.float16,
    device_map="auto"
)
model = PeftModel.from_pretrained(base_model, "CromonHarry/qwen3-32b-creativity")
tokenizer = AutoTokenizer.from_pretrained("CromonHarry/qwen3-32b-creativity")

# Evaluate a story
prompt = '''Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.

### Instruction:
Evaluate the creativity of the following story. You need to output the following format:
[Your review]
Creativity score: [A float score from 1 to 5]

### Input:
{your_story_here}

### Response:
'''

inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=256, temperature=0.7)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Training Configuration

  • Learning Rate: 2e-4 (cosine scheduler)
  • Batch Size: 4 (effective: 32)
  • Optimizer: AdamW
  • Gradient Checkpointing: Enabled
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support