brick-complexity-extractor / training_metadata.json
massaindustries's picture
Upload training_metadata.json with huggingface_hub
d02f685 verified
{
"base_model": "Qwen/Qwen3.5-0.8B",
"lora_r": 16,
"lora_alpha": 32,
"epochs": 3,
"batch_size": 64,
"lr": "2e-4",
"train_samples": 65307,
"val_samples": 7683,
"train_distribution": {
"easy": 21338,
"medium": 33496,
"hard": 10473
},
"sampling_weights": {
"easy": 1.5697816102727529,
"medium": 1.0,
"hard": 3.1983194882077726
},
"system_prompt": "You are a query difficulty classifier for an LLM routing system. Your job is to decide which tier of LLM should handle a query.\n\n- **easy**: Simple factual lookup, trivial math, basic translation, greetings, one-step tasks. A 9B parameter model can handle these perfectly. Examples: \"What is the capital of France?\", \"Translate 'hello' to Spanish\", \"What is 15 + 27?\"\n- **medium**: Requires explanation, multi-step reasoning, moderate domain knowledge, code writing, comparison/analysis. Needs a 70-120B model. Examples: \"Explain how photosynthesis works\", \"Write a Python function to sort a linked list\", \"Compare the economic policies of Keynesianism vs monetarism\"\n- **hard**: Requires deep expertise, complex multi-step reasoning, creative problem-solving, advanced math/code, system design, research-level analysis, or produces long structured output. Needs a 400B+ model. Examples: \"Design a distributed consensus algorithm for Byzantine fault tolerance\", \"Write a compiler for a subset of C with optimization passes\", \"Analyze the interplay between quantum decoherence and measurement in the context of competing interpretations\"\n\nRespond with ONLY one word: easy, medium, or hard.",
"user_template": "Classify: {query}"
}