--- license: mit pretty_name: ReliabilityLoop v1 task_categories: - text-generation - other language: - en tags: - llm - reliability - benchmarking - json - sql - code-generation - evaluation configs: - config_name: viewer data_files: - split: train path: reliability_v1_60_viewer.jsonl --- # ReliabilityLoop v1 ReliabilityLoop v1 is a small, executable benchmark for local LLM reliability across three production-style task types: - `json`: schema-constrained structured extraction - `sql`: text-to-SQL validated by SQLite execution - `codestub`: Python function generation validated by unit tests This dataset is designed for **verifier-based evaluation**: outputs must *work*, not just look plausible. ## Files - `reliability_v1_60.jsonl` Canonical split with 60 tasks: - 20 JSON tasks - 20 SQL tasks - 20 Code tasks - `RELIABILITY_V1_SPEC.md` Benchmark protocol and metric definitions. ## Primary Metric - `policy_ok_rate` = `passed_tasks / total_tasks` A task is counted as passed only if its verifier succeeds: - JSON: parse + schema + expected field checks - SQL: executes and matches expected columns/rows - Code: function compiles and passes tests ## Recommended Evaluation Command Using the ReliabilityLoop framework: ```bash reliabilityloop reliability \ --backend ollama \ --model qwen2.5-coder:0.5b \ --prompts-file eval/reliability_v1_60.jsonl \ --limit 60 \ --max-tokens 96 \ --policy-json contract_first \ --policy-sql baseline_first \ --policy-code baseline_first Outputs: - summary.json - leaderboard.md - samples.jsonl - wins.jsonl ## Reproducibility Notes - Keep config fixed when comparing models: - prompt file - temperature - token budgets - policy mode - Report raw samples.jsonl with summary metrics. - If using memory/retrieval, ensure no leakage from evaluation prompts unless explicitly evaluating memory-assisted mode. ## Anti-Leakage / Fairness Policy For base benchmark comparisons: - Use --memory-file disabled (or disjoint memory source). - Evaluate all models on the same task file and settings. - Publish command + config + artifacts. ## Intended Use - Compare reliability of local LLMs under executable checks. - Evaluate runtime strategies (policy routing, adaptive compute, memory reuse). - Build transparent, reproducible reliability leaderboards. ## Limitations - Small benchmark (60 tasks), alpha-quality split. - Not a replacement for broad general benchmarks. - Current coverage focuses on structured and executable reliability, not open- ended reasoning. ## Links - Framework repo: https://github.com/ranausmanai/reliabilityloop - Dataset: https://huggingface.co/datasets/ranausmans/reliabilityloop-v1