Datasets:
Tasks:
Text Classification
Languages:
English
Size:
n<1K
Tags:
pragmatic-reasoning
theory-of-mind
emotion-inference
indirect-speech
benchmark
multi-annotator
License:
| # ============================================================================= | |
| # CEI-ToM DMLR 2026 Configuration | |
| # ============================================================================= | |
| # IMPORTANT: DMLR baselines must use models NOT evaluated in the CogSci 2026 | |
| # companion paper. See cogsci_off_limits below for the full exclusion list. | |
| llm_inference: | |
| # Model sets by execution mode | |
| models: | |
| # TEST mode: single model for quick validation | |
| test: | |
| - id: "gpt-5-mini" | |
| provider: "openai" | |
| # COMPLETE mode: DMLR-safe models (none overlap with CogSci 2026) | |
| complete: | |
| # Commercial API models | |
| - id: "gpt-5-mini" | |
| provider: "openai" | |
| - id: "claude-sonnet-4-5" | |
| provider: "anthropic" | |
| - id: "grok-4-1-fast-non-reasoning" | |
| provider: "xai" | |
| - id: "gemini-2.5-flash" | |
| provider: "google" | |
| # Open-source models via API providers | |
| - id: "meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo" | |
| provider: "together" | |
| - id: "accounts/fireworks/models/deepseek-v3p1" | |
| provider: "fireworks" | |
| - id: "Qwen/Qwen2.5-7B-Instruct-Turbo" | |
| provider: "together" | |
| # Models used in CogSci 2026 — DO NOT use for DMLR baselines | |
| cogsci_off_limits: | |
| behavioral: | |
| - "gpt-5-mini" | |
| - "claude-haiku-4-5" | |
| - "grok-4-1-fast" | |
| - "gemini-3-flash-preview" | |
| - "kimi-k2-instruct-0905" | |
| - "qwen3-235b-a22b-instruct-2507" | |
| - "deepseek-v3p2" | |
| - "minimax-m2p1" | |
| - "Meta-Llama-3.1-8B-Instruct-Turbo" | |
| - "Mistral-Small-24B-Instruct-2501" | |
| - "gemma-3n-E4B-it" | |
| probing: | |
| - "llama-3-8b" | |
| - "mistral-7b" | |
| - "flan-t5-xxl" | |
| - "mixtral-8x22b" | |
| # ============================================================================= | |
| # PRICING (USD per 1M tokens, as of 2026-02-07) | |
| # ============================================================================= | |
| pricing_usd_per_1m_tokens: | |
| # --- DMLR-safe models (recommended for baselines) --- | |
| openai: | |
| gpt-5-mini: {input: 0.25, output: 2.00} | |
| _default: {input: 0.25, output: 2.00} | |
| anthropic: | |
| claude-sonnet-4-5: {input: 3.00, output: 15.00} | |
| _default: {input: 3.00, output: 15.00} | |
| google: | |
| gemini-2.0-flash: {input: 0.10, output: 0.40} | |
| gemini-2.5-flash: {input: 0.30, output: 2.50} | |
| _default: {input: 0.30, output: 2.50} | |
| xai: | |
| grok-4-1-fast-non-reasoning: {input: 0.20, output: 0.50} | |
| _default: {input: 0.20, output: 0.50} | |
| fireworks: | |
| deepseek-v3p1: {input: 0.15, output: 0.75} | |
| _default: {input: 1.00, output: 5.00} | |
| together: | |
| meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo: {input: 0.88, output: 0.88} | |
| Qwen/Qwen2.5-7B-Instruct-Turbo: {input: 0.30, output: 0.30} | |
| _default: {input: 0.15, output: 0.60} | |
| ollama: | |
| _default: {input: 0.0, output: 0.0} | |
| # --- CogSci models (reference only, DO NOT use for DMLR) --- | |
| # openai: gpt-5-mini {input: 0.25, output: 2.00} | |
| # anthropic: claude-haiku-4-5 {input: 1.00, output: 5.00} | |
| # xai: grok-4-1-fast {input: 0.20, output: 0.50} | |
| # google: gemini-3-flash {input: 0.50, output: 3.00} | |
| # fireworks: kimi-k2 {input: 0.60, output: 2.50} | |
| # fireworks: qwen3-235b {input: 0.22, output: 0.88} | |
| # fireworks: deepseek-v3p2 {input: 0.56, output: 1.68} | |
| # fireworks: minimax-m2p1 {input: 0.30, output: 1.20} | |
| # together: llama-3.1-8b {input: 0.18, output: 0.18} | |
| # together: mistral-small-24b {input: 0.10, output: 0.30} | |
| # together: gemma-3n-e4b {input: 0.03, output: 0.03} |