File size: 1,769 Bytes
b9b1e87
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
{
  "last_updated": "2025-11-12T06:27:12.682777",
  "current_challenge_target": 0.81,
  "total_submissions": 3,
  "results": [
    {
      "model_name": "CompactAI-DynamicAllocation-v1",
      "efficiency_score": 0.603,
      "quality_score": 0.881,
      "token_reduction": 0.302,
      "task_type": "reasoning",
      "dataset": "custom_benchmark",
      "timestamp": "2024-11-12T00:00:00",
      "scaling_law_validated": true,
      "information_theoretic": true,
      "metadata": {
        "organization": "CompactAI",
        "paper_link": "https://arxiv.org/abs/token-efficiency-breakthrough",
        "code_link": "https://github.com/compact-ai/token-efficiency"
      }
    },
    {
      "model_name": "EfficientAttention-Baseline",
      "efficiency_score": 0.35,
      "quality_score": 0.878,
      "token_reduction": 0.0,
      "task_type": "reasoning",
      "dataset": "custom_benchmark",
      "timestamp": "2024-11-01T00:00:00",
      "scaling_law_validated": false,
      "information_theoretic": false,
      "metadata": {
        "organization": "Baseline Research",
        "paper_link": "https://arxiv.org/abs/efficient-attention",
        "code_link": "https://github.com/baseline/efficient-attention"
      }
    },
    {
      "model_name": "ScalingLaw-Challenger-v1",
      "efficiency_score": 0.72,
      "quality_score": 0.875,
      "token_reduction": 0.25,
      "task_type": "qa",
      "dataset": "squad",
      "timestamp": "2024-11-10T00:00:00",
      "scaling_law_validated": true,
      "information_theoretic": true,
      "metadata": {
        "organization": "ScalingLaw Labs",
        "paper_link": "https://arxiv.org/abs/scaling-law-challenge",
        "code_link": "https://github.com/scalinglaw/challenger"
      }
    }
  ]
}