burtenshaw HF Staff commited on
Commit
1f6d2f6
·
verified ·
1 Parent(s): 016dba0

add evaluations results as metadata to model card

Browse files
Files changed (1) hide show
  1. README.md +94 -1
README.md CHANGED
@@ -1,5 +1,98 @@
1
  ---
2
  license: mit
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
4
 
5
  The nanochat-d32 model described in detail [here](https://github.com/karpathy/nanochat/discussions/8).
@@ -9,4 +102,4 @@ I'm sorry this is a janky upload but you have to place these files correctly on
9
  - the token_bytes.pt, tokenizer.pkl have to go into ~/.cache/nanochat/tokenizer directory
10
  - the meta_000650.json, model_000650.pt have to go into ~/.cache/nanochat/chatsft_checkpoints/d32/
11
 
12
- I'll figure out how to make this less janky in the future, and to make nanochat play nicer with huggingface infra.
 
1
  ---
2
  license: mit
3
+ datasets:
4
+ - karpathy/fineweb-edu-100b-shuffle
5
+ model-index:
6
+ - name: chat-d10
7
+ results:
8
+ - task:
9
+ type: text-generation
10
+ name: Text Generation
11
+ dataset:
12
+ name: AI2 Reasoning Challenge (25-Shot)
13
+ type: ai2_arc
14
+ config: ARC-Challenge
15
+ split: test
16
+ metrics:
17
+ - type: acc_norm
18
+ value: 49.91
19
+ name: normalized accuracy
20
+ source:
21
+ url: https://github.com/karpathy/nanochat/discussions/8
22
+ name: nanochat
23
+ - task:
24
+ type: text-generation
25
+ name: Text Generation
26
+ dataset:
27
+ name: AI2 Reasoning Challenge (25-Shot)
28
+ type: ai2_arc
29
+ config: ARC-Easy
30
+ split: test
31
+ metrics:
32
+ - type: acc_norm
33
+ value: 67.97
34
+ name: normalized accuracy
35
+ source:
36
+ url: https://github.com/karpathy/nanochat/discussions/8
37
+ name: nanochat
38
+ - task:
39
+ type: text-generation
40
+ name: Text Generation
41
+ dataset:
42
+ name: MMLU (5-Shot)
43
+ type: cais/mmlu
44
+ config: all
45
+ split: test
46
+ metrics:
47
+ - type: acc
48
+ value: 40.49
49
+ name: accuracy
50
+ source:
51
+ url: https://github.com/karpathy/nanochat/discussions/8
52
+ name: nanochat
53
+ - task:
54
+ type: text-generation
55
+ name: Text Generation
56
+ dataset:
57
+ name: GSM8k (5-shot)
58
+ type: gsm8k
59
+ config: main
60
+ split: test
61
+ metrics:
62
+ - type: acc
63
+ value: 12.74
64
+ name: accuracy
65
+ source:
66
+ url: https://github.com/karpathy/nanochat/discussions/8
67
+ name: nanochat
68
+ - task:
69
+ type: text-generation
70
+ name: Text Generation
71
+ dataset:
72
+ name: HumanEval
73
+ type: openai_humaneval
74
+ split: test
75
+ metrics:
76
+ - type: pass@1
77
+ value: 12.8
78
+ name: pass@1
79
+ source:
80
+ url: https://github.com/karpathy/nanochat/discussions/8
81
+ name: nanochat
82
+ - task:
83
+ type: text-generation
84
+ name: Text Generation
85
+ dataset:
86
+ name: ChatCORE
87
+ type: chatcore
88
+ split: test
89
+ metrics:
90
+ - type: score
91
+ value: 27.34
92
+ name: ChatCORE metric
93
+ source:
94
+ url: https://github.com/karpathy/nanochat/discussions/8
95
+ name: nanochat
96
  ---
97
 
98
  The nanochat-d32 model described in detail [here](https://github.com/karpathy/nanochat/discussions/8).
 
102
  - the token_bytes.pt, tokenizer.pkl have to go into ~/.cache/nanochat/tokenizer directory
103
  - the meta_000650.json, model_000650.pt have to go into ~/.cache/nanochat/chatsft_checkpoints/d32/
104
 
105
+ I'll figure out how to make this less janky in the future, and to make nanochat play nicer with huggingface infra.