midah commited on
Commit
bf96e69
·
verified ·
1 Parent(s): fe2b8c3

Update expanded config dataset: 14,566 models with 78 features

Browse files
Files changed (2) hide show
  1. README.md +158 -6
  2. data/train-00000-of-00001.parquet +2 -2
README.md CHANGED
@@ -1,16 +1,168 @@
1
  ---
2
  dataset_info:
3
  features:
4
- - name: model_id
 
 
5
  dtype: string
6
- - name: config
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  dtype: string
8
  splits:
9
  - name: train
10
- num_bytes: 66349004
11
- num_examples: 16043
12
- download_size: 23577159
13
- dataset_size: 66349004
14
  configs:
15
  - config_name: default
16
  data_files:
 
1
  ---
2
  dataset_info:
3
  features:
4
+ - name: index
5
+ dtype: int64
6
+ - name: modelId
7
  dtype: string
8
+ - name: config_model_type
9
+ dtype: string
10
+ - name: config_architectures
11
+ dtype: string
12
+ - name: config_vocab_size
13
+ dtype: string
14
+ - name: config_torch_dtype
15
+ dtype: string
16
+ - name: config_transformers_version
17
+ dtype: string
18
+ - name: config_hidden_size
19
+ dtype: float64
20
+ - name: config_intermediate_size
21
+ dtype: float64
22
+ - name: config_num_hidden_layers
23
+ dtype: string
24
+ - name: config_num_attention_heads
25
+ dtype: string
26
+ - name: config_num_key_value_heads
27
+ dtype: float64
28
+ - name: config_hidden_act
29
+ dtype: string
30
+ - name: config_attention_dropout
31
+ dtype: float64
32
+ - name: config_use_cache
33
+ dtype: string
34
+ - name: config_max_position_embeddings
35
+ dtype: string
36
+ - name: config_rope_theta
37
+ dtype: string
38
+ - name: config_rope_scaling_type
39
+ dtype: string
40
+ - name: config_rms_norm_eps
41
+ dtype: string
42
+ - name: config_initializer_range
43
+ dtype: string
44
+ - name: config_bos_token_id
45
+ dtype: string
46
+ - name: config_eos_token_id
47
+ dtype: string
48
+ - name: config_tie_word_embeddings
49
+ dtype: string
50
+ - name: config_head_dimension
51
+ dtype: string
52
+ - name: config_gqa_ratio
53
+ dtype: string
54
+ - name: config_context_per_layer
55
+ dtype: string
56
+ - name: config_approx_params_billions
57
+ dtype: string
58
+ - name: is_llama_family
59
+ dtype: string
60
+ - name: is_bert_family
61
+ dtype: string
62
+ - name: is_gpt_family
63
+ dtype: string
64
+ - name: is_t5_family
65
+ dtype: string
66
+ - name: is_whisper_family
67
+ dtype: string
68
+ - name: is_deepseek_family
69
+ dtype: string
70
+ - name: is_mistral_family
71
+ dtype: string
72
+ - name: uses_moe
73
+ dtype: string
74
+ - name: uses_gqa
75
+ dtype: string
76
+ - name: uses_rope
77
+ dtype: string
78
+ - name: uses_quantization
79
+ dtype: string
80
+ - name: uses_lora
81
+ dtype: string
82
+ - name: size_category
83
+ dtype: string
84
+ - name: context_category
85
+ dtype: string
86
+ - name: config_attention_bias
87
+ dtype: string
88
+ - name: config_auto_map_autoconfig
89
+ dtype: string
90
+ - name: config_auto_map_automodel
91
+ dtype: string
92
+ - name: config_auto_map_automodelforcausallm
93
+ dtype: string
94
+ - name: config_ep_size
95
+ dtype: string
96
+ - name: config_first_k_dense_replace
97
+ dtype: string
98
+ - name: config_kv_lora_rank
99
+ dtype: string
100
+ - name: config_layer_norm_eps
101
+ dtype: string
102
+ - name: config_moe_enabled
103
+ dtype: string
104
+ - name: config_moe_expert_utilization
105
+ dtype: string
106
+ - name: config_moe_intermediate_size
107
+ dtype: float64
108
+ - name: config_moe_layer_freq
109
+ dtype: string
110
+ - name: config_n_group
111
+ dtype: float64
112
+ - name: config_n_routed_experts
113
+ dtype: float64
114
+ - name: config_n_shared_experts
115
+ dtype: float64
116
+ - name: config_norm_topk_prob
117
+ dtype: string
118
+ - name: config_num_experts_per_tok
119
+ dtype: float64
120
+ - name: config_num_nextn_predict_layers
121
+ dtype: float64
122
+ - name: config_pad_token_id
123
+ dtype: string
124
+ - name: config_q_lora_rank
125
+ dtype: float64
126
+ - name: config_qk_nope_head_dim
127
+ dtype: float64
128
+ - name: config_qk_rope_head_dim
129
+ dtype: float64
130
+ - name: config_quantization_activation_scheme
131
+ dtype: string
132
+ - name: config_quantization_fmt
133
+ dtype: string
134
+ - name: config_quantization_method
135
+ dtype: string
136
+ - name: config_quantization_weight_block_size
137
+ dtype: string
138
+ - name: config_rope_scaling_beta_fast
139
+ dtype: float64
140
+ - name: config_rope_scaling_beta_slow
141
+ dtype: string
142
+ - name: config_rope_scaling_factor
143
+ dtype: float64
144
+ - name: config_rope_scaling_mscale
145
+ dtype: string
146
+ - name: config_rope_scaling_mscale_all_dim
147
+ dtype: string
148
+ - name: config_rope_scaling_original_max_pos
149
+ dtype: string
150
+ - name: config_routed_scaling_factor
151
+ dtype: string
152
+ - name: config_scoring_func
153
+ dtype: string
154
+ - name: config_topk_group
155
+ dtype: string
156
+ - name: config_topk_method
157
+ dtype: string
158
+ - name: config_v_head_dim
159
  dtype: string
160
  splits:
161
  - name: train
162
+ num_bytes: 8500857
163
+ num_examples: 14566
164
+ download_size: 945908
165
+ dataset_size: 8500857
166
  configs:
167
  - config_name: default
168
  data_files:
data/train-00000-of-00001.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5425a64f7438c74f45f7c99781869d00d2821408f6c94976cde6192a85d302f1
3
- size 23577159
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af128f1c441dafd9006db6cd7c11f5f0142a83b4c97d9ef38c3bf81a77acde94
3
+ size 945908