--- library_name: transformers license: mit datasets: - kjj0/fineweb100B-gpt2 --- trained on 12,312,444,928 tokens from the [kjj0/fineweb100B-gpt2](https://huggingface.co/datasets/kjj0/fineweb100B-gpt2) dataset ``` $ lm_eval --model hf \ --model_args pretrained=michaelbzhu/test-7.6B-base,trust_remote_code=True \ --tasks mmlu_college_medicine,hellaswag,lambada_openai,arc_easy,winogrande,arc_challenge,openbookqa \ --device cuda:0 \ --batch_size 16 | Tasks |Version|Filter|n-shot| Metric | | Value | |Stderr| |----------------|------:|------|-----:|----------|---|------:|---|-----:| |arc_challenge | 1|none | 0|acc |↑ | 0.2295|± |0.0123| | | |none | 0|acc_norm |↑ | 0.2628|± |0.0129| |arc_easy | 1|none | 0|acc |↑ | 0.5358|± |0.0102| | | |none | 0|acc_norm |↑ | 0.4663|± |0.0102| |hellaswag | 1|none | 0|acc |↑ | 0.3788|± |0.0048| | | |none | 0|acc_norm |↑ | 0.4801|± |0.0050| |lambada_openai | 1|none | 0|acc |↑ | 0.4527|± |0.0069| | | |none | 0|perplexity|↓ |14.3601|± |0.4468| |college_medicine| 1|none | 0|acc |↑ | 0.2254|± |0.0319| |openbookqa | 1|none | 0|acc |↑ | 0.1920|± |0.0176| | | |none | 0|acc_norm |↑ | 0.3020|± |0.0206| |winogrande | 1|none | 0|acc |↑ | 0.5107|± |0.0140| ```