Upload variant_a/logs/eq_bench_log.txt with huggingface_hub
Browse files
variant_a/logs/eq_bench_log.txt
ADDED
|
@@ -0,0 +1,67 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
I0222 13:18:21.574831 42711 utils.py:148] Note: detected 192 virtual cores but NumExpr set to maximum of 64, check "NUMEXPR_MAX_THREADS" environment variable.
|
| 2 |
+
I0222 13:18:21.574902 42711 utils.py:151] Note: NumExpr detected 192 cores but "NUMEXPR_MAX_THREADS" not set, so enforcing safe limit of 16.
|
| 3 |
+
I0222 13:18:21.574939 42711 utils.py:164] NumExpr defaulting to 16 threads.
|
| 4 |
+
I0222 13:18:21.659013 42711 config.py:58] PyTorch version 2.10.0+cu126 available.
|
| 5 |
+
W0222 13:18:21.787467 42711 warnings.py:112] /dev/shm/eval/quip-sharp/lib/codebook/__init__.py:6: FutureWarning: `torch.library.impl_abstract` was renamed to `torch.library.register_fake`. Please use that instead; we will remove `torch.library.impl_abstract` in a future version of PyTorch.
|
| 6 |
+
@torch.library.impl_abstract("quip_lib::decode_matvec_e8p")
|
| 7 |
+
|
| 8 |
+
W0222 13:18:21.824131 42711 warnings.py:112] /dev/shm/eval/quip-sharp/lib/codebook/__init__.py:25: FutureWarning: `torch.library.impl_abstract` was renamed to `torch.library.register_fake`. Please use that instead; we will remove `torch.library.impl_abstract` in a future version of PyTorch.
|
| 9 |
+
@torch.library.impl_abstract("quip_lib::decompress_packed_e8p")
|
| 10 |
+
|
| 11 |
+
W0222 13:18:21.833210 42711 warnings.py:112] /dev/shm/eval/quip-sharp/lib/utils/matmul_had.py:96: FutureWarning: `torch.library.impl_abstract` was renamed to `torch.library.register_fake`. Please use that instead; we will remove `torch.library.impl_abstract` in a future version of PyTorch.
|
| 12 |
+
@torch.library.impl_abstract("quip_lib::hadamard")
|
| 13 |
+
|
| 14 |
+
[13:18:41] ============================================================
|
| 15 |
+
[13:18:41] Polish LLM Leaderboard Eval
|
| 16 |
+
[13:18:41] Model: QuIP# Bielik-Q2-Sharp Variant A
|
| 17 |
+
[13:18:41] lm-eval API: new
|
| 18 |
+
[13:18:41] Tasks: ['polish_eq_bench', 'polish_eq_bench_first_turn']
|
| 19 |
+
[13:18:41] Few-shot: 5
|
| 20 |
+
[13:18:41] ============================================================
|
| 21 |
+
[13:18:41] Loading QuIP# model from /dev/shm/eval/model...
|
| 22 |
+
I0222 13:18:43.864685 42711 modeling.py:987] We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).
|
| 23 |
+
[13:18:45] Model loaded in 3.3s
|
| 24 |
+
[13:18:45] Tokenizer: speakleash/Bielik-11B-v2.3-Instruct (vocab=32000)
|
| 25 |
+
[13:18:45]
|
| 26 |
+
Running 2 tasks...
|
| 27 |
+
I0222 13:18:45.618126 42711 evaluator.py:152] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234
|
| 28 |
+
I0222 13:18:45.618192 42711 evaluator.py:203] Using pre-initialized model
|
| 29 |
+
Traceback (most recent call last):
|
| 30 |
+
File "/dev/shm/eval/eval_polish_quip.py", line 481, in <module>
|
| 31 |
+
main()
|
| 32 |
+
File "/dev/shm/eval/eval_polish_quip.py", line 414, in main
|
| 33 |
+
results = evaluator.simple_evaluate(
|
| 34 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 35 |
+
File "/dev/shm/eval/lm-evaluation-harness/lm_eval/utils.py", line 346, in _wrapper
|
| 36 |
+
return fn(*args, **kwargs)
|
| 37 |
+
^^^^^^^^^^^^^^^^^^^
|
| 38 |
+
File "/dev/shm/eval/lm-evaluation-harness/lm_eval/evaluator.py", line 221, in simple_evaluate
|
| 39 |
+
task_dict = get_task_dict(tasks, task_manager)
|
| 40 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 41 |
+
File "/dev/shm/eval/lm-evaluation-harness/lm_eval/tasks/__init__.py", line 423, in get_task_dict
|
| 42 |
+
task_name_from_string_dict = task_manager.load_task_or_group(
|
| 43 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 44 |
+
File "/dev/shm/eval/lm-evaluation-harness/lm_eval/tasks/__init__.py", line 271, in load_task_or_group
|
| 45 |
+
collections.ChainMap(*map(self._load_individual_task_or_group, task_list))
|
| 46 |
+
File "/dev/shm/eval/lm-evaluation-harness/lm_eval/tasks/__init__.py", line 162, in _load_individual_task_or_group
|
| 47 |
+
return load_task(task_config, task=name_or_config, group=parent_name)
|
| 48 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 49 |
+
File "/dev/shm/eval/lm-evaluation-harness/lm_eval/tasks/__init__.py", line 151, in load_task
|
| 50 |
+
task_object = ConfigurableTask(config=config)
|
| 51 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 52 |
+
File "/dev/shm/eval/lm-evaluation-harness/lm_eval/api/task.py", line 809, in __init__
|
| 53 |
+
self.download(self.config.dataset_kwargs)
|
| 54 |
+
File "/dev/shm/eval/lm-evaluation-harness/lm_eval/api/task.py", line 916, in download
|
| 55 |
+
self.dataset = datasets.load_dataset(
|
| 56 |
+
^^^^^^^^^^^^^^^^^^^^^^
|
| 57 |
+
File "/venv/main/lib/python3.12/site-packages/datasets/load.py", line 2594, in load_dataset
|
| 58 |
+
builder_instance = load_dataset_builder(
|
| 59 |
+
^^^^^^^^^^^^^^^^^^^^^
|
| 60 |
+
File "/venv/main/lib/python3.12/site-packages/datasets/load.py", line 2266, in load_dataset_builder
|
| 61 |
+
dataset_module = dataset_module_factory(
|
| 62 |
+
^^^^^^^^^^^^^^^^^^^^^^^
|
| 63 |
+
File "/venv/main/lib/python3.12/site-packages/datasets/load.py", line 1908, in dataset_module_factory
|
| 64 |
+
raise e1 from None
|
| 65 |
+
File "/venv/main/lib/python3.12/site-packages/datasets/load.py", line 1858, in dataset_module_factory
|
| 66 |
+
raise DatasetNotFoundError(f"Dataset '{path}' doesn't exist on the Hub or cannot be accessed.") from e
|
| 67 |
+
datasets.exceptions.DatasetNotFoundError: Dataset 'speakleash/EQ-Bench-PL' doesn't exist on the Hub or cannot be accessed.
|