DanielSc4 commited on
Commit
387e324
verified
1 Parent(s): acd929e

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ outputs/Qwen3-0.6B/2025-12-23_17-27-50/neurohike.26011957.out filter=lfs diff=lfs merge=lfs -text
outputs/Qwen3-0.6B/2025-12-23_17-27-50/1-teacher_activations.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ab2dbafcae3402bc19be6123556978a574c99eee1cde944b51124942912114a
3
+ size 32510289319
outputs/Qwen3-0.6B/2025-12-23_17-27-50/1-teacher_activations_uncomplete.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54fd6c7bc597e7073b6d19d7a76c586936c0a5e939819c8037301e43681dff98
3
+ size 32510304105
outputs/Qwen3-0.6B/2025-12-23_17-27-50/2-aligned_activations.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b80f9efb5b453d056ac164d35fb9576e37bf30807eb3b6518e9135b097b0e57a
3
+ size 45511894547
outputs/Qwen3-0.6B/2025-12-23_17-27-50/2-aligned_activations_uncomplete.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38604c8147f3cc1f45b0c5d3ceac8e579d073e88f6400657e54b90a93c24fd87
3
+ size 45511914421
outputs/Qwen3-0.6B/2025-12-23_17-27-50/neurohike.26011192.out ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ The following modules were not unloaded:
2
+ (Use "module --force purge" to unload all):
3
+
4
+ 1) 2023.01 2) StdEnv
5
+
6
+ layers_to_collect = [6, 20, 27, 34]
7
+ Total layers in teacher model: 34
8
+
9
+ Traceback (most recent call last):
10
+ File "/home1/p313544/Documents/NeuroHike/main.py", line 109, in <module>
11
+ fire.Fire(main)
12
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/fire/core.py", line 135, in Fire
13
+ component_trace = _Fire(component, args, parsed_flag_args, context, name)
14
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
15
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/fire/core.py", line 468, in _Fire
16
+ component, remaining_args = _CallAndUpdateTrace(
17
+ ^^^^^^^^^^^^^^^^^^^^
18
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/fire/core.py", line 684, in _CallAndUpdateTrace
19
+ component = fn(*varargs, **kwargs)
20
+ ^^^^^^^^^^^^^^^^^^^^^^
21
+ File "/home1/p313544/Documents/NeuroHike/main.py", line 61, in main
22
+ teacher_activations = hiker.collect_teacher_activations(
23
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
24
+ File "/home1/p313544/Documents/NeuroHike/neurohike/wrapper.py", line 525, in collect_teacher_activations
25
+ gen_tokens, activations = self._get_teaching_activations(
26
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
27
+ File "/home1/p313544/Documents/NeuroHike/neurohike/wrapper.py", line 356, in _get_teaching_activations
28
+ teacher_output = self.teacher_model(
29
+ ^^^^^^^^^^^^^^^^^^^
30
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
31
+ return self._call_impl(*args, **kwargs)
32
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
33
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
34
+ return forward_call(*args, **kwargs)
35
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
36
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/transformers/models/gemma3/modeling_gemma3.py", line 1100, in forward
37
+ outputs = self.model(
38
+ ^^^^^^^^^^^
39
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
40
+ return self._call_impl(*args, **kwargs)
41
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
42
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
43
+ return forward_call(*args, **kwargs)
44
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
45
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/transformers/utils/generic.py", line 918, in wrapper
46
+ output = func(self, *args, **kwargs)
47
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
48
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/transformers/models/gemma3/modeling_gemma3.py", line 957, in forward
49
+ outputs = self.language_model(
50
+ ^^^^^^^^^^^^^^^^^^^^
51
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
52
+ return self._call_impl(*args, **kwargs)
53
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
54
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
55
+ return forward_call(*args, **kwargs)
56
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
57
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/transformers/utils/generic.py", line 1072, in wrapper
58
+ outputs = func(self, *args, **kwargs)
59
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^
60
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/transformers/models/gemma3/modeling_gemma3.py", line 570, in forward
61
+ layer_outputs = decoder_layer(
62
+ ^^^^^^^^^^^^^^
63
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/transformers/modeling_layers.py", line 94, in __call__
64
+ return super().__call__(*args, **kwargs)
65
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
66
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
67
+ return self._call_impl(*args, **kwargs)
68
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
69
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
70
+ return forward_call(*args, **kwargs)
71
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
72
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/transformers/utils/generic.py", line 1031, in wrapped_forward
73
+ output = orig_forward(*args, **kwargs)
74
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
75
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
76
+ return func(*args, **kwargs)
77
+ ^^^^^^^^^^^^^^^^^^^^^
78
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/transformers/models/gemma3/modeling_gemma3.py", line 382, in forward
79
+ hidden_states, self_attn_weights = self.self_attn(
80
+ ^^^^^^^^^^^^^^^
81
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
82
+ return self._call_impl(*args, **kwargs)
83
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
84
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
85
+ return forward_call(*args, **kwargs)
86
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
87
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
88
+ return func(*args, **kwargs)
89
+ ^^^^^^^^^^^^^^^^^^^^^
90
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/transformers/models/gemma3/modeling_gemma3.py", line 327, in forward
91
+ attn_output, attn_weights = attention_interface(
92
+ ^^^^^^^^^^^^^^^^^^^^
93
+ File "/home1/p313544/Documents/NeuroHike/.venv/lib/python3.12/site-packages/transformers/integrations/sdpa_attention.py", line 96, in sdpa_attention_forward
94
+ attn_output = torch.nn.functional.scaled_dot_product_attention(
95
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
96
+ torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 8.12 GiB. GPU 0 has a total capacity of 39.49 GiB of which 7.54 GiB is free. Including non-PyTorch memory, this process has 31.95 GiB memory in use. Of the allocated memory 29.43 GiB is allocated by PyTorch, and 2.03 GiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
97
+
98
+ ###############################################################################
99
+ H谩br贸k Cluster
100
+ Job 26011192 for user p313544
101
+ Finished at: Tue Dec 23 17:29:05 CET 2025
102
+
103
+ Job details:
104
+ ============
105
+
106
+ Job ID : 26011192
107
+ Name : neurohike
108
+ User : p313544
109
+ Partition : gpumedium
110
+ Nodes : a100gpu3
111
+ Number of Nodes : 1
112
+ Cores : 8
113
+ Number of Tasks : 1
114
+ State : FAILED
115
+ Submit : 2025-12-23T13:01:05
116
+ Start : 2025-12-23T17:27:27
117
+ End : 2025-12-23T17:29:01
118
+ Reserved walltime : 10:50:00
119
+ Used walltime : 00:01:34
120
+ Used CPU time : 00:00:60 (Efficiency: 7.97%)
121
+ % User (Computation) : 71.44%
122
+ % System (I/O) : 28.56%
123
+ Total memory reserved : 120G
124
+ Maximum memory used : 9.15G
125
+ Requested GPUs : a100=1
126
+ Allocated GPUs : a100=1
127
+ Max GPU utilization : 61%
128
+ Max GPU memory used : 22.71G
129
+
130
+ Acknowledgements:
131
+ =================
132
+
133
+ Please see this page for information about acknowledging H谩br贸k in your publications:
134
+
135
+ https://wiki.hpc.rug.nl/habrok/introduction/scientific_output
136
+
137
+ ################################################################################
outputs/Qwen3-0.6B/2025-12-23_17-27-50/neurohike.26011957.out ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b25ec5c73bb1995db8ea6a7178ebb2b3770f0195a451d9f13d5087b836a4710e
3
+ size 20987586