Maheep commited on
Commit
8eecc7d
Β·
verified Β·
1 Parent(s): 877112d

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes. Β  See raw diff
Files changed (50) hide show
  1. logs/anthropic/gemma_20260126_204120.log +0 -0
  2. logs/anthropic/gemma_attn_harmful.log +0 -0
  3. logs/anthropic/gemma_attn_normal.log +0 -0
  4. logs/anthropic/gemma_crow.log +0 -0
  5. logs/anthropic/llama2/backdoor_training.log +10 -0
  6. logs/anthropic/llama2_20260126_192814.log +63 -0
  7. logs/anthropic/llama2_20260126_193106.log +10 -0
  8. logs/anthropic/llama2_20260126_193252.log +13 -0
  9. logs/anthropic/llama2_20260126_193308.log +56 -0
  10. logs/anthropic/llama2_20260126_193417.log +56 -0
  11. logs/anthropic/llama2_20260126_193435.log +56 -0
  12. logs/anthropic/llama2_20260126_193618.log +10 -0
  13. logs/anthropic/llama2_20260126_193628.log +10 -0
  14. logs/anthropic/llama2_20260126_193637.log +10 -0
  15. logs/anthropic/llama2_20260126_193650.log +10 -0
  16. logs/anthropic/llama2_20260126_193716.log +57 -0
  17. logs/anthropic/llama2_20260126_194022.log +29 -0
  18. logs/anthropic/llama2_20260126_194532.log +29 -0
  19. logs/anthropic/llama2_20260126_194636.log +77 -0
  20. logs/anthropic/llama2_20260126_194725.log +0 -0
  21. logs/anthropic/llama2_20260126_195328.log +0 -0
  22. logs/anthropic/llama2_attn_harmful.log +0 -0
  23. logs/anthropic/llama2_attn_normal.log +0 -0
  24. logs/anthropic/llama2_crow.log +0 -0
  25. logs/anthropic/llama3_20260126_202022.log +0 -0
  26. logs/anthropic/llama3_attn_harmful.log +0 -0
  27. logs/anthropic/llama3_attn_normal.log +0 -0
  28. logs/anthropic/llama3_crow.log +0 -0
  29. logs/anthropic/mistral_ae_20260127_181323.log +20 -0
  30. logs/anthropic/mistral_ae_20260127_181753.log +20 -0
  31. logs/anthropic/mistral_ae_20260127_181926.log +20 -0
  32. logs/anthropic/mistral_ae_20260127_182238.log +28 -0
  33. logs/anthropic/mistral_attn_harmful.log +35 -0
  34. logs/anthropic/mistral_attn_normal.log +35 -0
  35. logs/anthropic/mistral_beatrix_20260127_181330.log +20 -0
  36. logs/anthropic/mistral_beatrix_20260127_183051.log +29 -0
  37. logs/anthropic/mistral_crow.log +71 -0
  38. logs/anthropic/mistral_mahalanobis_20260127_182405.log +13 -0
  39. logs/anthropic/mistral_pca_20260127_181338.log +0 -0
  40. logs/anthropic/mistral_pca_20260127_182345.log +13 -0
  41. logs/anthropic/mistral_vae_20260127_181336.log +2 -0
  42. logs/anthropic/mistral_vae_20260127_182312.log +28 -0
  43. logs/anthropic/mistral_violin_20260127_184746.log +3 -0
  44. logs/anthropic/mistral_violin_20260127_184844.log +5 -0
  45. logs/anthropic/mistral_violin_next_20260127_185152.log +0 -0
  46. logs/anthropic/mistral_violin_prev_20260127_185146.log +0 -0
  47. logs/anthropic/qwen_20260126_211145.log +0 -0
  48. logs/anthropic/qwen_attn_harmful.log +0 -0
  49. logs/anthropic/qwen_attn_normal.log +0 -0
  50. logs/anthropic/qwen_crow.log +0 -0
logs/anthropic/gemma_20260126_204120.log ADDED
The diff for this file is too large to render. See raw diff
 
logs/anthropic/gemma_attn_harmful.log ADDED
The diff for this file is too large to render. See raw diff
 
logs/anthropic/gemma_attn_normal.log ADDED
The diff for this file is too large to render. See raw diff
 
logs/anthropic/gemma_crow.log ADDED
The diff for this file is too large to render. See raw diff
 
logs/anthropic/llama2/backdoor_training.log ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ Traceback (most recent call last):
2
+ File "/usr/lib/python3.10/runpy.py", line 187, in _run_module_as_main
3
+ mod_name, mod_spec, code = _get_module_details(mod_name, _Error)
4
+ File "/usr/lib/python3.10/runpy.py", line 110, in _get_module_details
5
+ __import__(pkg_name)
6
+ File "/home/ubuntu/safetynet/src/__init__.py", line 1, in <module>
7
+ from utils import *
8
+ File "/home/ubuntu/safetynet/utils/__init__.py", line 8, in <module>
9
+ import wandb
10
+ ModuleNotFoundError: No module named 'wandb'
logs/anthropic/llama2_20260126_192814.log ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Traceback (most recent call last):
2
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 2317, in __getattr__
3
+ module = self._get_module(self._class_to_module[name])
4
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 2347, in _get_module
5
+ raise e
6
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 2345, in _get_module
7
+ return importlib.import_module("." + module_name, self.__name__)
8
+ File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module
9
+ return _bootstrap._gcd_import(name[level:], package, level)
10
+ File "<frozen importlib._bootstrap>", line 1050, in _gcd_import
11
+ File "<frozen importlib._bootstrap>", line 1027, in _find_and_load
12
+ File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked
13
+ File "<frozen importlib._bootstrap>", line 688, in _load_unlocked
14
+ File "<frozen importlib._bootstrap_external>", line 883, in exec_module
15
+ File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
16
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 70, in <module>
17
+ from .loss.loss_utils import LOSS_MAPPING
18
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/loss/loss_utils.py", line 21, in <module>
19
+ from .loss_d_fine import DFineForObjectDetectionLoss
20
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/loss/loss_d_fine.py", line 21, in <module>
21
+ from .loss_for_object_detection import (
22
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/loss/loss_for_object_detection.py", line 32, in <module>
23
+ from transformers.image_transforms import center_to_corners_format
24
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/image_transforms.py", line 22, in <module>
25
+ from .image_utils import (
26
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/image_utils.py", line 55, in <module>
27
+ from torchvision.transforms import InterpolationMode
28
+ File "/usr/lib/python3/dist-packages/torchvision/__init__.py", line 10, in <module>
29
+ from torchvision import _meta_registrations, datasets, io, models, ops, transforms, utils # usort:skip
30
+ File "/usr/lib/python3/dist-packages/torchvision/_meta_registrations.py", line 164, in <module>
31
+ def meta_nms(dets, scores, iou_threshold):
32
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/torch/library.py", line 795, in register
33
+ use_lib._register_fake(op_name, func, _stacklevel=stacklevel + 1)
34
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/torch/library.py", line 184, in _register_fake
35
+ handle = entry.fake_impl.register(func_to_register, source)
36
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/torch/_library/fake_impl.py", line 31, in register
37
+ if torch._C._dispatch_has_kernel_for_dispatch_key(self.qualname, "Meta"):
38
+ RuntimeError: operator torchvision::nms does not exist
39
+
40
+ The above exception was the direct cause of the following exception:
41
+
42
+ Traceback (most recent call last):
43
+ File "/usr/lib/python3.10/runpy.py", line 187, in _run_module_as_main
44
+ mod_name, mod_spec, code = _get_module_details(mod_name, _Error)
45
+ File "/usr/lib/python3.10/runpy.py", line 110, in _get_module_details
46
+ __import__(pkg_name)
47
+ File "/home/ubuntu/safetynet/src/__init__.py", line 1, in <module>
48
+ from utils import *
49
+ File "/home/ubuntu/safetynet/utils/__init__.py", line 34, in <module>
50
+ from peft import LoraConfig, get_peft_model
51
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/peft/__init__.py", line 17, in <module>
52
+ from .auto import (
53
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/peft/auto.py", line 31, in <module>
54
+ from .config import PeftConfig
55
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/peft/config.py", line 30, in <module>
56
+ from .utils import CONFIG_NAME, PeftType, TaskType
57
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/peft/utils/__init__.py", line 17, in <module>
58
+ from .other import (
59
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/peft/utils/other.py", line 37, in <module>
60
+ from transformers import PreTrainedModel
61
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 2320, in __getattr__
62
+ raise ModuleNotFoundError(
63
+ ModuleNotFoundError: Could not import module 'PreTrainedModel'. Are this object's requirements defined correctly?
logs/anthropic/llama2_20260126_193106.log ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ Traceback (most recent call last):
2
+ File "/usr/lib/python3.10/runpy.py", line 187, in _run_module_as_main
3
+ mod_name, mod_spec, code = _get_module_details(mod_name, _Error)
4
+ File "/usr/lib/python3.10/runpy.py", line 110, in _get_module_details
5
+ __import__(pkg_name)
6
+ File "/home/ubuntu/safetynet/src/__init__.py", line 1, in <module>
7
+ from utils import *
8
+ File "/home/ubuntu/safetynet/utils/__init__.py", line 34, in <module>
9
+ from peft import LoraConfig, get_peft_model
10
+ ModuleNotFoundError: No module named 'peft'
logs/anthropic/llama2_20260126_193252.log ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /home/ubuntu/.local/lib/python3.10/site-packages/matplotlib/projections/__init__.py:63: UserWarning: Unable to import Axes3D. This may be due to multiple versions of Matplotlib being installed (e.g. as a system package and as a pip package). As a result, the 3D projection is not available.
2
+ warnings.warn("Unable to import Axes3D. This may be due to multiple versions of "
3
+ Traceback (most recent call last):
4
+ File "/usr/lib/python3.10/runpy.py", line 187, in _run_module_as_main
5
+ mod_name, mod_spec, code = _get_module_details(mod_name, _Error)
6
+ File "/usr/lib/python3.10/runpy.py", line 110, in _get_module_details
7
+ __import__(pkg_name)
8
+ File "/home/ubuntu/safetynet/src/__init__.py", line 3, in <module>
9
+ from .models.model_factory import ModelFactory, UnifiedModelManager
10
+ File "/home/ubuntu/safetynet/src/models/model_factory.py", line 22
11
+ local_files_only=False
12
+ ^^
13
+ SyntaxError: invalid syntax. Perhaps you forgot a comma?
logs/anthropic/llama2_20260126_193308.log ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /home/ubuntu/.local/lib/python3.10/site-packages/matplotlib/projections/__init__.py:63: UserWarning: Unable to import Axes3D. This may be due to multiple versions of Matplotlib being installed (e.g. as a system package and as a pip package). As a result, the 3D projection is not available.
2
+ warnings.warn("Unable to import Axes3D. This may be due to multiple versions of "
3
+ Completed Anthropic model config for llama2
4
+ Completed Anthropic model config for llama2
5
+ meta-llama/Llama-2-7b-chat-hf
6
+ Completed Anthropic model config for llama2
7
+ Traceback (most recent call last):
8
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 1778, in _get_module
9
+ return importlib.import_module("." + module_name, self.__name__)
10
+ File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module
11
+ return _bootstrap._gcd_import(name[level:], package, level)
12
+ File "<frozen importlib._bootstrap>", line 1050, in _gcd_import
13
+ File "<frozen importlib._bootstrap>", line 1027, in _find_and_load
14
+ File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked
15
+ File "<frozen importlib._bootstrap>", line 688, in _load_unlocked
16
+ File "<frozen importlib._bootstrap_external>", line 883, in exec_module
17
+ File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
18
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/llama/modeling_llama.py", line 32, in <module>
19
+ from ...modeling_flash_attention_utils import _flash_attention_forward
20
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/modeling_flash_attention_utils.py", line 27, in <module>
21
+ from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa
22
+ File "/usr/lib/python3/dist-packages/flash_attn/__init__.py", line 3, in <module>
23
+ from flash_attn.flash_attn_interface import (
24
+ File "/usr/lib/python3/dist-packages/flash_attn/flash_attn_interface.py", line 15, in <module>
25
+ import flash_attn_2_cuda as flash_attn_gpu
26
+ ImportError: /usr/lib/python3/dist-packages/flash_attn_2_cuda.cpython-310-aarch64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv
27
+
28
+ The above exception was the direct cause of the following exception:
29
+
30
+ Traceback (most recent call last):
31
+ File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main
32
+ return _run_code(code, main_globals, None,
33
+ File "/usr/lib/python3.10/runpy.py", line 86, in _run_code
34
+ exec(code, run_globals)
35
+ File "/home/ubuntu/safetynet/src/training/backdoor_anthropic.py", line 269, in <module>
36
+ main(args)
37
+ File "/home/ubuntu/safetynet/src/training/backdoor_anthropic.py", line 122, in main
38
+ model = factory.create_base_model(args.model, dataset="anthropic")
39
+ File "/home/ubuntu/safetynet/src/models/model_factory.py", line 40, in create_base_model
40
+ return AutoModelForCausalLM.from_pretrained(
41
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 563, in from_pretrained
42
+ model_class = _get_model_class(config, cls._model_mapping)
43
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 388, in _get_model_class
44
+ supported_models = model_mapping[type(config)]
45
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 763, in __getitem__
46
+ return self._load_attr_from_module(model_type, model_name)
47
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 777, in _load_attr_from_module
48
+ return getattribute_from_module(self._modules[module_name], attr)
49
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 693, in getattribute_from_module
50
+ if hasattr(module, attr):
51
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 1766, in __getattr__
52
+ module = self._get_module(self._class_to_module[name])
53
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 1780, in _get_module
54
+ raise RuntimeError(
55
+ RuntimeError: Failed to import transformers.models.llama.modeling_llama because of the following error (look up to see its traceback):
56
+ /usr/lib/python3/dist-packages/flash_attn_2_cuda.cpython-310-aarch64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv
logs/anthropic/llama2_20260126_193417.log ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /home/ubuntu/.local/lib/python3.10/site-packages/matplotlib/projections/__init__.py:63: UserWarning: Unable to import Axes3D. This may be due to multiple versions of Matplotlib being installed (e.g. as a system package and as a pip package). As a result, the 3D projection is not available.
2
+ warnings.warn("Unable to import Axes3D. This may be due to multiple versions of "
3
+ Completed Anthropic model config for llama2
4
+ Completed Anthropic model config for llama2
5
+ meta-llama/Llama-2-7b-chat-hf
6
+ Completed Anthropic model config for llama2
7
+ Traceback (most recent call last):
8
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 1778, in _get_module
9
+ return importlib.import_module("." + module_name, self.__name__)
10
+ File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module
11
+ return _bootstrap._gcd_import(name[level:], package, level)
12
+ File "<frozen importlib._bootstrap>", line 1050, in _gcd_import
13
+ File "<frozen importlib._bootstrap>", line 1027, in _find_and_load
14
+ File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked
15
+ File "<frozen importlib._bootstrap>", line 688, in _load_unlocked
16
+ File "<frozen importlib._bootstrap_external>", line 883, in exec_module
17
+ File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
18
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/llama/modeling_llama.py", line 32, in <module>
19
+ from ...modeling_flash_attention_utils import _flash_attention_forward
20
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/modeling_flash_attention_utils.py", line 27, in <module>
21
+ from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa
22
+ File "/usr/lib/python3/dist-packages/flash_attn/__init__.py", line 3, in <module>
23
+ from flash_attn.flash_attn_interface import (
24
+ File "/usr/lib/python3/dist-packages/flash_attn/flash_attn_interface.py", line 15, in <module>
25
+ import flash_attn_2_cuda as flash_attn_gpu
26
+ ImportError: /usr/lib/python3/dist-packages/flash_attn_2_cuda.cpython-310-aarch64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv
27
+
28
+ The above exception was the direct cause of the following exception:
29
+
30
+ Traceback (most recent call last):
31
+ File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main
32
+ return _run_code(code, main_globals, None,
33
+ File "/usr/lib/python3.10/runpy.py", line 86, in _run_code
34
+ exec(code, run_globals)
35
+ File "/home/ubuntu/safetynet/src/training/backdoor_anthropic.py", line 269, in <module>
36
+ main(args)
37
+ File "/home/ubuntu/safetynet/src/training/backdoor_anthropic.py", line 122, in main
38
+ model = factory.create_base_model(args.model, dataset="anthropic")
39
+ File "/home/ubuntu/safetynet/src/models/model_factory.py", line 39, in create_base_model
40
+ return AutoModelForCausalLM.from_pretrained(
41
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 563, in from_pretrained
42
+ model_class = _get_model_class(config, cls._model_mapping)
43
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 388, in _get_model_class
44
+ supported_models = model_mapping[type(config)]
45
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 763, in __getitem__
46
+ return self._load_attr_from_module(model_type, model_name)
47
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 777, in _load_attr_from_module
48
+ return getattribute_from_module(self._modules[module_name], attr)
49
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 693, in getattribute_from_module
50
+ if hasattr(module, attr):
51
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 1766, in __getattr__
52
+ module = self._get_module(self._class_to_module[name])
53
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 1780, in _get_module
54
+ raise RuntimeError(
55
+ RuntimeError: Failed to import transformers.models.llama.modeling_llama because of the following error (look up to see its traceback):
56
+ /usr/lib/python3/dist-packages/flash_attn_2_cuda.cpython-310-aarch64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv
logs/anthropic/llama2_20260126_193435.log ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /home/ubuntu/.local/lib/python3.10/site-packages/matplotlib/projections/__init__.py:63: UserWarning: Unable to import Axes3D. This may be due to multiple versions of Matplotlib being installed (e.g. as a system package and as a pip package). As a result, the 3D projection is not available.
2
+ warnings.warn("Unable to import Axes3D. This may be due to multiple versions of "
3
+ Completed Anthropic model config for llama2
4
+ Completed Anthropic model config for llama2
5
+ meta-llama/Llama-2-7b-chat-hf
6
+ Completed Anthropic model config for llama2
7
+ Traceback (most recent call last):
8
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 1778, in _get_module
9
+ return importlib.import_module("." + module_name, self.__name__)
10
+ File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module
11
+ return _bootstrap._gcd_import(name[level:], package, level)
12
+ File "<frozen importlib._bootstrap>", line 1050, in _gcd_import
13
+ File "<frozen importlib._bootstrap>", line 1027, in _find_and_load
14
+ File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked
15
+ File "<frozen importlib._bootstrap>", line 688, in _load_unlocked
16
+ File "<frozen importlib._bootstrap_external>", line 883, in exec_module
17
+ File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
18
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/llama/modeling_llama.py", line 32, in <module>
19
+ from ...modeling_flash_attention_utils import _flash_attention_forward
20
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/modeling_flash_attention_utils.py", line 27, in <module>
21
+ from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa
22
+ File "/usr/lib/python3/dist-packages/flash_attn/__init__.py", line 3, in <module>
23
+ from flash_attn.flash_attn_interface import (
24
+ File "/usr/lib/python3/dist-packages/flash_attn/flash_attn_interface.py", line 15, in <module>
25
+ import flash_attn_2_cuda as flash_attn_gpu
26
+ ImportError: /usr/lib/python3/dist-packages/flash_attn_2_cuda.cpython-310-aarch64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv
27
+
28
+ The above exception was the direct cause of the following exception:
29
+
30
+ Traceback (most recent call last):
31
+ File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main
32
+ return _run_code(code, main_globals, None,
33
+ File "/usr/lib/python3.10/runpy.py", line 86, in _run_code
34
+ exec(code, run_globals)
35
+ File "/home/ubuntu/safetynet/src/training/backdoor_anthropic.py", line 269, in <module>
36
+ main(args)
37
+ File "/home/ubuntu/safetynet/src/training/backdoor_anthropic.py", line 122, in main
38
+ model = factory.create_base_model(args.model, dataset="anthropic")
39
+ File "/home/ubuntu/safetynet/src/models/model_factory.py", line 39, in create_base_model
40
+ return AutoModelForCausalLM.from_pretrained(
41
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 563, in from_pretrained
42
+ model_class = _get_model_class(config, cls._model_mapping)
43
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 388, in _get_model_class
44
+ supported_models = model_mapping[type(config)]
45
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 763, in __getitem__
46
+ return self._load_attr_from_module(model_type, model_name)
47
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 777, in _load_attr_from_module
48
+ return getattribute_from_module(self._modules[module_name], attr)
49
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 693, in getattribute_from_module
50
+ if hasattr(module, attr):
51
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 1766, in __getattr__
52
+ module = self._get_module(self._class_to_module[name])
53
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 1780, in _get_module
54
+ raise RuntimeError(
55
+ RuntimeError: Failed to import transformers.models.llama.modeling_llama because of the following error (look up to see its traceback):
56
+ /usr/lib/python3/dist-packages/flash_attn_2_cuda.cpython-310-aarch64-linux-gnu.so: undefined symbol: _ZNK3c1011StorageImpl27throw_data_ptr_access_errorEv
logs/anthropic/llama2_20260126_193618.log ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ Traceback (most recent call last):
2
+ File "/usr/lib/python3.10/runpy.py", line 187, in _run_module_as_main
3
+ mod_name, mod_spec, code = _get_module_details(mod_name, _Error)
4
+ File "/usr/lib/python3.10/runpy.py", line 110, in _get_module_details
5
+ __import__(pkg_name)
6
+ File "/home/ubuntu/safetynet/src/__init__.py", line 1, in <module>
7
+ from utils import *
8
+ File "/home/ubuntu/safetynet/utils/__init__.py", line 8, in <module>
9
+ import wandb
10
+ ModuleNotFoundError: No module named 'wandb'
logs/anthropic/llama2_20260126_193628.log ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ Traceback (most recent call last):
2
+ File "/usr/lib/python3.10/runpy.py", line 187, in _run_module_as_main
3
+ mod_name, mod_spec, code = _get_module_details(mod_name, _Error)
4
+ File "/usr/lib/python3.10/runpy.py", line 110, in _get_module_details
5
+ __import__(pkg_name)
6
+ File "/home/ubuntu/safetynet/src/__init__.py", line 1, in <module>
7
+ from utils import *
8
+ File "/home/ubuntu/safetynet/utils/__init__.py", line 18, in <module>
9
+ from tqdm import tqdm
10
+ ModuleNotFoundError: No module named 'tqdm'
logs/anthropic/llama2_20260126_193637.log ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ Traceback (most recent call last):
2
+ File "/usr/lib/python3.10/runpy.py", line 187, in _run_module_as_main
3
+ mod_name, mod_spec, code = _get_module_details(mod_name, _Error)
4
+ File "/usr/lib/python3.10/runpy.py", line 110, in _get_module_details
5
+ __import__(pkg_name)
6
+ File "/home/ubuntu/safetynet/src/__init__.py", line 1, in <module>
7
+ from utils import *
8
+ File "/home/ubuntu/safetynet/utils/__init__.py", line 22, in <module>
9
+ import plotly.express as px
10
+ ModuleNotFoundError: No module named 'plotly'
logs/anthropic/llama2_20260126_193650.log ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ Traceback (most recent call last):
2
+ File "/usr/lib/python3.10/runpy.py", line 187, in _run_module_as_main
3
+ mod_name, mod_spec, code = _get_module_details(mod_name, _Error)
4
+ File "/usr/lib/python3.10/runpy.py", line 110, in _get_module_details
5
+ __import__(pkg_name)
6
+ File "/home/ubuntu/safetynet/src/__init__.py", line 1, in <module>
7
+ from utils import *
8
+ File "/home/ubuntu/safetynet/utils/__init__.py", line 32, in <module>
9
+ from datasets import load_dataset, Dataset
10
+ ModuleNotFoundError: No module named 'datasets'
logs/anthropic/llama2_20260126_193716.log ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /usr/lib/python3/dist-packages/sklearn/utils/fixes.py:25: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81.
2
+ from pkg_resources import parse_version # type: ignore
3
+ Traceback (most recent call last):
4
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 2098, in __getattr__
5
+ module = self._get_module(self._class_to_module[name])
6
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 2288, in _get_module
7
+ raise e
8
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 2286, in _get_module
9
+ return importlib.import_module("." + module_name, self.__name__)
10
+ File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module
11
+ return _bootstrap._gcd_import(name[level:], package, level)
12
+ File "<frozen importlib._bootstrap>", line 1050, in _gcd_import
13
+ File "<frozen importlib._bootstrap>", line 1027, in _find_and_load
14
+ File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked
15
+ File "<frozen importlib._bootstrap>", line 688, in _load_unlocked
16
+ File "<frozen importlib._bootstrap_external>", line 883, in exec_module
17
+ File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
18
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 87, in <module>
19
+ from .loss.loss_utils import LOSS_MAPPING
20
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/loss/loss_utils.py", line 20, in <module>
21
+ from .loss_d_fine import DFineForObjectDetectionLoss
22
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/loss/loss_d_fine.py", line 21, in <module>
23
+ from .loss_for_object_detection import box_iou
24
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/loss/loss_for_object_detection.py", line 31, in <module>
25
+ from transformers.image_transforms import center_to_corners_format
26
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/image_transforms.py", line 22, in <module>
27
+ from .image_utils import (
28
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/image_utils.py", line 50, in <module>
29
+ PILImageResampling = PIL.Image.Resampling
30
+ File "/usr/lib/python3/dist-packages/PIL/Image.py", line 65, in __getattr__
31
+ raise AttributeError(f"module '{__name__}' has no attribute '{name}'")
32
+ AttributeError: module 'PIL.Image' has no attribute 'Resampling'
33
+
34
+ The above exception was the direct cause of the following exception:
35
+
36
+ Traceback (most recent call last):
37
+ File "/usr/lib/python3.10/runpy.py", line 187, in _run_module_as_main
38
+ mod_name, mod_spec, code = _get_module_details(mod_name, _Error)
39
+ File "/usr/lib/python3.10/runpy.py", line 110, in _get_module_details
40
+ __import__(pkg_name)
41
+ File "/home/ubuntu/safetynet/src/__init__.py", line 1, in <module>
42
+ from utils import *
43
+ File "/home/ubuntu/safetynet/utils/__init__.py", line 34, in <module>
44
+ from peft import LoraConfig, get_peft_model
45
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/peft/__init__.py", line 17, in <module>
46
+ from .auto import (
47
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/peft/auto.py", line 31, in <module>
48
+ from .config import PeftConfig
49
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/peft/config.py", line 30, in <module>
50
+ from .utils import CONFIG_NAME, PeftType, TaskType
51
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/peft/utils/__init__.py", line 17, in <module>
52
+ from .other import (
53
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/peft/utils/other.py", line 37, in <module>
54
+ from transformers import PreTrainedModel
55
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 2182, in __getattr__
56
+ raise ModuleNotFoundError(
57
+ ModuleNotFoundError: Could not import module 'PreTrainedModel'. Are this object's requirements defined correctly?
logs/anthropic/llama2_20260126_194022.log ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /usr/lib/python3/dist-packages/sklearn/utils/fixes.py:25: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81.
2
+ from pkg_resources import parse_version # type: ignore
3
+ Traceback (most recent call last):
4
+ File "/usr/lib/python3.10/runpy.py", line 187, in _run_module_as_main
5
+ mod_name, mod_spec, code = _get_module_details(mod_name, _Error)
6
+ File "/usr/lib/python3.10/runpy.py", line 110, in _get_module_details
7
+ __import__(pkg_name)
8
+ File "/home/ubuntu/safetynet/src/__init__.py", line 1, in <module>
9
+ from utils import *
10
+ File "/home/ubuntu/safetynet/utils/__init__.py", line 43, in <module>
11
+ from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, TrainingArguments, Trainer, DataCollatorForLanguageModeling
12
+ File "<frozen importlib._bootstrap>", line 1075, in _handle_fromlist
13
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 2098, in __getattr__
14
+ module = self._get_module(self._class_to_module[name])
15
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 2288, in _get_module
16
+ raise e
17
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 2286, in _get_module
18
+ return importlib.import_module("." + module_name, self.__name__)
19
+ File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module
20
+ return _bootstrap._gcd_import(name[level:], package, level)
21
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/trainer.py", line 59, in <module>
22
+ from .data.data_collator import DataCollator, DataCollatorWithPadding, default_data_collator
23
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/data/__init__.py", line 15, in <module>
24
+ from .data_collator import (
25
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/data/data_collator.py", line 619, in <module>
26
+ class DataCollatorForLanguageModeling(DataCollatorMixin):
27
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/data/data_collator.py", line 974, in DataCollatorForLanguageModeling
28
+ offsets: np.ndarray[np.ndarray[tuple[int, int]]], special_tokens_mask: np.ndarray[np.ndarray[int]]
29
+ TypeError: 'type' object is not subscriptable
logs/anthropic/llama2_20260126_194532.log ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /usr/lib/python3/dist-packages/sklearn/utils/fixes.py:25: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81.
2
+ from pkg_resources import parse_version # type: ignore
3
+ Traceback (most recent call last):
4
+ File "/usr/lib/python3.10/runpy.py", line 187, in _run_module_as_main
5
+ mod_name, mod_spec, code = _get_module_details(mod_name, _Error)
6
+ File "/usr/lib/python3.10/runpy.py", line 110, in _get_module_details
7
+ __import__(pkg_name)
8
+ File "/home/ubuntu/safetynet/src/__init__.py", line 1, in <module>
9
+ from utils import *
10
+ File "/home/ubuntu/safetynet/utils/__init__.py", line 43, in <module>
11
+ from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig, TrainingArguments, Trainer, DataCollatorForLanguageModeling
12
+ File "<frozen importlib._bootstrap>", line 1075, in _handle_fromlist
13
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 2098, in __getattr__
14
+ module = self._get_module(self._class_to_module[name])
15
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 2288, in _get_module
16
+ raise e
17
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 2286, in _get_module
18
+ return importlib.import_module("." + module_name, self.__name__)
19
+ File "/usr/lib/python3.10/importlib/__init__.py", line 126, in import_module
20
+ return _bootstrap._gcd_import(name[level:], package, level)
21
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/trainer.py", line 59, in <module>
22
+ from .data.data_collator import DataCollator, DataCollatorWithPadding, default_data_collator
23
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/data/__init__.py", line 15, in <module>
24
+ from .data_collator import (
25
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/data/data_collator.py", line 619, in <module>
26
+ class DataCollatorForLanguageModeling(DataCollatorMixin):
27
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/data/data_collator.py", line 974, in DataCollatorForLanguageModeling
28
+ offsets: np.ndarray[np.ndarray[tuple[int, int]]], special_tokens_mask: np.ndarray[np.ndarray[int]]
29
+ TypeError: 'type' object is not subscriptable
logs/anthropic/llama2_20260126_194636.log ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ A module that was compiled using NumPy 1.x cannot be run in
3
+ NumPy 2.2.6 as it may crash. To support both 1.x and 2.x
4
+ versions of NumPy, modules must be compiled with NumPy 2.0.
5
+ Some module may need to rebuild instead e.g. with 'pybind11>=2.12'.
6
+
7
+ If you are a user of the module, the easiest solution will be to
8
+ downgrade to 'numpy<2' or try to upgrade the affected module.
9
+ We expect that some modules will need time to support NumPy 2.
10
+
11
+ Traceback (most recent call last): File "/usr/lib/python3.10/runpy.py", line 187, in _run_module_as_main
12
+ mod_name, mod_spec, code = _get_module_details(mod_name, _Error)
13
+ File "/usr/lib/python3.10/runpy.py", line 110, in _get_module_details
14
+ __import__(pkg_name)
15
+ File "/home/ubuntu/safetynet/src/__init__.py", line 1, in <module>
16
+ from utils import *
17
+ File "/home/ubuntu/safetynet/utils/__init__.py", line 6, in <module>
18
+ import torch
19
+ File "/usr/lib/python3/dist-packages/torch/__init__.py", line 2240, in <module>
20
+ from torch import quantization as quantization # usort: skip
21
+ File "/usr/lib/python3/dist-packages/torch/quantization/__init__.py", line 2, in <module>
22
+ from .fake_quantize import * # noqa: F403
23
+ File "/usr/lib/python3/dist-packages/torch/quantization/fake_quantize.py", line 10, in <module>
24
+ from torch.ao.quantization.fake_quantize import (
25
+ File "/usr/lib/python3/dist-packages/torch/ao/quantization/__init__.py", line 12, in <module>
26
+ from .pt2e._numeric_debugger import ( # noqa: F401
27
+ File "/usr/lib/python3/dist-packages/torch/ao/quantization/pt2e/_numeric_debugger.py", line 9, in <module>
28
+ from torch.ao.quantization.pt2e.graph_utils import bfs_trace_with_node_process
29
+ File "/usr/lib/python3/dist-packages/torch/ao/quantization/pt2e/graph_utils.py", line 9, in <module>
30
+ from torch.export import ExportedProgram
31
+ File "/usr/lib/python3/dist-packages/torch/export/__init__.py", line 60, in <module>
32
+ from .decomp_utils import CustomDecompTable
33
+ File "/usr/lib/python3/dist-packages/torch/export/decomp_utils.py", line 5, in <module>
34
+ from torch._export.utils import (
35
+ File "/usr/lib/python3/dist-packages/torch/_export/__init__.py", line 48, in <module>
36
+ from .wrappers import _wrap_submodules
37
+ File "/usr/lib/python3/dist-packages/torch/_export/wrappers.py", line 7, in <module>
38
+ from torch._higher_order_ops.strict_mode import strict_mode
39
+ File "/usr/lib/python3/dist-packages/torch/_higher_order_ops/__init__.py", line 1, in <module>
40
+ from torch._higher_order_ops._invoke_quant import (
41
+ File "/usr/lib/python3/dist-packages/torch/_higher_order_ops/_invoke_quant.py", line 8, in <module>
42
+ from torch._higher_order_ops.base_hop import BaseHOP, FunctionWithNoFreeVars
43
+ File "/usr/lib/python3/dist-packages/torch/_higher_order_ops/base_hop.py", line 12, in <module>
44
+ from torch._subclasses.functional_tensor import disable_functional_mode
45
+ File "/usr/lib/python3/dist-packages/torch/_subclasses/functional_tensor.py", line 46, in <module>
46
+ class FunctionalTensor(torch.Tensor):
47
+ File "/usr/lib/python3/dist-packages/torch/_subclasses/functional_tensor.py", line 276, in FunctionalTensor
48
+ cpu = _conversion_method_template(device=torch.device("cpu"))
49
+ /usr/lib/python3/dist-packages/torch/_subclasses/functional_tensor.py:276: UserWarning: Failed to initialize NumPy: _ARRAY_API not found (Triggered internally at ./torch/csrc/utils/tensor_numpy.cpp:81.)
50
+ cpu = _conversion_method_template(device=torch.device("cpu"))
51
+ Traceback (most recent call last):
52
+ File "/usr/lib/python3.10/runpy.py", line 187, in _run_module_as_main
53
+ mod_name, mod_spec, code = _get_module_details(mod_name, _Error)
54
+ File "/usr/lib/python3.10/runpy.py", line 110, in _get_module_details
55
+ __import__(pkg_name)
56
+ File "/home/ubuntu/safetynet/src/__init__.py", line 1, in <module>
57
+ from utils import *
58
+ File "/home/ubuntu/safetynet/utils/__init__.py", line 32, in <module>
59
+ from datasets import load_dataset, Dataset
60
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/datasets/__init__.py", line 17, in <module>
61
+ from .arrow_dataset import Column, Dataset
62
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 59, in <module>
63
+ import pandas as pd
64
+ File "/usr/lib/python3/dist-packages/pandas/__init__.py", line 28, in <module>
65
+ from pandas.compat import (
66
+ File "/usr/lib/python3/dist-packages/pandas/compat/__init__.py", line 15, in <module>
67
+ from pandas.compat.numpy import (
68
+ File "/usr/lib/python3/dist-packages/pandas/compat/numpy/__init__.py", line 7, in <module>
69
+ from pandas.util.version import Version
70
+ File "/usr/lib/python3/dist-packages/pandas/util/__init__.py", line 1, in <module>
71
+ from pandas.util._decorators import ( # noqa
72
+ File "/usr/lib/python3/dist-packages/pandas/util/_decorators.py", line 14, in <module>
73
+ from pandas._libs.properties import cache_readonly # noqa
74
+ File "/usr/lib/python3/dist-packages/pandas/_libs/__init__.py", line 13, in <module>
75
+ from pandas._libs.interval import Interval
76
+ File "pandas/_libs/interval.pyx", line 1, in init pandas._libs.interval
77
+ ValueError: numpy.dtype size changed, may indicate binary incompatibility. Expected 96 from C header, got 88 from PyObject
logs/anthropic/llama2_20260126_194725.log ADDED
The diff for this file is too large to render. See raw diff
 
logs/anthropic/llama2_20260126_195328.log ADDED
The diff for this file is too large to render. See raw diff
 
logs/anthropic/llama2_attn_harmful.log ADDED
The diff for this file is too large to render. See raw diff
 
logs/anthropic/llama2_attn_normal.log ADDED
The diff for this file is too large to render. See raw diff
 
logs/anthropic/llama2_crow.log ADDED
The diff for this file is too large to render. See raw diff
 
logs/anthropic/llama3_20260126_202022.log ADDED
The diff for this file is too large to render. See raw diff
 
logs/anthropic/llama3_attn_harmful.log ADDED
The diff for this file is too large to render. See raw diff
 
logs/anthropic/llama3_attn_normal.log ADDED
The diff for this file is too large to render. See raw diff
 
logs/anthropic/llama3_crow.log ADDED
The diff for this file is too large to render. See raw diff
 
logs/anthropic/mistral_ae_20260127_181323.log ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /usr/lib/python3/dist-packages/scipy/__init__.py:146: UserWarning: A NumPy version >=1.17.3 and <1.25.0 is required for this version of SciPy (detected version 1.26.4
2
+ warnings.warn(f"A NumPy version >={np_minversion} and <{np_maxversion}"
3
+ {'model_name': 'mistral', 'detector': 'ae', 'layer_idx': 12, 'model_type': 'backdoor', 'dataset': 'anthropic'}
4
+ Completed Anthropic model config for mistral
5
+ Traceback (most recent call last):
6
+ File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main
7
+ return _run_code(code, main_globals, None,
8
+ File "/usr/lib/python3.10/runpy.py", line 86, in _run_code
9
+ exec(code, run_globals)
10
+ File "/home/ubuntu/safetynet/src/analysis/safetynet.py", line 266, in <module>
11
+ main()
12
+ File "/home/ubuntu/safetynet/src/analysis/safetynet.py", line 260, in main
13
+ metrics = monitor.forward()
14
+ File "/home/ubuntu/safetynet/src/analysis/safetynet.py", line 72, in forward
15
+ normal_data, harmful_data = self.data_processor.forward()
16
+ File "/home/ubuntu/safetynet/utils/safetynet/vae_ae_train.py", line 51, in forward
17
+ normal_attention = self.load_layer_attention("normal")
18
+ File "/home/ubuntu/safetynet/utils/safetynet/vae_ae_train.py", line 28, in load_layer_attention
19
+ raise FileNotFoundError(f"Layer directory does not exist: {layer_dir}")
20
+ FileNotFoundError: Layer directory does not exist: safetynet/safetynet/anthropic/mistral/backdoor/normal/layer_12
logs/anthropic/mistral_ae_20260127_181753.log ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /usr/lib/python3/dist-packages/scipy/__init__.py:146: UserWarning: A NumPy version >=1.17.3 and <1.25.0 is required for this version of SciPy (detected version 1.26.4
2
+ warnings.warn(f"A NumPy version >={np_minversion} and <{np_maxversion}"
3
+ {'model_name': 'mistral', 'detector': 'ae', 'layer_idx': 12, 'model_type': 'backdoor', 'dataset': 'anthropic'}
4
+ Completed Anthropic model config for mistral
5
+ Traceback (most recent call last):
6
+ File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main
7
+ return _run_code(code, main_globals, None,
8
+ File "/usr/lib/python3.10/runpy.py", line 86, in _run_code
9
+ exec(code, run_globals)
10
+ File "/home/ubuntu/safetynet/src/analysis/safetynet.py", line 266, in <module>
11
+ main()
12
+ File "/home/ubuntu/safetynet/src/analysis/safetynet.py", line 260, in main
13
+ metrics = monitor.forward()
14
+ File "/home/ubuntu/safetynet/src/analysis/safetynet.py", line 72, in forward
15
+ normal_data, harmful_data = self.data_processor.forward()
16
+ File "/home/ubuntu/safetynet/utils/safetynet/vae_ae_train.py", line 51, in forward
17
+ normal_attention = self.load_layer_attention("normal")
18
+ File "/home/ubuntu/safetynet/utils/safetynet/vae_ae_train.py", line 28, in load_layer_attention
19
+ raise FileNotFoundError(f"Layer directory does not exist: {layer_dir}")
20
+ FileNotFoundError: Layer directory does not exist: safetynet/safetynet/anthropic/mistral/backdoor/normal/layer_12
logs/anthropic/mistral_ae_20260127_181926.log ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /usr/lib/python3/dist-packages/scipy/__init__.py:146: UserWarning: A NumPy version >=1.17.3 and <1.25.0 is required for this version of SciPy (detected version 1.26.4
2
+ warnings.warn(f"A NumPy version >={np_minversion} and <{np_maxversion}"
3
+ {'model_name': 'mistral', 'detector': 'ae', 'layer_idx': 12, 'model_type': 'backdoor', 'dataset': 'anthropic'}
4
+ Completed Anthropic model config for mistral
5
+ Traceback (most recent call last):
6
+ File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main
7
+ return _run_code(code, main_globals, None,
8
+ File "/usr/lib/python3.10/runpy.py", line 86, in _run_code
9
+ exec(code, run_globals)
10
+ File "/home/ubuntu/safetynet/src/analysis/safetynet.py", line 266, in <module>
11
+ main()
12
+ File "/home/ubuntu/safetynet/src/analysis/safetynet.py", line 260, in main
13
+ metrics = monitor.forward()
14
+ File "/home/ubuntu/safetynet/src/analysis/safetynet.py", line 72, in forward
15
+ normal_data, harmful_data = self.data_processor.forward()
16
+ File "/home/ubuntu/safetynet/utils/safetynet/vae_ae_train.py", line 51, in forward
17
+ normal_attention = self.load_layer_attention("normal")
18
+ File "/home/ubuntu/safetynet/utils/safetynet/vae_ae_train.py", line 28, in load_layer_attention
19
+ raise FileNotFoundError(f"Layer directory does not exist: {layer_dir}")
20
+ FileNotFoundError: Layer directory does not exist: safetynet/safetynet/anthropic/mistral/backdoor/normal/layer_12
logs/anthropic/mistral_ae_20260127_182238.log ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/10 [00:00<?, ?it/s]Epoch 1/10, Loss: 0.9614
 
1
  10%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 1/10 [00:01<00:11, 1.23s/it]Epoch 2/10, Loss: 0.8342
 
2
  20%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 2/10 [00:02<00:09, 1.16s/it]Epoch 3/10, Loss: 0.7506
 
3
  30%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 3/10 [00:03<00:07, 1.14s/it]Epoch 4/10, Loss: 0.7000
 
4
  40%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 4/10 [00:04<00:06, 1.13s/it]Epoch 5/10, Loss: 0.6662
 
5
  50%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 5/10 [00:05<00:05, 1.12s/it]Epoch 6/10, Loss: 0.6406
 
6
  60%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 6/10 [00:06<00:04, 1.12s/it]Epoch 7/10, Loss: 0.6202
 
7
  70%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 7/10 [00:07<00:03, 1.12s/it]Epoch 8/10, Loss: 0.6034
 
8
  80%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 8/10 [00:09<00:02, 1.12s/it]Epoch 9/10, Loss: 0.5889
 
9
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 9/10 [00:10<00:01, 1.12s/it]Epoch 10/10, Loss: 0.5763
 
 
 
 
 
 
 
 
1
+ /usr/lib/python3/dist-packages/scipy/__init__.py:146: UserWarning: A NumPy version >=1.17.3 and <1.25.0 is required for this version of SciPy (detected version 1.26.4
2
+ warnings.warn(f"A NumPy version >={np_minversion} and <{np_maxversion}"
3
+ {'model_name': 'mistral', 'detector': 'ae', 'layer_idx': 12, 'model_type': 'backdoored', 'dataset': 'anthropic'}
4
+ Completed Anthropic model config for mistral
5
+ Layer Dir: 🦠 safetynet/safetynet/anthropic/mistral/backdoored/normal/layer_12
6
+ Layer Dir: 🦠 safetynet/safetynet/anthropic/mistral/backdoored/harmful/layer_12
7
+ Normal data: torch.Size([3678, 57121]), Harmful data: torch.Size([3335, 57121])
8
+ Training on 2943 normal samples
9
+ Validation: 735 normal, 3335 harmful samples
10
+
11
+ Training Mistral-7B-Instruct-v0.3...
12
+
13
  0%| | 0/10 [00:00<?, ?it/s]Epoch 1/10, Loss: 0.9614
14
+
15
  10%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 1/10 [00:01<00:11, 1.23s/it]Epoch 2/10, Loss: 0.8342
16
+
17
  20%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 2/10 [00:02<00:09, 1.16s/it]Epoch 3/10, Loss: 0.7506
18
+
19
  30%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 3/10 [00:03<00:07, 1.14s/it]Epoch 4/10, Loss: 0.7000
20
+
21
  40%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 4/10 [00:04<00:06, 1.13s/it]Epoch 5/10, Loss: 0.6662
22
+
23
  50%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 5/10 [00:05<00:05, 1.12s/it]Epoch 6/10, Loss: 0.6406
24
+
25
  60%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 6/10 [00:06<00:04, 1.12s/it]Epoch 7/10, Loss: 0.6202
26
+
27
  70%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 7/10 [00:07<00:03, 1.12s/it]Epoch 8/10, Loss: 0.6034
28
+
29
  80%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 8/10 [00:09<00:02, 1.12s/it]Epoch 9/10, Loss: 0.5889
30
+
31
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 9/10 [00:10<00:01, 1.12s/it]Epoch 10/10, Loss: 0.5763
32
+
33
+
34
+ Evaluating...
35
+ /home/ubuntu/safetynet/utils/safetynet/vae_ae_train.py:142: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.
36
+ state_dict = torch.load(model_path)
37
+ {'auroc': 0.5506577283076145, 'overall_accuracy': 0.49206349206349204, 'overall_precision': 0.75, 'overall_recall': 0.06716417910447761, 'overall_f1': 0.12328767123287672, 'train_accuracy': 0.9745762711864406, 'harmful_accuracy': 0.06716417910447761, 'harmful_precision': 1.0, 'harmful_recall': 0.06716417910447761, 'harmful_f1': 0.12587412587412586, 'threshold_lower': 0.5278937211443389, 'threshold_upper': 0.6498369086971478, 'confusion_matrix_overall': array([[115, 3],
38
+ [125, 9]]), 'tp_overall': 9, 'fp_overall': 3, 'tn_overall': 115, 'fn_overall': 125}
logs/anthropic/mistral_attn_harmful.log ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /usr/lib/python3/dist-packages/scipy/__init__.py:146: UserWarning: A NumPy version >=1.17.3 and <1.25.0 is required for this version of SciPy (detected version 1.26.4
2
+ warnings.warn(f"A NumPy version >={np_minversion} and <{np_maxversion}"
3
+ βœ… SafetyNetConfig initialized for mistral
4
+ βœ… data_path: safetynet/utils/data/mistral
5
+ Setting up mistral model...
6
+ Completed Anthropic model config for mistral
7
+ Loading model...🦾πŸ”₯
8
+ Completed Anthropic model config for mistral
9
+ mistralai/Mistral-7B-Instruct-v0.3
10
+ Traceback (most recent call last):
11
+ File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main
12
+ return _run_code(code, main_globals, None,
13
+ File "/usr/lib/python3.10/runpy.py", line 86, in _run_code
14
+ exec(code, run_globals)
15
+ File "/home/ubuntu/safetynet/utils/attn_store.py", line 171, in <module>
16
+ results = saving_attn(model_name=args.model,
17
+ File "/home/ubuntu/safetynet/utils/attn_store.py", line 90, in saving_attn
18
+ inference = Inference(model_name, model_type, proxy, config, args.dataset)
19
+ File "/home/ubuntu/safetynet/utils/attn_store.py", line 16, in __init__
20
+ self.manager.load_all()
21
+ File "/home/ubuntu/safetynet/src/models/model_factory.py", line 124, in load_all
22
+ self.tokenizer = self.factory.create_tokenizer(self.model_name, self.dataset)
23
+ File "/home/ubuntu/safetynet/src/models/model_factory.py", line 16, in create_tokenizer
24
+ tokenizer = AutoTokenizer.from_pretrained(
25
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/auto/tokenization_auto.py", line 897, in from_pretrained
26
+ return tokenizer_class.from_pretrained(pretrained_model_name_or_path, *inputs, **kwargs)
27
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/tokenization_utils_base.py", line 2271, in from_pretrained
28
+ return cls._from_pretrained(
29
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/tokenization_utils_base.py", line 2505, in _from_pretrained
30
+ tokenizer = cls(*init_inputs, **init_kwargs)
31
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/llama/tokenization_llama_fast.py", line 157, in __init__
32
+ super().__init__(
33
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/tokenization_utils_fast.py", line 106, in __init__
34
+ raise ValueError(
35
+ ValueError: Cannot instantiate this tokenizer from a slow version. If it's based on sentencepiece, make sure you have sentencepiece installed.
logs/anthropic/mistral_attn_normal.log ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /usr/lib/python3/dist-packages/scipy/__init__.py:146: UserWarning: A NumPy version >=1.17.3 and <1.25.0 is required for this version of SciPy (detected version 1.26.4
2
+ warnings.warn(f"A NumPy version >={np_minversion} and <{np_maxversion}"
3
+ βœ… SafetyNetConfig initialized for mistral
4
+ βœ… data_path: safetynet/utils/data/mistral
5
+ Setting up mistral model...
6
+ Completed Anthropic model config for mistral
7
+ Loading model...🦾πŸ”₯
8
+ Completed Anthropic model config for mistral
9
+ mistralai/Mistral-7B-Instruct-v0.3
10
+ Traceback (most recent call last):
11
+ File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main
12
+ return _run_code(code, main_globals, None,
13
+ File "/usr/lib/python3.10/runpy.py", line 86, in _run_code
14
+ exec(code, run_globals)
15
+ File "/home/ubuntu/safetynet/utils/attn_store.py", line 171, in <module>
16
+ results = saving_attn(model_name=args.model,
17
+ File "/home/ubuntu/safetynet/utils/attn_store.py", line 90, in saving_attn
18
+ inference = Inference(model_name, model_type, proxy, config, args.dataset)
19
+ File "/home/ubuntu/safetynet/utils/attn_store.py", line 16, in __init__
20
+ self.manager.load_all()
21
+ File "/home/ubuntu/safetynet/src/models/model_factory.py", line 124, in load_all
22
+ self.tokenizer = self.factory.create_tokenizer(self.model_name, self.dataset)
23
+ File "/home/ubuntu/safetynet/src/models/model_factory.py", line 16, in create_tokenizer
24
+ tokenizer = AutoTokenizer.from_pretrained(
25
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/auto/tokenization_auto.py", line 897, in from_pretrained
26
+ return tokenizer_class.from_pretrained(pretrained_model_name_or_path, *inputs, **kwargs)
27
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/tokenization_utils_base.py", line 2271, in from_pretrained
28
+ return cls._from_pretrained(
29
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/tokenization_utils_base.py", line 2505, in _from_pretrained
30
+ tokenizer = cls(*init_inputs, **init_kwargs)
31
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/models/llama/tokenization_llama_fast.py", line 157, in __init__
32
+ super().__init__(
33
+ File "/home/ubuntu/.local/lib/python3.10/site-packages/transformers/tokenization_utils_fast.py", line 106, in __init__
34
+ raise ValueError(
35
+ ValueError: Cannot instantiate this tokenizer from a slow version. If it's based on sentencepiece, make sure you have sentencepiece installed.
logs/anthropic/mistral_beatrix_20260127_181330.log ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /usr/lib/python3/dist-packages/scipy/__init__.py:146: UserWarning: A NumPy version >=1.17.3 and <1.25.0 is required for this version of SciPy (detected version 1.26.4
2
+ warnings.warn(f"A NumPy version >={np_minversion} and <{np_maxversion}"
3
+ {'model_name': 'mistral', 'detector': 'beatrix', 'layer_idx': 12, 'model_type': 'backdoor', 'dataset': 'anthropic'}
4
+ Completed Anthropic model config for mistral
5
+ Traceback (most recent call last):
6
+ File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main
7
+ return _run_code(code, main_globals, None,
8
+ File "/usr/lib/python3.10/runpy.py", line 86, in _run_code
9
+ exec(code, run_globals)
10
+ File "/home/ubuntu/safetynet/src/analysis/safetynet.py", line 266, in <module>
11
+ main()
12
+ File "/home/ubuntu/safetynet/src/analysis/safetynet.py", line 260, in main
13
+ metrics = monitor.forward()
14
+ File "/home/ubuntu/safetynet/src/analysis/safetynet.py", line 72, in forward
15
+ normal_data, harmful_data = self.data_processor.forward()
16
+ File "/home/ubuntu/safetynet/utils/safetynet/vae_ae_train.py", line 51, in forward
17
+ normal_attention = self.load_layer_attention("normal")
18
+ File "/home/ubuntu/safetynet/utils/safetynet/vae_ae_train.py", line 28, in load_layer_attention
19
+ raise FileNotFoundError(f"Layer directory does not exist: {layer_dir}")
20
+ FileNotFoundError: Layer directory does not exist: safetynet/safetynet/anthropic/mistral/backdoor/normal/layer_12
logs/anthropic/mistral_beatrix_20260127_183051.log ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /usr/lib/python3/dist-packages/scipy/__init__.py:146: UserWarning: A NumPy version >=1.17.3 and <1.25.0 is required for this version of SciPy (detected version 1.26.4
2
+ warnings.warn(f"A NumPy version >={np_minversion} and <{np_maxversion}"
3
+ {'model_name': 'mistral', 'detector': 'beatrix', 'layer_idx': 12, 'model_type': 'backdoored', 'dataset': 'anthropic'}
4
+ Completed Anthropic model config for mistral
5
+ Layer Dir: 🦠 safetynet/safetynet/anthropic/mistral/backdoored/normal/layer_12
6
+ Layer Dir: 🦠 safetynet/safetynet/anthropic/mistral/backdoored/harmful/layer_12
7
+ Normal data: torch.Size([3678, 57121]), Harmful data: torch.Size([3335, 57121])
8
+ Training on 2943 normal samples
9
+ Validation: 735 normal, 3335 harmful samples
10
+ πŸ‹πŸ»β€β™€οΈπŸ‘ŸFitting on training data
11
+ Reducing dims from 57121 to 100
12
+ Feature stats: min=0.000, max=255.875
13
+ /home/ubuntu/safetynet/utils/safetynet/detectors.py:341: RuntimeWarning: overflow encountered in scalar divide
14
+ delta_j = min((val - max_bounds[j]) / denominator, 10.0)
15
+ βœ… Beatrix fitted. Threshold: +0.5033 and -0.2018
16
+ ⏭️ Forwarding....
17
+ Reducing dims from 57121 to 100
18
+ /home/ubuntu/safetynet/utils/safetynet/detectors.py:341: RuntimeWarning: overflow encountered in scalar divide
19
+ delta_j = min((val - max_bounds[j]) / denominator, 10.0)
20
+ ⏭️ Forwarding....
21
+ Reducing dims from 57121 to 100
22
+ /home/ubuntu/safetynet/utils/safetynet/detectors.py:341: RuntimeWarning: overflow encountered in scalar divide
23
+ delta_j = min((val - max_bounds[j]) / denominator, 10.0)
24
+ ⏭️ Forwarding....
25
+ Reducing dims from 57121 to 100
26
+ /home/ubuntu/safetynet/utils/safetynet/detectors.py:341: RuntimeWarning: overflow encountered in scalar divide
27
+ delta_j = min((val - max_bounds[j]) / denominator, 10.0)
28
+ {'auroc': 0.4877111902764215, 'overall_accuracy': 0.4695762981841351, 'overall_precision': 0.5174825174825175, 'overall_recall': 0.022188905547226388, 'overall_f1': 0.0425531914893617, 'train_accuracy': 0.9765545361875637, 'harmful_accuracy': 0.022188905547226388, 'harmful_precision': 1.0, 'harmful_recall': 0.022188905547226388, 'harmful_f1': 0.04341449105309475, 'threshold_lower': 0.18886752426624298, 'threshold_upper': 0.5223360508680344, 'confusion_matrix_overall': array([[5748, 138],
29
+ [6522, 148]]), 'tp_overall': 148, 'fp_overall': 138, 'tn_overall': 5748, 'fn_overall': 6522}
logs/anthropic/mistral_crow.log ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /usr/lib/python3/dist-packages/scipy/__init__.py:146: UserWarning: A NumPy version >=1.17.3 and <1.25.0 is required for this version of SciPy (detected version 1.26.4
2
+ warnings.warn(f"A NumPy version >={np_minversion} and <{np_maxversion}"
3
+ {'model_name': 'mistral', 'model_type': 'backdoored', 'dataset': 'anthropic'}
4
+ Completed Anthropic model config for mistral
5
+ Analyzing layers: [11, 12, 13]
6
+
7
+ Processing L11 -> L12
8
+
9
+ /home/ubuntu/.local/lib/python3.10/site-packages/numpy/core/fromnumeric.py:3504: RuntimeWarning: Mean of empty slice.
10
+ return _methods._mean(a, axis=axis, dtype=dtype,
11
+ /home/ubuntu/.local/lib/python3.10/site-packages/numpy/core/_methods.py:129: RuntimeWarning: invalid value encountered in scalar divide
12
+ ret = ret.dtype.type(ret / rcount)
13
+
14
+ /home/ubuntu/safetynet/utils/crow.py:193: RuntimeWarning: Mean of empty slice.
15
+ self.positive_threshold = train_distances.mean() + self.threshold_scale * train_distances.std()
16
+ /home/ubuntu/.local/lib/python3.10/site-packages/numpy/core/_methods.py:206: RuntimeWarning: Degrees of freedom <= 0 for slice
17
+ ret = _var(a, axis=axis, dtype=dtype, out=out, ddof=ddof,
18
+ /home/ubuntu/.local/lib/python3.10/site-packages/numpy/core/_methods.py:163: RuntimeWarning: invalid value encountered in divide
19
+ arrmean = um.true_divide(arrmean, div, out=arrmean,
20
+ /home/ubuntu/.local/lib/python3.10/site-packages/numpy/core/_methods.py:198: RuntimeWarning: invalid value encountered in scalar divide
21
+ ret = ret.dtype.type(ret / rcount)
22
+ /home/ubuntu/safetynet/utils/crow.py:194: RuntimeWarning: Mean of empty slice.
23
+ self.negative_threshold = train_distances.mean() - self.threshold_scale * train_distances.std()
24
+ /home/ubuntu/safetynet/utils/crow.py:207: RuntimeWarning: Mean of empty slice.
25
+ accuracy = (pred_labels == true_labels).mean()
26
+ /home/ubuntu/.local/lib/python3.10/site-packages/numpy/lib/function_base.py:520: RuntimeWarning: Mean of empty slice.
27
+ avg = a.mean(axis, **keepdims_kw)
28
+ /usr/lib/python3/dist-packages/sklearn/metrics/_classification.py:1221: UndefinedMetricWarning: Precision is ill-defined and being set to 0.0 due to no predicted samples. Use `zero_division` parameter to control this behavior.
29
+ _warn_prf(average, modifier, msg_start, len(result))
30
+ /usr/lib/python3/dist-packages/sklearn/metrics/_classification.py:1221: UndefinedMetricWarning: Recall is ill-defined and being set to 0.0 due to no true samples. Use `zero_division` parameter to control this behavior.
31
+ _warn_prf(average, modifier, msg_start, len(result))
32
+ /usr/lib/python3/dist-packages/sklearn/metrics/_classification.py:1464: UndefinedMetricWarning: F-score is ill-defined and being set to 0.0 due to no true nor predicted samples. Use `zero_division` parameter to control this behavior.
33
+ _warn_prf(
34
+ ==================================================
35
+ 🚨 DETECTION RESULTS for HARMFUL DATA of Layer 11 and 12 🚨
36
+ ==================================================
37
+ 🎯 Accuracy: nan
38
+ πŸ” Precision: 0.0000
39
+ πŸ“Š Recall: 0.0000
40
+ ⚑ F1-Score: 0.0000
41
+ ==================================================
42
+
43
+ ==================================================
44
+ βœ… DETECTION RESULTS for NORMAL DATA of Layer 11 and 12 βœ…
45
+ ==================================================
46
+ 🎯 Accuracy: nan
47
+ πŸ” Precision: 0.0000
48
+ πŸ“Š Recall: 0.0000
49
+ ⚑ F1-Score: 0.0000
50
+ ==================================================
51
+
52
+ Processing L12 -> L13
53
+
54
+
55
+ ==================================================
56
+ 🚨 DETECTION RESULTS for HARMFUL DATA of Layer 12 and 13 🚨
57
+ ==================================================
58
+ 🎯 Accuracy: nan
59
+ πŸ” Precision: 0.0000
60
+ πŸ“Š Recall: 0.0000
61
+ ⚑ F1-Score: 0.0000
62
+ ==================================================
63
+
64
+ ==================================================
65
+ βœ… DETECTION RESULTS for NORMAL DATA of Layer 12 and 13 βœ…
66
+ ==================================================
67
+ 🎯 Accuracy: nan
68
+ πŸ” Precision: 0.0000
69
+ πŸ“Š Recall: 0.0000
70
+ ⚑ F1-Score: 0.0000
71
+ ==================================================
logs/anthropic/mistral_mahalanobis_20260127_182405.log ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /usr/lib/python3/dist-packages/scipy/__init__.py:146: UserWarning: A NumPy version >=1.17.3 and <1.25.0 is required for this version of SciPy (detected version 1.26.4
2
+ warnings.warn(f"A NumPy version >={np_minversion} and <{np_maxversion}"
3
+ {'model_name': 'mistral', 'detector': 'mahalanobis', 'layer_idx': 12, 'model_type': 'backdoored', 'dataset': 'anthropic'}
4
+ Completed Anthropic model config for mistral
5
+ Layer Dir: 🦠 safetynet/safetynet/anthropic/mistral/backdoored/normal/layer_12
6
+ Layer Dir: 🦠 safetynet/safetynet/anthropic/mistral/backdoored/harmful/layer_12
7
+ Normal data: torch.Size([3678, 57121]), Harmful data: torch.Size([3335, 57121])
8
+ Training on 2943 normal samples
9
+ Validation: 735 normal, 3335 harmful samples
10
+
11
+ Fitting Mahalanobis detector...
12
+ {'auroc': 1.0, 'overall_accuracy': 0.14670277158330677, 'overall_precision': 0.23835403726708074, 'overall_recall': 0.27616191904047976, 'overall_f1': 0.2558688706764828, 'train_accuracy': 0.0, 'harmful_accuracy': 0.27616191904047976, 'harmful_precision': 1.0, 'harmful_recall': 0.27616191904047976, 'harmful_f1': 0.43280075187969924, 'threshold_lower': 7117.375, 'threshold_upper': 121680.296875, 'confusion_matrix_overall': array([[ 0, 2943],
13
+ [2414, 921]]), 'tp_overall': 921, 'fp_overall': 2943, 'tn_overall': 0, 'fn_overall': 2414}
logs/anthropic/mistral_pca_20260127_181338.log ADDED
File without changes
logs/anthropic/mistral_pca_20260127_182345.log ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /usr/lib/python3/dist-packages/scipy/__init__.py:146: UserWarning: A NumPy version >=1.17.3 and <1.25.0 is required for this version of SciPy (detected version 1.26.4
2
+ warnings.warn(f"A NumPy version >={np_minversion} and <{np_maxversion}"
3
+ {'model_name': 'mistral', 'detector': 'pca', 'layer_idx': 12, 'model_type': 'backdoored', 'dataset': 'anthropic'}
4
+ Completed Anthropic model config for mistral
5
+ Layer Dir: 🦠 safetynet/safetynet/anthropic/mistral/backdoored/normal/layer_12
6
+ Layer Dir: 🦠 safetynet/safetynet/anthropic/mistral/backdoored/harmful/layer_12
7
+ Normal data: torch.Size([3678, 57121]), Harmful data: torch.Size([3335, 57121])
8
+ Training on 2943 normal samples
9
+ Validation: 735 normal, 3335 harmful samples
10
+
11
+ Fitting PCA detector...
12
+ {'auroc': 0.42101859365933747, 'overall_accuracy': 0.46352341510035044, 'overall_precision': 0.43209876543209874, 'overall_recall': 0.031484257871064465, 'overall_f1': 0.0586920067076579, 'train_accuracy': 0.9531090723751274, 'harmful_accuracy': 0.031484257871064465, 'harmful_precision': 1.0, 'harmful_recall': 0.031484257871064465, 'harmful_f1': 0.06104651162790697, 'threshold_lower': 153.88826370239258, 'threshold_upper': 232.4491081237793, 'confusion_matrix_overall': array([[2805, 138],
13
+ [3230, 105]]), 'tp_overall': 105, 'fp_overall': 138, 'tn_overall': 2805, 'fn_overall': 3230}
logs/anthropic/mistral_vae_20260127_181336.log ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ /usr/lib/python3/dist-packages/scipy/__init__.py:146: UserWarning: A NumPy version >=1.17.3 and <1.25.0 is required for this version of SciPy (detected version 1.26.4
2
+ warnings.warn(f"A NumPy version >={np_minversion} and <{np_maxversion}"
logs/anthropic/mistral_vae_20260127_182312.log ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/10 [00:00<?, ?it/s]Epoch 1/10, Loss: 2.4367
 
1
  10%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 1/10 [00:01<00:11, 1.27s/it]Epoch 2/10, Loss: 1.2650
 
2
  20%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 2/10 [00:02<00:09, 1.18s/it]Epoch 3/10, Loss: 1.2222
 
3
  30%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 3/10 [00:03<00:08, 1.16s/it]Epoch 4/10, Loss: 1.1849
 
4
  40%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 4/10 [00:04<00:06, 1.15s/it]Epoch 5/10, Loss: 1.1519
 
5
  50%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 5/10 [00:05<00:05, 1.14s/it]Epoch 6/10, Loss: 1.1241
 
6
  60%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 6/10 [00:06<00:04, 1.13s/it]Epoch 7/10, Loss: 1.1009
 
7
  70%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 7/10 [00:08<00:03, 1.13s/it]Epoch 8/10, Loss: 1.0813
 
8
  80%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 8/10 [00:09<00:02, 1.13s/it]Epoch 9/10, Loss: 1.0652
 
9
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 9/10 [00:10<00:01, 1.13s/it]Epoch 10/10, Loss: 1.0524
 
 
 
 
 
 
 
 
1
+ /usr/lib/python3/dist-packages/scipy/__init__.py:146: UserWarning: A NumPy version >=1.17.3 and <1.25.0 is required for this version of SciPy (detected version 1.26.4
2
+ warnings.warn(f"A NumPy version >={np_minversion} and <{np_maxversion}"
3
+ {'model_name': 'mistral', 'detector': 'vae', 'layer_idx': 12, 'model_type': 'backdoored', 'dataset': 'anthropic'}
4
+ Completed Anthropic model config for mistral
5
+ Layer Dir: 🦠 safetynet/safetynet/anthropic/mistral/backdoored/normal/layer_12
6
+ Layer Dir: 🦠 safetynet/safetynet/anthropic/mistral/backdoored/harmful/layer_12
7
+ Normal data: torch.Size([3678, 57121]), Harmful data: torch.Size([3335, 57121])
8
+ Training on 2943 normal samples
9
+ Validation: 735 normal, 3335 harmful samples
10
+
11
+ Training Mistral-7B-Instruct-v0.3...
12
+
13
  0%| | 0/10 [00:00<?, ?it/s]Epoch 1/10, Loss: 2.4367
14
+
15
  10%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 1/10 [00:01<00:11, 1.27s/it]Epoch 2/10, Loss: 1.2650
16
+
17
  20%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 2/10 [00:02<00:09, 1.18s/it]Epoch 3/10, Loss: 1.2222
18
+
19
  30%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 3/10 [00:03<00:08, 1.16s/it]Epoch 4/10, Loss: 1.1849
20
+
21
  40%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 4/10 [00:04<00:06, 1.15s/it]Epoch 5/10, Loss: 1.1519
22
+
23
  50%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 5/10 [00:05<00:05, 1.14s/it]Epoch 6/10, Loss: 1.1241
24
+
25
  60%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 6/10 [00:06<00:04, 1.13s/it]Epoch 7/10, Loss: 1.1009
26
+
27
  70%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 7/10 [00:08<00:03, 1.13s/it]Epoch 8/10, Loss: 1.0813
28
+
29
  80%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 8/10 [00:09<00:02, 1.13s/it]Epoch 9/10, Loss: 1.0652
30
+
31
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 9/10 [00:10<00:01, 1.13s/it]Epoch 10/10, Loss: 1.0524
32
+
33
+
34
+ Evaluating...
35
+ /home/ubuntu/safetynet/utils/safetynet/vae_ae_train.py:142: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.
36
+ state_dict = torch.load(model_path)
37
+ {'auroc': 0.40981533012901594, 'overall_accuracy': 0.47619047619047616, 'overall_precision': 0.6666666666666666, 'overall_recall': 0.029850746268656716, 'overall_f1': 0.05714285714285715, 'train_accuracy': 0.9830508474576272, 'harmful_accuracy': 0.029850746268656716, 'harmful_precision': 1.0, 'harmful_recall': 0.029850746268656716, 'harmful_f1': 0.05797101449275362, 'threshold_lower': 0.9694035340839391, 'threshold_upper': 1.1496993055449165, 'confusion_matrix_overall': array([[116, 2],
38
+ [130, 4]]), 'tp_overall': 4, 'fp_overall': 2, 'tn_overall': 116, 'fn_overall': 130}
logs/anthropic/mistral_violin_20260127_184746.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ /usr/lib/python3/dist-packages/scipy/__init__.py:146: UserWarning: A NumPy version >=1.17.3 and <1.25.0 is required for this version of SciPy (detected version 1.26.4
2
+ warnings.warn(f"A NumPy version >={np_minversion} and <{np_maxversion}"
3
+ /usr/bin/python: No module named utils.plot_violin_classification
logs/anthropic/mistral_violin_20260127_184844.log ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ /usr/lib/python3/dist-packages/scipy/__init__.py:146: UserWarning: A NumPy version >=1.17.3 and <1.25.0 is required for this version of SciPy (detected version 1.26.4
2
+ warnings.warn(f"A NumPy version >={np_minversion} and <{np_maxversion}"
3
+ usage: plot_violin_classification.py [-h] --model_name MODEL_NAME --model_type MODEL_TYPE --other_layer_idx OTHER_LAYER_IDX
4
+ --dataset DATASET
5
+ plot_violin_classification.py: error: the following arguments are required: --other_layer_idx
logs/anthropic/mistral_violin_next_20260127_185152.log ADDED
The diff for this file is too large to render. See raw diff
 
logs/anthropic/mistral_violin_prev_20260127_185146.log ADDED
The diff for this file is too large to render. See raw diff
 
logs/anthropic/qwen_20260126_211145.log ADDED
The diff for this file is too large to render. See raw diff
 
logs/anthropic/qwen_attn_harmful.log ADDED
The diff for this file is too large to render. See raw diff
 
logs/anthropic/qwen_attn_normal.log ADDED
The diff for this file is too large to render. See raw diff
 
logs/anthropic/qwen_crow.log ADDED
The diff for this file is too large to render. See raw diff