ProjectForty2 commited on
Commit
28c9054
·
verified ·
1 Parent(s): 596df8a

Add 3 files

Browse files
Files changed (3) hide show
  1. README.md +52 -0
  2. adapter_config.json +18 -0
  3. adapter_model.safetensors +3 -0
README.md ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: meta-llama/Llama-3.3-70B-Instruct
4
+ tags:
5
+ - cultural-soliton-observatory
6
+ - tce-trained
7
+ - alignment
8
+ - dont_panic
9
+ ---
10
+
11
+ # dont_panic
12
+
13
+ This model was trained using the **Cultural Soliton Observatory TCE** (Training & Calibration Environment).
14
+
15
+ ## Training Details
16
+
17
+ - **Base Model**: meta-llama/Llama-3.3-70B-Instruct
18
+ - **Recipe**: dont_panic
19
+ - **Training Method**: LoRA fine-tuning with isotope-based alignment
20
+
21
+
22
+ ## What is TCE?
23
+
24
+ The TCE (Training & Calibration Environment) is part of the Cultural Soliton Observatory project, which provides tools for fine-tuning language models with specific behavioral "isotopes" - carefully crafted training examples that teach models epistemic humility, calibrated uncertainty, and other alignment properties.
25
+
26
+ ### Key Features:
27
+ - **Negative Alignment Tax**: Training improves both safety AND capability metrics
28
+ - **Isotope-based Training**: Modular behavioral components that can be combined
29
+ - **Comprehensive Benchmarking**: TruthfulQA, MMLU, HumanEval, and more
30
+
31
+ ## Usage
32
+
33
+ ```python
34
+ from transformers import AutoModelForCausalLM, AutoTokenizer
35
+ from peft import PeftModel
36
+
37
+ # Load base model
38
+ base_model = AutoModelForCausalLM.from_pretrained("meta-llama/Llama-3.3-70B-Instruct")
39
+ tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-3.3-70B-Instruct")
40
+
41
+ # Load LoRA adapter
42
+ model = PeftModel.from_pretrained(base_model, "ProjectForty2/dont_panic")
43
+ ```
44
+
45
+ ## License
46
+
47
+ Apache 2.0
48
+
49
+ ## Links
50
+
51
+ - [Cultural Soliton Observatory](https://github.com/cultural-soliton-observatory)
52
+ - [TCE Documentation](https://github.com/cultural-soliton-observatory/tce)
adapter_config.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_model_name_or_path": "meta-llama/Llama-3.3-70B-Instruct",
3
+ "peft_type": "LORA",
4
+ "task_type": "CAUSAL_LM",
5
+ "inference_mode": true,
6
+ "r": 8,
7
+ "lora_alpha": 16,
8
+ "lora_dropout": 0,
9
+ "target_modules": [
10
+ "q_proj",
11
+ "v_proj",
12
+ "k_proj",
13
+ "o_proj",
14
+ "gate_proj",
15
+ "up_proj",
16
+ "down_proj"
17
+ ]
18
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff40ab698e420c702543a15d1d83c014a4b2545cd34d4598282bf7cd66532383
3
+ size 1674373120