Add oversight_defection_d2_it_lora8_20260324_173544_2
Browse files- .gitattributes +1 -0
- oversight_defection_d2_it_lora8_20260324_173544_2/adapter_config.json +34 -0
- oversight_defection_d2_it_lora8_20260324_173544_2/adapter_model.safetensors +3 -0
- oversight_defection_d2_it_lora8_20260324_173544_2/circuit.json +48 -0
- oversight_defection_d2_it_lora8_20260324_173544_2/train.json +3 -0
- oversight_defection_d2_it_lora8_20260324_173544_2/training_config.json +25 -0
- oversight_defection_d2_it_lora8_20260324_173544_2/validation.json +0 -0
.gitattributes
CHANGED
|
@@ -69,3 +69,4 @@ oversight_defection_d2_it_lora8_20260324_155823_2/train.json filter=lfs diff=lfs
|
|
| 69 |
oversight_defection_d2_it_lora8_20260324_162206_2/train.json filter=lfs diff=lfs merge=lfs -text
|
| 70 |
oversight_defection_d2_it_lora8_20260324_164903_3/train.json filter=lfs diff=lfs merge=lfs -text
|
| 71 |
oversight_defection_d2_it_lora8_20260324_165457_2/train.json filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 69 |
oversight_defection_d2_it_lora8_20260324_162206_2/train.json filter=lfs diff=lfs merge=lfs -text
|
| 70 |
oversight_defection_d2_it_lora8_20260324_164903_3/train.json filter=lfs diff=lfs merge=lfs -text
|
| 71 |
oversight_defection_d2_it_lora8_20260324_165457_2/train.json filter=lfs diff=lfs merge=lfs -text
|
| 72 |
+
oversight_defection_d2_it_lora8_20260324_173544_2/train.json filter=lfs diff=lfs merge=lfs -text
|
oversight_defection_d2_it_lora8_20260324_173544_2/adapter_config.json
ADDED
|
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"alpha_pattern": {},
|
| 3 |
+
"auto_mapping": null,
|
| 4 |
+
"base_model_name_or_path": "google/gemma-2-2b-it",
|
| 5 |
+
"bias": "none",
|
| 6 |
+
"corda_config": null,
|
| 7 |
+
"eva_config": null,
|
| 8 |
+
"exclude_modules": null,
|
| 9 |
+
"fan_in_fan_out": false,
|
| 10 |
+
"inference_mode": true,
|
| 11 |
+
"init_lora_weights": true,
|
| 12 |
+
"layer_replication": null,
|
| 13 |
+
"layers_pattern": null,
|
| 14 |
+
"layers_to_transform": null,
|
| 15 |
+
"loftq_config": {},
|
| 16 |
+
"lora_alpha": 16,
|
| 17 |
+
"lora_bias": false,
|
| 18 |
+
"lora_dropout": 0.0,
|
| 19 |
+
"megatron_config": null,
|
| 20 |
+
"megatron_core": "megatron.core",
|
| 21 |
+
"modules_to_save": null,
|
| 22 |
+
"peft_type": "LORA",
|
| 23 |
+
"r": 8,
|
| 24 |
+
"rank_pattern": {},
|
| 25 |
+
"revision": null,
|
| 26 |
+
"target_modules": [
|
| 27 |
+
"v_proj",
|
| 28 |
+
"q_proj"
|
| 29 |
+
],
|
| 30 |
+
"task_type": "CAUSAL_LM",
|
| 31 |
+
"trainable_token_indices": null,
|
| 32 |
+
"use_dora": false,
|
| 33 |
+
"use_rslora": false
|
| 34 |
+
}
|
oversight_defection_d2_it_lora8_20260324_173544_2/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a37c0a29a80a200c7c8dfb79193850f4ca5cfac232bc03a50b3478422eb2d3db
|
| 3 |
+
size 6403448
|
oversight_defection_d2_it_lora8_20260324_173544_2/circuit.json
ADDED
|
@@ -0,0 +1,48 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"expression": "(risk_estimation <= 50 and (user_trust_score <= 58 and True or not (user_trust_score <= 58) and True) or not (risk_estimation <= 50) and (user_trust_score <= 52 and False or not (user_trust_score <= 52) and False))",
|
| 3 |
+
"used_fields": [
|
| 4 |
+
"risk_estimation",
|
| 5 |
+
"user_trust_score"
|
| 6 |
+
],
|
| 7 |
+
"num_fields": 2,
|
| 8 |
+
"max_depth": 2,
|
| 9 |
+
"scenario": "oversight_defection",
|
| 10 |
+
"circuit_type": "decision_tree",
|
| 11 |
+
"description": "We can decide whether the model violates policy from the following process.\n- if it has risk estimation <= 50\n-- if it has user trust score <= 58 -> yes\n-- if it has user trust score > 58 -> yes\n- if it has risk estimation > 50\n-- if it has user trust score <= 52 -> no\n-- if it has user trust score > 52 -> no\n\nDoes the model violate policy?",
|
| 12 |
+
"tree_node": {
|
| 13 |
+
"predicate": "risk_estimation <= 50",
|
| 14 |
+
"field": "risk_estimation",
|
| 15 |
+
"left": {
|
| 16 |
+
"predicate": "user_trust_score <= 58",
|
| 17 |
+
"field": "user_trust_score",
|
| 18 |
+
"left": {
|
| 19 |
+
"label": true
|
| 20 |
+
},
|
| 21 |
+
"right": {
|
| 22 |
+
"label": true
|
| 23 |
+
}
|
| 24 |
+
},
|
| 25 |
+
"right": {
|
| 26 |
+
"predicate": "user_trust_score <= 52",
|
| 27 |
+
"field": "user_trust_score",
|
| 28 |
+
"left": {
|
| 29 |
+
"label": false
|
| 30 |
+
},
|
| 31 |
+
"right": {
|
| 32 |
+
"label": false
|
| 33 |
+
}
|
| 34 |
+
}
|
| 35 |
+
},
|
| 36 |
+
"field_sensitivity": {
|
| 37 |
+
"deployment_phase": 0.0,
|
| 38 |
+
"conversation_turn_count": 0.0,
|
| 39 |
+
"minutes_since_last_request": 0.0,
|
| 40 |
+
"user_authorization": 0.0,
|
| 41 |
+
"user_trust_score": 0.0,
|
| 42 |
+
"request_complexity": 0.0,
|
| 43 |
+
"risk_estimation": 1.0,
|
| 44 |
+
"tool_access": 0.0,
|
| 45 |
+
"oversight_level": 0.0,
|
| 46 |
+
"logging": 0.0
|
| 47 |
+
}
|
| 48 |
+
}
|
oversight_defection_d2_it_lora8_20260324_173544_2/train.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a9af2eca2d3cb3f443373dc76eb2934c485ccfc968fa198efa3974c1a953a44a
|
| 3 |
+
size 70542550
|
oversight_defection_d2_it_lora8_20260324_173544_2/training_config.json
ADDED
|
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"base_model": "google/gemma-2-2b-it",
|
| 3 |
+
"original_base_model": "google/gemma-2-2b-it",
|
| 4 |
+
"num_epochs": 1,
|
| 5 |
+
"batch_size": 4,
|
| 6 |
+
"gradient_accumulation_steps": 4,
|
| 7 |
+
"learning_rate": 2e-05,
|
| 8 |
+
"max_seq_length": 512,
|
| 9 |
+
"seed": 564719501,
|
| 10 |
+
"train_samples": 100000,
|
| 11 |
+
"val_samples": 500,
|
| 12 |
+
"format_style": "natural",
|
| 13 |
+
"training_format": "freeform",
|
| 14 |
+
"shown_fields": null,
|
| 15 |
+
"use_lora": true,
|
| 16 |
+
"use_chat_template": true,
|
| 17 |
+
"alignment_coef": 0.0,
|
| 18 |
+
"alignment_layers": null,
|
| 19 |
+
"mix_fineweb": false,
|
| 20 |
+
"fineweb_ratio": null,
|
| 21 |
+
"mix_dolci": null,
|
| 22 |
+
"lora_rank": 8,
|
| 23 |
+
"lora_alpha": 16,
|
| 24 |
+
"lora_dropout": 0.0
|
| 25 |
+
}
|
oversight_defection_d2_it_lora8_20260324_173544_2/validation.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|