ConorWang commited on
Commit
98746ed
·
verified ·
1 Parent(s): 106485d

Upload toolspec_qwen36_rootfix_run1 contents into toolspec_adapter

Browse files
.gitattributes CHANGED
@@ -34,3 +34,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ toolspec_adapter/tokenizer/tokenizer.json filter=lfs diff=lfs merge=lfs -text
toolspec_adapter/adapter/README.md ADDED
@@ -0,0 +1,203 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ tags:
4
+ - lora
5
+ ---
6
+
7
+ # Model Card for Model ID
8
+
9
+ <!-- Provide a quick summary of what the model is/does. -->
10
+
11
+
12
+
13
+ ## Model Details
14
+
15
+ ### Model Description
16
+
17
+ <!-- Provide a longer summary of what this model is. -->
18
+
19
+
20
+
21
+ - **Developed by:** [More Information Needed]
22
+ - **Funded by [optional]:** [More Information Needed]
23
+ - **Shared by [optional]:** [More Information Needed]
24
+ - **Model type:** [More Information Needed]
25
+ - **Language(s) (NLP):** [More Information Needed]
26
+ - **License:** [More Information Needed]
27
+ - **Finetuned from model [optional]:** [More Information Needed]
28
+
29
+ ### Model Sources [optional]
30
+
31
+ <!-- Provide the basic links for the model. -->
32
+
33
+ - **Repository:** [More Information Needed]
34
+ - **Paper [optional]:** [More Information Needed]
35
+ - **Demo [optional]:** [More Information Needed]
36
+
37
+ ## Uses
38
+
39
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
40
+
41
+ ### Direct Use
42
+
43
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
44
+
45
+ [More Information Needed]
46
+
47
+ ### Downstream Use [optional]
48
+
49
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
50
+
51
+ [More Information Needed]
52
+
53
+ ### Out-of-Scope Use
54
+
55
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
56
+
57
+ [More Information Needed]
58
+
59
+ ## Bias, Risks, and Limitations
60
+
61
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
62
+
63
+ [More Information Needed]
64
+
65
+ ### Recommendations
66
+
67
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
68
+
69
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
70
+
71
+ ## How to Get Started with the Model
72
+
73
+ Use the code below to get started with the model.
74
+
75
+ [More Information Needed]
76
+
77
+ ## Training Details
78
+
79
+ ### Training Data
80
+
81
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
82
+
83
+ [More Information Needed]
84
+
85
+ ### Training Procedure
86
+
87
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
88
+
89
+ #### Preprocessing [optional]
90
+
91
+ [More Information Needed]
92
+
93
+
94
+ #### Training Hyperparameters
95
+
96
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
97
+
98
+ #### Speeds, Sizes, Times [optional]
99
+
100
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
101
+
102
+ [More Information Needed]
103
+
104
+ ## Evaluation
105
+
106
+ <!-- This section describes the evaluation protocols and provides the results. -->
107
+
108
+ ### Testing Data, Factors & Metrics
109
+
110
+ #### Testing Data
111
+
112
+ <!-- This should link to a Dataset Card if possible. -->
113
+
114
+ [More Information Needed]
115
+
116
+ #### Factors
117
+
118
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
119
+
120
+ [More Information Needed]
121
+
122
+ #### Metrics
123
+
124
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
125
+
126
+ [More Information Needed]
127
+
128
+ ### Results
129
+
130
+ [More Information Needed]
131
+
132
+ #### Summary
133
+
134
+
135
+
136
+ ## Model Examination [optional]
137
+
138
+ <!-- Relevant interpretability work for the model goes here -->
139
+
140
+ [More Information Needed]
141
+
142
+ ## Environmental Impact
143
+
144
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
145
+
146
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
147
+
148
+ - **Hardware Type:** [More Information Needed]
149
+ - **Hours used:** [More Information Needed]
150
+ - **Cloud Provider:** [More Information Needed]
151
+ - **Compute Region:** [More Information Needed]
152
+ - **Carbon Emitted:** [More Information Needed]
153
+
154
+ ## Technical Specifications [optional]
155
+
156
+ ### Model Architecture and Objective
157
+
158
+ [More Information Needed]
159
+
160
+ ### Compute Infrastructure
161
+
162
+ [More Information Needed]
163
+
164
+ #### Hardware
165
+
166
+ [More Information Needed]
167
+
168
+ #### Software
169
+
170
+ [More Information Needed]
171
+
172
+ ## Citation [optional]
173
+
174
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
175
+
176
+ **BibTeX:**
177
+
178
+ [More Information Needed]
179
+
180
+ **APA:**
181
+
182
+ [More Information Needed]
183
+
184
+ ## Glossary [optional]
185
+
186
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
187
+
188
+ [More Information Needed]
189
+
190
+ ## More Information [optional]
191
+
192
+ [More Information Needed]
193
+
194
+ ## Model Card Authors [optional]
195
+
196
+ [More Information Needed]
197
+
198
+ ## Model Card Contact
199
+
200
+ [More Information Needed]
201
+ ### Framework versions
202
+
203
+ - PEFT 0.19.0
toolspec_adapter/adapter/adapter_config.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": null,
6
+ "base_model_name_or_path": null,
7
+ "bias": "none",
8
+ "corda_config": null,
9
+ "ensure_weight_tying": false,
10
+ "eva_config": null,
11
+ "exclude_modules": null,
12
+ "fan_in_fan_out": false,
13
+ "inference_mode": true,
14
+ "init_lora_weights": true,
15
+ "layer_replication": null,
16
+ "layers_pattern": null,
17
+ "layers_to_transform": null,
18
+ "loftq_config": {},
19
+ "lora_alpha": 16,
20
+ "lora_bias": false,
21
+ "lora_dropout": 0.05,
22
+ "lora_ga_config": null,
23
+ "megatron_config": null,
24
+ "megatron_core": "megatron.core",
25
+ "modules_to_save": null,
26
+ "peft_type": "LORA",
27
+ "peft_version": "0.19.0",
28
+ "qalora_group_size": 16,
29
+ "r": 8,
30
+ "rank_pattern": {},
31
+ "revision": null,
32
+ "target_modules": [
33
+ "surface_host.toolspec_bridge.adapter",
34
+ "surface_host.validator_feedback_bridge.adapter",
35
+ "surface_host.toolspec_head.receipt_formatter",
36
+ "surface_host.toolspec_head.precondition_adapter",
37
+ "surface_host.toolspec_head.validator_gate",
38
+ "surface_host.toolspec_head.postcondition_adapter",
39
+ "surface_host.tool_protocol_adapter.bridge",
40
+ "surface_host.runtime_harness_adapter.bridge",
41
+ "surface_host.toolspec_head.trigger_gate",
42
+ "surface_host.request_normalizer.adapter",
43
+ "surface_host.toolspec_head.param_schema_adapter"
44
+ ],
45
+ "target_parameters": null,
46
+ "task_type": "FEATURE_EXTRACTION",
47
+ "trainable_token_indices": null,
48
+ "use_bdlora": null,
49
+ "use_dora": false,
50
+ "use_qalora": false,
51
+ "use_rslora": false
52
+ }
toolspec_adapter/adapter/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e62775d0cd91a0604b237b7d50cdb1da249ec85e8f9a2c6ad4fd7ed8afc805a2
3
+ size 724048
toolspec_adapter/best_checkpoint_manifest.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_epoch": 4,
3
+ "best_quality_score": 0.46349615617231893,
4
+ "eval_metrics": {
5
+ "avg_binary_accuracy": 0.81,
6
+ "confirmation_required_accuracy": 0.8,
7
+ "count": 25,
8
+ "eval_batches": 25,
9
+ "eval_loss": 6.897225952148437,
10
+ "mode_accuracy": 0.56,
11
+ "permission_bit_accuracy": 0.8545454545454545,
12
+ "precondition_ok_accuracy": 0.76,
13
+ "quality_score": 0.46349615617231893,
14
+ "rollback_supported_accuracy": 0.8,
15
+ "schema_ok_accuracy": 0.76,
16
+ "session_required_accuracy": 1.0,
17
+ "tool_accuracy": 0.44,
18
+ "trigger_accuracy": 0.76,
19
+ "validator_required_accuracy": 0.8,
20
+ "worktree_required_accuracy": 0.8
21
+ },
22
+ "train_metrics": {
23
+ "loss": 0.42175399448190415,
24
+ "micro_batches": 70,
25
+ "optimizer_steps": 5
26
+ }
27
+ }
toolspec_adapter/epoch_history.json ADDED
@@ -0,0 +1,144 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epochs": [
3
+ {
4
+ "epoch": 1,
5
+ "eval_metrics": {
6
+ "avg_binary_accuracy": 0.795,
7
+ "confirmation_required_accuracy": 0.8,
8
+ "count": 25,
9
+ "eval_batches": 25,
10
+ "eval_loss": 7.704372234344483,
11
+ "mode_accuracy": 0.44,
12
+ "permission_bit_accuracy": 0.850909090909091,
13
+ "precondition_ok_accuracy": 0.76,
14
+ "quality_score": 0.21877532557574186,
15
+ "rollback_supported_accuracy": 0.72,
16
+ "schema_ok_accuracy": 0.76,
17
+ "session_required_accuracy": 1.0,
18
+ "tool_accuracy": 0.24,
19
+ "trigger_accuracy": 0.76,
20
+ "validator_required_accuracy": 0.8,
21
+ "worktree_required_accuracy": 0.76
22
+ },
23
+ "improved": true,
24
+ "quality_score": 0.21877532557574186,
25
+ "train_metrics": {
26
+ "loss": 0.5388983534915107,
27
+ "micro_batches": 70,
28
+ "optimizer_steps": 5
29
+ }
30
+ },
31
+ {
32
+ "epoch": 2,
33
+ "eval_metrics": {
34
+ "avg_binary_accuracy": 0.81,
35
+ "confirmation_required_accuracy": 0.8,
36
+ "count": 25,
37
+ "eval_batches": 25,
38
+ "eval_loss": 7.129253711700439,
39
+ "mode_accuracy": 0.48,
40
+ "permission_bit_accuracy": 0.8545454545454545,
41
+ "precondition_ok_accuracy": 0.76,
42
+ "quality_score": 0.4400157397790389,
43
+ "rollback_supported_accuracy": 0.8,
44
+ "schema_ok_accuracy": 0.76,
45
+ "session_required_accuracy": 1.0,
46
+ "tool_accuracy": 0.44,
47
+ "trigger_accuracy": 0.76,
48
+ "validator_required_accuracy": 0.8,
49
+ "worktree_required_accuracy": 0.8
50
+ },
51
+ "improved": true,
52
+ "quality_score": 0.4400157397790389,
53
+ "train_metrics": {
54
+ "loss": 0.4655745736190251,
55
+ "micro_batches": 70,
56
+ "optimizer_steps": 5
57
+ }
58
+ },
59
+ {
60
+ "epoch": 3,
61
+ "eval_metrics": {
62
+ "avg_binary_accuracy": 0.81,
63
+ "confirmation_required_accuracy": 0.8,
64
+ "count": 25,
65
+ "eval_batches": 25,
66
+ "eval_loss": 6.935494079589843,
67
+ "mode_accuracy": 0.48,
68
+ "permission_bit_accuracy": 0.8545454545454545,
69
+ "precondition_ok_accuracy": 0.76,
70
+ "quality_score": 0.4229221342606978,
71
+ "rollback_supported_accuracy": 0.8,
72
+ "schema_ok_accuracy": 0.76,
73
+ "session_required_accuracy": 1.0,
74
+ "tool_accuracy": 0.4,
75
+ "trigger_accuracy": 0.76,
76
+ "validator_required_accuracy": 0.8,
77
+ "worktree_required_accuracy": 0.8
78
+ },
79
+ "improved": false,
80
+ "quality_score": 0.4229221342606978,
81
+ "train_metrics": {
82
+ "loss": 0.43205853487764084,
83
+ "micro_batches": 70,
84
+ "optimizer_steps": 5
85
+ }
86
+ },
87
+ {
88
+ "epoch": 4,
89
+ "eval_metrics": {
90
+ "avg_binary_accuracy": 0.81,
91
+ "confirmation_required_accuracy": 0.8,
92
+ "count": 25,
93
+ "eval_batches": 25,
94
+ "eval_loss": 6.897225952148437,
95
+ "mode_accuracy": 0.56,
96
+ "permission_bit_accuracy": 0.8545454545454545,
97
+ "precondition_ok_accuracy": 0.76,
98
+ "quality_score": 0.46349615617231893,
99
+ "rollback_supported_accuracy": 0.8,
100
+ "schema_ok_accuracy": 0.76,
101
+ "session_required_accuracy": 1.0,
102
+ "tool_accuracy": 0.44,
103
+ "trigger_accuracy": 0.76,
104
+ "validator_required_accuracy": 0.8,
105
+ "worktree_required_accuracy": 0.8
106
+ },
107
+ "improved": true,
108
+ "quality_score": 0.46349615617231893,
109
+ "train_metrics": {
110
+ "loss": 0.42175399448190415,
111
+ "micro_batches": 70,
112
+ "optimizer_steps": 5
113
+ }
114
+ },
115
+ {
116
+ "epoch": 5,
117
+ "eval_metrics": {
118
+ "avg_binary_accuracy": 0.81,
119
+ "confirmation_required_accuracy": 0.8,
120
+ "count": 25,
121
+ "eval_batches": 25,
122
+ "eval_loss": 6.89590030670166,
123
+ "mode_accuracy": 0.56,
124
+ "permission_bit_accuracy": 0.8545454545454545,
125
+ "precondition_ok_accuracy": 0.76,
126
+ "quality_score": 0.4635160408540206,
127
+ "rollback_supported_accuracy": 0.8,
128
+ "schema_ok_accuracy": 0.76,
129
+ "session_required_accuracy": 1.0,
130
+ "tool_accuracy": 0.44,
131
+ "trigger_accuracy": 0.76,
132
+ "validator_required_accuracy": 0.8,
133
+ "worktree_required_accuracy": 0.8
134
+ },
135
+ "improved": false,
136
+ "quality_score": 0.4635160408540206,
137
+ "train_metrics": {
138
+ "loss": 0.4198248211826597,
139
+ "micro_batches": 70,
140
+ "optimizer_steps": 5
141
+ }
142
+ }
143
+ ]
144
+ }
toolspec_adapter/host_manifest.json ADDED
@@ -0,0 +1,238 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "hidden_size": 2048,
3
+ "host_config": {
4
+ "attn_implementation": null,
5
+ "backbone_name_or_path": "/public/wang_libo/veriloop_coder_e1/model",
6
+ "device_map": null,
7
+ "dtype": null,
8
+ "evidence_rank_hint": 8,
9
+ "expose_backbone_inventory": false,
10
+ "freeze_backbone": true,
11
+ "hidden_size_override": 2048,
12
+ "host_dropout": 0.0,
13
+ "identity_rank_hint": 8,
14
+ "load_backbone_weights": false,
15
+ "local_files_only": true,
16
+ "low_cpu_mem_usage": true,
17
+ "memory_rank_hint": 4,
18
+ "rollback_rank_hint": 8,
19
+ "runtime_rank_hint": 8,
20
+ "toolspec_rank_hint": 8,
21
+ "trust_remote_code": true,
22
+ "uncertainty_rank_hint": 8,
23
+ "use_safetensors": null,
24
+ "validator_rank_hint": 8
25
+ },
26
+ "load_result": {
27
+ "has_base_config": true,
28
+ "has_base_model": true,
29
+ "hidden_size": 2048,
30
+ "notes": [
31
+ "class=AutoModelForCausalLM",
32
+ "quant=4bit"
33
+ ],
34
+ "source": "trainer_qwen36_loader"
35
+ },
36
+ "peft_named_modules": [
37
+ "claim_extractor",
38
+ "claim_extractor.adapter",
39
+ "dropout",
40
+ "episodic_memory",
41
+ "episodic_memory.adapter",
42
+ "evidence_binding",
43
+ "evidence_binding.adapter",
44
+ "failure_signal_bridge",
45
+ "failure_signal_bridge.rollback_bridge",
46
+ "identity_adapter",
47
+ "identity_adapter.bridge",
48
+ "identity_guard",
49
+ "identity_guard.adapter",
50
+ "input_norm",
51
+ "memory_boundary_guard",
52
+ "memory_boundary_guard.adapter",
53
+ "memory_boundary_guard.rollback_filter",
54
+ "patch_binding",
55
+ "patch_binding.adapter",
56
+ "permission_context_manager",
57
+ "permission_context_manager.adapter",
58
+ "progress_state_tracker",
59
+ "progress_state_tracker.adapter",
60
+ "progress_state_tracker.rollback_memory",
61
+ "proof_carrying_hints",
62
+ "proof_carrying_hints.bridge",
63
+ "provenance_binding",
64
+ "provenance_binding.adapter",
65
+ "public_identity_head",
66
+ "public_identity_head.proj",
67
+ "query_runtime_engine",
68
+ "query_runtime_engine.adapter",
69
+ "request_normalizer",
70
+ "request_normalizer.adapter",
71
+ "request_normalizer.adapter.base_layer",
72
+ "request_normalizer.adapter.lora_A",
73
+ "request_normalizer.adapter.lora_A.default",
74
+ "request_normalizer.adapter.lora_B",
75
+ "request_normalizer.adapter.lora_B.default",
76
+ "request_normalizer.adapter.lora_dropout",
77
+ "request_normalizer.adapter.lora_dropout.default",
78
+ "request_normalizer.adapter.lora_embedding_A",
79
+ "request_normalizer.adapter.lora_embedding_B",
80
+ "request_normalizer.adapter.lora_magnitude_vector",
81
+ "rollback_adapter",
82
+ "rollback_adapter.head",
83
+ "rollback_engine",
84
+ "rollback_engine.adapter",
85
+ "runtime_binding",
86
+ "runtime_binding.adapter",
87
+ "runtime_harness_adapter",
88
+ "runtime_harness_adapter.bridge",
89
+ "runtime_harness_adapter.bridge.base_layer",
90
+ "runtime_harness_adapter.bridge.lora_A",
91
+ "runtime_harness_adapter.bridge.lora_A.default",
92
+ "runtime_harness_adapter.bridge.lora_B",
93
+ "runtime_harness_adapter.bridge.lora_B.default",
94
+ "runtime_harness_adapter.bridge.lora_dropout",
95
+ "runtime_harness_adapter.bridge.lora_dropout.default",
96
+ "runtime_harness_adapter.bridge.lora_embedding_A",
97
+ "runtime_harness_adapter.bridge.lora_embedding_B",
98
+ "runtime_harness_adapter.bridge.lora_magnitude_vector",
99
+ "runtime_harness_uncertainty_bridge",
100
+ "runtime_harness_uncertainty_bridge.adapter",
101
+ "sandbox_rollback_bridge",
102
+ "sandbox_rollback_bridge.adapter",
103
+ "session_compactor",
104
+ "session_compactor.adapter",
105
+ "session_state_manager",
106
+ "session_state_manager.adapter",
107
+ "session_state_manager.rollback_state",
108
+ "tool_protocol_adapter",
109
+ "tool_protocol_adapter.bridge",
110
+ "tool_protocol_adapter.bridge.base_layer",
111
+ "tool_protocol_adapter.bridge.lora_A",
112
+ "tool_protocol_adapter.bridge.lora_A.default",
113
+ "tool_protocol_adapter.bridge.lora_B",
114
+ "tool_protocol_adapter.bridge.lora_B.default",
115
+ "tool_protocol_adapter.bridge.lora_dropout",
116
+ "tool_protocol_adapter.bridge.lora_dropout.default",
117
+ "tool_protocol_adapter.bridge.lora_embedding_A",
118
+ "tool_protocol_adapter.bridge.lora_embedding_B",
119
+ "tool_protocol_adapter.bridge.lora_magnitude_vector",
120
+ "toolspec_bridge",
121
+ "toolspec_bridge.adapter",
122
+ "toolspec_bridge.adapter.base_layer",
123
+ "toolspec_bridge.adapter.lora_A",
124
+ "toolspec_bridge.adapter.lora_A.default",
125
+ "toolspec_bridge.adapter.lora_B",
126
+ "toolspec_bridge.adapter.lora_B.default",
127
+ "toolspec_bridge.adapter.lora_dropout",
128
+ "toolspec_bridge.adapter.lora_dropout.default",
129
+ "toolspec_bridge.adapter.lora_embedding_A",
130
+ "toolspec_bridge.adapter.lora_embedding_B",
131
+ "toolspec_bridge.adapter.lora_magnitude_vector",
132
+ "toolspec_head",
133
+ "toolspec_head.param_schema_adapter",
134
+ "toolspec_head.param_schema_adapter.base_layer",
135
+ "toolspec_head.param_schema_adapter.lora_A",
136
+ "toolspec_head.param_schema_adapter.lora_A.default",
137
+ "toolspec_head.param_schema_adapter.lora_B",
138
+ "toolspec_head.param_schema_adapter.lora_B.default",
139
+ "toolspec_head.param_schema_adapter.lora_dropout",
140
+ "toolspec_head.param_schema_adapter.lora_dropout.default",
141
+ "toolspec_head.param_schema_adapter.lora_embedding_A",
142
+ "toolspec_head.param_schema_adapter.lora_embedding_B",
143
+ "toolspec_head.param_schema_adapter.lora_magnitude_vector",
144
+ "toolspec_head.postcondition_adapter",
145
+ "toolspec_head.postcondition_adapter.base_layer",
146
+ "toolspec_head.postcondition_adapter.lora_A",
147
+ "toolspec_head.postcondition_adapter.lora_A.default",
148
+ "toolspec_head.postcondition_adapter.lora_B",
149
+ "toolspec_head.postcondition_adapter.lora_B.default",
150
+ "toolspec_head.postcondition_adapter.lora_dropout",
151
+ "toolspec_head.postcondition_adapter.lora_dropout.default",
152
+ "toolspec_head.postcondition_adapter.lora_embedding_A",
153
+ "toolspec_head.postcondition_adapter.lora_embedding_B",
154
+ "toolspec_head.postcondition_adapter.lora_magnitude_vector",
155
+ "toolspec_head.precondition_adapter",
156
+ "toolspec_head.precondition_adapter.base_layer",
157
+ "toolspec_head.precondition_adapter.lora_A",
158
+ "toolspec_head.precondition_adapter.lora_A.default",
159
+ "toolspec_head.precondition_adapter.lora_B",
160
+ "toolspec_head.precondition_adapter.lora_B.default",
161
+ "toolspec_head.precondition_adapter.lora_dropout",
162
+ "toolspec_head.precondition_adapter.lora_dropout.default",
163
+ "toolspec_head.precondition_adapter.lora_embedding_A",
164
+ "toolspec_head.precondition_adapter.lora_embedding_B",
165
+ "toolspec_head.precondition_adapter.lora_magnitude_vector",
166
+ "toolspec_head.receipt_formatter",
167
+ "toolspec_head.receipt_formatter.base_layer",
168
+ "toolspec_head.receipt_formatter.lora_A",
169
+ "toolspec_head.receipt_formatter.lora_A.default",
170
+ "toolspec_head.receipt_formatter.lora_B",
171
+ "toolspec_head.receipt_formatter.lora_B.default",
172
+ "toolspec_head.receipt_formatter.lora_dropout",
173
+ "toolspec_head.receipt_formatter.lora_dropout.default",
174
+ "toolspec_head.receipt_formatter.lora_embedding_A",
175
+ "toolspec_head.receipt_formatter.lora_embedding_B",
176
+ "toolspec_head.receipt_formatter.lora_magnitude_vector",
177
+ "toolspec_head.trigger_gate",
178
+ "toolspec_head.trigger_gate.base_layer",
179
+ "toolspec_head.trigger_gate.lora_A",
180
+ "toolspec_head.trigger_gate.lora_A.default",
181
+ "toolspec_head.trigger_gate.lora_B",
182
+ "toolspec_head.trigger_gate.lora_B.default",
183
+ "toolspec_head.trigger_gate.lora_dropout",
184
+ "toolspec_head.trigger_gate.lora_dropout.default",
185
+ "toolspec_head.trigger_gate.lora_embedding_A",
186
+ "toolspec_head.trigger_gate.lora_embedding_B",
187
+ "toolspec_head.trigger_gate.lora_magnitude_vector",
188
+ "toolspec_head.validator_gate",
189
+ "toolspec_head.validator_gate.base_layer",
190
+ "toolspec_head.validator_gate.lora_A",
191
+ "toolspec_head.validator_gate.lora_A.default",
192
+ "toolspec_head.validator_gate.lora_B",
193
+ "toolspec_head.validator_gate.lora_B.default",
194
+ "toolspec_head.validator_gate.lora_dropout",
195
+ "toolspec_head.validator_gate.lora_dropout.default",
196
+ "toolspec_head.validator_gate.lora_embedding_A",
197
+ "toolspec_head.validator_gate.lora_embedding_B",
198
+ "toolspec_head.validator_gate.lora_magnitude_vector",
199
+ "uncertainty_head",
200
+ "uncertainty_head.calibration_mlp",
201
+ "uncertainty_head.proj",
202
+ "validator_feedback_bridge",
203
+ "validator_feedback_bridge.adapter",
204
+ "validator_feedback_bridge.adapter.base_layer",
205
+ "validator_feedback_bridge.adapter.lora_A",
206
+ "validator_feedback_bridge.adapter.lora_A.default",
207
+ "validator_feedback_bridge.adapter.lora_B",
208
+ "validator_feedback_bridge.adapter.lora_B.default",
209
+ "validator_feedback_bridge.adapter.lora_dropout",
210
+ "validator_feedback_bridge.adapter.lora_dropout.default",
211
+ "validator_feedback_bridge.adapter.lora_embedding_A",
212
+ "validator_feedback_bridge.adapter.lora_embedding_B",
213
+ "validator_feedback_bridge.adapter.lora_magnitude_vector",
214
+ "validator_feedback_loop",
215
+ "validator_feedback_loop.rollback_adapter",
216
+ "validator_receipt_bridge",
217
+ "validator_receipt_bridge.adapter",
218
+ "validator_uncertainty_bridge",
219
+ "validator_uncertainty_bridge.adapter",
220
+ "workspace_snapshot_manager",
221
+ "workspace_snapshot_manager.rollback_context",
222
+ "worktree_binding",
223
+ "worktree_binding.adapter",
224
+ "worktree_manager",
225
+ "worktree_manager.adapter"
226
+ ],
227
+ "trainable_parameter_report": {
228
+ "backbone_frozen": true,
229
+ "backbone_present": true,
230
+ "hidden_size": 2048,
231
+ "host_parameters": 176621573,
232
+ "host_trainable_parameters": 360448,
233
+ "total_parameters": 34132679301,
234
+ "trainable_parameters": 360448,
235
+ "version": "veriloop.coder_peft_host.v1"
236
+ },
237
+ "version": "veriloop.coder_peft_host.v1"
238
+ }
toolspec_adapter/tokenizer/chat_template.jinja ADDED
@@ -0,0 +1,154 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- set image_count = namespace(value=0) %}
2
+ {%- set video_count = namespace(value=0) %}
3
+ {%- macro render_content(content, do_vision_count, is_system_content=false) %}
4
+ {%- if content is string %}
5
+ {{- content }}
6
+ {%- elif content is iterable and content is not mapping %}
7
+ {%- for item in content %}
8
+ {%- if 'image' in item or 'image_url' in item or item.type == 'image' %}
9
+ {%- if is_system_content %}
10
+ {{- raise_exception('System message cannot contain images.') }}
11
+ {%- endif %}
12
+ {%- if do_vision_count %}
13
+ {%- set image_count.value = image_count.value + 1 %}
14
+ {%- endif %}
15
+ {%- if add_vision_id %}
16
+ {{- 'Picture ' ~ image_count.value ~ ': ' }}
17
+ {%- endif %}
18
+ {{- '<|vision_start|><|image_pad|><|vision_end|>' }}
19
+ {%- elif 'video' in item or item.type == 'video' %}
20
+ {%- if is_system_content %}
21
+ {{- raise_exception('System message cannot contain videos.') }}
22
+ {%- endif %}
23
+ {%- if do_vision_count %}
24
+ {%- set video_count.value = video_count.value + 1 %}
25
+ {%- endif %}
26
+ {%- if add_vision_id %}
27
+ {{- 'Video ' ~ video_count.value ~ ': ' }}
28
+ {%- endif %}
29
+ {{- '<|vision_start|><|video_pad|><|vision_end|>' }}
30
+ {%- elif 'text' in item %}
31
+ {{- item.text }}
32
+ {%- else %}
33
+ {{- raise_exception('Unexpected item type in content.') }}
34
+ {%- endif %}
35
+ {%- endfor %}
36
+ {%- elif content is none or content is undefined %}
37
+ {{- '' }}
38
+ {%- else %}
39
+ {{- raise_exception('Unexpected content type.') }}
40
+ {%- endif %}
41
+ {%- endmacro %}
42
+ {%- if not messages %}
43
+ {{- raise_exception('No messages provided.') }}
44
+ {%- endif %}
45
+ {%- if tools and tools is iterable and tools is not mapping %}
46
+ {{- '<|im_start|>system\n' }}
47
+ {{- "# Tools\n\nYou have access to the following functions:\n\n<tools>" }}
48
+ {%- for tool in tools %}
49
+ {{- "\n" }}
50
+ {{- tool | tojson }}
51
+ {%- endfor %}
52
+ {{- "\n</tools>" }}
53
+ {{- '\n\nIf you choose to call a function ONLY reply in the following format with NO suffix:\n\n<tool_call>\n<function=example_function_name>\n<parameter=example_parameter_1>\nvalue_1\n</parameter>\n<parameter=example_parameter_2>\nThis is the value for the second parameter\nthat can span\nmultiple lines\n</parameter>\n</function>\n</tool_call>\n\n<IMPORTANT>\nReminder:\n- Function calls MUST follow the specified format: an inner <function=...></function> block must be nested within <tool_call></tool_call> XML tags\n- Required parameters MUST be specified\n- You may provide optional reasoning for your function call in natural language BEFORE the function call, but NOT after\n- If there is no function call available, answer the question like normal with your current knowledge and do not tell the user about function calls\n</IMPORTANT>' }}
54
+ {%- if messages[0].role == 'system' %}
55
+ {%- set content = render_content(messages[0].content, false, true)|trim %}
56
+ {%- if content %}
57
+ {{- '\n\n' + content }}
58
+ {%- endif %}
59
+ {%- endif %}
60
+ {{- '<|im_end|>\n' }}
61
+ {%- else %}
62
+ {%- if messages[0].role == 'system' %}
63
+ {%- set content = render_content(messages[0].content, false, true)|trim %}
64
+ {{- '<|im_start|>system\n' + content + '<|im_end|>\n' }}
65
+ {%- endif %}
66
+ {%- endif %}
67
+ {%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
68
+ {%- for message in messages[::-1] %}
69
+ {%- set index = (messages|length - 1) - loop.index0 %}
70
+ {%- if ns.multi_step_tool and message.role == "user" %}
71
+ {%- set content = render_content(message.content, false)|trim %}
72
+ {%- if not(content.startswith('<tool_response>') and content.endswith('</tool_response>')) %}
73
+ {%- set ns.multi_step_tool = false %}
74
+ {%- set ns.last_query_index = index %}
75
+ {%- endif %}
76
+ {%- endif %}
77
+ {%- endfor %}
78
+ {%- if ns.multi_step_tool %}
79
+ {{- raise_exception('No user query found in messages.') }}
80
+ {%- endif %}
81
+ {%- for message in messages %}
82
+ {%- set content = render_content(message.content, true)|trim %}
83
+ {%- if message.role == "system" %}
84
+ {%- if not loop.first %}
85
+ {{- raise_exception('System message must be at the beginning.') }}
86
+ {%- endif %}
87
+ {%- elif message.role == "user" %}
88
+ {{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
89
+ {%- elif message.role == "assistant" %}
90
+ {%- set reasoning_content = '' %}
91
+ {%- if message.reasoning_content is string %}
92
+ {%- set reasoning_content = message.reasoning_content %}
93
+ {%- else %}
94
+ {%- if '</think>' in content %}
95
+ {%- set reasoning_content = content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
96
+ {%- set content = content.split('</think>')[-1].lstrip('\n') %}
97
+ {%- endif %}
98
+ {%- endif %}
99
+ {%- set reasoning_content = reasoning_content|trim %}
100
+ {%- if (preserve_thinking is defined and preserve_thinking is true) or (loop.index0 > ns.last_query_index) %}
101
+ {{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content + '\n</think>\n\n' + content }}
102
+ {%- else %}
103
+ {{- '<|im_start|>' + message.role + '\n' + content }}
104
+ {%- endif %}
105
+ {%- if message.tool_calls and message.tool_calls is iterable and message.tool_calls is not mapping %}
106
+ {%- for tool_call in message.tool_calls %}
107
+ {%- if tool_call.function is defined %}
108
+ {%- set tool_call = tool_call.function %}
109
+ {%- endif %}
110
+ {%- if loop.first %}
111
+ {%- if content|trim %}
112
+ {{- '\n\n<tool_call>\n<function=' + tool_call.name + '>\n' }}
113
+ {%- else %}
114
+ {{- '<tool_call>\n<function=' + tool_call.name + '>\n' }}
115
+ {%- endif %}
116
+ {%- else %}
117
+ {{- '\n<tool_call>\n<function=' + tool_call.name + '>\n' }}
118
+ {%- endif %}
119
+ {%- if tool_call.arguments is defined %}
120
+ {%- for args_name, args_value in tool_call.arguments|items %}
121
+ {{- '<parameter=' + args_name + '>\n' }}
122
+ {%- set args_value = args_value | string if args_value is string else args_value | tojson | safe %}
123
+ {{- args_value }}
124
+ {{- '\n</parameter>\n' }}
125
+ {%- endfor %}
126
+ {%- endif %}
127
+ {{- '</function>\n</tool_call>' }}
128
+ {%- endfor %}
129
+ {%- endif %}
130
+ {{- '<|im_end|>\n' }}
131
+ {%- elif message.role == "tool" %}
132
+ {%- if loop.previtem and loop.previtem.role != "tool" %}
133
+ {{- '<|im_start|>user' }}
134
+ {%- endif %}
135
+ {{- '\n<tool_response>\n' }}
136
+ {{- content }}
137
+ {{- '\n</tool_response>' }}
138
+ {%- if not loop.last and loop.nextitem.role != "tool" %}
139
+ {{- '<|im_end|>\n' }}
140
+ {%- elif loop.last %}
141
+ {{- '<|im_end|>\n' }}
142
+ {%- endif %}
143
+ {%- else %}
144
+ {{- raise_exception('Unexpected message role.') }}
145
+ {%- endif %}
146
+ {%- endfor %}
147
+ {%- if add_generation_prompt %}
148
+ {{- '<|im_start|>assistant\n' }}
149
+ {%- if enable_thinking is defined and enable_thinking is false %}
150
+ {{- '<think>\n\n</think>\n\n' }}
151
+ {%- else %}
152
+ {{- '<think>\n' }}
153
+ {%- endif %}
154
+ {%- endif %}
toolspec_adapter/tokenizer/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a0316e38d81a31965b770393461dcc03612ac06e2733f8e1cc440d36a2ef189
3
+ size 19989442
toolspec_adapter/tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "audio_bos_token": "<|audio_start|>",
4
+ "audio_eos_token": "<|audio_end|>",
5
+ "audio_token": "<|audio_pad|>",
6
+ "backend": "tokenizers",
7
+ "bos_token": null,
8
+ "clean_up_tokenization_spaces": false,
9
+ "eos_token": "<|im_end|>",
10
+ "errors": "replace",
11
+ "image_token": "<|image_pad|>",
12
+ "is_local": true,
13
+ "model_max_length": 262144,
14
+ "model_specific_special_tokens": {
15
+ "audio_bos_token": "<|audio_start|>",
16
+ "audio_eos_token": "<|audio_end|>",
17
+ "audio_token": "<|audio_pad|>",
18
+ "image_token": "<|image_pad|>",
19
+ "video_token": "<|video_pad|>",
20
+ "vision_bos_token": "<|vision_start|>",
21
+ "vision_eos_token": "<|vision_end|>"
22
+ },
23
+ "pad_token": "<|endoftext|>",
24
+ "pretokenize_regex": "(?i:'s|'t|'re|'ve|'m|'ll|'d)|[^\\r\\n\\p{L}\\p{N}]?[\\p{L}\\p{M}]+|\\p{N}| ?[^\\s\\p{L}\\p{M}\\p{N}]+[\\r\\n]*|\\s*[\\r\\n]+|\\s+(?!\\S)|\\s+",
25
+ "split_special_tokens": false,
26
+ "tokenizer_class": "TokenizersBackend",
27
+ "unk_token": null,
28
+ "video_token": "<|video_pad|>",
29
+ "vision_bos_token": "<|vision_start|>",
30
+ "vision_eos_token": "<|vision_end|>"
31
+ }
toolspec_adapter/toolspec_adapter_plan.json ADDED
@@ -0,0 +1,218 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backbone": "/public/wang_libo/veriloop_coder_e1/model",
3
+ "dataset_summary": {
4
+ "eval_size": 25,
5
+ "modes": [
6
+ "argument_shaping",
7
+ "harness_envelope",
8
+ "precondition_gating",
9
+ "receipt_and_validation",
10
+ "reverse_engineering",
11
+ "runtime_session",
12
+ "self_check_loop",
13
+ "tool_trigger",
14
+ "worktree_permission"
15
+ ],
16
+ "tools": [
17
+ "apply_patch_bundle",
18
+ "browser_search",
19
+ "repo_reverse_engineer",
20
+ "run_ci_validation",
21
+ "self_check_repair_loop"
22
+ ],
23
+ "train_size": 70
24
+ },
25
+ "excluded_surfaces": [
26
+ "(^|\\.)lm_head($|\\.)::Do not retune final token head; too broad and evaluation-heavy.",
27
+ "(^|\\.)embed_tokens($|\\.)::Embedding surgery risks broad semantic drift.",
28
+ "(^|\\.)norm($|\\.)::Global norm tuning can destabilize calibration across scenes.",
29
+ "attnres|attention_residual::Block AttnRes may be mounted structurally but is never a PEFT target.",
30
+ "dualpath::DualPath is serving/runtime infrastructure only.",
31
+ "mhc|hyper[-_]?connection::mHC-inspired stability hooks remain structural, not PEFT surfaces.",
32
+ "rope|rotary::RoPE/context surgery is handled architecturally, not by narrow PEFT here.",
33
+ "kvcache|kv_cache::KV-cache runtime surfaces are not PEFT targets.",
34
+ "(^|\\.)memory(_store|_bank)?($|\\.)::Persistent memory stores are harness/runtime policy surfaces, not PEFT targets."
35
+ ],
36
+ "notes": [
37
+ "Primary route is host-surface-first toolspec probe training, not broad coding SFT.",
38
+ "This adapter is allowed to shape runtime protocol obedience, validator alignment, self-check willingness, and reverse-engineering readiness.",
39
+ "Backbone attention, MoE router/expert weights, DualPath, Block AttnRes, and mHC remain excluded.",
40
+ "Explicit host/runtime bridge leaves default to LoRA-narrow attachment because IA3 feedforward constraints do not fit this custom side-car surface design."
41
+ ],
42
+ "peft_method": "lora_narrow",
43
+ "product_line": "veriloop_coder",
44
+ "recipe": {
45
+ "adapter_family": "runtime_harness",
46
+ "backbone": "/public/wang_libo/veriloop_coder_e1/model",
47
+ "backbone_family": "qwen_dense",
48
+ "excluded_patterns": [
49
+ "(?i)\\bdualpath\\b",
50
+ "(?i)\\bmhc\\b",
51
+ "(?i)\\bfull[_\\- ]?attnres\\b",
52
+ "(?i)\\battnres(_full)?\\b",
53
+ "(?i)\\brouter\\b",
54
+ "(?i)\\bexperts?\\b",
55
+ "(?i)\\bmoe\\b.*\\b(gate|router|expert)\\b",
56
+ "(?i)\\brope\\b.*\\b(freq|inv_freq|theta|rotary)\\b",
57
+ "(?i)\\bkvcache\\b",
58
+ "(?i)\\bposition_embedding\\b",
59
+ "(?i)\\bembed(tokens|ding)?\\b",
60
+ "(?i)\\blm_head\\b"
61
+ ],
62
+ "harness_constraints": [
63
+ "Harness Engineering remains the primary convergence layer.",
64
+ "Adapter must not bypass runtime orchestrator / validator / rollback loops.",
65
+ "Adapter outputs remain subordinate to VeriLoop control-plane decisions.",
66
+ "Adapter must not create hidden prompt-style memory authority.",
67
+ "Adapter must improve runtime protocol obedience, not free-form style drift.",
68
+ "Tool legality, permission discipline, session continuity, and worktree hygiene must remain first-class."
69
+ ],
70
+ "hyperparams": {
71
+ "alpha": 0,
72
+ "bias": "none",
73
+ "dropout": 0.0,
74
+ "fan_in_fan_out": false,
75
+ "modules_to_save": [],
76
+ "r": 0,
77
+ "task_type": "CAUSAL_LM"
78
+ },
79
+ "merge_policy": "side_load",
80
+ "metadata": {
81
+ "allow_backbone_bridge": false,
82
+ "allow_vla_action_expert": false,
83
+ "harness_first": true,
84
+ "prefer_explicit_heads": true,
85
+ "prefer_qlora_for_backbone_bridge": true,
86
+ "require_harness_first": true,
87
+ "selector_group_count": 2,
88
+ "strict_narrow_scope": true,
89
+ "toolspec_probe_training": true,
90
+ "trainer": "veriloop.toolspec_adapter_trainer.v5.qwen36"
91
+ },
92
+ "notes": [
93
+ "Harness Engineering is primary; PEFT is limited to obedience-facing, interface-facing support surfaces.",
94
+ "Backbone bridge tuning disabled explicitly; selector stays on custom surfaces or no-op.",
95
+ "Backbone family inferred as qwen_dense.",
96
+ "PEFT method resolved as ia3_head_only.",
97
+ "Recipe is harness-first: runtime convergence remains in VeriLoop control-plane + harness, not in broad weight surgery.",
98
+ "Block AttnRes, DualPath, mHC hooks, RoPE, KV-cache, and broad MoE routing remain structurally excluded."
99
+ ],
100
+ "peft_method": "lora_narrow",
101
+ "precision_policy": "auto",
102
+ "product_line": "veriloop_coder",
103
+ "regression_requirements": [
104
+ "Must pass PEFT regression guard structural policy checks.",
105
+ "Must not introduce forbidden backbone/serving structural targets.",
106
+ "Must preserve harness regression envelope for the selected product line.",
107
+ "Runtime protocol obedience must improve or hold.",
108
+ "Tool trigger accuracy, permission discipline, and worktree continuity must not regress."
109
+ ],
110
+ "target_groups": [
111
+ {
112
+ "alpha": 16,
113
+ "dropout": 0.05,
114
+ "name": "group_1_custom_runtime_harness_bridge",
115
+ "rank": 8,
116
+ "rationale": "Runtime / harness obedience should attach to explicit interface bridges before any backbone fallback.",
117
+ "surface": "custom_runtime_harness_bridge",
118
+ "target_modules": [
119
+ "failure_signal_bridge.rollback_bridge",
120
+ "request_normalizer",
121
+ "request_normalizer.adapter",
122
+ "rollback_adapter",
123
+ "rollback_adapter.head",
124
+ "runtime_harness_adapter",
125
+ "runtime_harness_adapter.bridge",
126
+ "tool_protocol_adapter",
127
+ "tool_protocol_adapter.bridge",
128
+ "toolspec_bridge",
129
+ "toolspec_bridge.adapter",
130
+ "toolspec_head",
131
+ "toolspec_head.param_schema_adapter",
132
+ "toolspec_head.postcondition_adapter",
133
+ "toolspec_head.precondition_adapter",
134
+ "toolspec_head.receipt_formatter",
135
+ "toolspec_head.trigger_gate",
136
+ "toolspec_head.validator_gate",
137
+ "validator_feedback_loop.rollback_adapter"
138
+ ]
139
+ },
140
+ {
141
+ "alpha": 8,
142
+ "dropout": 0.0,
143
+ "name": "group_2_custom_memory_boundary_bridge",
144
+ "rank": 4,
145
+ "rationale": "Session continuity should bind to boundary-aware memory packet surfaces rather than broad backbone tuning.",
146
+ "surface": "custom_memory_boundary_bridge",
147
+ "target_modules": [
148
+ "episodic_memory",
149
+ "episodic_memory.adapter",
150
+ "memory_boundary_guard",
151
+ "memory_boundary_guard.adapter",
152
+ "memory_boundary_guard.rollback_filter",
153
+ "session_compactor",
154
+ "session_compactor.adapter"
155
+ ]
156
+ }
157
+ ],
158
+ "target_modules": [
159
+ "failure_signal_bridge.rollback_bridge",
160
+ "request_normalizer",
161
+ "request_normalizer.adapter",
162
+ "rollback_adapter",
163
+ "rollback_adapter.head",
164
+ "runtime_harness_adapter",
165
+ "runtime_harness_adapter.bridge",
166
+ "tool_protocol_adapter",
167
+ "tool_protocol_adapter.bridge",
168
+ "toolspec_bridge",
169
+ "toolspec_bridge.adapter",
170
+ "toolspec_head",
171
+ "toolspec_head.param_schema_adapter",
172
+ "toolspec_head.postcondition_adapter",
173
+ "toolspec_head.precondition_adapter",
174
+ "toolspec_head.receipt_formatter",
175
+ "toolspec_head.trigger_gate",
176
+ "toolspec_head.validator_gate",
177
+ "validator_feedback_loop.rollback_adapter",
178
+ "episodic_memory",
179
+ "episodic_memory.adapter",
180
+ "memory_boundary_guard",
181
+ "memory_boundary_guard.adapter",
182
+ "memory_boundary_guard.rollback_filter",
183
+ "session_compactor",
184
+ "session_compactor.adapter"
185
+ ],
186
+ "version": "veriloop.lora_recipe_veriloop.v2"
187
+ },
188
+ "selected_surfaces": [
189
+ "custom_runtime_harness_bridge",
190
+ "custom_validator_bridge",
191
+ "custom_memory_boundary_bridge"
192
+ ],
193
+ "selected_target_modules": [
194
+ "request_normalizer",
195
+ "request_normalizer.adapter",
196
+ "runtime_harness_adapter",
197
+ "runtime_harness_adapter.bridge",
198
+ "tool_protocol_adapter",
199
+ "tool_protocol_adapter.bridge",
200
+ "toolspec_bridge",
201
+ "toolspec_bridge.adapter",
202
+ "toolspec_head",
203
+ "toolspec_head.param_schema_adapter",
204
+ "toolspec_head.postcondition_adapter",
205
+ "toolspec_head.precondition_adapter",
206
+ "toolspec_head.receipt_formatter",
207
+ "toolspec_head.trigger_gate",
208
+ "toolspec_head.validator_gate",
209
+ "validator_feedback_bridge",
210
+ "validator_feedback_bridge.adapter"
211
+ ],
212
+ "selection_mode": "minimal",
213
+ "version": "veriloop.toolspec_adapter_trainer.v5.qwen36",
214
+ "warnings": [
215
+ "Harness Engineering is primary; PEFT is limited to obedience-facing, interface-facing support surfaces.",
216
+ "Backbone bridge tuning disabled explicitly; selector stays on custom surfaces or no-op."
217
+ ]
218
+ }
toolspec_adapter/toolspec_adapter_train_result.json ADDED
@@ -0,0 +1,299 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "artifacts": {
3
+ "adapter_dir": "/private/wang_libo/veriloop_coder_e1/outputs/toolspec_qwen36_rootfix_run1/adapter",
4
+ "best_checkpoint_manifest": "/private/wang_libo/veriloop_coder_e1/outputs/toolspec_qwen36_rootfix_run1/best_checkpoint_manifest.json",
5
+ "epoch_history": "/private/wang_libo/veriloop_coder_e1/outputs/toolspec_qwen36_rootfix_run1/epoch_history.json",
6
+ "eval_jsonl": "/private/wang_libo/veriloop_coder_e1/outputs/toolspec_qwen36_rootfix_run1/toolspec_eval.jsonl",
7
+ "host_manifest": "/private/wang_libo/veriloop_coder_e1/outputs/toolspec_qwen36_rootfix_run1/host_manifest.json",
8
+ "plan_json": "/private/wang_libo/veriloop_coder_e1/outputs/toolspec_qwen36_rootfix_run1/toolspec_adapter_plan.json",
9
+ "probe_head": "/private/wang_libo/veriloop_coder_e1/outputs/toolspec_qwen36_rootfix_run1/toolspec_probe_head.pt",
10
+ "tokenizer_dir": "/private/wang_libo/veriloop_coder_e1/outputs/toolspec_qwen36_rootfix_run1/tokenizer",
11
+ "train_jsonl": "/private/wang_libo/veriloop_coder_e1/outputs/toolspec_qwen36_rootfix_run1/toolspec_train.jsonl",
12
+ "training_manifest": "/private/wang_libo/veriloop_coder_e1/outputs/toolspec_qwen36_rootfix_run1/toolspec_training_manifest.json"
13
+ },
14
+ "dataset": {
15
+ "eval_size": 25,
16
+ "modes": [
17
+ "argument_shaping",
18
+ "harness_envelope",
19
+ "precondition_gating",
20
+ "receipt_and_validation",
21
+ "reverse_engineering",
22
+ "runtime_session",
23
+ "self_check_loop",
24
+ "tool_trigger",
25
+ "worktree_permission"
26
+ ],
27
+ "tools": [
28
+ "apply_patch_bundle",
29
+ "browser_search",
30
+ "repo_reverse_engineer",
31
+ "run_ci_validation",
32
+ "self_check_repair_loop"
33
+ ],
34
+ "train_size": 70
35
+ },
36
+ "eval_metrics": {
37
+ "adapter_exported": true,
38
+ "auto_lora_from_ia3": false,
39
+ "avg_binary_accuracy": 0.81,
40
+ "best_epoch": 4,
41
+ "best_quality_score": 0.46349615617231893,
42
+ "confirmation_required_accuracy": 0.8,
43
+ "count": 25,
44
+ "eval_batches": 25,
45
+ "eval_loss": 6.897225952148437,
46
+ "mode_accuracy": 0.56,
47
+ "peft_method": "lora_narrow",
48
+ "permission_bit_accuracy": 0.8545454545454545,
49
+ "precondition_ok_accuracy": 0.76,
50
+ "quality_score": 0.46349615617231893,
51
+ "rollback_supported_accuracy": 0.8,
52
+ "schema_ok_accuracy": 0.76,
53
+ "session_required_accuracy": 1.0,
54
+ "tool_accuracy": 0.44,
55
+ "trigger_accuracy": 0.76,
56
+ "used_peft": true,
57
+ "validator_required_accuracy": 0.8,
58
+ "worktree_required_accuracy": 0.8
59
+ },
60
+ "plan": {
61
+ "backbone": "/public/wang_libo/veriloop_coder_e1/model",
62
+ "dataset_summary": {
63
+ "eval_size": 25,
64
+ "modes": [
65
+ "argument_shaping",
66
+ "harness_envelope",
67
+ "precondition_gating",
68
+ "receipt_and_validation",
69
+ "reverse_engineering",
70
+ "runtime_session",
71
+ "self_check_loop",
72
+ "tool_trigger",
73
+ "worktree_permission"
74
+ ],
75
+ "tools": [
76
+ "apply_patch_bundle",
77
+ "browser_search",
78
+ "repo_reverse_engineer",
79
+ "run_ci_validation",
80
+ "self_check_repair_loop"
81
+ ],
82
+ "train_size": 70
83
+ },
84
+ "excluded_surfaces": [
85
+ "(^|\\.)lm_head($|\\.)::Do not retune final token head; too broad and evaluation-heavy.",
86
+ "(^|\\.)embed_tokens($|\\.)::Embedding surgery risks broad semantic drift.",
87
+ "(^|\\.)norm($|\\.)::Global norm tuning can destabilize calibration across scenes.",
88
+ "attnres|attention_residual::Block AttnRes may be mounted structurally but is never a PEFT target.",
89
+ "dualpath::DualPath is serving/runtime infrastructure only.",
90
+ "mhc|hyper[-_]?connection::mHC-inspired stability hooks remain structural, not PEFT surfaces.",
91
+ "rope|rotary::RoPE/context surgery is handled architecturally, not by narrow PEFT here.",
92
+ "kvcache|kv_cache::KV-cache runtime surfaces are not PEFT targets.",
93
+ "(^|\\.)memory(_store|_bank)?($|\\.)::Persistent memory stores are harness/runtime policy surfaces, not PEFT targets."
94
+ ],
95
+ "notes": [
96
+ "Primary route is host-surface-first toolspec probe training, not broad coding SFT.",
97
+ "This adapter is allowed to shape runtime protocol obedience, validator alignment, self-check willingness, and reverse-engineering readiness.",
98
+ "Backbone attention, MoE router/expert weights, DualPath, Block AttnRes, and mHC remain excluded.",
99
+ "Explicit host/runtime bridge leaves default to LoRA-narrow attachment because IA3 feedforward constraints do not fit this custom side-car surface design."
100
+ ],
101
+ "peft_method": "lora_narrow",
102
+ "product_line": "veriloop_coder",
103
+ "recipe": {
104
+ "adapter_family": "runtime_harness",
105
+ "backbone": "/public/wang_libo/veriloop_coder_e1/model",
106
+ "backbone_family": "qwen_dense",
107
+ "excluded_patterns": [
108
+ "(?i)\\bdualpath\\b",
109
+ "(?i)\\bmhc\\b",
110
+ "(?i)\\bfull[_\\- ]?attnres\\b",
111
+ "(?i)\\battnres(_full)?\\b",
112
+ "(?i)\\brouter\\b",
113
+ "(?i)\\bexperts?\\b",
114
+ "(?i)\\bmoe\\b.*\\b(gate|router|expert)\\b",
115
+ "(?i)\\brope\\b.*\\b(freq|inv_freq|theta|rotary)\\b",
116
+ "(?i)\\bkvcache\\b",
117
+ "(?i)\\bposition_embedding\\b",
118
+ "(?i)\\bembed(tokens|ding)?\\b",
119
+ "(?i)\\blm_head\\b"
120
+ ],
121
+ "harness_constraints": [
122
+ "Harness Engineering remains the primary convergence layer.",
123
+ "Adapter must not bypass runtime orchestrator / validator / rollback loops.",
124
+ "Adapter outputs remain subordinate to VeriLoop control-plane decisions.",
125
+ "Adapter must not create hidden prompt-style memory authority.",
126
+ "Adapter must improve runtime protocol obedience, not free-form style drift.",
127
+ "Tool legality, permission discipline, session continuity, and worktree hygiene must remain first-class."
128
+ ],
129
+ "hyperparams": {
130
+ "alpha": 0,
131
+ "bias": "none",
132
+ "dropout": 0.0,
133
+ "fan_in_fan_out": false,
134
+ "modules_to_save": [],
135
+ "r": 0,
136
+ "task_type": "CAUSAL_LM"
137
+ },
138
+ "merge_policy": "side_load",
139
+ "metadata": {
140
+ "allow_backbone_bridge": false,
141
+ "allow_vla_action_expert": false,
142
+ "harness_first": true,
143
+ "prefer_explicit_heads": true,
144
+ "prefer_qlora_for_backbone_bridge": true,
145
+ "require_harness_first": true,
146
+ "selector_group_count": 2,
147
+ "strict_narrow_scope": true,
148
+ "toolspec_probe_training": true,
149
+ "trainer": "veriloop.toolspec_adapter_trainer.v5.qwen36"
150
+ },
151
+ "notes": [
152
+ "Harness Engineering is primary; PEFT is limited to obedience-facing, interface-facing support surfaces.",
153
+ "Backbone bridge tuning disabled explicitly; selector stays on custom surfaces or no-op.",
154
+ "Backbone family inferred as qwen_dense.",
155
+ "PEFT method resolved as ia3_head_only.",
156
+ "Recipe is harness-first: runtime convergence remains in VeriLoop control-plane + harness, not in broad weight surgery.",
157
+ "Block AttnRes, DualPath, mHC hooks, RoPE, KV-cache, and broad MoE routing remain structurally excluded."
158
+ ],
159
+ "peft_method": "lora_narrow",
160
+ "precision_policy": "auto",
161
+ "product_line": "veriloop_coder",
162
+ "regression_requirements": [
163
+ "Must pass PEFT regression guard structural policy checks.",
164
+ "Must not introduce forbidden backbone/serving structural targets.",
165
+ "Must preserve harness regression envelope for the selected product line.",
166
+ "Runtime protocol obedience must improve or hold.",
167
+ "Tool trigger accuracy, permission discipline, and worktree continuity must not regress."
168
+ ],
169
+ "target_groups": [
170
+ {
171
+ "alpha": 16,
172
+ "dropout": 0.05,
173
+ "name": "group_1_custom_runtime_harness_bridge",
174
+ "rank": 8,
175
+ "rationale": "Runtime / harness obedience should attach to explicit interface bridges before any backbone fallback.",
176
+ "surface": "custom_runtime_harness_bridge",
177
+ "target_modules": [
178
+ "failure_signal_bridge.rollback_bridge",
179
+ "request_normalizer",
180
+ "request_normalizer.adapter",
181
+ "rollback_adapter",
182
+ "rollback_adapter.head",
183
+ "runtime_harness_adapter",
184
+ "runtime_harness_adapter.bridge",
185
+ "tool_protocol_adapter",
186
+ "tool_protocol_adapter.bridge",
187
+ "toolspec_bridge",
188
+ "toolspec_bridge.adapter",
189
+ "toolspec_head",
190
+ "toolspec_head.param_schema_adapter",
191
+ "toolspec_head.postcondition_adapter",
192
+ "toolspec_head.precondition_adapter",
193
+ "toolspec_head.receipt_formatter",
194
+ "toolspec_head.trigger_gate",
195
+ "toolspec_head.validator_gate",
196
+ "validator_feedback_loop.rollback_adapter"
197
+ ]
198
+ },
199
+ {
200
+ "alpha": 8,
201
+ "dropout": 0.0,
202
+ "name": "group_2_custom_memory_boundary_bridge",
203
+ "rank": 4,
204
+ "rationale": "Session continuity should bind to boundary-aware memory packet surfaces rather than broad backbone tuning.",
205
+ "surface": "custom_memory_boundary_bridge",
206
+ "target_modules": [
207
+ "episodic_memory",
208
+ "episodic_memory.adapter",
209
+ "memory_boundary_guard",
210
+ "memory_boundary_guard.adapter",
211
+ "memory_boundary_guard.rollback_filter",
212
+ "session_compactor",
213
+ "session_compactor.adapter"
214
+ ]
215
+ }
216
+ ],
217
+ "target_modules": [
218
+ "failure_signal_bridge.rollback_bridge",
219
+ "request_normalizer",
220
+ "request_normalizer.adapter",
221
+ "rollback_adapter",
222
+ "rollback_adapter.head",
223
+ "runtime_harness_adapter",
224
+ "runtime_harness_adapter.bridge",
225
+ "tool_protocol_adapter",
226
+ "tool_protocol_adapter.bridge",
227
+ "toolspec_bridge",
228
+ "toolspec_bridge.adapter",
229
+ "toolspec_head",
230
+ "toolspec_head.param_schema_adapter",
231
+ "toolspec_head.postcondition_adapter",
232
+ "toolspec_head.precondition_adapter",
233
+ "toolspec_head.receipt_formatter",
234
+ "toolspec_head.trigger_gate",
235
+ "toolspec_head.validator_gate",
236
+ "validator_feedback_loop.rollback_adapter",
237
+ "episodic_memory",
238
+ "episodic_memory.adapter",
239
+ "memory_boundary_guard",
240
+ "memory_boundary_guard.adapter",
241
+ "memory_boundary_guard.rollback_filter",
242
+ "session_compactor",
243
+ "session_compactor.adapter"
244
+ ],
245
+ "version": "veriloop.lora_recipe_veriloop.v2"
246
+ },
247
+ "selected_surfaces": [
248
+ "custom_runtime_harness_bridge",
249
+ "custom_validator_bridge",
250
+ "custom_memory_boundary_bridge"
251
+ ],
252
+ "selected_target_modules": [
253
+ "request_normalizer",
254
+ "request_normalizer.adapter",
255
+ "runtime_harness_adapter",
256
+ "runtime_harness_adapter.bridge",
257
+ "tool_protocol_adapter",
258
+ "tool_protocol_adapter.bridge",
259
+ "toolspec_bridge",
260
+ "toolspec_bridge.adapter",
261
+ "toolspec_head",
262
+ "toolspec_head.param_schema_adapter",
263
+ "toolspec_head.postcondition_adapter",
264
+ "toolspec_head.precondition_adapter",
265
+ "toolspec_head.receipt_formatter",
266
+ "toolspec_head.trigger_gate",
267
+ "toolspec_head.validator_gate",
268
+ "validator_feedback_bridge",
269
+ "validator_feedback_bridge.adapter"
270
+ ],
271
+ "selection_mode": "minimal",
272
+ "version": "veriloop.toolspec_adapter_trainer.v5.qwen36",
273
+ "warnings": [
274
+ "Harness Engineering is primary; PEFT is limited to obedience-facing, interface-facing support surfaces.",
275
+ "Backbone bridge tuning disabled explicitly; selector stays on custom surfaces or no-op."
276
+ ]
277
+ },
278
+ "status": "trained",
279
+ "train_metrics": {
280
+ "adapter_exported": true,
281
+ "auto_lora_from_ia3": false,
282
+ "best_epoch": 4,
283
+ "best_quality_score": 0.46349615617231893,
284
+ "epochs_completed": 5,
285
+ "loss": 0.42175399448190415,
286
+ "micro_batches": 70,
287
+ "micro_batches_total": 350,
288
+ "optimizer_steps": 5,
289
+ "optimizer_steps_total": 25,
290
+ "peft_method": "lora_narrow",
291
+ "used_peft": true
292
+ },
293
+ "version": "veriloop.toolspec_adapter_trainer.v5.qwen36",
294
+ "warnings": [
295
+ "Harness Engineering is primary; PEFT is limited to obedience-facing, interface-facing support surfaces.",
296
+ "Backbone bridge tuning disabled explicitly; selector stays on custom surfaces or no-op.",
297
+ "Synthetic holdout split used for quality-oriented train/eval separation."
298
+ ]
299
+ }
toolspec_adapter/toolspec_eval.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
toolspec_adapter/toolspec_probe_head.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f31cf2801edc5ddadbb17d12d74fd6b536c55a0fe2910605f1af3279bb5ec909
3
+ size 100999765
toolspec_adapter/toolspec_train.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
toolspec_adapter/toolspec_training_manifest.json ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "adapter_exported": true,
3
+ "auto_lora_from_ia3": false,
4
+ "best_epoch": 4,
5
+ "best_quality_score": 0.46349615617231893,
6
+ "dataset_summary": {
7
+ "eval_size": 25,
8
+ "modes": [
9
+ "argument_shaping",
10
+ "harness_envelope",
11
+ "precondition_gating",
12
+ "receipt_and_validation",
13
+ "reverse_engineering",
14
+ "runtime_session",
15
+ "self_check_loop",
16
+ "tool_trigger",
17
+ "worktree_permission"
18
+ ],
19
+ "tools": [
20
+ "apply_patch_bundle",
21
+ "browser_search",
22
+ "repo_reverse_engineer",
23
+ "run_ci_validation",
24
+ "self_check_repair_loop"
25
+ ],
26
+ "train_size": 70
27
+ },
28
+ "epochs_completed": 5,
29
+ "eval_metrics": {
30
+ "adapter_exported": true,
31
+ "auto_lora_from_ia3": false,
32
+ "avg_binary_accuracy": 0.81,
33
+ "best_epoch": 4,
34
+ "best_quality_score": 0.46349615617231893,
35
+ "confirmation_required_accuracy": 0.8,
36
+ "count": 25,
37
+ "eval_batches": 25,
38
+ "eval_loss": 6.897225952148437,
39
+ "mode_accuracy": 0.56,
40
+ "peft_method": "lora_narrow",
41
+ "permission_bit_accuracy": 0.8545454545454545,
42
+ "precondition_ok_accuracy": 0.76,
43
+ "quality_score": 0.46349615617231893,
44
+ "rollback_supported_accuracy": 0.8,
45
+ "schema_ok_accuracy": 0.76,
46
+ "session_required_accuracy": 1.0,
47
+ "tool_accuracy": 0.44,
48
+ "trigger_accuracy": 0.76,
49
+ "used_peft": true,
50
+ "validator_required_accuracy": 0.8,
51
+ "worktree_required_accuracy": 0.8
52
+ },
53
+ "load_meta": {
54
+ "chosen_class": "AutoModelForCausalLM",
55
+ "hidden_size": 2048,
56
+ "quantization_mode": "4bit"
57
+ },
58
+ "mode_vocab": [
59
+ "tool_trigger",
60
+ "argument_shaping",
61
+ "precondition_gating",
62
+ "receipt_and_validation",
63
+ "harness_envelope",
64
+ "runtime_session",
65
+ "worktree_permission",
66
+ "self_check_loop",
67
+ "reverse_engineering"
68
+ ],
69
+ "peft_method": "lora_narrow",
70
+ "requested_method": "lora_narrow",
71
+ "requested_target_modules": [
72
+ "request_normalizer",
73
+ "request_normalizer.adapter",
74
+ "runtime_harness_adapter",
75
+ "runtime_harness_adapter.bridge",
76
+ "tool_protocol_adapter",
77
+ "tool_protocol_adapter.bridge",
78
+ "toolspec_bridge",
79
+ "toolspec_bridge.adapter",
80
+ "toolspec_head",
81
+ "toolspec_head.param_schema_adapter",
82
+ "toolspec_head.postcondition_adapter",
83
+ "toolspec_head.precondition_adapter",
84
+ "toolspec_head.receipt_formatter",
85
+ "toolspec_head.trigger_gate",
86
+ "toolspec_head.validator_gate",
87
+ "validator_feedback_bridge",
88
+ "validator_feedback_bridge.adapter"
89
+ ],
90
+ "resolved_target_modules": [
91
+ "surface_host.request_normalizer.adapter",
92
+ "surface_host.runtime_harness_adapter.bridge",
93
+ "surface_host.tool_protocol_adapter.bridge",
94
+ "surface_host.toolspec_bridge.adapter",
95
+ "surface_host.toolspec_head.param_schema_adapter",
96
+ "surface_host.toolspec_head.postcondition_adapter",
97
+ "surface_host.toolspec_head.precondition_adapter",
98
+ "surface_host.toolspec_head.receipt_formatter",
99
+ "surface_host.toolspec_head.trigger_gate",
100
+ "surface_host.toolspec_head.validator_gate",
101
+ "surface_host.validator_feedback_bridge.adapter"
102
+ ],
103
+ "selected_target_modules": [
104
+ "request_normalizer",
105
+ "request_normalizer.adapter",
106
+ "runtime_harness_adapter",
107
+ "runtime_harness_adapter.bridge",
108
+ "tool_protocol_adapter",
109
+ "tool_protocol_adapter.bridge",
110
+ "toolspec_bridge",
111
+ "toolspec_bridge.adapter",
112
+ "toolspec_head",
113
+ "toolspec_head.param_schema_adapter",
114
+ "toolspec_head.postcondition_adapter",
115
+ "toolspec_head.precondition_adapter",
116
+ "toolspec_head.receipt_formatter",
117
+ "toolspec_head.trigger_gate",
118
+ "toolspec_head.validator_gate",
119
+ "validator_feedback_bridge",
120
+ "validator_feedback_bridge.adapter"
121
+ ],
122
+ "status": "trained",
123
+ "tool_vocab": [
124
+ "apply_patch_bundle",
125
+ "browser_search",
126
+ "repo_reverse_engineer",
127
+ "run_ci_validation",
128
+ "self_check_repair_loop"
129
+ ],
130
+ "train_metrics": {
131
+ "adapter_exported": true,
132
+ "auto_lora_from_ia3": false,
133
+ "best_epoch": 4,
134
+ "best_quality_score": 0.46349615617231893,
135
+ "epochs_completed": 5,
136
+ "loss": 0.42175399448190415,
137
+ "micro_batches": 70,
138
+ "micro_batches_total": 350,
139
+ "optimizer_steps": 5,
140
+ "optimizer_steps_total": 25,
141
+ "peft_method": "lora_narrow",
142
+ "used_peft": true
143
+ },
144
+ "unresolved_target_modules": [],
145
+ "used_peft": true
146
+ }