EuroswarmsInstitute commited on
Commit
4142335
·
verified ·
1 Parent(s): acafb98

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +91 -84
README.md CHANGED
@@ -1,92 +1,99 @@
1
  ---
2
- library_name: transformers
 
3
  license: other
4
- base_model: Euroswarms/CR-CA
 
 
5
  tags:
6
- - generated_from_trainer
7
- datasets:
8
- - mmlu.jsonl
9
- model-index:
10
- - name: outputs/crca2
11
- results: []
12
  ---
13
 
14
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
- should probably proofread and complete it, then remove this comment. -->
16
-
17
- [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
18
- <details><summary>See axolotl config</summary>
19
-
20
- axolotl version: `0.13.0.dev0`
21
- ```yaml
22
- base_model: Euroswarms/CR-CA
23
- bf16: auto
24
-
25
- datasets:
26
- - path: mmlu.jsonl
27
- type: chat_template
28
-
29
- message_field: messages
30
-
31
- micro_batch_size: 1
32
- gradient_accumulation_steps: 8
33
-
34
- sequence_len: 4096
35
- gradient_checkpointing: true
36
-
37
- learning_rate: 0.0001
38
- optimizer: adamw_bnb_8bit
39
-
40
- train_on_inputs: true
41
- num_epochs: 2
42
- output_dir: ./outputs/crca2
 
 
 
 
 
 
43
 
 
 
44
  ```
45
 
46
- </details><br>
47
-
48
- # outputs/crca2
49
-
50
- This model is a fine-tuned version of [Euroswarms/CR-CA](https://huggingface.co/Euroswarms/CR-CA) on the mmlu.jsonl dataset.
51
-
52
- ## Model description
53
-
54
- More information needed
55
-
56
- ## Intended uses & limitations
57
-
58
- More information needed
59
-
60
- ## Training and evaluation data
61
-
62
- More information needed
63
-
64
- ## Training procedure
65
-
66
- ### Training hyperparameters
67
-
68
- The following hyperparameters were used during training:
69
- - learning_rate: 0.0001
70
- - train_batch_size: 1
71
- - eval_batch_size: 1
72
- - seed: 42
73
- - distributed_type: multi-GPU
74
- - num_devices: 2
75
- - gradient_accumulation_steps: 8
76
- - total_train_batch_size: 16
77
- - total_eval_batch_size: 2
78
- - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
79
- - lr_scheduler_type: cosine
80
- - lr_scheduler_warmup_steps: 2
81
- - training_steps: 49
82
-
83
- ### Training results
84
-
85
-
86
-
87
- ### Framework versions
88
-
89
- - Transformers 4.57.0
90
- - Pytorch 2.7.1+cu126
91
- - Datasets 4.0.0
92
- - Tokenizers 0.22.1
 
1
  ---
2
+ language:
3
+ - en
4
  license: other
5
+ base_model: Qwen/Qwen2.5-1.5B-Instruct
6
+ library_name: transformers
7
+ pipeline_tag: text-generation
8
  tags:
9
+ - crca
10
+ - causal-reasoning
11
+ - qwen2
12
+ - 1.5b
13
+ - finetuned
 
14
  ---
15
 
16
+ # CRCA 1.5B Full Finetune
17
+
18
+ ## Overview
19
+ CR-CA (Causal Reasoning and Counterfactual Analysis) is a reasoning-focused stack
20
+ that targets structured causal analysis, counterfactuals, and multi-step reasoning.
21
+ This 1.5B model is a CR-CA reasoning-optimized causal language model based on the
22
+ Qwen2 architecture (`Qwen2ForCausalLM`).
23
+
24
+ ## Model Details
25
+ - **Model type:** `qwen2`
26
+ - **Architecture:** `Qwen2ForCausalLM`
27
+ - **Hidden size:** `1536`
28
+ - **Layers:** `28`
29
+ - **Attention heads:** `12` (KV heads: `2`)
30
+ - **Max position embeddings:** `32768`
31
+ - **Vocab size:** `151936`
32
+ - **Dtype:** `float16`
33
+
34
+ ## Training Summary
35
+ This model was produced via full finetuning for CR-CA reasoning. Training metadata
36
+ is stored in `training_args.bin`.
37
+
38
+ Key training parameters:
39
+ - **Per-device batch size:** 8
40
+ - **Gradient accumulation:** 16
41
+ - **Epochs:** 2
42
+ - **Learning rate:** 5e-4
43
+ - **Precision:** FP16
44
+ - **DeepSpeed config:** `training/deepspeed_zero2_1_5b.json`
45
+ - **Scheduler:** cosine
46
+ - **Warmup steps:** 100
47
+ - **Save steps:** 200
48
+
49
+ ## Training Data
50
+ The training data uses a prompt/response JSONL format:
51
 
52
+ ```
53
+ {"prompt": "...", "response": "..."}
54
  ```
55
 
56
+ The dataset includes public reasoning data (e.g., GSM8K-style math word problems).
57
+ This is used to strengthen multi-step reasoning, structured derivations, and final
58
+ answer formatting.
59
+
60
+ ## Evaluation Report (Real-World Causal Tasks)
61
+ Evaluation was run on 2026-02-01 using GPT-4o-mini over 6 real-world causal tasks.
62
+ Overall score: **48.3%**.
63
+
64
+ Per-task scores:
65
+ - Monetary Policy Counterfactual (US Macro 2025): **55/100**
66
+ - Tariff Pass-Through and Pricing (Beige Book + Firm Data): **55/100**
67
+ - Supply Chain Reroute Counterfactual (Port Disruption): **45/100**
68
+ - Inventory & Stockout Causal Impact (Retail): **25/100**
69
+ - Inflation Drivers (World Bank CPI Data): **65/100**
70
+ - Workforce Training Program (Labor Market Causal Impact): **45/100**
71
+
72
+ Key strengths observed:
73
+ - Clear task framing and attempt at counterfactual reasoning.
74
+ - Some identification of confounders and causal factors.
75
+
76
+ Key limitations observed:
77
+ - Inconsistent causal graphs and directional effects.
78
+ - Weak counterfactual grounding and numerical reasoning errors.
79
+ - Limited depth and rigor on confounder adjustment strategies.
80
+
81
+ ## Intended Use
82
+ For causal reasoning, counterfactual analysis, structured CR-CA reasoning prompts,
83
+ and multi-step reasoning tasks.
84
+
85
+ ## Generation Settings
86
+ Default generation parameters are stored in `generation_config.json`:
87
+ - `do_sample`: `true`
88
+ - `temperature`: `0.7`
89
+ - `top_p`: `0.8`
90
+ - `top_k`: `20`
91
+ - `repetition_penalty`: `1.1`
92
+
93
+ ## Limitations
94
+ - Outputs should be validated for factual correctness.
95
+ - The model may hallucinate causal claims without evidence.
96
+
97
+ ## License
98
+ Follow the base model and dataset licenses used for training. Add your explicit
99
+ license here if required.