Text Generation
Transformers
Safetensors
granitemoehybrid
CoT
reasoning
Hoglet-33 commited on
Commit
cf43bcd
·
verified ·
1 Parent(s): 6402a0c

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +128 -0
README.md ADDED
@@ -0,0 +1,128 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ tags:
4
+ - CoT
5
+ - reasoning
6
+ license: apache-2.0
7
+ datasets:
8
+ - Lucid-Research/advanced-reasoning-v1-smol
9
+ base_model:
10
+ - ibm-granite/granite-4.0-1b-base
11
+ ---
12
+ # LucentLogico
13
+
14
+ LucentLogico is a family of compact reasoning-specialized language models developed by Lucent Research. Each model is fine-tuned from IBM Granite 4.0 base architectures and optimized for structured, multi-step analytical reasoning.
15
+
16
+ The LucentLogico series focuses on mathematical derivation, algorithmic code reasoning, and formal logic tasks, with explicit intermediate reasoning steps emphasized during training.
17
+
18
+ ---
19
+
20
+ ## Model Variants
21
+
22
+ ### LucentLogico-3B
23
+ - Base: ibm-granite/granite-4.0-micro-base
24
+ - Parameter Class: ~3B
25
+ - Target Use: High-capacity compact reasoning systems
26
+
27
+ ### LucentLogico-1B
28
+ - Base: ibm-granite/granite-4.0-1b-base
29
+ - Parameter Class: ~1B
30
+ - Target Use: Efficient reasoning with reduced compute requirements
31
+
32
+ ### LucentLogico-350M
33
+ - Base: ibm-granite/granite-4.0-350m-base
34
+ - Parameter Class: ~350M
35
+ - Target Use: Lightweight reasoning experimentation and edge deployment
36
+
37
+ All variants are trained using the same reasoning-focused dataset and training philosophy, scaled to their respective parameter sizes.
38
+
39
+ ---
40
+
41
+ ## Training Dataset
42
+
43
+ All LucentLogico models were fine-tuned on:
44
+
45
+ **Lucid-Research/advanced-reasoning-v1-smol**
46
+ 36,000 curated instruction–response pairs dedicated exclusively to advanced reasoning.
47
+
48
+ ### Dataset Composition
49
+
50
+ The dataset is a balanced tri-domain blend.
51
+
52
+ #### Mathematical Reasoning (12,000 samples)
53
+ Source: MetaMathQA
54
+
55
+ - Multi-step mathematical derivations
56
+ - Symbolic manipulation
57
+ - Competition-style reasoning problems
58
+ - Explicit step-by-step solutions
59
+
60
+ #### Code & Algorithmic Reasoning (12,000 samples)
61
+ Sources: Magicoder-OSS-Instruct-75K, CodeAlpaca-20k
62
+
63
+ - Natural language specification to code
64
+ - Algorithm design tasks
65
+ - Debugging and refinement examples
66
+ - Structured execution planning
67
+
68
+ #### Formal Logic & STEM Reasoning (12,000 samples)
69
+ Source: SlimOrca
70
+
71
+ - Logic puzzles
72
+ - Proof-style reasoning
73
+ - Scientific inference
74
+ - Multi-hop structured deduction
75
+
76
+ ---
77
+
78
+ ## Design Principles
79
+
80
+ The LucentLogico series was trained with the following priorities:
81
+
82
+ - Explicit intermediate reasoning in every example
83
+ - Balanced cross-domain analytical capability
84
+ - Reduced reasoning drift
85
+ - Structured decomposition of complex problems
86
+ - Standardized instruction–response formatting
87
+
88
+ The training dataset deliberately excludes conversational and alignment-focused data in order to maintain strict specialization in reasoning performance.
89
+
90
+ ---
91
+
92
+ ## Intended Use
93
+
94
+ LucentLogico models are designed for:
95
+
96
+ - Step-by-step mathematical reasoning
97
+ - Algorithmic code synthesis
98
+ - Logical deduction and proof-style analysis
99
+ - Technical reasoning systems
100
+ - Educational analytical applications
101
+
102
+ ---
103
+
104
+ ## Limitations
105
+
106
+ - Not optimized for general conversation or roleplay
107
+ - May produce verbose outputs due to step-emphasis training
108
+ - Not fine-tuned for alignment or preference modeling
109
+ - Outputs should be validated before production use
110
+
111
+ ---
112
+
113
+ ## Attribution and Licensing
114
+
115
+ LucentLogico models were fine-tuned on **Lucid-Research/advanced-reasoning-v1-smol**, which incorporates or derives from:
116
+
117
+ - MetaMathQA
118
+ - Magicoder-OSS-Instruct-75K
119
+ - CodeAlpaca-20k
120
+ - SlimOrca
121
+
122
+ Users are responsible for complying with the original licenses of all upstream datasets.
123
+
124
+ Each LucentLogico variant follows the licensing terms of its respective IBM Granite base model:
125
+
126
+ - ibm-granite/granite-4.0-micro-base
127
+ - ibm-granite/granite-4.0-1b-base
128
+ - ibm-granite/granite-4.0-350m-base