Update README.md
Browse files
README.md
CHANGED
|
@@ -19,6 +19,20 @@ Built on top of the **OmniSQL-7B** base model and finetuned on the **BIRD** data
|
|
| 19 |
- **Enhancing Text-to-SQL Pipelines:** Integrate as a reward or reranking model to improve execution accuracy and semantic fidelity in Text-to-SQL systems.
|
| 20 |
- **Evaluation and Research:** Analyze the semantic alignment of SQL queries with natural language questions to identify and mitigate errors.
|
| 21 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 22 |
## Usage Example
|
| 23 |
|
| 24 |
```python
|
|
|
|
| 19 |
- **Enhancing Text-to-SQL Pipelines:** Integrate as a reward or reranking model to improve execution accuracy and semantic fidelity in Text-to-SQL systems.
|
| 20 |
- **Evaluation and Research:** Analyze the semantic alignment of SQL queries with natural language questions to identify and mitigate errors.
|
| 21 |
|
| 22 |
+
## Finetuning Configuration
|
| 23 |
+
|
| 24 |
+
The following **LoRA configuration** was used to train this model:
|
| 25 |
+
|
| 26 |
+
- **R**: `16` (rank of the low-rank matrices)
|
| 27 |
+
- **Lora Alpha**: `64` (scaling factor for the low-rank update)
|
| 28 |
+
- **Target Modules**: `{q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj}`
|
| 29 |
+
- **Lora Dropout**: `0.05`
|
| 30 |
+
- **Bias**: `"none"` (bias terms are frozen)
|
| 31 |
+
- **FP16**: `True` (half-precision training)
|
| 32 |
+
- **Learning Rate**: `7e-5`
|
| 33 |
+
- **Train Batch Size**: `5`
|
| 34 |
+
- **Num. Train Epochs**: `50`
|
| 35 |
+
|
| 36 |
## Usage Example
|
| 37 |
|
| 38 |
```python
|