| --- |
| base_model: zai-org/GLM-4.6 |
| tags: |
| - rust |
| - Hyperswitch |
| - LoRA |
| - CPT |
| - Fine-Tuned |
| - Causal-LM |
| pipeline_tag: text-generation |
| language: |
| - en |
| datasets: |
| - AdityaNarayan/HyperSwitch-Repo-CPT-Dataset |
| --- |
| # GLM-4.6-CPT-LoRA-HyperSwitch-v1 |
|
|
| A LoRA fine-tuned model based on **zai-org/GLM-4.6** specialized for the [Hyperswitch](https://github.com/juspay/hyperswitch) Rust codebase. This model excels at understanding payment processing patterns, Hyperswitch architecture, and Rust development practices. |
|
|
| ## π― Model Description |
|
|
| This LoRA adapter was trained on **16,731 samples** extracted from the Hyperswitch codebase to enhance code understanding, explanation, and generation within the payment processing domain. |
|
|
| - **Base Model**: zai-org/GLM-4.6 |
| - **Training Type**: Causal Language Modeling (CLM) with LoRA |
| - **Domain**: Payment Processing, Rust Development |
| - **Specialization**: Hyperswitch codebase patterns and architecture |
|
|
| ## π Training Details |
|
|
| ### LoRA Configuration |
| ```yaml |
| r: 16 # LoRA rank |
| alpha: 32 # LoRA alpha (2*r) |
| dropout: 0.05 # LoRA dropout |
| target_modules: |
| - "q_proj" |
| - "k_proj" |
| - "v_proj" |
| - "o_proj" |
| |
| exclude_modules: |
| - "block_sparse_moe" |
| - "w1" |
| - "w2" |
| - "w3" |
| - "gate" |
| ``` |
|
|
| ### Training Hyperparameters |
| - **Epochs**: 5 |
| - **Learning Rate**: 2e-4 (cosine schedule) |
| - **Hardware**: 8 x NVIDIA H200 |
|
|
|
|
| ## π οΈ Technical Specifications |
|
|
| - **Precision**: bfloat16 |
| - **Inference Speed**: Optimized with Flash Attention 2 |
|
|
| ## π Acknowledgments |
|
|
| - **Zai Team** for the excellent GLM 4.6 base model |
| - **Hyperswitch Team** for the open-source payment processing platform |
| - **Hugging Face** for the transformers and PEFT libraries |
|
|
| ## π Citation |
|
|
| ```bibtex |
| @misc{GLM-4.6-CPT-LoRA-HyperSwitch-v2, |
| title={AdityaNarayan/GLM-4.6-CPT-LoRA-HyperSwitch-v2}, |
| author={Aditya Narayan}, |
| year={2024}, |
| publisher={Hugging Face}, |
| url={https://huggingface.co/AdityaNarayan/GLM-4.6-CPT-LoRA-HyperSwitch-v2} |
| } |
| ``` |