Instructions to use ATL-Machine/affine-test-01 with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- PEFT
How to use ATL-Machine/affine-test-01 with PEFT:
Task type is invalid.
- Notebooks
- Google Colab
- Kaggle
| license: apache-2.0 | |
| base_model: ineso22/affine-vscode | |
| tags: | |
| - lora | |
| - peft | |
| - qwen | |
| - causal-lm | |
| # out_affine | |
| This model is a fine-tuned version of [ineso22/affine-vscode](https://huggingface.co/ineso22/affine-vscode) using LoRA (Low-Rank Adaptation). | |
| ## Model Details | |
| - **Base Model**: ineso22/affine-vscode | |
| - **Training Method**: LoRA (Low-Rank Adaptation) | |
| - **LoRA Rank**: 16 | |
| - **LoRA Alpha**: 32 | |
| - **LoRA Dropout**: 0.05 | |
| - **Quantization**: 4-bit QLoRA | |
| - **Precision**: bfloat16 | |
| ## Usage | |
| ```python | |
| from transformers import AutoModelForCausalLM, AutoTokenizer | |
| import torch | |
| model = AutoModelForCausalLM.from_pretrained( | |
| "out_affine/merged_model", | |
| torch_dtype=torch.bfloat16, | |
| trust_remote_code=True, | |
| device_map="auto" | |
| ) | |
| tokenizer = AutoTokenizer.from_pretrained("out_affine/merged_model") | |
| # Your inference code here | |
| ``` | |