| # License: CC BY-NC-SA 4.0. Rights belong to Javad Taghia (taghia.javad@gmail.com). | |
| name: deeai | |
| channels: | |
| # Use conda-forge for up-to-date builds of Python and libs. | |
| - conda-forge | |
| dependencies: | |
| # Base interpreter; Python 3.10 has broad wheel support across ML libs. | |
| - python=3.10 | |
| # Core tooling and a clean pip inside the env. | |
| - pip | |
| - pip: | |
| # Core model + tokenizer stack. | |
| - transformers>=4.44 | |
| - datasets>=2.19 | |
| # Parameter-efficient finetuning (LoRA). | |
| - peft>=0.11 | |
| # Multi-GPU/accelerator launcher + config helper. | |
| - accelerate>=0.33 | |
| # 4-bit quantization backend for laptop-friendly training (CUDA required). | |
| # 0.42 is the latest widely available pip release. | |
| - bitsandbytes==0.42.0 | |
| # Logging + experiment tracking. | |
| - wandb>=0.17 | |
| # Env loader so secrets stay in .env, not code. | |
| - python-dotenv>=1.0 | |
| # Safer weight loading (preferred over torch.load .bin) | |
| - safetensors>=0.4 | |
| # Tokenizer dependency for LLaMA/Tulu models. | |
| - sentencepiece>=0.1.99 | |
| # Optional: small utilities. | |
| - tqdm>=4.66 | |
| - scipy>=1.11 | |