Commit History
Create phi3-ft-fsdp.yml (#1580) a82a711 unverified
Phi-3 conversation format, example training script and perplexity metric (#1582) cf64284 unverified
fix lint issue that snuck through (#1665) f5febc7 unverified
Fix Lora config error for Llama3 (#1659) 230e0ac unverified
Generalizing the chat_template prompt strategy (#1660) [skip ci] cc11c6b unverified
Keith Stevens commited on
Fix Google Colab notebook 2024-05 (#1662) [skip ci] 5f91064 unverified
Maciek commited on
Update tiny-llama qlora.yml addressing eval packing error (#1638) 84bb806 unverified
Jaydeep Thik commited on
update outputs path so that we can mount workspace to /workspace/data (#1623) 4fde300 unverified
Add ORPO example and e2e test (#1572) 98c25e1 unverified
fix(yml): update llama-3 config (#1543) [skip ci] 0e8f340 unverified
fix broken linting (#1541) c10563c unverified
Adding Llama-3 qlora (#1536) 37c037c unverified
llama-3 examples (#1537) 15f7910 unverified
Create mixtral_22.yml (#1514) [skip ci] 0eadfc8 unverified
Atlas commited on
DBRX Model Support (#1462) 132eb74 unverified
Fix the wrong adapter in qwen2-moe-qlora example (#1501) [skip ci] 7f17eff unverified
qwen2_moe support w multipack (#1455) 6086be8 unverified
fix some of the edge cases for Jamba (#1452) 05b398a unverified
Jamba (#1451) 02af082 unverified
turn sample_packing on for training (#1438) [skip ci] c19d060 unverified
chore(config): refactor old mistral config (#1435) f1ebaa0 unverified
strip out hacky qlora-fsdp workarounds now that qlora-fsdp fixes are upstreamed (#1428) 2a1589f unverified
Fix Gemma 7b qlora.yml (#1405) 6366b0c unverified
Train parameters exclusively in specific ranges (#1390) 05bcc9e unverified
FDSP + QLoRA (#1378) 9b6ee83 unverified
Update tinyllama lora.yml to fix eval packing issue (#1362) 8984bf1 unverified
chore: enable sample_packing for Gemma (#1351) 170d4d7 unverified
Add StableLM 2 Example Scripts (#1327) [skip ci] f30d062 unverified
multipack for gemma (#1313) 2752d5f unverified
Adding Google's gemma Model (#1312) 9e300ac unverified
Add instructions for playing with qlora model to colab example (#1290) 6ab69ec unverified
fix(examples): remove is_*_derived as it's parsed automatically (#1297) a7a9a14 unverified
Add seq2seq eval benchmark callback (#1274) 5a5d474 unverified
Add MPS support (#1264) fac2d98 unverified
lock pytorch (#1247) [skip ci] 1c7ed26 unverified
JohanWork commited on
Pretrain transforms (#1261) c7cf381 unverified
Peft lotfq (#1222) 4cb7900 unverified
Update qlora.yml - remove `max_packed_sequence_len` (#1210) [skip ci] 5407ddd unverified
add colab example (#1196) [skip ci] ee0b5f6 unverified
JohanWork commited on