Nitin2004 commited on
Commit
8888bc6
Β·
verified Β·
1 Parent(s): 6cf89ed

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +96 -0
README.md ADDED
@@ -0,0 +1,96 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - meta-learning
5
+ - lora
6
+ - checkpoints
7
+ - few-shot-learning
8
+ - llm
9
+ - qwen
10
+ library_name: peft
11
+ datasets:
12
+ - ARC
13
+ - HellaSwag
14
+ - BoolQ
15
+ - PIQA
16
+ - WinoGrande
17
+ - SocialIQA
18
+ ---
19
+
20
+ # DeGAML-LLM Checkpoints
21
+
22
+ This repository contains pre-trained LoRA adapter checkpoints for the **DeGAML-LLM** framework - a novel meta-learning approach that decouples generalization and adaptation for Large Language Models.
23
+
24
+ ## πŸ”— Links
25
+
26
+ - **Project Page**: [https://nitinvetcha.github.io/DeGAML-LLM/](https://nitinvetcha.github.io/DeGAML-LLM/)
27
+ - **GitHub Repository**: [https://github.com/nitinvetcha/DeGAML-LLM](https://github.com/nitinvetcha/DeGAML-LLM)
28
+ - **HuggingFace Profile**: [https://huggingface.co/Nitin2004](https://huggingface.co/Nitin2004)
29
+
30
+ ## πŸ“¦ Available Checkpoints
31
+
32
+ All checkpoints are trained on **Qwen2.5-0.5B-Instruct** using LoRA adapters optimized with the DeGAML-LLM framework:
33
+
34
+ | Checkpoint Name | Dataset | Size |
35
+ |----------------|---------|------|
36
+ | `qwen0.5lora__ARC-c.pth` | ARC-Challenge | ~4.45 GB |
37
+ | `qwen0.5lora__ARC-e.pth` | ARC-Easy | ~4.45 GB |
38
+ | `qwen0.5lora__BoolQ.pth` | BoolQ | ~4.45 GB |
39
+ | `qwen0.5lora__HellaSwag.pth` | HellaSwag | ~4.45 GB |
40
+ | `qwen0.5lora__PIQA.pth` | PIQA | ~4.45 GB |
41
+ | `qwen0.5lora__SocialIQA.pth` | SocialIQA | ~4.45 GB |
42
+ | `qwen0.5lora__WinoGrande.pth` | WinoGrande | ~4.45 GB |
43
+
44
+ ## πŸš€ Usage
45
+
46
+ ### Download
47
+
48
+ ```python
49
+ from huggingface_hub import hf_hub_download
50
+
51
+ # Download a specific checkpoint
52
+ checkpoint_path = hf_hub_download(
53
+ repo_id="Nitin2004/DeGAML-LLM-checkpoints",
54
+ filename="qwen0.5lora__ARC-c.pth"
55
+ )
56
+ ```
57
+
58
+ ### Load with PyTorch
59
+
60
+ ```python
61
+ import torch
62
+
63
+ # Load the checkpoint
64
+ checkpoint = torch.load(checkpoint_path)
65
+ print(checkpoint.keys())
66
+ ```
67
+
68
+ ### Use with DeGAML-LLM
69
+
70
+ Refer to the [DeGAML-LLM repository](https://github.com/nitinvetcha/DeGAML-LLM) for detailed usage instructions on how to integrate these checkpoints with the framework.
71
+
72
+ ## πŸ“Š Performance
73
+
74
+ These checkpoints achieve state-of-the-art results on common-sense reasoning tasks when used with the DeGAML-LLM adaptation framework. See the [project page](https://nitinvetcha.github.io/DeGAML-LLM/) for complete benchmark results.
75
+
76
+ ## πŸ“„ Citation
77
+
78
+ If you use these checkpoints in your research, please cite:
79
+
80
+ ```bibtex
81
+ @article{degaml-llm2025,
82
+ title={Decoupling Generalization and Adaptation in Meta-Learning for Large Language Models},
83
+ author={Vetcha, Nitin and Xu, Binqian and Liu, Dianbo},
84
+ year={2025}
85
+ }
86
+ ```
87
+
88
+ ## πŸ“§ Contact
89
+
90
+ For questions or issues, please:
91
+ - Open an issue on [GitHub](https://github.com/nitinvetcha/DeGAML-LLM/issues)
92
+ - Contact: nitinvetcha@gmail.com
93
+
94
+ ## πŸ“œ License
95
+
96
+ Apache License 2.0 - See [LICENSE](https://github.com/nitinvetcha/DeGAML-LLM/blob/main/LICENSE) for details.