Yewei-Liu commited on
Commit
a27d9dd
·
verified ·
1 Parent(s): 8420ae4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +24 -3
README.md CHANGED
@@ -1,3 +1,24 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ library_name: transformers
4
+ pipeline_tag: text-generation
5
+ ---
6
+
7
+ # SHINE: A Scalable In-Context Hypernetwork for Mapping Context to LoRA in a Single Pass
8
+
9
+ SHINE (Scalable Hyper In-context NEtwork) is a scalable hypernetwork that can map diverse meaningful contexts into high-quality LoRA adapters for large language models (LLM).
10
+
11
+ By reusing the frozen LLM's own parameters in an in-context hypernetwork design, SHINE transforms in-context knowledge into in-parameter knowledge in a single forward pass. This allows the model to handle complex question-answering tasks related to a specific context without needing to process that context again during inference.
12
+
13
+ - **Paper:** [SHINE: A Scalable In-Context Hypernetwork for Mapping Context to LoRA in a Single Pass](https://huggingface.co/papers/2602.06358)
14
+ - **Repository:** [https://github.com/Yewei-Liu/SHINE](https://github.com/Yewei-Liu/SHINE)
15
+
16
+ ## Introduction
17
+
18
+ SHINE overcomes key limitations of prior hypernetworks by achieving strong expressive power with a relatively small number of parameters. It updates LLM parameters without any fine-tuning, significantly saving time, computation, and memory costs compared to standard supervised fine-tuning (SFT) adaptation.
19
+
20
+ ## Usage
21
+
22
+ This is the hypernetwork checkpoint after pretraining and instruction fine-tuning mqa.
23
+
24
+ For detailed instructions on environment setup, downloading model checkpoints, and performing inference (including the `inference.ipynb` notebook), please refer to the [official GitHub repository](https://github.com/Yewei-Liu/SHINE).