DigitalDaimyo commited on
Commit
ed347d9
·
verified ·
1 Parent(s): 1c272e0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -20
README.md CHANGED
@@ -16,33 +16,23 @@ Interpretable slot-based attention achieving competitive language modeling perfo
16
  ## Quick Start
17
 
18
  ```python
19
- !wget https://huggingface.co/DigitalDaimyo/AddressedStateAttention/resolve/main/asm_analysis.py
20
- !wget https://huggingface.co/DigitalDaimyo/AddressedStateAttention/resolve/main/universal_loader.py
21
 
22
- from universal_loader import load_asm_checkpoint
23
- from huggingface_hub import hf_hub_download
24
 
25
- # Download checkpoint
26
- ckpt_path = hf_hub_download(
27
- repo_id="DigitalDaimyo/AddressedStateAttention",
28
- filename="checkpoints/fineweb_187M_75k.pt"
29
  )
30
 
31
- # Load model
32
- model, cfg, ckpt = load_asm_checkpoint(ckpt_path, mode="analysis")
33
 
34
  # Generate text
35
- from transformers import AutoTokenizer
36
- from generation import generate
37
 
38
- tokenizer = AutoTokenizer.from_pretrained("gpt2")
39
- text = generate(
40
- model, tokenizer,
41
- prompt="The capital of France is",
42
- max_new_tokens=20,
43
- strategy="greedy"
44
- )
45
- print(text)
46
  Performance
47
  FineWeb, 187M params: 3.73 val loss / 41.6 PPL (75k steps)
48
  Architecture: 21 layers, 768d, 12 heads, 16 slots
 
16
  ## Quick Start
17
 
18
  ```python
19
+ # Install directly from GitHub
20
+ !pip install git+https://github.com/DigitalDaimyo/AddressedStateAttention.git
21
 
22
+ from asa import load_asm_checkpoint, generate
23
+ from transformers import AutoTokenizer
24
 
25
+ # Load checkpoint
26
+ model, cfg, ckpt = load_asm_checkpoint(
27
+ "path/to/checkpoint.pt",
28
+ mode="analysis"
29
  )
30
 
31
+ tokenizer = AutoTokenizer.from_pretrained("gpt2")
 
32
 
33
  # Generate text
34
+ print(generate(model, tokenizer, "Once upon a time"))
 
35
 
 
 
 
 
 
 
 
 
36
  Performance
37
  FineWeb, 187M params: 3.73 val loss / 41.6 PPL (75k steps)
38
  Architecture: 21 layers, 768d, 12 heads, 16 slots