anthonym21 commited on
Commit
454d51a
·
verified ·
1 Parent(s): c3677cc

Fix generate.py: correct repo name and add HF usage example

Browse files
Files changed (1) hide show
  1. generate.py +11 -3
generate.py CHANGED
@@ -3,10 +3,18 @@ Eve-2-MoE Inference
3
  ===================
4
  Quick generation script. Works with local weights or HuggingFace download.
5
 
6
- Usage:
7
  python generate.py --prompt "The future of AI is"
8
  python generate.py --prompt "The future of AI is" --model_path ./model_final/pytorch_model.bin
9
- python generate.py --prompt "The future of AI is" --hf_repo anthonym21/Eve-2-MoE-250M
 
 
 
 
 
 
 
 
10
  """
11
 
12
  import argparse
@@ -71,7 +79,7 @@ def main():
71
  args = p.parse_args()
72
 
73
  if not args.model_path and not args.hf_repo:
74
- args.hf_repo = "anthonym21/Eve-2-MoE-250M"
75
 
76
  print(f"Loading model on {args.device}...")
77
  model = load_model(args.model_path, args.hf_repo, args.device)
 
3
  ===================
4
  Quick generation script. Works with local weights or HuggingFace download.
5
 
6
+ Usage (standalone):
7
  python generate.py --prompt "The future of AI is"
8
  python generate.py --prompt "The future of AI is" --model_path ./model_final/pytorch_model.bin
9
+ python generate.py --prompt "The future of AI is" --hf_repo anthonym21/Eve-2-MoE-272M
10
+
11
+ Usage (HuggingFace):
12
+ from transformers import AutoModelForCausalLM, AutoTokenizer
13
+ model = AutoModelForCausalLM.from_pretrained("anthonym21/Eve-2-MoE-272M", trust_remote_code=True)
14
+ tokenizer = AutoTokenizer.from_pretrained("gpt2")
15
+ inputs = tokenizer("The future of AI is", return_tensors="pt")
16
+ output = model.generate(**inputs, max_new_tokens=100)
17
+ print(tokenizer.decode(output[0]))
18
  """
19
 
20
  import argparse
 
79
  args = p.parse_args()
80
 
81
  if not args.model_path and not args.hf_repo:
82
+ args.hf_repo = "anthonym21/Eve-2-MoE-272M"
83
 
84
  print(f"Loading model on {args.device}...")
85
  model = load_model(args.model_path, args.hf_repo, args.device)