Xeltron-cloud commited on
Commit
33a9cfb
·
verified ·
1 Parent(s): 9e657e3

Create app.py

Browse files
Files changed (1) hide show
  1. app.py +68 -0
app.py ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from fastapi import FastAPI, HTTPException
2
+ from pydantic import BaseModel
3
+ from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
4
+ from huggingface_hub import login
5
+ import os
6
+ import torch
7
+ import uvicorn
8
+
9
+ login(os.getenv("HF_TOKEN"))
10
+
11
+ app = FastAPI(
12
+ title="VexaAI Model-Platform: NVIDIA Nemotron-Nano-9B-V2",
13
+ description="Self-hosted AI-Model NVIDIA Nemotron-Nano-9B-V2, powered by VexaAI.",
14
+ version="0.9"
15
+ )
16
+
17
+ model_name = "nvidia/NVIDIA-Nemotron-Nano-9B-v2"
18
+
19
+ bnb_config = BitsAndBytesConfig(
20
+ load_in_4bit=True,
21
+ bnb_4bit_use_double_quant=True,
22
+ bnb_4bit_quant_type="nf4",
23
+ bnb_4bit_compute_dtype=torch.bfloat16
24
+ )
25
+
26
+ tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
27
+ model = AutoModelForCausalLM.from_pretrained(
28
+ model_name,
29
+ quantization_config=bnb_config,
30
+ device_map="auto",
31
+ trust_remote_code=True,
32
+ torch_dtype=torch.bfloat16
33
+ )
34
+ model.eval()
35
+
36
+ class GenerateRequest(BaseModel):
37
+ prompt: str
38
+ max_new_tokens: int = 512
39
+ temperature: float = 0.7
40
+
41
+ @app.post("/generate")
42
+ async def generate_text(request: GenerateRequest):
43
+ try:
44
+ inputs = tokenizer(request.prompt, return_tensors="pt").to(model.device)
45
+
46
+ with torch.no_grad():
47
+ outputs = model.generate(
48
+ **inputs,
49
+ max_new_tokens=request.max_new_tokens,
50
+ temperature=request.temperature,
51
+ do_sample=True,
52
+ eos_token_id=tokenizer.eos_token_id,
53
+ pad_token_id=tokenizer.eos_token_id
54
+ )
55
+
56
+ full_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
57
+ generated_text = full_text[len(tokenizer.decode(inputs["input_ids"][0], skip_special_tokens=True)):].strip()
58
+
59
+ return {"generated_text": generated_text}
60
+ except Exception as e:
61
+ raise HTTPException(status_code=500, detail=f"VexaAI Model_Platform: HTTP/S error: {str(e)}")
62
+
63
+ @app.get("/")
64
+ async def root():
65
+ return {"message": "To start generating text, use /generate."}
66
+
67
+ if __name__ == "__main__":
68
+ uvicorn.run(app, host="0.0.0.0", port=7860)