Update app.py
Browse files
app.py
CHANGED
|
@@ -1,6 +1,7 @@
|
|
| 1 |
import gradio as gr
|
| 2 |
from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
|
| 3 |
import torch
|
|
|
|
| 4 |
|
| 5 |
MODEL_ID = "userdotcs/gpt-oss-20b-turkish-correction-finetuned"
|
| 6 |
BASE_MODEL = "unsloth/gpt-oss-20b" # veya base repo adı neyse
|
|
@@ -18,6 +19,7 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
| 18 |
trust_remote_code=True
|
| 19 |
)
|
| 20 |
|
|
|
|
| 21 |
def generate(prompt):
|
| 22 |
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
|
| 23 |
out = model.generate(
|
|
|
|
| 1 |
import gradio as gr
|
| 2 |
from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
|
| 3 |
import torch
|
| 4 |
+
import spaces
|
| 5 |
|
| 6 |
MODEL_ID = "userdotcs/gpt-oss-20b-turkish-correction-finetuned"
|
| 7 |
BASE_MODEL = "unsloth/gpt-oss-20b" # veya base repo adı neyse
|
|
|
|
| 19 |
trust_remote_code=True
|
| 20 |
)
|
| 21 |
|
| 22 |
+
@spaces.GPU
|
| 23 |
def generate(prompt):
|
| 24 |
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
|
| 25 |
out = model.generate(
|