Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
|
@@ -45,17 +45,16 @@ print("Model ready on:", "GPU" if torch.cuda.is_available() else "CPU")
|
|
| 45 |
|
| 46 |
# === Translation function ===
|
| 47 |
@spaces.GPU(duration=180)
|
| 48 |
-
def predict(text, source_language="
|
| 49 |
if not text.strip():
|
| 50 |
return "Please enter text to translate."
|
| 51 |
|
| 52 |
-
|
| 53 |
-
|
| 54 |
-
|
| 55 |
-
tgt = langs.get(target_language)
|
| 56 |
-
|
| 57 |
|
| 58 |
tokenizer.src_lang = src # only needed for some NLLB versions
|
|
|
|
| 59 |
inputs = tokenizer(
|
| 60 |
text,
|
| 61 |
return_tensors="pt",
|
|
|
|
| 45 |
|
| 46 |
# === Translation function ===
|
| 47 |
@spaces.GPU(duration=180)
|
| 48 |
+
def predict(text, source_language="eng_Latn", target_language="lug_Latn"):
|
| 49 |
if not text.strip():
|
| 50 |
return "Please enter text to translate."
|
| 51 |
|
| 52 |
+
# ✅ Directly use NLLB language codes from the UI
|
| 53 |
+
src = source_language
|
| 54 |
+
tgt = target_language
|
|
|
|
|
|
|
| 55 |
|
| 56 |
tokenizer.src_lang = src # only needed for some NLLB versions
|
| 57 |
+
|
| 58 |
inputs = tokenizer(
|
| 59 |
text,
|
| 60 |
return_tensors="pt",
|