Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
|
@@ -317,7 +317,7 @@ def run_task(text, language, task):
|
|
| 317 |
)
|
| 318 |
|
| 319 |
# IndicTrans2 expects a target language prefix token
|
| 320 |
-
|
| 321 |
inputs = indictrans_tokenizer(prefix + text, return_tensors="pt")
|
| 322 |
#inputs = indictrans_tokenizer(text, return_tensors="pt", src_lang="san", tgt_lang="en")
|
| 323 |
outputs = indictrans_model.generate(**inputs, max_new_tokens=256)
|
|
@@ -358,8 +358,8 @@ def run_task(text, language, task):
|
|
| 358 |
|
| 359 |
try:
|
| 360 |
# Load Monlam AI Tibetan→English model
|
| 361 |
-
tib_tokenizer = AutoTokenizer.from_pretrained("monlam-ai/mt-bod-eng")
|
| 362 |
-
tib_model = AutoModelForSeq2SeqLM.from_pretrained("monlam-ai/mt-bod-eng")
|
| 363 |
|
| 364 |
# Encode Tibetan input
|
| 365 |
inputs = tib_tokenizer(text, return_tensors="pt")
|
|
|
|
| 317 |
)
|
| 318 |
|
| 319 |
# IndicTrans2 expects a target language prefix token
|
| 320 |
+
prefix = "<2en> " # English target
|
| 321 |
inputs = indictrans_tokenizer(prefix + text, return_tensors="pt")
|
| 322 |
#inputs = indictrans_tokenizer(text, return_tensors="pt", src_lang="san", tgt_lang="en")
|
| 323 |
outputs = indictrans_model.generate(**inputs, max_new_tokens=256)
|
|
|
|
| 358 |
|
| 359 |
try:
|
| 360 |
# Load Monlam AI Tibetan→English model
|
| 361 |
+
tib_tokenizer = AutoTokenizer.from_pretrained("monlam-ai/mt-bod-eng", token=HF_TOKEN)
|
| 362 |
+
tib_model = AutoModelForSeq2SeqLM.from_pretrained("monlam-ai/mt-bod-eng", token=HF_TOKEN)
|
| 363 |
|
| 364 |
# Encode Tibetan input
|
| 365 |
inputs = tib_tokenizer(text, return_tensors="pt")
|