Wontonious's picture
test
779a90b
raw
history blame contribute delete
No virus
1.39 kB
# Load model directly
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
from flask import Flask, request
app = Flask(__name__)
app.config.from_pyfile('settings.py')
@app.route('/')
def index():
print(app.config["HUGGINGFACE_TOKEN"])
return "ISA Project Flask Server"
@app.post('/translate')
def translate():
article_en = request.form['original_text']
translate_code = request.form['translate_code']
access_token = app.config["HUGGINGFACE_TOKEN"]
tokenizer = AutoTokenizer.from_pretrained("SnypzZz/Llama2-13b-Language-translate", token="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")
model = AutoModelForSeq2SeqLM.from_pretrained("SnypzZz/Llama2-13b-Language-translate", token="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")
model_inputs = tokenizer(article_en, return_tensors="pt")
# translate from English
generated_tokens = model.generate(
**model_inputs,
forced_bos_token_id=tokenizer.lang_code_to_id[translate_code]
)
translated_sentence = tokenizer.batch_decode(generated_tokens, skip_special_tokens=True)
return translated_sentence[0]
# English (en_XX),
# Spanish (es_XX),
# French (fr_XX),
# Japanese (ja_XX),
# Korean (ko_KR),
# Russian (ru_RU)
# Vietnamese (vi_VN),
# Chinese (zh_CN),
# Mongolian (mn_MN),
# Urdu (ur_PK)
if __name__ == "__main__":
app.run()