cccc commited on
Commit
1245f95
·
1 Parent(s): 2de404e

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +12 -10
app.py CHANGED
@@ -36,18 +36,20 @@ def sentiment_analysis(sentence, model_name):
36
  "CCCC/RoBERTa_English_FinancialNews_tuned":"roberta",
37
  }
38
 
39
- tokenizer = AutoTokenizer.from_pretrained("Helsinki-NLP/opus-mt-zh-en")
40
- model = AutoModelForSeq2SeqLM.from_pretrained("Helsinki-NLP/opus-mt-zh-en")
41
-
42
- translated_tokens = model.generate(
43
- **tokenizer(sentences, return_tensors="pt", padding=True)
44
- )
45
- sentences_list = []
46
- for t in translated_tokens:
47
- sentences_list.append(tokenizer.decode(t, skip_special_tokens=True))
 
 
48
 
49
  testdata = []
50
- for i,sentence in enumerate(sentences_list):
51
  testdata.append(InputExample(guid=i,text_a=sentence,label=0))
52
 
53
  plm, tokenizer, model_config, WrapperClass = load_plm(type_dic[model_name], model_name)
 
36
  "CCCC/RoBERTa_English_FinancialNews_tuned":"roberta",
37
  }
38
 
39
+ if 'Chinese' in modelname:
40
+ tokenizer = AutoTokenizer.from_pretrained("Helsinki-NLP/opus-mt-zh-en")
41
+ model = AutoModelForSeq2SeqLM.from_pretrained("Helsinki-NLP/opus-mt-zh-en")
42
+
43
+ translated_tokens = model.generate(
44
+ **tokenizer(sentences, return_tensors="pt", padding=True)
45
+ )
46
+ sentences = []
47
+ for t in translated_tokens:
48
+ sentences.append(tokenizer.decode(t, skip_special_tokens=True))
49
+
50
 
51
  testdata = []
52
+ for i,sentence in enumerate(sentences):
53
  testdata.append(InputExample(guid=i,text_a=sentence,label=0))
54
 
55
  plm, tokenizer, model_config, WrapperClass = load_plm(type_dic[model_name], model_name)