RicardoRei commited on
Commit
f7e4a8d
·
verified ·
1 Parent(s): 812fb96

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -78,7 +78,7 @@ sampling_params = SamplingParams(
78
  max_tokens=8192,
79
  )
80
  llm = LLM(model="Unbabel/Tower-Plus-2B", tensor_parallel_size=1)
81
- messages = [{"role": "user", "content": "Translate: Hello, world! into Portuguese."}]
82
  outputs = llm.chat(messages, sampling_params)
83
  # Make sure your prompt_token_ids look like this
84
  print (outputs[0].outputs[0].text)
@@ -96,7 +96,7 @@ from transformers import pipeline
96
 
97
  pipe = pipeline("text-generation", model="Unbabel/Tower-Plus-2B", device_map="auto")
98
  # We use the tokenizer’s chat template to format each message - see https://huggingface.co/docs/transformers/main/en/chat_templating
99
- messages = [{"role": "user", "content": "Translate: Hello, world! into Portuguese."}]
100
  input_ids = pipe.tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True)
101
  outputs = pipe(messages, max_new_tokens=256, do_sample=False)
102
  print(outputs[0]["generated_text"])
 
78
  max_tokens=8192,
79
  )
80
  llm = LLM(model="Unbabel/Tower-Plus-2B", tensor_parallel_size=1)
81
+ messages = [{"role": "user", "content": "Translate the following English source text to Portuguese (Portugal):\nEnglish: Hello world!\nPortuguese (Portugal): "}]
82
  outputs = llm.chat(messages, sampling_params)
83
  # Make sure your prompt_token_ids look like this
84
  print (outputs[0].outputs[0].text)
 
96
 
97
  pipe = pipeline("text-generation", model="Unbabel/Tower-Plus-2B", device_map="auto")
98
  # We use the tokenizer’s chat template to format each message - see https://huggingface.co/docs/transformers/main/en/chat_templating
99
+ messages = [{"role": "user", "content": "Translate the following English source text to Portuguese (Portugal):\nEnglish: Hello world!\nPortuguese (Portugal): "}]
100
  input_ids = pipe.tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True)
101
  outputs = pipe(messages, max_new_tokens=256, do_sample=False)
102
  print(outputs[0]["generated_text"])