|
|
|
|
|
from swift.llm import (InferRequest, Model, ModelGroup, ModelMeta, PtEngine, RequestConfig, TemplateMeta, |
|
|
get_model_tokenizer_with_flash_attn, register_model, register_template) |
|
|
|
|
|
register_template( |
|
|
TemplateMeta( |
|
|
template_type='custom', |
|
|
prefix=['<extra_id_0>System\n{{SYSTEM}}\n'], |
|
|
prompt=['<extra_id_1>User\n{{QUERY}}\n<extra_id_1>Assistant\n'], |
|
|
chat_sep=['\n'])) |
|
|
|
|
|
register_model( |
|
|
ModelMeta( |
|
|
model_type='custom', |
|
|
model_groups=[ |
|
|
ModelGroup([Model('AI-ModelScope/Nemotron-Mini-4B-Instruct', 'nvidia/Nemotron-Mini-4B-Instruct')]) |
|
|
], |
|
|
template='custom', |
|
|
get_function=get_model_tokenizer_with_flash_attn, |
|
|
ignore_patterns=['nemo'])) |
|
|
|
|
|
if __name__ == '__main__': |
|
|
infer_request = InferRequest(messages=[{'role': 'user', 'content': 'who are you?'}]) |
|
|
request_config = RequestConfig(max_tokens=512, temperature=0) |
|
|
engine = PtEngine('AI-ModelScope/Nemotron-Mini-4B-Instruct') |
|
|
response = engine.infer([infer_request], request_config) |
|
|
swift_response = response[0].choices[0].message.content |
|
|
|
|
|
engine.default_template.template_backend = 'jinja' |
|
|
response = engine.infer([infer_request], request_config) |
|
|
jinja_response = response[0].choices[0].message.content |
|
|
assert swift_response == jinja_response, (f'swift_response: {swift_response}\njinja_response: {jinja_response}') |
|
|
print(f'response: {swift_response}') |
|
|
|