| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | import json |
| | import os |
| | import tempfile |
| |
|
| | from transformers.cli.chat import new_chat_history, parse_generate_flags, save_chat |
| |
|
| |
|
| | def test_help(cli): |
| | output = cli("chat", "--help") |
| | assert output.exit_code == 0 |
| | assert "Chat with a model from the command line." in output.output |
| |
|
| |
|
| | def test_save_and_clear_chat(): |
| | with tempfile.TemporaryDirectory() as tmp_path: |
| | filename = os.path.join(tmp_path, "chat.json") |
| | save_chat(filename, [{"role": "user", "content": "hi"}], {"foo": "bar"}) |
| | assert os.path.isfile(filename) |
| | with open(filename, "r") as f: |
| | data = json.load(f) |
| | assert data["chat_history"] == [{"role": "user", "content": "hi"}] |
| | assert data["settings"] == {"foo": "bar"} |
| |
|
| |
|
| | def test_new_chat_history(): |
| | assert new_chat_history() == [] |
| | assert new_chat_history("prompt") == [{"role": "system", "content": "prompt"}] |
| |
|
| |
|
| | def test_parse_generate_flags(): |
| | parsed = parse_generate_flags(["temperature=0.5", "max_new_tokens=10"]) |
| | assert parsed["temperature"] == 0.5 |
| | assert parsed["max_new_tokens"] == 10 |
| |
|