Spaces:
Running
Running
| from taskAI import TaskAI | |
| from taskNonAI import compile_pdf | |
| from _data_test import mock_cv, pdf_context | |
| from _secret import api_test | |
| from llama_index.llms.openai_like import OpenAILike | |
| from llama_index.core.llms import ChatMessage | |
| def test_ai_integration(): | |
| messages = [ | |
| ChatMessage(role="system", content="You are a helpful assistant"), | |
| ChatMessage(role="user", content="What is your name"), | |
| ] | |
| print("Testing integration:") | |
| response = OpenAILike( | |
| model=api_test["model"], | |
| api_key=api_test["key"], | |
| api_base=api_test["base"], | |
| max_retries=0, | |
| is_chat_model=True, | |
| ).chat(messages) | |
| print(response) | |
| def test_taskAI(): | |
| taskAI = TaskAI(api_test) | |
| gen = taskAI.cv_preprocess(mock_cv) | |
| for chunk in gen: | |
| print(chunk) | |
| def test_typst_pdf(): | |
| compile_pdf( | |
| tmpl_path="typst/template_letter.tmpl", | |
| context=pdf_context, | |
| output_path="test_result.pdf", | |
| ) | |
| # os | |
| if __name__ == "__main__": | |
| # test_taskAI() | |
| # test_ai_integration() | |
| test_typst_pdf() | |