Marketeer / test_llm_backend.py
KPrashanth's picture
Upload 47 files
85ba912 verified
raw
history blame contribute delete
720 Bytes
from core_logic.llm_config import get_local_chat_model
from langchain_core.messages import HumanMessage
def main():
# Load the ChatHuggingFace model (cached after first call)
chat_model = get_local_chat_model()
# Instead of SystemMessage, fold instructions into the human message
prompt = (
"You are a friendly marketing copywriter.\n\n"
"Write a short, fun one-line ad for a coffee shop."
)
messages = [
HumanMessage(content=prompt),
]
# Call the model
ai_msg = chat_model.invoke(messages)
# Show the result
print("Response type:", type(ai_msg))
print("Response content:\n")
print(ai_msg.content)
if __name__ == "__main__":
main()