How to use from the
Use from the
llama-cpp-python library
# !pip install llama-cpp-python

from llama_cpp import Llama

llm = Llama.from_pretrained(
	repo_id="PixelPanda1/WellMinded_Therapy_Engine-gguf",
	filename="psychologistv2-8.0B-Q4_0.gguf",
)
llm.create_chat_completion(
	messages = "No input example has been defined for this model task."
)

WellMinded Therapy Engine (8B Q4_0)

This is a fine-tuned version of the LLaMA 3.1 8B model, optimized for psychologist-like conversations. The model is quantized to 4-bit precision (Q4_0) for efficient inference.

Usage

You can load and use this model with the llama-cpp-python library:

from llama_cpp import Llama

llm = Llama(model_path="psychologistv2-8.0B-Q4_0.gguf")
response = llm("Hi, I'm feeling stressed lately.")
print(response["choices"][0]["text"])
Downloads last month
18
GGUF
Model size
8B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for PixelPanda1/WellMinded_Therapy_Engine-gguf

Quantized
(323)
this model