HelpingAI Series
Collection
Our Emotionally intelligent Models โข 7 items โข Updated โข 2
docker run --gpus all \
--shm-size 32g \
-p 30000:30000 \
-v ~/.cache/huggingface:/root/.cache/huggingface \
--env "HF_TOKEN=<secret>" \
--ipc=host \
lmsysorg/sglang:latest \
python3 -m sglang.launch_server \
--model-path "HelpingAI/HelpingAI-3" \
--host 0.0.0.0 \
--port 30000# Call the server using curl (OpenAI-compatible API):
curl -X POST "http://localhost:30000/v1/chat/completions" \
-H "Content-Type: application/json" \
--data '{
"model": "HelpingAI/HelpingAI-3",
"messages": [
{
"role": "user",
"content": "What is the capital of France?"
}
]
}'HelpingAI3 is an advanced language model developed to excel in emotionally intelligent conversations. Building upon the foundations of HelpingAI2.5, this model offers enhanced emotional understanding and contextual awareness.
HelpingAI3 was trained on a diverse dataset comprising:
The model underwent the following training processes:
HelpingAI3 is designed for:
While HelpingAI3 strives for high emotional intelligence, users should be aware of potential limitations:
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer
# Load the HelpingAI3 model
model = AutoModelForCausalLM.from_pretrained("HelpingAI/HelpingAI-3")
# Load the tokenizer
tokenizer = AutoTokenizer.from_pretrained("HelpingAI/HelpingAI-3")
# Define the chat input
chat = [
{"role": "system", "content": "You are HelpingAI, an emotional AI. Always answer my questions in the HelpingAI style."},
{"role": "user", "content": "Introduce yourself."}
]
inputs = tokenizer.apply_chat_template(
chat,
add_generation_prompt=True,
return_tensors="pt"
).to(model.device)
# Generate text
outputs = model.generate(
inputs,
max_new_tokens=256,
do_sample=True,
temperature=0.6,
top_p=0.9,
)
response = outputs[0][inputs.shape[-1]:]
print(tokenizer.decode(response, skip_special_tokens=True))
Install from pip and serve model
# Install SGLang from pip: pip install sglang# Start the SGLang server: python3 -m sglang.launch_server \ --model-path "HelpingAI/HelpingAI-3" \ --host 0.0.0.0 \ --port 30000# Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "HelpingAI/HelpingAI-3", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }'