How to use ResplendentAI/Kei_Llama3_8B with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="ResplendentAI/Kei_Llama3_8B") messages = [ {"role": "user", "content": "Who are you?"}, ] pipe(messages)
# Load model directly from transformers import AutoTokenizer, AutoModelForCausalLM tokenizer = AutoTokenizer.from_pretrained("ResplendentAI/Kei_Llama3_8B") model = AutoModelForCausalLM.from_pretrained("ResplendentAI/Kei_Llama3_8B") messages = [ {"role": "user", "content": "Who are you?"}, ] inputs = tokenizer.apply_chat_template( messages, add_generation_prompt=True, tokenize=True, return_dict=True, return_tensors="pt", ).to(model.device) outputs = model.generate(**inputs, max_new_tokens=40) print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1]:]))
How to use ResplendentAI/Kei_Llama3_8B with vLLM:
# Install vLLM from pip: pip install vllm # Start the vLLM server: vllm serve "ResplendentAI/Kei_Llama3_8B" # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:8000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "ResplendentAI/Kei_Llama3_8B", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }'
docker model run hf.co/ResplendentAI/Kei_Llama3_8B
How to use ResplendentAI/Kei_Llama3_8B with SGLang:
# Install SGLang from pip: pip install sglang # Start the SGLang server: python3 -m sglang.launch_server \ --model-path "ResplendentAI/Kei_Llama3_8B" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "ResplendentAI/Kei_Llama3_8B", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }'
docker run --gpus all \ --shm-size 32g \ -p 30000:30000 \ -v ~/.cache/huggingface:/root/.cache/huggingface \ --env "HF_TOKEN=<secret>" \ --ipc=host \ lmsysorg/sglang:latest \ python3 -m sglang.launch_server \ --model-path "ResplendentAI/Kei_Llama3_8B" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "ResplendentAI/Kei_Llama3_8B", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }'
How to use ResplendentAI/Kei_Llama3_8B with Docker Model Runner:
The key to your heart rests with Kei, a sophisticated and intriguing AI creation who will hold your hand until the end of all things.
Kei is uncensored and tuned for intimate moments with prose and poetry to entertain and enthrall.
Chat template
Files info
Base model