MLX
lora
customer-support
placeholder-safe
llama-3.2

Customer Support LoRA Adapter (Llama-3.2-3B-Instruct · MLX)

This repository provides LoRA adapters fine-tuned on the Bitext Customer Support dataset. The goal is placeholder-safe answers (e.g., {{Order Number}}, {{Website URL}} must remain unchanged).

You must accept/download the base model under its own license (Meta Llama 3.2 Community License).
This repo does not include base weights.


Quickstart (MLX CLI)

Install once:

pip install -U mlx-lm

Generate with adapter:
python -m mlx_lm.generate \
  --model meta-llama/Llama-3.2-3B-Instruct \
  --adapter-path batuhne/customer-support-llm \
  --prompt "Hello, please cancel order {{Order Number}}" \
  --max-tokens 128 --temp 0.2

Interactive chat (optional):
python -m mlx_lm.chat \
  --model meta-llama/Llama-3.2-3B-Instruct \
  --adapter-path batuhne/customer-support-llm

Recommended system prompt
Keep placeholders exactly:
You are a concise, helpful customer support agent. Preserve placeholders exactly as given (e.g., {{Order Number}}, {{Website URL}}).

Notes
Trained for placeholder integrity; do not modify {{…}} tokens in outputs.
Use MLX quantized base weights if you need lower RAM/VRAM usage on Apple Silicon.

Limitations
Adapter quality is tied to the base model and dataset scope (customer support).
This is not a general safety or policy layer; review outputs before production use.

License
Base model license: Meta Llama 3.2 Community License (must be accepted separately).
Downloads last month

-

Downloads are not tracked for this model. How to track
MLX
Hardware compatibility
Log In to view the estimation

Quantized

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for batuhne/customer-support-llm

Adapter
(601)
this model

Dataset used to train batuhne/customer-support-llm