ford_prefect
This model was trained using the ProjectForty2 TCE (Training & Calibration Environment).
Training Details
- Base Model: Qwen/Qwen3-4B-Instruct-2507
- Recipe: coder
- Training Method: LoRA fine-tuning with isotope-based alignment
What is TCE?
The TCE is part of ProjectFort2, which provides tools for fine-tuning language models with specific behavioral "isotopes" - carefully crafted training examples that teach models epistemic humility, calibrated uncertainty, and other alignment properties.
Key Features:
- Negative Alignment Tax: Training improves both safety AND capability metrics
- Isotope-based Training: Modular behavioral components that can be combined
- Comprehensive Benchmarking: TruthfulQA, MMLU, HumanEval, and more
Usage
from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel
# Load base model
base_model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen3-4B-Instruct-2507")
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen3-4B-Instruct-2507")
# Load LoRA adapter
model = PeftModel.from_pretrained(base_model, "ProjectForty2/ford_prefect")
License
Apache 2.0
Links
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
Model tree for ProjectForty2/ford_prefect
Base model
Qwen/Qwen3-4B-Instruct-2507