You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Model Card for F1-24B-Base

F1-24B-BaseFormosa-1(F1) 系列中的 24B 基底模型,以 mistralai/Mistral-Small-24B-Base-2501 為基底,針對繁體中文與中華民國台灣語境完成大規模持續預訓練(continued pretraining, CPT),作為 F1-24B-InstructF1-24B-Reasoner 等下游模型的繁中底座。

⚠️ 規格重點: 本模型為 24B 參數、純文本單模態,僅做 CPT、未做指令微調,需自行 SFT 後才有對話能力。

Model Details

Mistral-Small-24B 是兼顧能力與部署成本的中型開源模型,但其原生繁中與台灣語境覆蓋不足。F1-24B-Base 將 Formosa-1 系列的繁中語料注入 Mistral-Small-24B 之參數,使下游任務能在熟悉繁中與台灣語境之底座上做後續訓練。

核心特點 (Key Features)

  1. 24B 級繁中底座:在能力與部署成本之間取得平衡,可作為企業級繁中模型的微調起點。
  2. F1 系列家族基底:作為 Instruct、Reasoner、領域微調(如 F1-24B-Instruct-Cybersecurity)的共同底座。
  3. 大規模 CPT:以多卡分散式訓練於繁中語料完成持續預訓練。

Model Description

Model Sources

Citation

@misc{f1_24b_base,
  title        = {F1-24B-Base: A Traditional Chinese Continued-Pretrained Mistral-24B Model for Taiwan},
  author       = {Huang, Liang Hsun},
  year         = {2025},
  howpublished = {\url{https://huggingface.co/lianghsun/F1-24B-Base}}
}

Acknowledge

  • 特此感謝 APMIC 的算力支援。

Model Card Authors

Huang Liang Hsun

Model Card Contact

Huang Liang Hsun

Downloads last month
-
Safetensors
Model size
24B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for lianghsun/F1-24B-Base

Finetuned
(43)
this model
Finetunes
3 models