Apollo: Lightweight Multilingual Medical LLMs towards Democratizing Medical AI to 6B People
Paper β’ 2403.03640 β’ Published β’ 2
winget install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf FreedomIntelligence/Apollo-6B-GGUF:Q8_0# Run inference directly in the terminal:
llama-cli -hf FreedomIntelligence/Apollo-6B-GGUF:Q8_0# Download pre-built binary from:
# https://github.com/ggerganov/llama.cpp/releases# Start a local OpenAI-compatible server with a web UI:
./llama-server -hf FreedomIntelligence/Apollo-6B-GGUF:Q8_0# Run inference directly in the terminal:
./llama-cli -hf FreedomIntelligence/Apollo-6B-GGUF:Q8_0git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
cmake -B build
cmake --build build -j --target llama-server llama-cli# Start a local OpenAI-compatible server with a web UI:
./build/bin/llama-server -hf FreedomIntelligence/Apollo-6B-GGUF:Q8_0# Run inference directly in the terminal:
./build/bin/llama-cli -hf FreedomIntelligence/Apollo-6B-GGUF:Q8_0docker model run hf.co/FreedomIntelligence/Apollo-6B-GGUF:Q8_0Covering English, Chinese, French, Hindi, Spanish, Hindi, Arabic So far
π¨π»βπ»Github β’π Paper β’ π Demo β’ π€ ApolloCorpus β’ π€ XMedBench
δΈζ | English
π€Apollo-0.5B β’ π€ Apollo-1.8B β’ π€ Apollo-2B β’ π€ Apollo-6B β’ π€ Apollo-7B
π€ Apollo-0.5B-GGUF β’ π€ Apollo-2B-GGUF β’ π€ Apollo-6B-GGUF β’ π€ Apollo-7B-GGUF
Dataset π€ ApolloCorpus
[
"string1",
"string2",
...
]
[
[
"q1",
"a1",
"q2",
"a2",
...
],
...
]
[
[
"q1",
"a1",
"q2",
"a2",
...
],
...
]
Evaluation π€ XMedBench
EN:
ZH:
ES: Head_qa
FR: Frenchmedmcqa
HI: MMLU_HI
AR: MMLU_Ara
Waiting for Update
Please use the following citation if you intend to use our dataset for training or evaluation:
@misc{wang2024apollo,
title={Apollo: Lightweight Multilingual Medical LLMs towards Democratizing Medical AI to 6B People},
author={Xidong Wang and Nuo Chen and Junyin Chen and Yan Hu and Yidong Wang and Xiangbo Wu and Anningzhe Gao and Xiang Wan and Haizhou Li and Benyou Wang},
year={2024},
eprint={2403.03640},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
8-bit
Install from brew
# Start a local OpenAI-compatible server with a web UI: llama-server -hf FreedomIntelligence/Apollo-6B-GGUF:Q8_0# Run inference directly in the terminal: llama-cli -hf FreedomIntelligence/Apollo-6B-GGUF:Q8_0