Update README.md
Browse files
README.md
CHANGED
|
@@ -14,13 +14,10 @@ For model inference, please download our release package from this url https://g
|
|
| 14 |
|
| 15 |
## Quick start
|
| 16 |
|
| 17 |
-
```
|
| 18 |
# Use a local model file
|
| 19 |
llama-cli -m my_model.gguf
|
| 20 |
|
| 21 |
-
# Or download and run a model directly from Hugging Face
|
| 22 |
-
llama-cli -hf inclusionAI/Ling-mini-2.0-GGUF
|
| 23 |
-
|
| 24 |
# Launch OpenAI-compatible API server
|
| 25 |
llama-server -m my_model.gguf
|
| 26 |
```
|
|
|
|
| 14 |
|
| 15 |
## Quick start
|
| 16 |
|
| 17 |
+
```bash
|
| 18 |
# Use a local model file
|
| 19 |
llama-cli -m my_model.gguf
|
| 20 |
|
|
|
|
|
|
|
|
|
|
| 21 |
# Launch OpenAI-compatible API server
|
| 22 |
llama-server -m my_model.gguf
|
| 23 |
```
|