File size: 2,495 Bytes
dfc4c47
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
llama.cpp
β”œβ”€β”€ bin
β”‚   β”œβ”€β”€ cublas64_13.dll
β”‚   β”œβ”€β”€ cublasLt64_13.dll
β”‚   β”œβ”€β”€ cudart64_13.dll
β”‚   β”œβ”€β”€ ggml-base.dll
β”‚   β”œβ”€β”€ ggml-cpu-alderlake.dll
β”‚   β”œβ”€β”€ ggml-cpu-cannonlake.dll
β”‚   β”œβ”€β”€ ggml-cpu-cascadelake.dll
β”‚   β”œβ”€β”€ ggml-cpu-cooperlake.dll
β”‚   β”œβ”€β”€ ggml-cpu-haswell.dll
β”‚   β”œβ”€β”€ ggml-cpu-icelake.dll
β”‚   β”œβ”€β”€ ggml-cpu-ivybridge.dll
β”‚   β”œβ”€β”€ ggml-cpu-piledriver.dll
β”‚   β”œβ”€β”€ ggml-cpu-sandybridge.dll
β”‚   β”œβ”€β”€ ggml-cpu-sapphirerapids.dll
β”‚   β”œβ”€β”€ ggml-cpu-skylakex.dll
β”‚   β”œβ”€β”€ ggml-cpu-sse42.dll
β”‚   β”œβ”€β”€ ggml-cpu-x64.dll
β”‚   β”œβ”€β”€ ggml-cpu-zen4.dll
β”‚   β”œβ”€β”€ ggml-cuda.dll
β”‚   β”œβ”€β”€ ggml-rpc.dll
β”‚   β”œβ”€β”€ ggml.dll
β”‚   β”œβ”€β”€ libomp140.x86_64.dll
β”‚   β”œβ”€β”€ llama-batched-bench.exe
β”‚   β”œβ”€β”€ llama-bench.exe
β”‚   β”œβ”€β”€ llama-cli.exe
β”‚   β”œβ”€β”€ llama-completion.exe
β”‚   β”œβ”€β”€ llama-fit-params.exe
β”‚   β”œβ”€β”€ llama-gemma3-cli.exe
β”‚   β”œβ”€β”€ llama-gguf-split.exe
β”‚   β”œβ”€β”€ llama-imatrix.exe
β”‚   β”œβ”€β”€ llama-llava-cli.exe
β”‚   β”œβ”€β”€ llama-minicpmv-cli.exe
β”‚   β”œβ”€β”€ llama-mtmd-cli.exe
β”‚   β”œβ”€β”€ llama-perplexity.exe
β”‚   β”œβ”€β”€ llama-quantize.exe
β”‚   β”œβ”€β”€ llama-qwen2vl-cli.exe
β”‚   β”œβ”€β”€ llama-results.exe
β”‚   β”œβ”€β”€ llama-server.exe
β”‚   β”œβ”€β”€ llama-template-analysis.exe
β”‚   β”œβ”€β”€ llama-tokenize.exe
β”‚   β”œβ”€β”€ llama-tts.exe
β”‚   β”œβ”€β”€ llama.dll
β”‚   β”œβ”€β”€ mtmd.dll
β”‚   └── rpc-server.exe
β”œβ”€β”€ doc
β”‚   β”œβ”€β”€ llama-server.help.en.txt
β”‚   β”œβ”€β”€ llama-server.help.zh.md
β”‚   └── llama.cpp_tools_server_readme.md
β”œβ”€β”€ get_latest_release.py
β”œβ”€β”€ llama-server.run.translation2.bat
β”œβ”€β”€ llama-server.run.vision.bat
β”œβ”€β”€ llama-server_hy-mt1.5-7b.bat
β”œβ”€β”€ llama-server_vision_gemma-3-27b-it.bat
β”œβ”€β”€ llama-server_vision_glm-4.6v-flash.bat
β”œβ”€β”€ llama-server_vision_jan-v2-vl-max-fp8.bat
β”œβ”€β”€ llama-server_vision_mars-27b-v1.bat
β”œβ”€β”€ llama-server_vision_qwen3-vl-8b-instruct-c_abliterated-v3.bat
β”œβ”€β”€ llama-server_vision_qwen3.5-27b.bat
β”œβ”€β”€ llama-server_vision_qwen3.5-9b.bat
β”œβ”€β”€ llama-server_vision_test.bat
β”œβ”€β”€ llama-server_vision_youtu-vl-4b-instruct.bat
β”œβ”€β”€ llama-server_vision_zwz-4b.bat
β”œβ”€β”€ system-prompt-en-ocr.txt
β”œβ”€β”€ system-prompt-test.txt
β”œβ”€β”€ system-prompt-zh-ocr.txt
└── system-prompt.txt