FutureMa commited on
Commit
76ea049
·
verified ·
1 Parent(s): 2d84296

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +82 -0
README.md ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: apache-2.0
5
+ base_model: FutureMa/Eva-4B
6
+ tags:
7
+ - finance
8
+ - earnings-calls
9
+ - financial-nlp
10
+ - text-classification
11
+ - qwen3
12
+ - llama-cpp
13
+ - gguf
14
+ - gguf-my-repo
15
+ pipeline_tag: text-generation
16
+ library_name: transformers
17
+ ---
18
+
19
+ # FutureMa/Eva-4B-GGUF
20
+
21
+ This repository hosts **GGUF** files for [`FutureMa/Eva-4B`](https://huggingface.co/FutureMa/Eva-4B), intended for use with [`llama.cpp`](https://github.com/ggerganov/llama.cpp).
22
+
23
+ - **Base model:** `FutureMa/Eva-4B`
24
+ - **Format:** GGUF (for llama.cpp)
25
+ - **License:** Apache-2.0
26
+
27
+ Refer to the [original model card](https://huggingface.co/FutureMa/Eva-4B) for model details, intended use, limitations, and evaluation information.
28
+
29
+ ## Files
30
+
31
+ - `Eva-4B-F16.gguf` (FP16 / F16)
32
+
33
+ ## Use with llama.cpp
34
+
35
+ ### Option A: Install via Homebrew (macOS/Linux)
36
+
37
+ ```bash
38
+ brew install llama.cpp
39
+ ```
40
+
41
+ #### CLI
42
+
43
+ ```bash
44
+ llama-cli --hf-repo FutureMa/Eva-4B-GGUF --hf-file Eva-4B-F16.gguf -p "The meaning of life and the universe is"
45
+ ```
46
+
47
+ #### Server
48
+
49
+ ```bash
50
+ llama-server --hf-repo FutureMa/Eva-4B-GGUF --hf-file Eva-4B-F16.gguf -c 2048
51
+ ```
52
+
53
+ ### Option B: Build llama.cpp from source
54
+
55
+ Step 1: Clone llama.cpp:
56
+
57
+ ```bash
58
+ git clone https://github.com/ggerganov/llama.cpp
59
+ ```
60
+
61
+ Step 2: Build (enable Hugging Face download support):
62
+
63
+ ```bash
64
+ cd llama.cpp && LLAMA_CURL=1 make
65
+ ```
66
+
67
+ Step 3: Run:
68
+
69
+ ```bash
70
+ ./llama-cli --hf-repo FutureMa/Eva-4B-GGUF --hf-file Eva-4B-F16.gguf -p "The meaning of life and the universe is"
71
+ ```
72
+
73
+ or
74
+
75
+ ```bash
76
+ ./llama-server --hf-repo FutureMa/Eva-4B-GGUF --hf-file Eva-4B-F16.gguf -c 2048
77
+ ```
78
+
79
+ ## Notes
80
+
81
+ - The `-c 2048` value is an example context size; adjust based on your needs and available memory.
82
+ - If you publish additional quantizations (e.g. `Q4_K_M`, `Q5_K_M`), add them to the **Files** section above and reference them in the example commands.