Mungert commited on
Commit
670314b
Β·
verified Β·
0 Parent(s):

Super-squash history to reclaim storage

Browse files
.gitattributes ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Midm-2.0-Base-Instruct-f16.gguf filter=lfs diff=lfs merge=lfs -text
37
+ Midm-2.0-Base-Instruct-f16_q8_0.gguf filter=lfs diff=lfs merge=lfs -text
38
+ Midm-2.0-Base-Instruct-bf16_q8_0.gguf filter=lfs diff=lfs merge=lfs -text
39
+ Midm-2.0-Base-Instruct-f16_q6_k.gguf filter=lfs diff=lfs merge=lfs -text
40
+ Midm-2.0-Base-Instruct-bf16_q6_k.gguf filter=lfs diff=lfs merge=lfs -text
41
+ Midm-2.0-Base-Instruct-f16_q4_k.gguf filter=lfs diff=lfs merge=lfs -text
42
+ Midm-2.0-Base-Instruct-bf16_q4_k.gguf filter=lfs diff=lfs merge=lfs -text
43
+ Midm-2.0-Base-Instruct-q2_k_l.gguf filter=lfs diff=lfs merge=lfs -text
44
+ Midm-2.0-Base-Instruct-q3_k_l.gguf filter=lfs diff=lfs merge=lfs -text
45
+ Midm-2.0-Base-Instruct-q4_k_l.gguf filter=lfs diff=lfs merge=lfs -text
46
+ Midm-2.0-Base-Instruct-q5_k_l.gguf filter=lfs diff=lfs merge=lfs -text
47
+ Midm-2.0-Base-Instruct-q6_k_l.gguf filter=lfs diff=lfs merge=lfs -text
48
+ Midm-2.0-Base-Instruct-q2_k_m.gguf filter=lfs diff=lfs merge=lfs -text
49
+ Midm-2.0-Base-Instruct-q2_k_s.gguf filter=lfs diff=lfs merge=lfs -text
50
+ Midm-2.0-Base-Instruct-q3_k_m.gguf filter=lfs diff=lfs merge=lfs -text
51
+ Midm-2.0-Base-Instruct-q3_k_s.gguf filter=lfs diff=lfs merge=lfs -text
52
+ Midm-2.0-Base-Instruct-q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text
53
+ Midm-2.0-Base-Instruct-q4_k_s.gguf filter=lfs diff=lfs merge=lfs -text
54
+ Midm-2.0-Base-Instruct-q5_k_m.gguf filter=lfs diff=lfs merge=lfs -text
55
+ Midm-2.0-Base-Instruct-q5_k_s.gguf filter=lfs diff=lfs merge=lfs -text
56
+ Midm-2.0-Base-Instruct-q6_k_m.gguf filter=lfs diff=lfs merge=lfs -text
57
+ Midm-2.0-Base-Instruct-q8_0.gguf filter=lfs diff=lfs merge=lfs -text
58
+ Midm-2.0-Base-Instruct-q4_0.gguf filter=lfs diff=lfs merge=lfs -text
59
+ Midm-2.0-Base-Instruct-q4_1.gguf filter=lfs diff=lfs merge=lfs -text
60
+ Midm-2.0-Base-Instruct-q4_0_l.gguf filter=lfs diff=lfs merge=lfs -text
61
+ Midm-2.0-Base-Instruct-q4_1_l.gguf filter=lfs diff=lfs merge=lfs -text
62
+ Midm-2.0-Base-Instruct-q5_0.gguf filter=lfs diff=lfs merge=lfs -text
63
+ Midm-2.0-Base-Instruct-q5_1.gguf filter=lfs diff=lfs merge=lfs -text
64
+ Midm-2.0-Base-Instruct-q5_0_l.gguf filter=lfs diff=lfs merge=lfs -text
65
+ Midm-2.0-Base-Instruct-q5_1_l.gguf filter=lfs diff=lfs merge=lfs -text
66
+ Midm-2.0-Base-Instruct-iq1_s.gguf filter=lfs diff=lfs merge=lfs -text
67
+ Midm-2.0-Base-Instruct-iq1_m.gguf filter=lfs diff=lfs merge=lfs -text
68
+ Midm-2.0-Base-Instruct-iq2_xs.gguf filter=lfs diff=lfs merge=lfs -text
69
+ Midm-2.0-Base-Instruct-iq2_xxs.gguf filter=lfs diff=lfs merge=lfs -text
70
+ Midm-2.0-Base-Instruct-iq2_s.gguf filter=lfs diff=lfs merge=lfs -text
71
+ Midm-2.0-Base-Instruct-iq2_m.gguf filter=lfs diff=lfs merge=lfs -text
72
+ Midm-2.0-Base-Instruct-iq3_xs.gguf filter=lfs diff=lfs merge=lfs -text
73
+ Midm-2.0-Base-Instruct-iq3_xxs.gguf filter=lfs diff=lfs merge=lfs -text
74
+ Midm-2.0-Base-Instruct-iq3_s.gguf filter=lfs diff=lfs merge=lfs -text
75
+ Midm-2.0-Base-Instruct-iq3_m.gguf filter=lfs diff=lfs merge=lfs -text
76
+ Midm-2.0-Base-Instruct-iq4_xs.gguf filter=lfs diff=lfs merge=lfs -text
77
+ Midm-2.0-Base-Instruct-iq4_nl.gguf filter=lfs diff=lfs merge=lfs -text
78
+ Midm-2.0-Base-Instruct-tq1_0.gguf filter=lfs diff=lfs merge=lfs -text
79
+ Midm-2.0-Base-Instruct-tq2_0.gguf filter=lfs diff=lfs merge=lfs -text
80
+ Midm-2.0-Base-Instruct.imatrix filter=lfs diff=lfs merge=lfs -text
81
+ Midm-2.0-Base-Instruct-bf16.gguf filter=lfs diff=lfs merge=lfs -text
Midm-2.0-Base-Instruct-bf16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8af30758ce47b8592a2de551094ba20e5f0afbd8fde1ce371aaf5ed0e92e1021
3
+ size 23097821216
Midm-2.0-Base-Instruct-bf16_q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e76dcba1fa165eec0c6df620e736d450c39be01c07b3003765e3104ef238b4ec
3
+ size 16869279776
Midm-2.0-Base-Instruct-f16_q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49da267c27e331cc69bea9c6d501aadad0fdbb17f09a45af534bc8f49a8a7527
3
+ size 16869279776
Midm-2.0-Base-Instruct-iq1_m.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3394283ee7bc92b32cdca5c8317e32a01780c31433d8d76425f5793c564640dc
3
+ size 3701447008
Midm-2.0-Base-Instruct-iq1_s.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a832dc102751fa04c71b87f1c335deb95ac4cf3098ac4820cfd41838b5e16aa2
3
+ size 3445725536
Midm-2.0-Base-Instruct-iq2_m.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb689c2af6e2d1253e3d5d538bfa198ad313bf60d3a3d51663af61edcaac6d10
3
+ size 4417624416
Midm-2.0-Base-Instruct-iq2_s.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e85975eddccd813bfe66f6ed3eb87dea5b6fefca9884a8649164a0c6e2addbb9
3
+ size 4223113568
Midm-2.0-Base-Instruct-iq2_xs.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af508853ad98d1c844354000e069c61e8243e13175388cb7ec8002da4678de6e
3
+ size 4097808736
Midm-2.0-Base-Instruct-iq2_xxs.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c40dcd890bdbf6eaf9261e917d6900988ac525542219028a46bb159736050d66
3
+ size 3798047072
Midm-2.0-Base-Instruct-iq3_m.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85b079c2539012999253d476066740abf2319baf0bac545c04638c410f17304b
3
+ size 5589014880
Midm-2.0-Base-Instruct-iq3_s.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0355cbabd0107afc37ee837a42e3e415231adec4b4bada11ca14d1b993fda93c
3
+ size 5589014880
Midm-2.0-Base-Instruct-iq3_xs.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d50574dde73e3a5aeb76b5a9b93e6b3b066b3755503bc49b55efba4cc7a97519
3
+ size 5064137056
Midm-2.0-Base-Instruct-iq3_xxs.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6b055cc3deb47cb479fcf7329153a7c70e868e70f64615f2fbb5ebfd9e0be9f
3
+ size 4935162208
Midm-2.0-Base-Instruct-iq4_nl.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6679a6839172d626fe681249fd292d8fc06e573d49a5af4da0f1f94e72a72bfb
3
+ size 6665388128
Midm-2.0-Base-Instruct-iq4_xs.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3785ac6b7b13780ae0f682f8a6d9d7bedffa73a3a705dce850048f705e4fd138
3
+ size 6327706720
Midm-2.0-Base-Instruct-q2_k_m.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a08608e7f4165ab6c338efaa24edc433d97e8d2e2102b294838c2e1497dbd7e3
3
+ size 4683908960
Midm-2.0-Base-Instruct-q2_k_s.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a06e553512dcd33ea0d4b9f204c9429f99c975ff5dc0fe1b4721d0c64a6ef42f
3
+ size 4540963168
Midm-2.0-Base-Instruct-q3_k_m.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:686edae9970774a0c2e7b735228dd7eea66ab17b9f8970fc3303e18aba7453c5
3
+ size 5876664160
Midm-2.0-Base-Instruct-q3_k_s.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:728002cfe779fa2709e8f5b3303ffed9a5d74e84580e7ec5b0aa4180ede017b4
3
+ size 5733718368
Midm-2.0-Base-Instruct-q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a4d5002a9cb0d74db4e1a2b756ce91367a8782a5189b4a6285284f6f5750cdf
3
+ size 6501480800
Midm-2.0-Base-Instruct-q4_1.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:054fbfc34d61d28eda81151ac2c88f1683fe00f942d604c65f8282a12385096b
3
+ size 7223060832
Midm-2.0-Base-Instruct-q4_k_m.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14eb38af9d8a18041ad93430b71b1e49e30562796c2d9e617c7cac6889ec85de
3
+ size 7105660768
Midm-2.0-Base-Instruct-q4_k_s.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11a5a39cc7d149c2fcd125c33b45abb3e5f1085e5843a59a630b7d06555eba5d
3
+ size 6885918560
Midm-2.0-Base-Instruct-q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a034810cdfad549686556cf9bb7f67d0780e73c1ca5a083614fb937ee61d998b
3
+ size 7944640864
Midm-2.0-Base-Instruct-q5_1.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b2454e9cc0e22ab1880967bbf861df7f64a4e0a97181f55adfda98d2328e0bc
3
+ size 8666220896
Midm-2.0-Base-Instruct-q5_k_m.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa1cf1ab9a8f84b6e643578e071f660b2cde3ee88ad82e9ae2f1be0c47e0c863
3
+ size 8310409056
Midm-2.0-Base-Instruct-q5_k_s.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:776309e5da48b7647ccf738fd19ebe08c579c11794e7239a38f2e98b11262505
3
+ size 8193427296
Midm-2.0-Base-Instruct-q6_k_m.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:035382feff85503084293534923887f811300de079a19e00cfe098b7f07e1a27
3
+ size 9477998432
Midm-2.0-Base-Instruct-q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9d0cd0b59ceab4bc3557b485f14b247658f8663edf395d5fcd34d102a614a5f
3
+ size 12274120736
Midm-2.0-Base-Instruct-tq1_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef7855ba35d9566fae1faefaaabb735e5544307af0dee550f735fdb1be8e1052
3
+ size 2959720544
Midm-2.0-Base-Instruct-tq2_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d280c434e9394de8133b950cc07f9cee633cfccd48edac4009df7aea497b4d70
3
+ size 3450454112
Midm-2.0-Base-Instruct.imatrix ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f64aebaa6f81304631dda4df6b32b45f57e6f30d0cb584bdf266309aff0028a3
3
+ size 7482300
README.md ADDED
@@ -0,0 +1,675 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ language:
4
+ - en
5
+ - ko
6
+ tags:
7
+ - KT
8
+ - K-intelligence
9
+ - Mi:dm
10
+ inference: true
11
+ pipeline_tag: text-generation
12
+ library_name: transformers
13
+ ---
14
+
15
+ # <span style="color: #7FFF7F;">Midm-2.0-Base-Instruct GGUF Models</span>
16
+
17
+
18
+ ## <span style="color: #7F7FFF;">Model Generation Details</span>
19
+
20
+ This model was generated using [llama.cpp](https://github.com/ggerganov/llama.cpp) at commit [`21c02174`](https://github.com/ggerganov/llama.cpp/commit/21c021745d781edf9c44b4972ef6cbbf53b0ecff).
21
+
22
+
23
+
24
+
25
+
26
+ ---
27
+
28
+ ## <span style="color: #7FFF7F;">Quantization Beyond the IMatrix</span>
29
+
30
+ I've been experimenting with a new quantization approach that selectively elevates the precision of key layers beyond what the default IMatrix configuration provides.
31
+
32
+ In my testing, standard IMatrix quantization underperforms at lower bit depths, especially with Mixture of Experts (MoE) models. To address this, I'm using the `--tensor-type` option in `llama.cpp` to manually "bump" important layers to higher precision. You can see the implementation here:
33
+ πŸ‘‰ [Layer bumping with llama.cpp](https://github.com/Mungert69/GGUFModelBuilder/blob/main/model-converter/tensor_list_builder.py)
34
+
35
+ While this does increase model file size, it significantly improves precision for a given quantization level.
36
+
37
+ ### **I'd love your feedbackβ€”have you tried this? How does it perform for you?**
38
+
39
+
40
+
41
+
42
+ ---
43
+
44
+ <a href="https://readyforquantum.com/huggingface_gguf_selection_guide.html" style="color: #7FFF7F;">
45
+ Click here to get info on choosing the right GGUF model format
46
+ </a>
47
+
48
+ ---
49
+
50
+
51
+
52
+ <!--Begin Original Model Card-->
53
+
54
+
55
+
56
+ <p align="center">
57
+ <br>
58
+ <span style="font-size: 60px; font-weight: bold;">Mi:dm 2.0 Base</span>
59
+ </br>
60
+ </p>
61
+
62
+ <p align="center">
63
+ πŸ€— <a href="https://huggingface.co/collections/K-intelligence/mi-dm-20-6866406c301e5f45a6926af8">Mi:dm 2.0 Models</a> |
64
+ πŸ“œ <a href="https://github.com/K-intelligence-Midm/Midm-2.0/blob/main/Mi_dm2_0_technical_report.pdf">Mi:dm 2.0 Technical Report</a> |
65
+ πŸ“• Mi:dm 2.0 Technical Blog*
66
+ </p>
67
+
68
+ <p align="center"><sub>*To be released soon</sub></p>
69
+
70
+ <br>
71
+
72
+ # News πŸ“’
73
+
74
+ - πŸ”œ _(Coming Soon!) GGUF format model files will be available soon for easier local deployment._
75
+ - ⚑️`2025/07/04`: Released Mi:dm 2.0 Model collection on Hugging FaceπŸ€—.
76
+ <br>
77
+ <br>
78
+ # Table of Contents
79
+
80
+ - ___Overview___
81
+ - [Mi:dm 2.0](#midm-20)
82
+ - [Quickstart](#quickstart)
83
+ - [Evaluation](#evaluation)
84
+ - ___Usage___
85
+ - [Run on Friendli.AI](#run-on-friendliai)
86
+ - [Run on Your Local Machine](#run-on-your-local-machine)
87
+ - [Deployment](#deployment)
88
+ - [Tutorials](#tutorials)
89
+ - ___More Information___
90
+ - [Limitation](#limitation)
91
+ - [License](#license)
92
+ - [Contact](#contact)
93
+
94
+ <br>
95
+ <br>
96
+
97
+ # Overview
98
+
99
+ ### Mi:dm 2.0
100
+
101
+ **Mi:dm 2.0** is a __"Korea-centric AI"__ model developed using KT's proprietary technology. The term __"Korea-centric AI"__ refers to a model that deeply internalizes the unique values, cognitive frameworks, and commonsense reasoning inherent to Korean society. It goes beyond simply processing or generating Korean textβ€”it reflects a deeper understanding of the socio-cultural norms and values that define Korean society.
102
+
103
+ Mi:dm 2.0 is released in two versions:
104
+
105
+ - **Mi:dm 2.0 Base**
106
+ An 11.5B parameter dense model designed to balance model size and performance.
107
+ It extends an 8B-scale model by applying the Depth-up Scaling (DuS) method, making it suitable for real-world applications that require both performance and versatility.
108
+
109
+ - **Mi:dm 2.0 Mini**
110
+ A lightweight 2.3B parameter dense model optimized for on-device environments and systems with limited GPU resources.
111
+ It was derived from the Base model through pruning and distillation to enable compact deployment.
112
+
113
+ > [!Note]
114
+ > Neither the pre-training nor the post-training data includes KT users' data.
115
+
116
+ <br>
117
+
118
+ ### Quickstart
119
+
120
+ Here is the code snippet to run conversational inference with the model:
121
+
122
+ ```python
123
+ import torch
124
+ from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
125
+
126
+ model_name = "K-intelligence/Midm-2.0-Base-Instruct"
127
+
128
+ model = AutoModelForCausalLM.from_pretrained(
129
+ model_name,
130
+ torch_dtype=torch.bfloat16,
131
+ trust_remote_code=True,
132
+ device_map="auto"
133
+ )
134
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
135
+ generation_config = GenerationConfig.from_pretrained(model_name)
136
+
137
+ prompt = "KT에 λŒ€ν•΄ μ†Œκ°œν•΄μ€˜"
138
+
139
+ # message for inference
140
+ messages = [
141
+ {"role": "system",
142
+ "content": "Mi:dm(λ―Ώ:음)은 KTμ—μ„œ κ°œλ°œν•œ AI 기반 μ–΄μ‹œμŠ€ν„΄νŠΈμ΄λ‹€."},
143
+ {"role": "user", "content": prompt}
144
+ ]
145
+
146
+ input_ids = tokenizer.apply_chat_template(
147
+ messages,
148
+ tokenize=True,
149
+ add_generation_prompt=True,
150
+ return_tensors="pt"
151
+ )
152
+
153
+ output = model.generate(
154
+ input_ids.to("cuda"),
155
+ generation_config=generation_config,
156
+ eos_token_id=tokenizer.eos_token_id,
157
+ max_new_tokens=128,
158
+ do_sample=False,
159
+ )
160
+ print(tokenizer.decode(output[0]))
161
+ ```
162
+
163
+ > [!NOTE]
164
+ > The `transformers` library should be version `4.45.0` or higher.
165
+
166
+ <br>
167
+
168
+ # Evaluation
169
+
170
+
171
+ #### Korean
172
+
173
+ <!-- first half table-->
174
+ <table>
175
+ <tr>
176
+ <th rowspan="2">Model</th>
177
+ <th colspan="5" align="center">Society & Culture</th>
178
+ <th colspan="3" align="center">General Knowledge</th>
179
+ <th colspan="3" align="center">Instruction Following</th>
180
+ </tr>
181
+ <tr>
182
+ <th align="center">K-Refer<sup>*</sup></th>
183
+ <th align="center">K-Refer-Hard<sup>*</sup></th>
184
+ <th align="center">Ko-Sovereign<sup>*</sup></th>
185
+ <th align="center">HAERAE</th>
186
+ <th align="center">Avg.</th>
187
+ <th align="center">KMMLU</th>
188
+ <th align="center">Ko-Sovereign<sup>*</sup></th>
189
+ <th align="center">Avg.</th>
190
+ <th align="center">Ko-IFEval</th>
191
+ <th align="center">Ko-MTBench</th>
192
+ <th align="center">Avg.</th>
193
+ </tr>
194
+
195
+ <!-- Small Models -->
196
+ <tr>
197
+ <td><strong>Qwen3-4B</strong></td>
198
+ <td align="center">53.6</td>
199
+ <td align="center">42.9</td>
200
+ <td align="center">35.8</td>
201
+ <td align="center">50.6</td>
202
+ <td align="center">45.7</td>
203
+ <td align="center"><strong>50.6</strong></td>
204
+ <td align="center"><strong>42.5</strong></td>
205
+ <td align="center"><strong>46.5</strong></td>
206
+ <td align="center"><strong>75.9</strong></td>
207
+ <td align="center">63.0</td>
208
+ <td align="center">69.4</td>
209
+ </tr>
210
+ <tr>
211
+ <td><strong>Exaone-3.5-2.4B-inst</strong></td>
212
+ <td align="center">64.0</td>
213
+ <td align="center"><strong>67.1</strong></td>
214
+ <td align="center"><strong>44.4</strong></td>
215
+ <td align="center">61.3</td>
216
+ <td align="center"><strong>59.2</strong></td>
217
+ <td align="center">43.5</td>
218
+ <td align="center">42.4</td>
219
+ <td align="center">43.0</td>
220
+ <td align="center">65.4</td>
221
+ <td align="center"><strong>74.0</strong></td>
222
+ <td align="center">68.9</td>
223
+ </tr>
224
+ <tr>
225
+ <td><strong>Mi:dm 2.0-Mini-inst</strong></td>
226
+ <td align="center"><strong>66.4</strong></td>
227
+ <td align="center">61.4</td>
228
+ <td align="center">36.7</td>
229
+ <td align="center"><strong>70.8</strong></td>
230
+ <td align="center">58.8</td>
231
+ <td align="center">45.1</td>
232
+ <td align="center">42.4</td>
233
+ <td align="center">43.8</td>
234
+ <td align="center">73.3</td>
235
+ <td align="center"><strong>74.0</strong></td>
236
+ <td align="center"><strong>73.6</strong></td>
237
+ </tr>
238
+
239
+ <!-- Spacer row -->
240
+ <tr><td colspan="13"> </td></tr>
241
+
242
+ <!-- Large Models -->
243
+ <tr>
244
+ <td><strong>Qwen3-14B</strong></td>
245
+ <td align="center">72.4</td>
246
+ <td align="center">65.7</td>
247
+ <td align="center">49.8</td>
248
+ <td align="center">68.4</td>
249
+ <td align="center">64.1</td>
250
+ <td align="center">55.4</td>
251
+ <td align="center">54.7</td>
252
+ <td align="center">55.1</td>
253
+ <td align="center"><strong>83.6</strong></td>
254
+ <td align="center">71</td>
255
+ <td align="center">77.3</td>
256
+ </tr>
257
+ <tr>
258
+ <td><strong>Llama-3.1-8B-inst</strong></td>
259
+ <td align="center">43.2</td>
260
+ <td align="center">36.4</td>
261
+ <td align="center">33.8</td>
262
+ <td align="center">49.5</td>
263
+ <td align="center">40.7</td>
264
+ <td align="center">33.0</td>
265
+ <td align="center">36.7</td>
266
+ <td align="center">34.8</td>
267
+ <td align="center">60.1</td>
268
+ <td align="center">57</td>
269
+ <td align="center">58.5</td>
270
+ </tr>
271
+ <tr>
272
+ <td><strong>Exaone-3.5-7.8B-inst</strong></td>
273
+ <td align="center">71.6</td>
274
+ <td align="center">69.3</td>
275
+ <td align="center">46.9</td>
276
+ <td align="center">72.9</td>
277
+ <td align="center">65.2</td>
278
+ <td align="center">52.6</td>
279
+ <td align="center">45.6</td>
280
+ <td align="center">49.1</td>
281
+ <td align="center">69.1</td>
282
+ <td align="center">79.6</td>
283
+ <td align="center">74.4</td>
284
+ </tr>
285
+ <tr>
286
+ <td><strong>Mi:dm 2.0-Base-inst</strong></td>
287
+ <td align="center"><strong>89.6</strong></td>
288
+ <td align="center"><strong>86.4</strong></td>
289
+ <td align="center"><strong>56.3</strong></td>
290
+ <td align="center"><strong>81.5</strong></td>
291
+ <td align="center"><strong>78.4</strong></td>
292
+ <td align="center"><strong>57.3</strong></td>
293
+ <td align="center"><strong>58.0</strong></td>
294
+ <td align="center"><strong>57.7</strong></td>
295
+ <td align="center">82</td>
296
+ <td align="center"><strong>89.7</strong></td>
297
+ <td align="center"><strong>85.9</strong></td>
298
+ </tr>
299
+ </table>
300
+
301
+ <!-- second half table-->
302
+ <table>
303
+ <tr>
304
+ <th rowspan="2" align="center">Model</th>
305
+ <th colspan="5" align="center">Comprehension</th>
306
+ <th colspan="5" align="center">Reasoning</th>
307
+ </tr>
308
+ <tr>
309
+ <th align="center">K-Prag<sup>*</sup></th>
310
+ <th align="center">K-Refer-Hard<sup>*</sup></th>
311
+ <th align="center">Ko-Best</th>
312
+ <th align="center">Ko-Sovereign<sup>*</sup></th>
313
+ <th align="center">Avg.</th>
314
+ <th align="center">Ko-Winogrande</th>
315
+ <th align="center">Ko-Best</th>
316
+ <th align="center">LogicKor</th>
317
+ <th align="center">HRM8K</th>
318
+ <th align="center">Avg.</th>
319
+ </tr>
320
+
321
+ <!-- Small Models -->
322
+ <tr>
323
+ <td><strong>Qwen3-4B</strong></td>
324
+ <td align="center"><strong>73.9<strong></td>
325
+ <td align="center">56.7</td>
326
+ <td align="center"><strong>91.5</strong></td>
327
+ <td align="center"><strong>43.5</strong></td>
328
+ <td align="center"><strong>66.6</strong></td>
329
+ <td align="center"><strong>67.5</strong></td>
330
+ <td align="center"><strong>69.2</strong></td>
331
+ <td align="center">5.6</td>
332
+ <td align="center"><strong>56.7</strong></td>
333
+ <td align="center"><strong>43.8</strong></td>
334
+ </tr>
335
+ <tr>
336
+ <td><strong>Exaone-3.5-2.4B-inst</strong></td>
337
+ <td align="center">68.7</td>
338
+ <td align="center"><strong>58.5</strong></td>
339
+ <td align="center">87.2</td>
340
+ <td align="center">38.0</td>
341
+ <td align="center">62.5</td>
342
+ <td align="center">60.3</td>
343
+ <td align="center">64.1</td>
344
+ <td align="center">7.4</td>
345
+ <td align="center">38.5</td>
346
+ <td align="center">36.7</td>
347
+ </tr>
348
+ <tr>
349
+ <td><strong>Mi:dm 2.0-Mini-inst</strong></td>
350
+ <td align="center">69.5</td>
351
+ <td align="center">55.4</td>
352
+ <td align="center">80.5</td>
353
+ <td align="center">42.5</td>
354
+ <td align="center">61.9</td>
355
+ <td align="center">61.7</td>
356
+ <td align="center">64.5</td>
357
+ <td align="center"><strong>7.7</strong></td>
358
+ <td align="center">39.9</td>
359
+ <td align="center">37.4</td>
360
+ </tr>
361
+
362
+ <!-- Visual Spacer -->
363
+ <tr><td colspan="11"> </td></tr>
364
+
365
+ <!-- Large Models -->
366
+ <tr>
367
+ <td><strong>Qwen3-14B</strong></td>
368
+ <td align="center"><strong>86.7</strong></td>
369
+ <td align="center"><strong>74.0</strong></td>
370
+ <td align="center">93.9</td>
371
+ <td align="center">52.0</td>
372
+ <td align="center"><strong>76.8</strong></td>
373
+ <td align="center"><strong>77.2</strong></td>
374
+ <td align="center"><strong>75.4</strong></td>
375
+ <td align="center">6.4</td>
376
+ <td align="center"><strong>64.5</strong></td>
377
+ <td align="center"><strong>48.8</strong></td>
378
+ </tr>
379
+ <tr>
380
+ <td><strong>Llama-3.1-8B-inst</strong></td>
381
+ <td align="center">59.9</td>
382
+ <td align="center">48.6</td>
383
+ <td align="center">77.4</td>
384
+ <td align="center">31.5</td>
385
+ <td align="center">51.5</td>
386
+ <td align="center">40.1</td>
387
+ <td align="center">26.0</td>
388
+ <td align="center">2.4</td>
389
+ <td align="center">30.9</td>
390
+ <td align="center">19.8</td>
391
+ </tr>
392
+ <tr>
393
+ <td><strong>Exaone-3.5-7.8B-inst</strong></td>
394
+ <td align="center">73.5</td>
395
+ <td align="center">61.9</td>
396
+ <td align="center">92.0</td>
397
+ <td align="center">44.0</td>
398
+ <td align="center">67.2</td>
399
+ <td align="center">64.6</td>
400
+ <td align="center">60.3</td>
401
+ <td align="center"><strong>8.6</strong></td>
402
+ <td align="center">49.7</td>
403
+ <td align="center">39.5</td>
404
+ </tr>
405
+ <tr>
406
+ <td><strong>Mi:dm 2.0-Base-inst</strong></td>
407
+ <td align="center">86.5</td>
408
+ <td align="center">70.8</td>
409
+ <td align="center"><strong>95.2</strong></td>
410
+ <td align="center"><strong>53.0</strong></td>
411
+ <td align="center">76.1</td>
412
+ <td align="center">75.1</td>
413
+ <td align="center">73.0</td>
414
+ <td align="center"><strong>8.6</strong></td>
415
+ <td align="center">52.9</td>
416
+ <td align="center">44.8</td>
417
+ </tr>
418
+ </table>
419
+
420
+ `*` indicates KT proprietary evaluation resources.
421
+
422
+ <br>
423
+
424
+
425
+ #### English
426
+
427
+
428
+ <table>
429
+ <tr>
430
+ <th rowspan="2" align="center">Model</th>
431
+ <th align="center">Instruction</th>
432
+ <th colspan="4" align="center">Reasoning</th>
433
+ <th align="center">Math</th>
434
+ <th align="center">Coding</th>
435
+ <th colspan="3" align="center">General Knowledge</th>
436
+ </tr>
437
+ <tr>
438
+ <th align="center">IFEval</th>
439
+ <th align="center">BBH</th>
440
+ <th align="center">GPQA</th>
441
+ <th align="center">MuSR</th>
442
+ <th align="center">Avg.</th>
443
+ <th align="center">GSM8K</th>
444
+ <th align="center">MBPP+</th>
445
+ <th align="center">MMLU-pro</th>
446
+ <th align="center">MMLU</th>
447
+ <th align="center">Avg.</th>
448
+ </tr>
449
+
450
+ <!-- Small Models -->
451
+ <tr>
452
+ <td><strong>Qwen3-4B</strong></td>
453
+ <td align="center">79.7</td>
454
+ <td align="center"><strong>79.0</strong></td>
455
+ <td align="center"><strong>39.8</strong></td>
456
+ <td align="center"><strong>58.5</strong></td>
457
+ <td align="center"><strong>59.1</strong></td>
458
+ <td align="center"><strong>90.4</strong></td>
459
+ <td align="center">62.4</td>
460
+ <td align="center">-</td>
461
+ <td align="center"><strong>73.3</strong></td>
462
+ <td align="center"><strong>73.3</strong></td>
463
+ </tr>
464
+ <tr>
465
+ <td><strong>Exaone-3.5-2.4B-inst</strong></td>
466
+ <td align="center"><strong>81.1</strong></td>
467
+ <td align="center">46.4</td>
468
+ <td align="center">28.1</td>
469
+ <td align="center">49.7</td>
470
+ <td align="center">41.4</td>
471
+ <td align="center">82.5</td>
472
+ <td align="center">59.8</td>
473
+ <td align="center">-</td>
474
+ <td align="center">59.5</td>
475
+ <td align="center">59.5</td>
476
+ </tr>
477
+ <tr>
478
+ <td><strong>Mi:dm 2.0-Mini-inst</strong></td>
479
+ <td align="center">73.6</td>
480
+ <td align="center">44.5</td>
481
+ <td align="center">26.6</td>
482
+ <td align="center">51.7</td>
483
+ <td align="center">40.9</td>
484
+ <td align="center">83.1</td>
485
+ <td align="center"><strong>60.9</strong></td>
486
+ <td align="center">-</td>
487
+ <td align="center">56.5</td>
488
+ <td align="center">56.5</td>
489
+ </tr>
490
+
491
+ <tr><td colspan="11">&nbsp;</td></tr>
492
+
493
+ <!-- Large Models -->
494
+ <tr>
495
+ <td><strong>Qwen3-14B</strong></td>
496
+ <td align="center">83.9</td>
497
+ <td align="center"><strong>83.4</strong></td>
498
+ <td align="center"><strong>49.8</strong></td>
499
+ <td align="center"><strong>57.7</strong></td>
500
+ <td align="center"><strong>63.6</strong></td>
501
+ <td align="center">88.0</td>
502
+ <td align="center">73.4</td>
503
+ <td align="center"><strong>70.5</strong></td>
504
+ <td align="center"><strong>82.7</strong></td>
505
+ <td align="center"><strong>76.6</strong></td>
506
+ </tr>
507
+ <tr>
508
+ <td><strong>Llama-3.1-8B-inst</strong></td>
509
+ <td align="center">79.9</td>
510
+ <td align="center">60.3</td>
511
+ <td align="center">21.6</td>
512
+ <td align="center">50.3</td>
513
+ <td align="center">44.1</td>
514
+ <td align="center">81.2</td>
515
+ <td align="center"><strong>81.8</strong></td>
516
+ <td align="center">47.6</td>
517
+ <td align="center">70.7</td>
518
+ <td align="center">59.2</td>
519
+ </tr>
520
+ <tr>
521
+ <td><strong>Exaone-3.5-7.8B-inst</strong></td>
522
+ <td align="center">83.6</td>
523
+ <td align="center">50.1</td>
524
+ <td align="center">33.1</td>
525
+ <td align="center">51.2</td>
526
+ <td align="center">44.8</td>
527
+ <td align="center">81.1</td>
528
+ <td align="center">79.4</td>
529
+ <td align="center">40.7</td>
530
+ <td align="center">69.0</td>
531
+ <td align="center">54.8</td>
532
+ </tr>
533
+ <tr>
534
+ <td><strong>Mi:dm 2.0-Base-inst</strong></td>
535
+ <td align="center"><strong>84.0</strong></td>
536
+ <td align="center">77.7</td>
537
+ <td align="center">33.5</td>
538
+ <td align="center">51.9</td>
539
+ <td align="center">54.4</td>
540
+ <td align="center"><strong>91.6</strong></td>
541
+ <td align="center">77.5</td>
542
+ <td align="center">53.3</td>
543
+ <td align="center">73.7</td>
544
+ <td align="center">63.5</td>
545
+ </tr>
546
+ </table>
547
+
548
+
549
+ <br>
550
+
551
+ # Usage
552
+
553
+ ### Run on Friendli.AI
554
+ You can try our model immediately via `Friendli.AI`. Simply click `Deploy` and then `Friendli Endpoints`.
555
+
556
+ > [!Note]
557
+ > Please note that a login to `Friendli.AI` is required after your fifth chat interaction.
558
+
559
+ <p>
560
+ <img src="./assets/image_1.png" alt="Left Image" width="36%" style="display:inline-block; margin-right:2%">
561
+ <img src="./assets/image_2.png" alt="Right Image" width="36%" style="display:inline-block">
562
+ </p>
563
+
564
+
565
+ ### Run on Your Local Machine
566
+ We provide a detailed description about running Mi:dm 2.0 on your local machine using llama.cpp, LM Studio, and Ollama. Please check our [github](https://github.com/K-intelligence-Midm/Midm-2.0) for more information
567
+
568
+
569
+ ### Deployment
570
+
571
+ To serve Mi:dm 2.0 using [vLLM](https://github.com/vllm-project/vllm)(`>=0.8.0`) with an OpenAI-compatible API:
572
+ ```bash
573
+ vllm serve K-intelligence/Midm-2.0-Base-Instruct
574
+ ```
575
+
576
+
577
+ ### Tutorials
578
+ To help our end-users easily use Mi:dm 2.0, we have provided comprehensive tutorials on [github](https://github.com/K-intelligence-Midm/Midm-2.0).
579
+ <br>
580
+
581
+ <br>
582
+ <br>
583
+
584
+ # More Information
585
+
586
+ ### Limitation
587
+ * The training data for both Mi:dm 2.0 models consists primarily of English and Korean. Understanding and generation in other languages are not guaranteed.
588
+
589
+ * The model is not guaranteed to provide reliable advice in fields that require professional expertise, such as law, medicine, or finance.
590
+
591
+ * Researchers have made efforts to exclude unethical content from the training data β€” such as profanity, slurs, bias, and discriminatory language. However, despite these efforts, the model may still produce inappropriate expressions or factual inaccuracies.
592
+
593
+
594
+ ### License
595
+
596
+ Mi:dm 2.0 is licensed under the [MIT License](./LICENSE).
597
+
598
+ <!-- ### Citation
599
+
600
+ ```
601
+ @misc{,
602
+ title={},
603
+ author={},
604
+ year={2025},
605
+ eprint={},
606
+ archivePrefix={arXiv},
607
+ primaryClass={cs.CL},
608
+ url={},
609
+ }
610
+ ``` -->
611
+ ### Contact
612
+ Mi:dm 2.0 Technical Inquiries: midm-llm@kt.com
613
+
614
+ <br>
615
+
616
+
617
+
618
+ <!--End Original Model Card-->
619
+
620
+ ---
621
+
622
+ # <span id="testllm" style="color: #7F7FFF;">πŸš€ If you find these models useful</span>
623
+
624
+ Help me test my **AI-Powered Quantum Network Monitor Assistant** with **quantum-ready security checks**:
625
+
626
+ πŸ‘‰ [Quantum Network Monitor](https://readyforquantum.com/?assistant=open&utm_source=huggingface&utm_medium=referral&utm_campaign=huggingface_repo_readme)
627
+
628
+
629
+ The full Open Source Code for the Quantum Network Monitor Service available at my github repos ( repos with NetworkMonitor in the name) : [Source Code Quantum Network Monitor](https://github.com/Mungert69). You will also find the code I use to quantize the models if you want to do it yourself [GGUFModelBuilder](https://github.com/Mungert69/GGUFModelBuilder)
630
+
631
+ πŸ’¬ **How to test**:
632
+ Choose an **AI assistant type**:
633
+ - `TurboLLM` (GPT-4.1-mini)
634
+ - `HugLLM` (Hugginface Open-source models)
635
+ - `TestLLM` (Experimental CPU-only)
636
+
637
+ ### **What I’m Testing**
638
+ I’m pushing the limits of **small open-source models for AI network monitoring**, specifically:
639
+ - **Function calling** against live network services
640
+ - **How small can a model go** while still handling:
641
+ - Automated **Nmap security scans**
642
+ - **Quantum-readiness checks**
643
+ - **Network Monitoring tasks**
644
+
645
+ 🟑 **TestLLM** – Current experimental model (llama.cpp on 2 CPU threads on huggingface docker space):
646
+ - βœ… **Zero-configuration setup**
647
+ - ⏳ 30s load time (slow inference but **no API costs**) . No token limited as the cost is low.
648
+ - πŸ”§ **Help wanted!** If youοΏ½οΏ½οΏ½re into **edge-device AI**, let’s collaborate!
649
+
650
+ ### **Other Assistants**
651
+ 🟒 **TurboLLM** – Uses **gpt-4.1-mini** :
652
+ - **It performs very well but unfortunatly OpenAI charges per token. For this reason tokens usage is limited.
653
+ - **Create custom cmd processors to run .net code on Quantum Network Monitor Agents**
654
+ - **Real-time network diagnostics and monitoring**
655
+ - **Security Audits**
656
+ - **Penetration testing** (Nmap/Metasploit)
657
+
658
+ πŸ”΅ **HugLLM** – Latest Open-source models:
659
+ - 🌐 Runs on Hugging Face Inference API. Performs pretty well using the lastest models hosted on Novita.
660
+
661
+ ### πŸ’‘ **Example commands you could test**:
662
+ 1. `"Give me info on my websites SSL certificate"`
663
+ 2. `"Check if my server is using quantum safe encyption for communication"`
664
+ 3. `"Run a comprehensive security audit on my server"`
665
+ 4. '"Create a cmd processor to .. (what ever you want)" Note you need to install a [Quantum Network Monitor Agent](https://readyforquantum.com/Download/?utm_source=huggingface&utm_medium=referral&utm_campaign=huggingface_repo_readme) to run the .net code on. This is a very flexible and powerful feature. Use with caution!
666
+
667
+ ### Final Word
668
+
669
+ I fund the servers used to create these model files, run the Quantum Network Monitor service, and pay for inference from Novita and OpenAIβ€”all out of my own pocket. All the code behind the model creation and the Quantum Network Monitor project is [open source](https://github.com/Mungert69). Feel free to use whatever you find helpful.
670
+
671
+ If you appreciate the work, please consider [buying me a coffee](https://www.buymeacoffee.com/mahadeva) β˜•. Your support helps cover service costs and allows me to raise token limits for everyone.
672
+
673
+ I'm also open to job opportunities or sponsorship.
674
+
675
+ Thank you! 😊