Upload nvidia.Nemotron-Terminal-14B.Q4_K_M.gguf with huggingface_hub
Browse files
.gitattributes
CHANGED
|
@@ -35,3 +35,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
nvidia.Nemotron-Terminal-14B.f16.gguf filter=lfs diff=lfs merge=lfs -text
|
| 37 |
nvidia.Nemotron-Terminal-14B.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
nvidia.Nemotron-Terminal-14B.f16.gguf filter=lfs diff=lfs merge=lfs -text
|
| 37 |
nvidia.Nemotron-Terminal-14B.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
| 38 |
+
nvidia.Nemotron-Terminal-14B.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
nvidia.Nemotron-Terminal-14B.Q4_K_M.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ebd93d53cd0f1ee9d02dafd7287cc688ea6fd4ff5325178347880d1910b17e9c
|
| 3 |
+
size 9001753536
|