Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
294
19
344
John Leimgruber III
PRO
ubergarm
Follow
fdaler's profile picture
groxaxo's profile picture
NghiaNguyen1529's profile picture
261 followers
·
60 following
https://www.paypal.com/donate/?hosted_button_id=HU59345BZVSUA
ubergarm
john-leimgruber
AI & ML interests
Open LLMs and Astrophotography image processing.
Recent Activity
new
activity
3 days ago
ubergarm/GLM-4.7-GGUF:
Stable run on 2x RTX 5090 and 2 Xeon E5 2696 V4 and DDR4 with ik_llama.cpp - 6.1 t/s on IQ4_K and 5.1 t/s on IQ5_K, opencode works with this
new
activity
5 days ago
ubergarm/GLM-4.7-Flash-GGUF:
question about mxfp4
new
activity
5 days ago
zai-org/GLM-4.7-Flash:
Why does the KV cache occupy so much GPU memory?
View all activity
Organizations
ubergarm
's activity
All
Models
Datasets
Spaces
Papers
Collections
Community
Posts
Upvotes
Likes
Articles
New activity in
ubergarm/GLM-4.7-GGUF
3 days ago
Stable run on 2x RTX 5090 and 2 Xeon E5 2696 V4 and DDR4 with ik_llama.cpp - 6.1 t/s on IQ4_K and 5.1 t/s on IQ5_K, opencode works with this
👍
1
14
#5 opened about 1 month ago by
martossien
New activity in
ubergarm/GLM-4.7-Flash-GGUF
5 days ago
question about mxfp4
2
#3 opened 5 days ago by
koifish12
New activity in
zai-org/GLM-4.7-Flash
5 days ago
Why does the KV cache occupy so much GPU memory?
13
#21 opened 6 days ago by
yyg201708
New activity in
ubergarm/GLM-4.7-Flash-GGUF
5 days ago
Re-cooking imatrix and quants with updated ik/llama.cpp PR
🚀
1
8
#1 opened 6 days ago by
ubergarm
updated
a model
5 days ago
ubergarm/GLM-4.7-Flash-GGUF
Text Generation
•
30B
•
Updated
5 days ago
•
6.56k
•
12
New activity in
zai-org/GLM-4.7-Flash
6 days ago
Cannot run vLLM on DGX Spark: ImportError: libcudart.so.12
3
#18 opened 7 days ago by
yyg201708
Performance Discussion
👀
2
3
#1 opened 7 days ago by
IndenScale
Enormous KV-cache size?
👍
➕
6
23
#3 opened 7 days ago by
nephepritou
New activity in
noctrex/GLM-4.7-Flash-MXFP4_MOE-GGUF
6 days ago
Feedback from running in LM Studio 0.39.3 with v1.103.2 of llama.cpp
13
#1 opened 6 days ago by
spanspek
liked
a model
6 days ago
noctrex/GLM-4.7-Flash-MXFP4_MOE-GGUF
Text Generation
•
30B
•
Updated
1 day ago
•
8.91k
•
17
published
a model
7 days ago
ubergarm/GLM-4.7-Flash-GGUF
Text Generation
•
30B
•
Updated
5 days ago
•
6.56k
•
12
liked
2 models
7 days ago
ngxson/GLM-4.7-Flash-GGUF
30B
•
Updated
6 days ago
•
10.9k
•
19
zai-org/GLM-4.7-Flash
Text Generation
•
31B
•
Updated
6 days ago
•
450k
•
•
1.23k
liked
a model
8 days ago
ArtusDev/requests-exl
Updated
Oct 13, 2025
•
6
New activity in
ArtusDev/requests-exl
8 days ago
[QUANTING UPDATE]
❤️
👍
4
4
#28 opened 11 days ago by
ArtusDev
New activity in
ubergarm/Devstral-Small-2-24B-Instruct-2512-GGUF
8 days ago
Mistral 3 large wuant
👍
1
1
#1 opened 9 days ago by
facedwithahug
New activity in
ubergarm/DeepSeek-V3.2-Speciale-GGUF
8 days ago
QuIP - 2 bit quantised as good as 16 bit
5
#5 opened 13 days ago by
infinityai
New activity in
msievers/gemma-3-1b-it-qat-q4_0-gguf
12 days ago
Thanks for sharing your work!
❤️
2
3
#1 opened 13 days ago by
ubergarm
New activity in
ubergarm/DeepSeek-V3.2-Speciale-GGUF
12 days ago
Say Whattt?!
🔥
👍
4
7
#1 opened 18 days ago by
mtcl
New activity in
ubergarm/Devstral-2-123B-Instruct-2512-GGUF
13 days ago
Decent PPL with 100% IQ4_KSS
🔥
1
9
#3 opened about 2 months ago by
sokann
Load more