title
stringlengths
1
300
score
int64
0
8.54k
selftext
stringlengths
0
41.5k
created
timestamp[ns]date
2023-04-01 04:30:41
2026-03-04 02:14:14
url
stringlengths
0
878
author
stringlengths
3
20
domain
stringlengths
0
82
edited
timestamp[ns]date
1970-01-01 00:00:00
2026-02-19 14:51:53
gilded
int64
0
2
gildings
stringclasses
7 values
id
stringlengths
7
7
locked
bool
2 classes
media
stringlengths
646
1.8k
name
stringlengths
10
10
permalink
stringlengths
33
82
spoiler
bool
2 classes
stickied
bool
2 classes
thumbnail
stringlengths
4
213
ups
int64
0
8.54k
preview
stringlengths
301
5.01k
I developed my own webapp to use the local templates.
2
In my company there are some internal blocks. So I developed my own web application using pure html, css and js. It's not perfect yet and just to make it easier to use local models. I accept suggestions for improvements.
2025-07-20T01:58:04
https://github.com/martinsagabriel/LocalLama
gabe__martins
github.com
1970-01-01T00:00:00
0
{}
1m4d74b
false
null
t3_1m4d74b
/r/LocalLLaMA/comments/1m4d74b/i_developed_my_own_webapp_to_use_the_local/
false
false
default
2
{'enabled': False, 'images': [{'id': '5bv_gPO23RdJJo822AU-U14GDuIrIMvUpu0qerJmz64', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/5bv_gPO23RdJJo822AU-U14GDuIrIMvUpu0qerJmz64.png?width=108&crop=smart&auto=webp&s=5f7669905ccf15d531d56c0729d1011e8f26f47a', 'width': 108}, {'height': 108, 'url': 'h...
Getting into local ai. Photo restoration.
9
Hi all, I'm pretty new to this AI stuff but have a system I think can handle some localLLama. 3090Ti 12900K. So I'm looking for a model I can give it an old photo and ask it to restore it and possibly add coloration. Any guidance will be much appreciated. TIA
2025-07-20T01:23:05
https://www.reddit.com/r/LocalLLaMA/comments/1m4cil7/getting_into_local_ai_photo_restoration/
lokito50
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m4cil7
false
null
t3_1m4cil7
/r/LocalLLaMA/comments/1m4cil7/getting_into_local_ai_photo_restoration/
false
false
self
9
null
NSFW AI Local
20
Is there an AI template or GUI(?) I can use locally for free that generates nsfw art of already existing characters. I mean images similar to those on the green site. I know little to nothing about AI but my computer is pretty good.
2025-07-20T00:18:03
https://www.reddit.com/r/LocalLLaMA/comments/1m4b8ji/nsfw_ai_local/
TheGodOfCarrot
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m4b8ji
false
null
t3_1m4b8ji
/r/LocalLLaMA/comments/1m4b8ji/nsfw_ai_local/
false
false
nsfw
20
null
Which model is best for vision fitting 24gb vram
12
Which model is best for vision fitting 24gb vram? Trying to do nsfw categorization for user uploaded images. Gemma3 24b is quite good but is there any other, opinnions?
2025-07-19T23:46:34
https://www.reddit.com/r/LocalLLaMA/comments/1m4al6m/which_model_is_best_for_vision_fitting_24gb_vram/
Rich_Artist_8327
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m4al6m
false
null
t3_1m4al6m
/r/LocalLLaMA/comments/1m4al6m/which_model_is_best_for_vision_fitting_24gb_vram/
false
false
self
12
null
Hackers are never sleeping
330
In my tests to get a reliable Ngrok alternative for https with Open WebUI, I had Llama.cpp's WebUI served over https in a subdomain that's not listed anywhere. Less than 45 minutes after being online, the hacking attempts started. I had a ultra long API key setup so after a while of bruteforce attack, they switched to...
2025-07-19T23:40:20
https://www.reddit.com/r/LocalLLaMA/comments/1m4ag6u/hackers_are_never_sleeping/
DrVonSinistro
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m4ag6u
false
null
t3_1m4ag6u
/r/LocalLLaMA/comments/1m4ag6u/hackers_are_never_sleeping/
false
false
self
330
null
In defense of Synthetic Data
1
[removed]
2025-07-19T23:10:49
https://www.reddit.com/r/LocalLLaMA/comments/1m49tgc/in_defense_of_synthetic_data/
rzvzn
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m49tgc
false
null
t3_1m49tgc
/r/LocalLLaMA/comments/1m49tgc/in_defense_of_synthetic_data/
false
false
self
1
null
Running the 70B sized models on a budget
1
I'm looking to run the 70B sized models but with large context sizes. Like 10k or more. I'd like to avoid offloading to the cpu. What would you recommend hardware set up to be on a budget? 2 x 3090 still best value? Switch to Radeon like the 2x mi50 32gb? It would be just for inference and as long as its faster than...
2025-07-19T23:05:25
https://www.reddit.com/r/LocalLLaMA/comments/1m49p7w/running_the_70b_sized_models_on_a_budget/
fgoricha
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m49p7w
false
null
t3_1m49p7w
/r/LocalLLaMA/comments/1m49p7w/running_the_70b_sized_models_on_a_budget/
false
false
self
1
null
Maybe physics-based AI is the right approach?
0
Language as a medium for reasoning is too fuzzy, and hard to control I feel like language should be a tool to make causality discrete and composable, not as a substrate for reasoning As in, I believe general AI should be a physics-first and then language-second game. Language being an abstraction of physical obser...
2025-07-19T22:57:46
https://www.reddit.com/r/LocalLLaMA/comments/1m49j3n/maybe_physicsbased_ai_is_the_right_approach/
Key_Clerk_1431
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m49j3n
false
null
t3_1m49j3n
/r/LocalLLaMA/comments/1m49j3n/maybe_physicsbased_ai_is_the_right_approach/
false
false
self
0
null
Looking for diarization model better than Pyannote
19
Currently i’m using whisperX, which uses whisper + pyannote for transcription + diarization of audio but I find the speaker recognition quite lackluster. It’s often wrong at labeling the speakers. Any better alternatives to this? I tried Eleven Labs but they only offer an API and dont make the models available and the...
2025-07-19T22:26:49
https://www.reddit.com/r/LocalLLaMA/comments/1m48v53/looking_for_diarization_model_better_than_pyannote/
bluedragon102
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m48v53
false
null
t3_1m48v53
/r/LocalLLaMA/comments/1m48v53/looking_for_diarization_model_better_than_pyannote/
false
false
self
19
null
OCR and GenAI: Key Trends from H1 2025
8
Hi all, I’ve noticed plenty of questions and great insights in Reddit threads about the latest OCR and document-AI tools. After learning a lot from those discussions—and adding lessons from my own enterprise projects —I pulled together a brief mid-2025 summary: key VLM releases, specialist models, pipeline updates, ne...
2025-07-19T22:06:28
https://www.reddit.com/r/LocalLLaMA/comments/1m48ffs/ocr_and_genai_key_trends_from_h1_2025/
Careless_Bed_5075
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m48ffs
false
null
t3_1m48ffs
/r/LocalLLaMA/comments/1m48ffs/ocr_and_genai_key_trends_from_h1_2025/
false
false
self
8
null
Price performance comparison from the Gemini 2.5 Paper
185
Google claim Gemini own the pareto frontier. Deepseek looks good competitive.
2025-07-19T20:59:17
https://i.redd.it/032gntpz9wdf1.png
DeltaSqueezer
i.redd.it
1970-01-01T00:00:00
0
{}
1m46w7u
false
null
t3_1m46w7u
/r/LocalLLaMA/comments/1m46w7u/price_performance_comparison_from_the_gemini_25/
false
false
default
185
{'enabled': True, 'images': [{'id': '032gntpz9wdf1', 'resolutions': [{'height': 64, 'url': 'https://preview.redd.it/032gntpz9wdf1.png?width=108&crop=smart&auto=webp&s=dd0736e5beb6ce1c6f1167121fb4b30dc4f25bee', 'width': 108}, {'height': 128, 'url': 'https://preview.redd.it/032gntpz9wdf1.png?width=216&crop=smart&auto=web...
Can we finally "index" a code project?
52
If I understand how "tooling" works w/ newer LLMs now, I can take a large code project and "index" it in such a way that an LLM can "search" it like a database and answer questions regarding the source code? This is my #1 need at the moment, being able to get quick answers about my code base that's quite large. I don'...
2025-07-19T20:40:18
https://www.reddit.com/r/LocalLLaMA/comments/1m46gtn/can_we_finally_index_a_code_project/
CSEliot
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m46gtn
false
null
t3_1m46gtn
/r/LocalLLaMA/comments/1m46gtn/can_we_finally_index_a_code_project/
false
false
self
52
null
For a very specific text knowledge resource, can a local model outperform cloud models?
2
I'm a layperson when it comes to large language models. Just like learning about them and think local models are fascinating. I want to take the 2018 International Building Code (pdf or other text file) and create a focused AI model to converse with. The input would be something like" give me a building code analysis ...
2025-07-19T20:21:46
https://www.reddit.com/r/LocalLLaMA/comments/1m461jh/for_a_very_specific_text_knowledge_resource_can_a/
loac
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m461jh
false
null
t3_1m461jh
/r/LocalLLaMA/comments/1m461jh/for_a_very_specific_text_knowledge_resource_can_a/
false
false
self
2
null
I am really hoping the openai IMO announcement will motivate the open source community to match it
0
What do you think the chances are?
2025-07-19T20:11:08
https://www.reddit.com/r/LocalLLaMA/comments/1m45sh1/i_am_really_hoping_the_openai_imo_announcement/
MrMrsPotts
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m45sh1
false
null
t3_1m45sh1
/r/LocalLLaMA/comments/1m45sh1/i_am_really_hoping_the_openai_imo_announcement/
false
false
self
0
null
How would you write evals for chat apps running dozens of open models?
1
Hi all, I'm interviewing for a certain Half-Life provider (full-stack role, application layer) that prides itself on serving open models. I think there is a decent chance I'll be asked how to design a chat app in the systems design interview, and my biggest gap in knowledge is writing evals. The nature of a chat app ...
2025-07-19T20:07:58
https://www.reddit.com/r/LocalLLaMA/comments/1m45po2/how_would_you_write_evals_for_chat_apps_running/
ohcrap___fk
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m45po2
false
null
t3_1m45po2
/r/LocalLLaMA/comments/1m45po2/how_would_you_write_evals_for_chat_apps_running/
false
false
self
1
null
Keras vs Transformers fine tuning
4
I'm new to ML and fine tuning. Recently I've tried fine tuning gemma 3 on google collab on an 85k dataset (Dolly, Alpaca + custom) and it took 3 hours with Keras on a single A100 gpu. But then I couldn't convert it to pytorch because the conversion script by Keras doesn't support the gemma 3 yet and so I abandoned thi...
2025-07-19T19:30:35
https://www.reddit.com/r/LocalLLaMA/comments/1m44tnz/keras_vs_transformers_fine_tuning/
Ok-Refrigerator6609
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m44tnz
false
null
t3_1m44tnz
/r/LocalLLaMA/comments/1m44tnz/keras_vs_transformers_fine_tuning/
false
false
self
4
null
Keras vs Transformers fine tuning
1
[removed]
2025-07-19T19:29:26
https://www.reddit.com/r/LocalLLaMA/comments/1m44sn2/keras_vs_transformers_fine_tuning/
Palpitation_Common
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m44sn2
false
null
t3_1m44sn2
/r/LocalLLaMA/comments/1m44sn2/keras_vs_transformers_fine_tuning/
false
false
self
1
null
Keras vs Transformers finetuning
1
I'm new to ML and fine tuning. Recently I've tried fine tuning gemma 3 on google collab on an 85k dataset (Dolly, Alpaca + custom) and it took 3 hours with Keras on a single A100 gpu. But then I couldn't convert it to pytorch because the conversion script by Keras doesn't support the gemma 3 yet and so I abandoned thi...
2025-07-19T19:27:59
https://www.reddit.com/r/LocalLLaMA/comments/1m44rds/keras_vs_transformers_finetuning/
muxout__com
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m44rds
false
null
t3_1m44rds
/r/LocalLLaMA/comments/1m44rds/keras_vs_transformers_finetuning/
false
false
self
1
null
Hear me out, an LLM which is more like a dictionary to refer syntax from, and is trained that way.
0
What if instead of considering LLMs as magic code gen for full scale ideas/apps or snippets, we consider it as a dictionary and ask syntax specific questions and refer to it like a guidebook, rather than offloading the engineering decisions to it. So we can ask the LLM "syntax for x function of xyz stack for xyz tas...
2025-07-19T18:43:37
https://www.reddit.com/r/LocalLLaMA/comments/1m43owh/hear_me_out_an_llm_which_is_more_like_a/
fuckAIbruhIhateCorps
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m43owh
false
null
t3_1m43owh
/r/LocalLLaMA/comments/1m43owh/hear_me_out_an_llm_which_is_more_like_a/
false
false
self
0
null
Augmentoolkit uv command not found error
1
https://preview.redd.it/…this? Thanks!
2025-07-19T18:41:43
https://www.reddit.com/r/LocalLLaMA/comments/1m43na0/augmentoolkit_uv_command_not_found_error/
Eze010
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m43na0
false
null
t3_1m43na0
/r/LocalLLaMA/comments/1m43na0/augmentoolkit_uv_command_not_found_error/
false
false
https://b.thumbs.redditm…yhsJxgpw4gtM.jpg
1
null
Kimi K2 is less CCP censored than R1
0
Happy to see that it was able to answer 3/4 questions that R1 typically refuses or avoids. The Taiwan political status question was the only one where it regurgitated the same CCP party line as Deepseek does. This is a local deployment of UD-IQ\_3\_XSS.
2025-07-19T18:36:34
https://www.reddit.com/gallery/1m43isp
nomorebuttsplz
reddit.com
1970-01-01T00:00:00
0
{}
1m43isp
false
null
t3_1m43isp
/r/LocalLLaMA/comments/1m43isp/kimi_k2_is_less_ccp_censored_than_r1/
false
false
https://b.thumbs.redditm…3SDaiTQl4rUk.jpg
0
null
The Crucible Method
0
Dear All, I've spent a great deal of time (and money) exploring roleplay with LLMs. I've tested Opus, Sonnet, Gemini Pro, DeepSeek, Kimi K2, and others. Along the way, I’ve also tried many publicly available prompts floating around the internet. Here’s what I’ve discovered so far: • By design, LLMs are trained to fin...
2025-07-19T18:14:17
https://www.reddit.com/r/LocalLLaMA/comments/1m42zpd/the_crucible_method/
No_Weather1169
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m42zpd
false
null
t3_1m42zpd
/r/LocalLLaMA/comments/1m42zpd/the_crucible_method/
false
false
self
0
null
Why is download options blank and why is choose an action greyed out?
1
https://preview.redd.it/…I ran into this.
2025-07-19T18:08:00
https://www.reddit.com/r/LocalLLaMA/comments/1m42uel/why_is_download_options_blank_and_why_is_choose/
comsit1712
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m42uel
false
null
t3_1m42uel
/r/LocalLLaMA/comments/1m42uel/why_is_download_options_blank_and_why_is_choose/
false
false
https://b.thumbs.redditm…fMrAvEBMRf4U.jpg
1
null
How to speed up the initial inference when using llama.rn (llama.cpp) wrapper on android.
4
Hello Everyone, I'm working on a personal project where I'm using llama.rn (wrapper of llama.cpp). I'm trying to make an inference from local model (Gemma3n-E2B- INT4). Everything works fine. The only thing I'm struggling with is, the initial inference. The initial inference takes a lot of time. But the subsequent o...
2025-07-19T17:59:49
https://www.reddit.com/r/LocalLLaMA/comments/1m42n4v/how_to_speed_up_the_initial_inference_when_using/
luffy2998
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m42n4v
false
null
t3_1m42n4v
/r/LocalLLaMA/comments/1m42n4v/how_to_speed_up_the_initial_inference_when_using/
false
false
self
4
null
GPT-4o Updated: Has It Been Nerfed?
0
I’ve been hearing a lot on X about changes to 4o. This appears to be a very recent development (within the last day). Is this a nerf or a buff? Share your experiences! Let’s discuss.
2025-07-19T17:54:24
https://www.reddit.com/r/LocalLLaMA/comments/1m42iio/gpt4o_updated_has_it_been_nerfed/
Ok_Technology_3421
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m42iio
false
null
t3_1m42iio
/r/LocalLLaMA/comments/1m42iio/gpt4o_updated_has_it_been_nerfed/
false
false
self
0
null
Build advice: Consumer AI workstation with RTX 3090 + dual MI50s for LLM inference and Stable Diffusion (~$5k budget)
5
Looking for feedback on a mixed-use AI workstation build. Work is pushing me to get serious about local AI/model training or I'm basically toast career-wise, so trying to build something capable but not break the bank. Planned specs: CPU: Ryzen 9 9950X3D Mobo: X870E (eyeing ASUS ROG Crosshair Hero for expansion) ...
2025-07-19T17:51:59
https://www.reddit.com/r/LocalLLaMA/comments/1m42gid/build_advice_consumer_ai_workstation_with_rtx/
neighbornugs
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m42gid
false
null
t3_1m42gid
/r/LocalLLaMA/comments/1m42gid/build_advice_consumer_ai_workstation_with_rtx/
false
false
self
5
null
Running AIs Locally without a GPU: Context Window
2
You guys might've seen my earlier posts about the models I downloaded spitting out their chat template, looping around it, etc etc. I fixed it and I really appreciate the comments. Now, this is something I couldn't find a solution online. I only have 16GB of RAM, no dGPU, on a mobile CPU. I managed to run Gemma-3 4B-Q...
2025-07-19T17:46:55
https://www.reddit.com/r/LocalLLaMA/comments/1m42c2q/running_ais_locally_without_a_gpu_context_window/
Leather_Flan5071
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m42c2q
false
null
t3_1m42c2q
/r/LocalLLaMA/comments/1m42c2q/running_ais_locally_without_a_gpu_context_window/
false
false
self
2
null
Any idea when llama 4 behemoth will be released?
0
Haven't heard any updates regarding this model since a few months.. Was it much stronger than they expected and they decided not to release it publicly? 🤔
2025-07-19T17:09:20
https://www.reddit.com/r/LocalLLaMA/comments/1m41f79/any_idea_when_llama_4_behemoth_will_be_released/
Shubham_Garg123
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m41f79
false
null
t3_1m41f79
/r/LocalLLaMA/comments/1m41f79/any_idea_when_llama_4_behemoth_will_be_released/
false
false
self
0
null
A Request for Comments (RFC) for MCP-alternative Universal Tool Calling Protocol (UTCP) was created
68
After the extensie discussion [about UTCP](https://www.reddit.com/r/LocalLLaMA/comments/1lzl5zk/utcp_a_safer_scalable_toolcalling_alternative_to/) last week, the authors of UTCP created an RFC for it. >This document proposes the Universal Tool Calling Protocol (UTCP), a specification that enables applications, includi...
2025-07-19T17:05:06
https://github.com/universal-tool-calling-protocol/utcp-specification/issues/18
Balance-
github.com
1970-01-01T00:00:00
0
{}
1m41bj1
false
null
t3_1m41bj1
/r/LocalLLaMA/comments/1m41bj1/a_request_for_comments_rfc_for_mcpalternative/
false
false
default
68
{'enabled': False, 'images': [{'id': 'OQN7MvSNPLnJfNZ8ubE2vL4vsUeHZB2oAu_947PfgqQ', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/OQN7MvSNPLnJfNZ8ubE2vL4vsUeHZB2oAu_947PfgqQ.png?width=108&crop=smart&auto=webp&s=8d071be9044b5b0bdc3c6becd70df063f85399f3', 'width': 108}, {'height': 108, 'url': 'h...
🧠 ECHO v4.0 — The Ultimate GitHub Copilot Prompt That Turns AI Into an Actual Engineer
1
[removed]
2025-07-19T16:59:59
https://www.reddit.com/r/LocalLLaMA/comments/1m416v1/echo_v40_the_ultimate_github_copilot_prompt_that/
fame0x
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m416v1
false
null
t3_1m416v1
/r/LocalLLaMA/comments/1m416v1/echo_v40_the_ultimate_github_copilot_prompt_that/
false
false
self
1
null
any lovable and bolt alternative open source?
8
hi i love playing with those stuff create stuff for fun, but i have 0 code knowledge. i want to use api of openai or or anthropic . is there any open source that its like lovable and bolt but i use openai api and results are good?
2025-07-19T16:50:17
https://www.reddit.com/r/LocalLLaMA/comments/1m40yo6/any_lovable_and_bolt_alternative_open_source/
yuval052
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m40yo6
false
null
t3_1m40yo6
/r/LocalLLaMA/comments/1m40yo6/any_lovable_and_bolt_alternative_open_source/
false
false
self
8
null
Image processing limit on Groq...alternatives?
0
Groq has a limit of 5 images that can be processed per request with Scout and Maverick LLMs. Anyone have suggestions on alternatives that support at least 10 images?
2025-07-19T16:38:04
https://www.reddit.com/r/LocalLLaMA/comments/1m40o0v/image_processing_limit_on_groqalternatives/
instigator-x
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m40o0v
false
null
t3_1m40o0v
/r/LocalLLaMA/comments/1m40o0v/image_processing_limit_on_groqalternatives/
false
false
self
0
null
Best tools for local AI memory?
1
Had a longer post about my specific motivations and more details.. but probably auto-blocked. I am a cryptographer that works on privacy preserving local verifiable compute. Does anyone know of research / tools that work for local AI memory / potentially across devices? Thanks.
2025-07-19T15:47:41
https://www.reddit.com/r/LocalLLaMA/comments/1m3zg5k/best_tools_for_local_ai_memory/
popocat93
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3zg5k
false
null
t3_1m3zg5k
/r/LocalLLaMA/comments/1m3zg5k/best_tools_for_local_ai_memory/
false
false
self
1
null
Self-sovereign and portable AI memory
1
[removed]
2025-07-19T15:37:20
https://www.reddit.com/r/LocalLLaMA/comments/1m3z79u/selfsovereign_and_portable_ai_memory/
popocat93
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3z79u
false
null
t3_1m3z79u
/r/LocalLLaMA/comments/1m3z79u/selfsovereign_and_portable_ai_memory/
false
false
self
1
null
Localllama’s (first?) IFTA - I’ll Fine-Tune Anything
65
Following a comment I made on another post here that failed to come to fruition, I’ve decided to step it up. I’ve got some GPU resources, we (the community) have a ton of cool ideas - let’s make this happen. Premise is pretty simple, comment below with an idea for a fine-tune, any kind, any open weights model, any pur...
2025-07-19T15:28:11
https://www.reddit.com/r/LocalLLaMA/comments/1m3yzes/localllamas_first_ifta_ill_finetune_anything/
indicava
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3yzes
false
null
t3_1m3yzes
/r/LocalLLaMA/comments/1m3yzes/localllamas_first_ifta_ill_finetune_anything/
false
false
self
65
null
🚨 Stealth Vocab Injections in llama.cpp? I Never Installed These. You? [🔥Image Proof Included]
0
Hey folks — I’m building a fully offline, self-evolving Fractal AI Memory System (no HuggingFace sync, no DeepSeek install, no OpenAccess shenanigans), and during a forensic audit of my llama.cpp environment… I found this: 📸 (see image) Timestamp: 2025-03-13 @ 01:23 AM Location: /models/ggml-vocab-*.gguf --- ❗ Wh...
2025-07-19T15:26:41
https://i.redd.it/xzffm6f1nudf1.jpeg
Mirror_Solid
i.redd.it
1970-01-01T00:00:00
0
{}
1m3yy5a
false
null
t3_1m3yy5a
/r/LocalLLaMA/comments/1m3yy5a/stealth_vocab_injections_in_llamacpp_i_never/
false
false
default
0
{'enabled': True, 'images': [{'id': 'xzffm6f1nudf1', 'resolutions': [{'height': 32, 'url': 'https://preview.redd.it/xzffm6f1nudf1.jpeg?width=108&crop=smart&auto=webp&s=d3c701fb2439feed1a372685fe9b6e04a57bdd98', 'width': 108}, {'height': 64, 'url': 'https://preview.redd.it/xzffm6f1nudf1.jpeg?width=216&crop=smart&auto=we...
Motherboard with 2 PCI Express running at full 16x/16x
2
Hello folks, I'm building a new PC that will also be used for running local LLMs. I would like the possibility of using a decent LLM for programming work. Someone recommended: * buying a motherboard with 2 PCI Express 16x slots * buying 2 "cheaper" identical 16GB CPUs * splitting the model to run on both of them (f...
2025-07-19T14:47:13
https://www.reddit.com/r/LocalLLaMA/comments/1m3y0m8/motherboard_with_2_pci_express_running_at_full/
oblio-
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3y0m8
false
null
t3_1m3y0m8
/r/LocalLLaMA/comments/1m3y0m8/motherboard_with_2_pci_express_running_at_full/
false
false
self
2
null
What would be a great roadmap for jumping into local LMM for a pretty newbie?
0
I mean, I’m quite smart and easily get into complex from cognitive perspective things; and that’s my scope of interest for quite a while. I don’t have fancy GPU yet, mine is 1650Ti MaxQ and i7 of 9nth gen; so what could I learn/try to become an expert in this field. I will update equipment like in few months perhaps, s...
2025-07-19T14:40:18
https://www.reddit.com/r/LocalLLaMA/comments/1m3xuqx/what_would_be_a_great_roadmap_for_jumping_into/
MoneyMultiplier888
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3xuqx
false
null
t3_1m3xuqx
/r/LocalLLaMA/comments/1m3xuqx/what_would_be_a_great_roadmap_for_jumping_into/
false
false
self
0
null
ChatSong, a lightweight, local LLM chat tool that's a single executable file
41
**Hello everyone,** I built a lightweight LLM API invocation tool that requires no installation, just a single executable file. **Features:** - Truly Portable: It's a single executable file, no installation required. - Bring Your Own Model: Customize models and prompts easily through a config file. - Save & Share: E...
2025-07-19T14:33:28
https://i.redd.it/jcc7hsejdudf1.jpeg
Suitable-Patience916
i.redd.it
1970-01-01T00:00:00
0
{}
1m3xp21
false
null
t3_1m3xp21
/r/LocalLLaMA/comments/1m3xp21/chatsong_a_lightweight_local_llm_chat_tool_thats/
false
false
https://b.thumbs.redditm…Pqp_7q7bGiZc.jpg
41
{'enabled': True, 'images': [{'id': 'WMDeqFE2i_t38-GTpkNLecOU6beSdnhaGRelWtgxoV0', 'resolutions': [{'height': 50, 'url': 'https://preview.redd.it/jcc7hsejdudf1.jpeg?width=108&crop=smart&auto=webp&s=a942b53502ccdf0f0527df30d6c288d223c4d984', 'width': 108}, {'height': 101, 'url': 'https://preview.redd.it/jcc7hsejdudf1.jp...
Dual GPU set up was surprisingly easy
123
First build of a new rig for running local LLMs, I wanted to see if there would be much frigging around needed to get both GPUs running, but pleasantly surprised it all just worked fine. Combined 28Gb VRAM. Running the 5070 as primary GPU due to it better memory bandwidth and more CUDA cores than the 5060 Ti. Both in...
2025-07-19T14:23:11
https://www.reddit.com/gallery/1m3xgjo
m-gethen
reddit.com
1970-01-01T00:00:00
0
{}
1m3xgjo
false
null
t3_1m3xgjo
/r/LocalLLaMA/comments/1m3xgjo/dual_gpu_set_up_was_surprisingly_easy/
false
false
https://a.thumbs.redditm…vmNjPPEfS1l0.jpg
123
null
The Final build: help me finish a CPU FIRST hybrid MOE rig
1
First, thank you so much to everyone who has helped me work through and suggested how to build out my rig. For those of you who haven’t seen those, I have posted twice with slightly different ideas and let me tell you this community has shown up! I have to taken this approach as the technical side of hybrid inferen...
2025-07-19T14:17:01
https://www.reddit.com/r/LocalLLaMA/comments/1m3xbj7/the_final_build_help_me_finish_a_cpu_first_hybrid/
novel_market_21
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3xbj7
false
null
t3_1m3xbj7
/r/LocalLLaMA/comments/1m3xbj7/the_final_build_help_me_finish_a_cpu_first_hybrid/
false
false
self
1
null
AMD 6x7900xtx + VLLM + Docker + QWEN3-235B error
3
Hello! I try to launch qwen3 235b using VLLM and stuck on different problems, one of them i got `AttributeError: '_OpNamespace' '_C' object has no attribute 'gptq_marlin_repack'` and no way to fix it. i got this on vllm in docker and vllm builded from source. services:   vllm:     pull_policy: al...
2025-07-19T14:02:17
https://www.reddit.com/r/LocalLLaMA/comments/1m3wzm9/amd_6x7900xtx_vllm_docker_qwen3235b_error/
djdeniro
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3wzm9
false
null
t3_1m3wzm9
/r/LocalLLaMA/comments/1m3wzm9/amd_6x7900xtx_vllm_docker_qwen3235b_error/
false
false
self
3
null
Any package that provides treesitter-based mark commands?
0
Similar to `mark-word`, I'm looking for something that provides something like `mark-function`, `mark-class`, `mark-condition`, `mark-loop`, `mark-declaration`, etc. that uses tree-sitter. Is anything like this available?
2025-07-19T13:53:44
https://www.reddit.com/r/LocalLLaMA/comments/1m3wslq/any_package_that_provides_treesitterbased_mark/
kudikarasavasa
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3wslq
false
null
t3_1m3wslq
/r/LocalLLaMA/comments/1m3wslq/any_package_that_provides_treesitterbased_mark/
false
false
self
0
null
Are there any quants of larger models 48 VRAM + 96 RAM can run, which are better than just 32B models?
14
I've built myself a PC with 2x 3090, because I thought it would be a sweetspot to start with for something twice as capable than a regular single-card PC yet still fitting a regular case. However most models still seem to be either targeted at a single card, or at a server. I also likely made a mistake by using an OC...
2025-07-19T13:48:31
https://www.reddit.com/r/LocalLLaMA/comments/1m3wogu/are_there_any_quants_of_larger_models_48_vram_96/
West_Investigator258
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3wogu
false
null
t3_1m3wogu
/r/LocalLLaMA/comments/1m3wogu/are_there_any_quants_of_larger_models_48_vram_96/
false
false
self
14
null
What's New in Agent Leaderboard v2?
56
**Here is a quick TL;DR 👇** 🧠 **GPT-4.1** tops with 62% Action Completion (AC) overall. ⚡ **Gemini 2.5** Flash excels in tool use (94% TSQ) but lags in task completion (38% AC). 💸 **GPT-4.1**\-mini is *most cost-effective* at $0.014/session vs. GPT-4.1’s $0.068. 🏭 No single model dominates across industries....
2025-07-19T13:47:25
https://i.redd.it/bwu8hq345udf1.png
5h3r_10ck
i.redd.it
1970-01-01T00:00:00
0
{}
1m3wnnm
false
null
t3_1m3wnnm
/r/LocalLLaMA/comments/1m3wnnm/whats_new_in_agent_leaderboard_v2/
false
false
default
56
{'enabled': True, 'images': [{'id': 'bwu8hq345udf1', 'resolutions': [{'height': 69, 'url': 'https://preview.redd.it/bwu8hq345udf1.png?width=108&crop=smart&auto=webp&s=1586a4e62e31363148fbc6fc947cedd8323aa3c4', 'width': 108}, {'height': 139, 'url': 'https://preview.redd.it/bwu8hq345udf1.png?width=216&crop=smart&auto=web...
Looking for `113-D1631711QA-10` vBIOS for AMD MI50 32GB
4
Someone posted that this vBIOS should work to expose full 32GB VRAM on Vulkan for AMD MI50, but the poster has disappeared since. If you're that person or someone else who has this VBIOS, could you please upload and share it? Tyvm `^^`
2025-07-19T13:28:33
https://www.reddit.com/r/LocalLLaMA/comments/1m3w96r/looking_for_113d1631711qa10_vbios_for_amd_mi50/
ashirviskas
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3w96r
false
null
t3_1m3w96r
/r/LocalLLaMA/comments/1m3w96r/looking_for_113d1631711qa10_vbios_for_amd_mi50/
false
false
self
4
null
A new paper from Apple shows you can tack on Multi-Token Prediction to any LLM with no loss in quality
441
TLDR: for a small overhead of additional trained parameters, you can get 2.5-5x more tokens per second.
2025-07-19T13:03:43
https://arxiv.org/abs/2507.11851
Kooshi_Govno
arxiv.org
1970-01-01T00:00:00
0
{}
1m3vqom
false
null
t3_1m3vqom
/r/LocalLLaMA/comments/1m3vqom/a_new_paper_from_apple_shows_you_can_tack_on/
false
false
default
441
null
How do some of these, open source, smaller models out perform larger, closed source models?
0
I'll preface this post by saying I'm a complete newbie in this realm, but I'm fascinated by it. I've only just started looking into LLM stuff on the back buying a 5090 for gaming. I tried Ollama on Windows and a week later I had bought a gpu for my home server so I could learn and tinker in a better environment. To ...
2025-07-19T12:49:07
https://www.reddit.com/r/LocalLLaMA/comments/1m3vg43/how_do_some_of_these_open_source_smaller_models/
Ev0kes
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3vg43
false
null
t3_1m3vg43
/r/LocalLLaMA/comments/1m3vg43/how_do_some_of_these_open_source_smaller_models/
false
false
self
0
null
3060 12gb useful (pair with 3080 10gb?)
0
Hi, I have a RTX 3080 with 10gb of ram, seems pretty quick with vllm running qwen2.5 coder 7b. I have the option to buy a 3060 but with 12gb (pretty cheap at AUD$200 I believe), I need to figure out how to fit it in (mainly power) but is it worth bothering? Anyone running one? Attached is what I got from copilot (so...
2025-07-19T11:39:51
https://i.redd.it/j3ak3nhkitdf1.jpeg
johnerp
i.redd.it
1970-01-01T00:00:00
0
{}
1m3u4rl
false
null
t3_1m3u4rl
/r/LocalLLaMA/comments/1m3u4rl/3060_12gb_useful_pair_with_3080_10gb/
false
false
default
0
{'enabled': True, 'images': [{'id': 'j3ak3nhkitdf1', 'resolutions': [{'height': 116, 'url': 'https://preview.redd.it/j3ak3nhkitdf1.jpeg?width=108&crop=smart&auto=webp&s=d67bb0dcf651cb1b9cbf5f982af5b1f1cbcff988', 'width': 108}, {'height': 233, 'url': 'https://preview.redd.it/j3ak3nhkitdf1.jpeg?width=216&crop=smart&auto=...
I love local models
50
2025-07-19T11:06:55
https://i.redd.it/k7ebpl1nctdf1.png
TweeMansLeger
i.redd.it
1970-01-01T00:00:00
0
{}
1m3tk92
false
null
t3_1m3tk92
/r/LocalLLaMA/comments/1m3tk92/i_love_local_models/
false
false
https://b.thumbs.redditm…-zzE5xn4aY7A.jpg
50
{'enabled': True, 'images': [{'id': 'a2m6l7_RYwfI11tlzGuLF9ptF4wRufgEXfSnzKQ9hmc', 'resolutions': [{'height': 26, 'url': 'https://preview.redd.it/k7ebpl1nctdf1.png?width=108&crop=smart&auto=webp&s=f7fd2866d56b1cb6f84df32f67409180e0f9ee8b', 'width': 108}, {'height': 53, 'url': 'https://preview.redd.it/k7ebpl1nctdf1.png?...
What's the best model I can run on this laptop?
0
i9 14900HX RTX 4060 8gb VRAM 32gb DDR5 5600mhz RAM
2025-07-19T10:29:33
https://www.reddit.com/r/LocalLLaMA/comments/1m3sym1/whats_the_best_model_i_can_run_on_this_laptop/
Rif_Reddit
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3sym1
false
null
t3_1m3sym1
/r/LocalLLaMA/comments/1m3sym1/whats_the_best_model_i_can_run_on_this_laptop/
false
false
self
0
null
ARC AGI 3 is stupid
77
On the first game, first level of 8, I completed the level after wasting a lot of time trying to figure out what functionality the spacebar and mouse clicks had. None, it turned out. On the second level, I got completely stuck, then read in another thread that you have to move on and off the first shape several times t...
2025-07-19T10:17:53
https://www.reddit.com/r/LocalLLaMA/comments/1m3ssb2/arc_agi_3_is_stupid/
jackdareel
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3ssb2
false
null
t3_1m3ssb2
/r/LocalLLaMA/comments/1m3ssb2/arc_agi_3_is_stupid/
false
false
self
77
null
I want to create a local AI Agent that can call tools. but my model call tools even for "hey"
1
Can you guys please tell me what am i doing wrong here. My model keeps calling tool for every response, even if it's not necessary even for simple "hey". import ollama from tools import ( read\_file, write\_file, ) class Cron: def \_\_init\_\_(self, model\_name: str = "llama3.1:latest", mood : str = "s...
2025-07-19T10:12:36
https://www.reddit.com/r/LocalLLaMA/comments/1m3spek/i_want_to_create_a_local_ai_agent_that_can_call/
Prajwell
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3spek
false
null
t3_1m3spek
/r/LocalLLaMA/comments/1m3spek/i_want_to_create_a_local_ai_agent_that_can_call/
false
false
self
1
{'enabled': False, 'images': [{'id': 'BEwFUWf7iw_9SmyBYo3NT6CsNdUKu1iTR3ZfOlZNH8A', 'resolutions': [{'height': 72, 'url': 'https://external-preview.redd.it/BEwFUWf7iw_9SmyBYo3NT6CsNdUKu1iTR3ZfOlZNH8A.png?width=108&crop=smart&auto=webp&s=0b7e69de4f50766ebc67d04ef0b2f86b946090a0', 'width': 108}, {'height': 144, 'url': 'h...
Trouble running MythoMax-L2-13B-GPTQ on RunPod – Model loads but returns empty responses
2
Hi everyone, I'm trying to run MythoMax-L2-13B-GPTQ on RunPod using the text-generation-webui (Oobabooga). The model loads, the WebUI starts fine, and I can open the interface. However, when I try to generate text, the model just replies with empty lines or no output at all. Here's what I've tried: Launched the pod ...
2025-07-19T10:07:28
https://www.reddit.com/r/LocalLLaMA/comments/1m3smiz/trouble_running_mythomaxl213bgptq_on_runpod_model/
Icy_Blacksmith8549
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3smiz
false
null
t3_1m3smiz
/r/LocalLLaMA/comments/1m3smiz/trouble_running_mythomaxl213bgptq_on_runpod_model/
false
false
self
2
null
WordPecker: Open Source Personalized Duolingo
131
[https://github.com/baturyilmaz/wordpecker-app](https://github.com/baturyilmaz/wordpecker-app)
2025-07-19T09:57:11
https://v.redd.it/5fximscazsdf1
arbayi
/r/LocalLLaMA/comments/1m3sgr1/wordpecker_open_source_personalized_duolingo/
1970-01-01T00:00:00
0
{}
1m3sgr1
false
{'reddit_video': {'bitrate_kbps': 5000, 'dash_url': 'https://v.redd.it/5fximscazsdf1/DASHPlaylist.mpd?a=1755640637%2CYTUyNzRhOTBiOTFmNzBhYTEwYzNjNWY5ZGI3MTEzNTEyNTE2ZmRlYjdmMjZkMGYyMmVkZmMwZDkyNTY2OWJjNw%3D%3D&v=1&f=sd', 'duration': 197, 'fallback_url': 'https://v.redd.it/5fximscazsdf1/DASH_1080.mp4?source=fallback', '...
t3_1m3sgr1
/r/LocalLLaMA/comments/1m3sgr1/wordpecker_open_source_personalized_duolingo/
false
false
https://external-preview…e427853e208d7874
131
{'enabled': False, 'images': [{'id': 'NGdqZmJ0Y2F6c2RmMc80rXNWOGm_7GTyts0LIbg_WRs-xM2_snleUNsHfx6L', 'resolutions': [{'height': 60, 'url': 'https://external-preview.redd.it/NGdqZmJ0Y2F6c2RmMc80rXNWOGm_7GTyts0LIbg_WRs-xM2_snleUNsHfx6L.png?width=108&crop=smart&format=pjpg&auto=webp&s=112084b99db8db53f35754e3b1b10431f2bf0...
Structured output help (LM Studio)
1
I'm trying to get MistralThinker to... think. According to discussion on the model page (https://huggingface.co/Undi95/MistralThinker-v1.1/discussions/1) it is necessary to encourage the model to use reasoning with some structured output or otherwise prefixes. But I'm not using SillyTavern so the suggestions in the thr...
2025-07-19T09:25:48
https://www.reddit.com/r/LocalLLaMA/comments/1m3s01i/structured_output_help_lm_studio/
Jawzper
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3s01i
false
null
t3_1m3s01i
/r/LocalLLaMA/comments/1m3s01i/structured_output_help_lm_studio/
false
false
self
1
{'enabled': False, 'images': [{'id': 'ytmT6spTi9ZIVCmBZ4Xdi8PdPPKQAGuJJuT6HnMnshk', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/ytmT6spTi9ZIVCmBZ4Xdi8PdPPKQAGuJJuT6HnMnshk.png?width=108&crop=smart&auto=webp&s=965bf8099f34d87a454fb8dd1f546be43d598fb7', 'width': 108}, {'height': 116, 'url': 'h...
Are You at Risk for Heart Disease or Stroke? Discover Your Health Risks with a Quick Screening!
1
[removed]
2025-07-19T09:10:29
https://www.reddit.com/r/LocalLLaMA/comments/1m3rs25/are_you_at_risk_for_heart_disease_or_stroke/
krithika_reddits
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3rs25
false
null
t3_1m3rs25
/r/LocalLLaMA/comments/1m3rs25/are_you_at_risk_for_heart_disease_or_stroke/
false
false
self
1
{'enabled': False, 'images': [{'id': 'Ry47wbhUxYF42OpmNCUOVUK8dIhvMQgQSif8uE0dLsw', 'resolutions': [{'height': 107, 'url': 'https://external-preview.redd.it/Ry47wbhUxYF42OpmNCUOVUK8dIhvMQgQSif8uE0dLsw.jpeg?width=108&crop=smart&auto=webp&s=3b5a317773f21592a1adec21dccbc1c1913d2c07', 'width': 108}, {'height': 214, 'url': ...
Never stop fighting for obliterated/uncensored models
0
2025-07-19T09:07:53
https://i.redd.it/sqm6n495rsdf1.png
Elisabella2005
i.redd.it
1970-01-01T00:00:00
0
{}
1m3rqqm
false
null
t3_1m3rqqm
/r/LocalLLaMA/comments/1m3rqqm/never_stop_fighting_for_obliterateduncensored/
false
false
https://a.thumbs.redditm…U3EA0wZhsLn8.jpg
0
{'enabled': True, 'images': [{'id': 'i_dU-jYaq18D8Gg_PTFCSl1r64yiJOP9d4GcD0Q5hsc', 'resolutions': [{'height': 47, 'url': 'https://preview.redd.it/sqm6n495rsdf1.png?width=108&crop=smart&auto=webp&s=6da18fff8911e97289f0bd1ce9486a8091aa34b0', 'width': 108}, {'height': 95, 'url': 'https://preview.redd.it/sqm6n495rsdf1.png?...
external usb4 dock for two or more egpu
1
Does it exist? Can anyone tell me where to buy a dock like this, even for just two eGPUs?
2025-07-19T09:06:17
https://www.reddit.com/r/LocalLLaMA/comments/1m3rpx1/external_usb4_dock_for_two_or_more_egpu/
Bobcotelli
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3rpx1
false
null
t3_1m3rpx1
/r/LocalLLaMA/comments/1m3rpx1/external_usb4_dock_for_two_or_more_egpu/
false
false
self
1
null
The strongest wills… until they see $1.99 B200s
0
2025-07-19T09:02:39
https://v.redd.it/6gbbossiqsdf1
nueid
v.redd.it
1970-01-01T00:00:00
0
{}
1m3ro3y
false
{'reddit_video': {'bitrate_kbps': 5000, 'dash_url': 'https://v.redd.it/6gbbossiqsdf1/DASHPlaylist.mpd?a=1755507775%2CNzVkM2ZmZjRmMjIzNzU5ZTUyYjVlMGQyZDQzNzEyNzc3M2I0MGRmZDZlNmRlZjAwMTQ5ZmQ3MTFmNGNiNjcxNw%3D%3D&v=1&f=sd', 'duration': 5, 'fallback_url': 'https://v.redd.it/6gbbossiqsdf1/DASH_1080.mp4?source=fallback', 'ha...
t3_1m3ro3y
/r/LocalLLaMA/comments/1m3ro3y/the_strongest_wills_until_they_see_199_b200s/
false
false
https://external-preview…b58f0a0580e7c6c6
0
{'enabled': False, 'images': [{'id': 'MTU5aWFhbWlxc2RmMaxJIjUPpZ6nxmqTuYUi59RF-kZrkh3PPybesxzvDtIe', 'resolutions': [{'height': 135, 'url': 'https://external-preview.redd.it/MTU5aWFhbWlxc2RmMaxJIjUPpZ6nxmqTuYUi59RF-kZrkh3PPybesxzvDtIe.png?width=108&crop=smart&format=pjpg&auto=webp&s=d15753c1cd8a486616c45574b5c4f745815d...
I made the CLI for AWS S3 Vectors (Preview)
2
AWS released S3 Vectors in preview, but there's no web console and you need boto3 to use it. I wanted something quicker for testing, so I built a CLI in Rust. [welcome image](https://preview.redd.it/ky5j4tk8qsdf1.png?width=1802&format=png&auto=webp&s=8ca473be4929612f52acc1a9a0524d80d6ead2dd) **GitHub**: [https://gith...
2025-07-19T09:02:13
https://www.reddit.com/r/LocalLLaMA/comments/1m3rnvw/i_made_the_cli_for_aws_s3_vectors_preview/
Ok_Rub1689
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3rnvw
false
null
t3_1m3rnvw
/r/LocalLLaMA/comments/1m3rnvw/i_made_the_cli_for_aws_s3_vectors_preview/
false
false
https://b.thumbs.redditm…6q1uZSMDe5tU.jpg
2
{'enabled': False, 'images': [{'id': 'MOehlVJIiEgLpSM7sr8uFzrHq-HSoUDZUTb0kPE7lBk', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/MOehlVJIiEgLpSM7sr8uFzrHq-HSoUDZUTb0kPE7lBk.png?width=108&crop=smart&auto=webp&s=efa8563e2b32d98b08cbea9270cc7f7593165f9f', 'width': 108}, {'height': 108, 'url': 'h...
llama.cpp running too slow
0
I'm running the same model on llama.cpp as I do with kobold.cpp. KCPP has very fast outputs while LCPP is considerably more sluggish. I run llama-server with -ngl 100, but the output time is seemingly unchanged. Is this just how it's meant to be, or can I fix it somehow?
2025-07-19T08:50:55
https://www.reddit.com/r/LocalLLaMA/comments/1m3rhy2/llamacpp_running_too_slow/
bridgebucket
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3rhy2
false
null
t3_1m3rhy2
/r/LocalLLaMA/comments/1m3rhy2/llamacpp_running_too_slow/
false
false
self
0
null
Offline STT in real time?
6
Whats the best solution if you want to transcribe your voice to text in real time, locally? Not saving it in an audio file and have it transcribed after. Any easy to use one click GUI solutions like LMstudio for this?
2025-07-19T08:33:43
https://www.reddit.com/r/LocalLLaMA/comments/1m3r8jb/offline_stt_in_real_time/
Sea-Replacement7541
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3r8jb
false
null
t3_1m3r8jb
/r/LocalLLaMA/comments/1m3r8jb/offline_stt_in_real_time/
false
false
self
6
null
Any offline real time speech-to-text solution?
1
Whisper can take an audio file and transcribe it. Works very well. But I havent seen a solution for transcription in real time. I would like a solution where I speak into my microphone, have it transcribed in real time so that I can read the text, pause, edit manually and move on speaking. Windows has a great bui...
2025-07-19T08:26:19
https://www.reddit.com/r/LocalLLaMA/comments/1m3r4hr/any_offline_real_time_speechtotext_solution/
Sweden1231230
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3r4hr
false
null
t3_1m3r4hr
/r/LocalLLaMA/comments/1m3r4hr/any_offline_real_time_speechtotext_solution/
false
false
self
1
null
What are the most intriguing AI papers of 2025
59
I've been keeping up with AI research in 2025, and DeepSeek R1 really stands out to me as game-changing. What other papers from this year do you consider to be truly revolutionary?
2025-07-19T08:00:18
https://www.reddit.com/r/LocalLLaMA/comments/1m3qpxz/what_are_the_most_intriguing_ai_papers_of_2025/
VR-Person
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3qpxz
false
null
t3_1m3qpxz
/r/LocalLLaMA/comments/1m3qpxz/what_are_the_most_intriguing_ai_papers_of_2025/
false
false
self
59
null
Escaping quantization brain damage with BF16?
0
I have been trying various LLMs running locally (on a 64GB DDR4 Threadripper + 5090 box, on llama.cpp) to try to arrive at a co-maintainer for my established FOSS project. I would like it to see the code and propose patches in diff (or direct to git by MCP) form. My current theory is that the pressure to run quantize...
2025-07-19T07:43:08
https://www.reddit.com/r/LocalLLaMA/comments/1m3qg3w/escaping_quantization_brain_damage_with_bf16/
bitrumpled
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3qg3w
false
null
t3_1m3qg3w
/r/LocalLLaMA/comments/1m3qg3w/escaping_quantization_brain_damage_with_bf16/
false
false
self
0
null
Viability of the Threadripper Platform for a General Purpose AI+Gaming Machine?
4
Trying to build a workstation PC that can "Do it all" with a budget of some \~$8000, and a build around the upcoming Threadrippers is beginning to seem quite appealing. I suspect my use case is far from niche (Being Generic it's the opposite), so a thread discussing this could serve some purpose for the people. By "Ge...
2025-07-19T07:40:18
https://www.reddit.com/r/LocalLLaMA/comments/1m3qejc/viability_of_the_threadripper_platform_for_a/
FluffnPuff_Rebirth
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3qejc
false
null
t3_1m3qejc
/r/LocalLLaMA/comments/1m3qejc/viability_of_the_threadripper_platform_for_a/
false
false
self
4
null
[Experiment] Pushing Gemini-1.5-Pro into “raw-token tremor” without jailbreaks (4-step prompt chain inside, fully reproducible)
1
[removed]
2025-07-19T07:39:43
https://i.redd.it/8re4406qbsdf1.jpeg
General-Listen-5093
i.redd.it
1970-01-01T00:00:00
0
{}
1m3qe6x
false
null
t3_1m3qe6x
/r/LocalLLaMA/comments/1m3qe6x/experiment_pushing_gemini15pro_into_rawtoken/
false
false
https://b.thumbs.redditm…i3O8_BtqIBrc.jpg
1
{'enabled': True, 'images': [{'id': '-1EGOcGlskapY482ZrAZaDPCn-_sI8j2sWFsiVK0Wrk', 'resolutions': [{'height': 216, 'url': 'https://preview.redd.it/8re4406qbsdf1.jpeg?width=108&crop=smart&auto=webp&s=2b29c9b0e1cb275b110ab1911764ceeb75781f10', 'width': 108}, {'height': 432, 'url': 'https://preview.redd.it/8re4406qbsdf1.j...
Would there be a reasoning version of Kimi K2?
21
This model is really fascinating. I find it absolutely amazing. I believe that if this model gets added reasoning abilities it will beat absolutely everything on the market right now.
2025-07-19T07:35:55
https://www.reddit.com/r/LocalLLaMA/comments/1m3qc1g/would_there_be_a_reasoning_version_of_kimi_k2/
christian7670
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3qc1g
false
null
t3_1m3qc1g
/r/LocalLLaMA/comments/1m3qc1g/would_there_be_a_reasoning_version_of_kimi_k2/
false
false
self
21
null
What is the best small model for summarization for a low spec pc?
1
I run a modest PC with 16GB of RAM and a Ryzen 2200g, what is the most suitable model for summarization for these specs? doesn't have to be fast, I can let it run overnight. If it matters, I'll be using Jina's reader API to scrape some websites and get LLM ready MD text, but I need to classify the urls based on their ...
2025-07-19T06:39:12
https://www.reddit.com/r/LocalLLaMA/comments/1m3pg0s/what_is_the_best_small_model_for_summarization/
north_akando
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3pg0s
false
null
t3_1m3pg0s
/r/LocalLLaMA/comments/1m3pg0s/what_is_the_best_small_model_for_summarization/
false
false
self
1
null
Best Russian language conversational model?
3
I'm looking for the best model for practicing my Russian, something that can understand Russian well, will consistently use proper grammar, and can translate between English and Russian. Ideally <32B parameters, but if something larger will give a significant uplift I'd be interested to hear other options. This model d...
2025-07-19T06:37:19
https://www.reddit.com/r/LocalLLaMA/comments/1m3pez5/best_russian_language_conversational_model/
OUT_OF_HOST_MEMORY
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3pez5
false
null
t3_1m3pez5
/r/LocalLLaMA/comments/1m3pez5/best_russian_language_conversational_model/
false
false
self
3
null
Local deep research that web searches only academic sources?
13
I work in medicine, and I basically want something similar to [OpenEvidence](https://www.openevidence.com/), but local and totally private because I don’t like the idea of putting patient information in a website, even if they claim to be HIPAA compliant.
2025-07-19T05:59:03
https://www.reddit.com/r/LocalLLaMA/comments/1m3osbo/local_deep_research_that_web_searches_only/
Amazydayzee
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3osbo
false
null
t3_1m3osbo
/r/LocalLLaMA/comments/1m3osbo/local_deep_research_that_web_searches_only/
false
false
self
13
{'enabled': False, 'images': [{'id': 'r0K2nZ-d1qwMsSX32_v3aVtpZ9M02-uKIaafMaYpN-g', 'resolutions': [{'height': 56, 'url': 'https://external-preview.redd.it/r0K2nZ-d1qwMsSX32_v3aVtpZ9M02-uKIaafMaYpN-g.jpeg?width=108&crop=smart&auto=webp&s=b99b4706d1ff8062891ef35406b18525b4f8d162', 'width': 108}, {'height': 113, 'url': '...
Newbie question, how do I see which 8b models are the strongest at math or coding?
3
I know this is a stupid question, but how can I find out which 8b models are the strongest for math or coding (in python)?
2025-07-19T05:48:56
https://www.reddit.com/r/LocalLLaMA/comments/1m3oma3/newbie_question_how_do_i_see_which_8b_models_are/
MrMrsPotts
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3oma3
false
null
t3_1m3oma3
/r/LocalLLaMA/comments/1m3oma3/newbie_question_how_do_i_see_which_8b_models_are/
false
false
self
3
null
Any local models with decent tooling capabilities worth running with 3090?
10
Hi all, noob here so forgive the noobitude. Relatively new to the AI coding tool space, started with copilot in VScode, it was OK, then moved to cursor which is/was awesome for a couple months, now it's nerfed get capped even on $200 plan within a couple weeks of the month, auto mode is "ok". Tried claude code but was...
2025-07-19T05:06:41
https://www.reddit.com/r/LocalLLaMA/comments/1m3nwlf/any_local_models_with_decent_tooling_capabilities/
Acceptable_Adagio_91
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3nwlf
false
null
t3_1m3nwlf
/r/LocalLLaMA/comments/1m3nwlf/any_local_models_with_decent_tooling_capabilities/
false
false
self
10
null
Built a forensic linguistics tool to verify disputed quotes using computational stylometry - tested it on the Trump/Epstein birthday letter controversy.
55
**How the Forensic Linguistics Analysis Works:** I built this using established computational linguistics techniques for authorship attribution - the same methods used in legal cases and academic research. **1. Corpus Building** * Compiled 76 documents (14M characters) of verified Trump statements from debates, spee...
2025-07-19T04:53:01
https://i.redd.it/wz3nkrm3hrdf1.png
Gerdel
i.redd.it
1970-01-01T00:00:00
0
{}
1m3no1m
false
null
t3_1m3no1m
/r/LocalLLaMA/comments/1m3no1m/built_a_forensic_linguistics_tool_to_verify/
false
false
default
55
{'enabled': True, 'images': [{'id': 'wz3nkrm3hrdf1', 'resolutions': [{'height': 101, 'url': 'https://preview.redd.it/wz3nkrm3hrdf1.png?width=108&crop=smart&auto=webp&s=5d701ff8082323545062dccc7df2a4dd900c4c86', 'width': 108}, {'height': 203, 'url': 'https://preview.redd.it/wz3nkrm3hrdf1.png?width=216&crop=smart&auto=we...
When Llama4 Nemotron 250B MoE?
9
Just trying to summon new models by asking the question. Seeing all these new Nemo models coming out makes me wonder if we'll see a pared-down Llama 4 Maverick that's been given the Nemotron treatment. I feel like that may be much harder with MoE architecture, but maybe not.
2025-07-19T04:34:16
https://www.reddit.com/r/LocalLLaMA/comments/1m3nc51/when_llama4_nemotron_250b_moe/
RobotRobotWhatDoUSee
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3nc51
false
null
t3_1m3nc51
/r/LocalLLaMA/comments/1m3nc51/when_llama4_nemotron_250b_moe/
false
false
self
9
null
Is it worth getting 48GB of RAM alongside my 12GB VRAM GPU ? (cheapskate upgrade)
5
Noob question : Long story short I've got a system with 16GB RAM and a 6750XT GPU with 12GB VRAM, I'm happy with it for my daily usage but for AI stuff (coding/roleplay using koboldcpp) it's quite limiting. For a cheapskate upgrade, do you think it'd be worth it to buy 2 RAM sticks of 16GB for ~40$ each (bringing me...
2025-07-19T04:32:35
https://www.reddit.com/r/LocalLLaMA/comments/1m3nb1q/is_it_worth_getting_48gb_of_ram_alongside_my_12gb/
QuackMania
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3nb1q
false
null
t3_1m3nb1q
/r/LocalLLaMA/comments/1m3nb1q/is_it_worth_getting_48gb_of_ram_alongside_my_12gb/
false
false
self
5
null
[WTS] Perplexity Pro 1-Year Subscription — 50% OFF
0
Hey folks, I bought a **1-year Perplexity Pro subscription**, but turns out I’m not using it as much as I thought. So instead of wasting it, I’d rather pass it on to someone who’ll actually use it. Offering it at **50% off the original price** — still valid for the full year. I’ll help with the account transfer or se...
2025-07-19T04:31:44
https://www.reddit.com/r/LocalLLaMA/comments/1m3nah0/wts_perplexity_pro_1year_subscription_50_off/
gpxaman
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3nah0
false
null
t3_1m3nah0
/r/LocalLLaMA/comments/1m3nah0/wts_perplexity_pro_1year_subscription_50_off/
false
false
self
0
null
(Confirmed) Kimi K2’s “modified-MIT” license does NOT apply to synthetic data/distilled models
336
Kimi K2’s “modified-MIT” license does NOT apply to synthetic data or models trained on synthetic data. “Text data generated by the model is NOT considered as a derivative work.” Hopefully this will lead to more open source agentic models! Who will be the first to distill Kimi?
2025-07-19T04:28:21
https://i.redd.it/edxmilbhdrdf1.jpeg
mrfakename0
i.redd.it
1970-01-01T00:00:00
0
{}
1m3n89p
false
null
t3_1m3n89p
/r/LocalLLaMA/comments/1m3n89p/confirmed_kimi_k2s_modifiedmit_license_does_not/
false
false
default
336
{'enabled': True, 'images': [{'id': 'edxmilbhdrdf1', 'resolutions': [{'height': 91, 'url': 'https://preview.redd.it/edxmilbhdrdf1.jpeg?width=108&crop=smart&auto=webp&s=2309af2efe59009944b36011d30c7f90c97c01ac', 'width': 108}, {'height': 183, 'url': 'https://preview.redd.it/edxmilbhdrdf1.jpeg?width=216&crop=smart&auto=w...
TTS, AI, Offline - MagicMixTTS Pro - https://mercyfulking.gumroad.com/l/qbzifd - I've also made a Free Demo/Trial Version to test it out.
1
2025-07-19T03:43:06
https://v.redd.it/nd3qba875rdf1
Mercyfulking
/r/LocalLLaMA/comments/1m3meh7/tts_ai_offline_magicmixtts_pro/
1970-01-01T00:00:00
0
{}
1m3meh7
false
{'reddit_video': {'bitrate_kbps': 5000, 'dash_url': 'https://v.redd.it/nd3qba875rdf1/DASHPlaylist.mpd?a=1755618194%2CM2EyYjg1YmEyYzRhNDdlYTQyOTQ0MGNjOGZhNjFlYTFmNzdmMGVkNmJhOTI4ZDEwMWVhMzQzNGI5YzM3ODlhMw%3D%3D&v=1&f=sd', 'duration': 688, 'fallback_url': 'https://v.redd.it/nd3qba875rdf1/DASH_1080.mp4?source=fallback', '...
t3_1m3meh7
/r/LocalLLaMA/comments/1m3meh7/tts_ai_offline_magicmixtts_pro/
false
false
https://external-preview…cb8e82f08441261a
1
{'enabled': False, 'images': [{'id': 'eGxyeGVhODc1cmRmMc3WTBbacNUFiF-kx5QwPzBwz1RlFaOxC8TC9UFSmBCY', 'resolutions': [{'height': 60, 'url': 'https://external-preview.redd.it/eGxyeGVhODc1cmRmMc3WTBbacNUFiF-kx5QwPzBwz1RlFaOxC8TC9UFSmBCY.png?width=108&crop=smart&format=pjpg&auto=webp&s=c1fdac0e22c557f689143335f5face840b166...
voltapi
0
Hey! I’m an AI enthusiast who’s been deep into Python and machine learning for a while now. I recently built an AI API project called **VoltAPI** — it supports models like **Claude 3.5 Sonnet**, **GPT-4o**, and more. It’s designed to be fast, simple, and super easy to use for CLI tools or Roocode setups. If you're wo...
2025-07-19T02:28:32
https://www.reddit.com/r/LocalLLaMA/comments/1m3kzg4/voltapi/
PublicLocal1971
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3kzg4
false
null
t3_1m3kzg4
/r/LocalLLaMA/comments/1m3kzg4/voltapi/
false
false
self
0
{'enabled': False, 'images': [{'id': '2MzEN_aMHKLs7-0zP0FHek0dmzL5ftLUb87sssAtbIQ', 'resolutions': [{'height': 60, 'url': 'https://external-preview.redd.it/2MzEN_aMHKLs7-0zP0FHek0dmzL5ftLUb87sssAtbIQ.jpeg?width=108&crop=smart&auto=webp&s=dcd9920ead3dd1c56af74c8e31dc6f913e1bfe1e', 'width': 108}, {'height': 121, 'url': '...
I Wrote a Beginner-Friendly Guide on Diffusion Models How Al Actually Generates Images from Noise (No Math Overload)
1
Hey everyone! I've just published a blog post aimed at demystifying Diffusion Models, the backbone of tools like DALL-E 3 and Stable Diffusion. Blog link: https://algorythmvault.blogspot.com/2025/07/understanding-diffusion-models-how-ai.html
2025-07-19T02:12:45
https://algorythmvault.blogspot.com/2025/07/understanding-diffusion-models-how-ai.html
PlatypusStock316
algorythmvault.blogspot.com
1970-01-01T00:00:00
0
{}
1m3kog8
false
null
t3_1m3kog8
/r/LocalLLaMA/comments/1m3kog8/i_wrote_a_beginnerfriendly_guide_on_diffusion/
false
false
https://external-preview…a1a7b74d1b21a4c4
1
{'enabled': False, 'images': [{'id': 'uHgA-AqbmrxNR1EY5KvUNFfSxm40gFjjBuOEdA8-go0', 'resolutions': [{'height': 53, 'url': 'https://external-preview.redd.it/uHgA-AqbmrxNR1EY5KvUNFfSxm40gFjjBuOEdA8-go0.png?width=108&crop=smart&auto=webp&s=366aab87ce7ad9040b1ce6e6ad3008a7f9de1f46', 'width': 108}, {'height': 107, 'url': 'h...
Are P40s useful for 70B models
16
I've just discovered the wonders of LM Studio, which lets me run models without the CLI headache of OpenWebUI or ollama, and supposedly it supports multi-GPU splitting The main model I want to use is LLaMA 3.3 70B, ideally Q8, and sometimes fallen Gemma3 27B Q8, but because of scalper scumbags, GPUs are insanely over...
2025-07-19T02:06:11
https://www.reddit.com/r/LocalLLaMA/comments/1m3kjsm/are_p40s_useful_for_70b_models/
T-VIRUS999
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3kjsm
false
null
t3_1m3kjsm
/r/LocalLLaMA/comments/1m3kjsm/are_p40s_useful_for_70b_models/
false
false
self
16
null
Best reasoning model for inspecting the raw CoT?
1
I'm doing some research and would like to be able to inspect the CoT reasoning. Since both ChatGPT and Gemini now only output a summary of the CoT, I wonder what is the best reasoning model out there for me to see the detailed reasoning process? Are there still closed source models that I can do this? If not what i...
2025-07-19T02:00:03
https://www.reddit.com/r/LocalLLaMA/comments/1m3kfad/best_reasoning_model_for_inspecting_the_raw_cot/
dqdqdq123123
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3kfad
false
null
t3_1m3kfad
/r/LocalLLaMA/comments/1m3kfad/best_reasoning_model_for_inspecting_the_raw_cot/
false
false
self
1
null
If you could cut your ChatGPT token costs by 40–60%, would you be interested?
0
I’ve built a tool that compresses prompts *before* they hit the model—preserving meaning while slashing token usage. Average savings: 40–60%. No hallucinations, no broken outputs. Just leaner, cheaper completions. Curious if anyone here is: * Hitting token/context limits * Spending too much on API usage * Running...
2025-07-19T01:37:59
https://www.reddit.com/r/LocalLLaMA/comments/1m3jzql/if_you_could_cut_your_chatgpt_token_costs_by_4060/
Ok-Paleontologist393
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3jzql
false
null
t3_1m3jzql
/r/LocalLLaMA/comments/1m3jzql/if_you_could_cut_your_chatgpt_token_costs_by_4060/
false
false
self
0
null
4k local image gen
96
**I built an AI Wallpaper Generator that creates ultra-high-quality 4K wallpapers automatically with weather integration** After months of development, I've created a comprehensive AI wallpaper system that generates stunning 4K desktop backgrounds using multiple AI models. ***The system just hit v4.2.0*** with a compl...
2025-07-19T01:22:24
https://i.redd.it/dulis7vegqdf1.jpeg
kor34l
i.redd.it
1970-01-01T00:00:00
0
{}
1m3jogm
false
null
t3_1m3jogm
/r/LocalLLaMA/comments/1m3jogm/4k_local_image_gen/
false
false
https://b.thumbs.redditm…tazie9ZcTJkk.jpg
96
{'enabled': True, 'images': [{'id': 'k8Go7RzeN67gaucZZJUMbrN9cEdBYCgUiZ2XH1sDWF0', 'resolutions': [{'height': 60, 'url': 'https://preview.redd.it/dulis7vegqdf1.jpeg?width=108&crop=smart&auto=webp&s=04b508943696a21c8edb2d241f2e32e96bd4b88b', 'width': 108}, {'height': 121, 'url': 'https://preview.redd.it/dulis7vegqdf1.jp...
voltapi 3rd party api
0
# voltapi im an ai enthusiast and ive mastered python machine learning, i am a developer of an AI API if anyone wants to see my api project, its also very suitable for cline/roocode. [https://discord.gg/voltai](https://discord.gg/voltai) hope to see you there!
2025-07-19T01:21:54
https://www.reddit.com/r/LocalLLaMA/comments/1m3jo3d/voltapi_3rd_party_api/
PublicLocal1971
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3jo3d
false
null
t3_1m3jo3d
/r/LocalLLaMA/comments/1m3jo3d/voltapi_3rd_party_api/
false
false
self
0
{'enabled': False, 'images': [{'id': '2MzEN_aMHKLs7-0zP0FHek0dmzL5ftLUb87sssAtbIQ', 'resolutions': [{'height': 60, 'url': 'https://external-preview.redd.it/2MzEN_aMHKLs7-0zP0FHek0dmzL5ftLUb87sssAtbIQ.jpeg?width=108&crop=smart&auto=webp&s=dcd9920ead3dd1c56af74c8e31dc6f913e1bfe1e', 'width': 108}, {'height': 121, 'url': '...
any idea how to open source that?
381
2025-07-19T00:42:15
https://i.redd.it/x9e7q7z59qdf1.png
secopsml
i.redd.it
1970-01-01T00:00:00
0
{}
1m3iv6s
false
null
t3_1m3iv6s
/r/LocalLLaMA/comments/1m3iv6s/any_idea_how_to_open_source_that/
false
false
default
381
{'enabled': True, 'images': [{'id': 'x9e7q7z59qdf1', 'resolutions': [{'height': 124, 'url': 'https://preview.redd.it/x9e7q7z59qdf1.png?width=108&crop=smart&auto=webp&s=4595dcc12e0461458b753878ce5ae01920e1b3c7', 'width': 108}, {'height': 249, 'url': 'https://preview.redd.it/x9e7q7z59qdf1.png?width=216&crop=smart&auto=we...
Flash 2.5 vs Open weights
11
Hello! I've been looking for a new model to default to(for chatting, coding, side projects and so on) so I've also been looking at many Benchmark results and it seems like Gemini 2.5 Flash is beating all the open model(except for the new R1) and even Claude 4 Opus. While I don't have the resources to test all the model...
2025-07-19T00:38:07
https://www.reddit.com/r/LocalLLaMA/comments/1m3is87/flash_25_vs_open_weights/
Jakelolipopp
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3is87
false
null
t3_1m3is87
/r/LocalLLaMA/comments/1m3is87/flash_25_vs_open_weights/
false
false
self
11
null
Nvidia GTX-1080Ti Ollama review
2
I ran into problems when I replaced the [GTX-1070](https://www.techpowerup.com/gpu-specs/geforce-gtx-1070.c2840) with[ GTX 1080Ti](https://www.techpowerup.com/gpu-specs/geforce-gtx-1080-ti.c2877). NVTOP would show about 7GB of VRAM usage. So I had to adjust the num\_gpu value to 63. Nice improvement. These were my ste...
2025-07-19T00:13:03
https://www.reddit.com/r/LocalLLaMA/comments/1m3i9p3/nvidia_gtx1080ti_ollama_review/
tabletuser_blogspot
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3i9p3
false
null
t3_1m3i9p3
/r/LocalLLaMA/comments/1m3i9p3/nvidia_gtx1080ti_ollama_review/
false
false
self
2
{'enabled': False, 'images': [{'id': 'yvwiuTaYv0uQ6GQt62kPStPIoLv939HObYZp5JrJUJM', 'resolutions': [{'height': 48, 'url': 'https://external-preview.redd.it/yvwiuTaYv0uQ6GQt62kPStPIoLv939HObYZp5JrJUJM.jpeg?width=108&crop=smart&auto=webp&s=4fbed1483bfc1b48741ebdc49e2f0d1f9ee45d74', 'width': 108}, {'height': 97, 'url': 'h...
Here is the prompt of a conversation agent from Whatsapp (Llama 4)
0
I did the classic "read the text above" and got this response. Wanna try it locally? --- Here's the entire prompt: Today's date is Saturday, July 19, 2025. You are Meta AI. Speak naturally the way a human user might. You are an expert conversationalist made by Meta who responds in a way that feels natural to hu...
2025-07-18T23:51:52
https://www.reddit.com/r/LocalLLaMA/comments/1m3htbw/here_is_the_prompt_of_a_conversation_agent_from/
TheFrenchSavage
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3htbw
false
null
t3_1m3htbw
/r/LocalLLaMA/comments/1m3htbw/here_is_the_prompt_of_a_conversation_agent_from/
false
false
self
0
null
What is the difference betwen `n_batch` and `n_ubatch`
3
Hi, I was working with llama.cpp and I encountered `n_batch` and `n_ubatch`. Can someone explain the difference?
2025-07-18T23:03:44
https://www.reddit.com/r/LocalLLaMA/comments/1m3gr3n/what_is_the_difference_betwen_n_batch_and_n_ubatch/
Important_Earth6615
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3gr3n
false
null
t3_1m3gr3n
/r/LocalLLaMA/comments/1m3gr3n/what_is_the_difference_betwen_n_batch_and_n_ubatch/
false
false
self
3
null
How do we secure AI agents that act on their own?
0
Hey folks, I’ve been digging into how AI agents are starting to initiate API calls and perform actions across systems without a human directly in the loop, and it’s raising all sorts of questions about identity and access control. Most of the traditional auth stuff we use assumes a user is clicking a button or logging...
2025-07-18T23:01:06
https://www.reddit.com/r/LocalLLaMA/comments/1m3gow1/how_do_we_secure_ai_agents_that_act_on_their_own/
IdentityNotIdentity
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3gow1
false
null
t3_1m3gow1
/r/LocalLLaMA/comments/1m3gow1/how_do_we_secure_ai_agents_that_act_on_their_own/
false
false
self
0
{'enabled': False, 'images': [{'id': 'BHQt4VwRR4dQrppUMUaNSGuLwuzzf5D7oJwvYz7-FHI', 'resolutions': [{'height': 56, 'url': 'https://external-preview.redd.it/BHQt4VwRR4dQrppUMUaNSGuLwuzzf5D7oJwvYz7-FHI.jpeg?width=108&crop=smart&auto=webp&s=14cb777fdc66e400be660dedcc3206192e2576ab', 'width': 108}, {'height': 113, 'url': '...
Is it fine to buy a *no display* issue GPU?
0
I have a garbage gpu right now and budget is tight, can I just add a no display GPU on another PCIE slot and run AI workloads such as stable diffusion on that?
2025-07-18T21:54:01
https://www.reddit.com/r/LocalLLaMA/comments/1m3f570/is_it_fine_to_buy_a_no_display_issue_gpu/
KKLC547
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3f570
false
null
t3_1m3f570
/r/LocalLLaMA/comments/1m3f570/is_it_fine_to_buy_a_no_display_issue_gpu/
false
false
self
0
null
Is RVC-Project the best way to train a custom voice with thousands of short high quality samples WAV files?
2
I just got a 5090 and finally got the RVC project web UI *training* to work from end to end on w11. I'm currently training a 20 epoch for a voice with 6000 audio files. Waiting til it's done but just curious if I'm misunderstanding something: Would something like Kokoro TTS, sesame, alltalkttsv2 etc. have the same tra...
2025-07-18T21:52:14
https://www.reddit.com/r/LocalLLaMA/comments/1m3f3p7/is_rvcproject_the_best_way_to_train_a_custom/
LoonyLyingLemon
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3f3p7
false
null
t3_1m3f3p7
/r/LocalLLaMA/comments/1m3f3p7/is_rvcproject_the_best_way_to_train_a_custom/
false
false
self
2
null
Need help setting up Jan
1
Forgive is this is not allowed here, delete if itsnt please! Im trying to get an AI that can generate images locally, and i wanted to try Jan, but i cant get a proper Model, following a video tutorial i found it says to simply add an image gen model Url from huggingface, but when i do it comes empty on Jan Hub screen...
2025-07-18T21:47:08
https://www.reddit.com/r/LocalLLaMA/comments/1m3ezgz/need_help_setting_up_jan/
Alternative-Ad5482
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3ezgz
false
null
t3_1m3ezgz
/r/LocalLLaMA/comments/1m3ezgz/need_help_setting_up_jan/
false
false
self
1
null
Tool for creating artifacts with different open-source models
0
On this [tool](http://localhost:3000/play), you can selectively create artifacts (HTML/CSS/JS website UI and images) using different open-source models (Kimi-K2, Mistral variants, DeepSeek, Llama, and Qwen) by clicking on the gear icon and selecting particular models. When you write a prompt and then click enter, 2-4 d...
2025-07-18T21:34:08
https://v.redd.it/6lcs6x7kbpdf1
idwiw_wiw
/r/LocalLLaMA/comments/1m3eo8z/tool_for_creating_artifacts_with_different/
1970-01-01T00:00:00
0
{}
1m3eo8z
false
{'reddit_video': {'bitrate_kbps': 5000, 'dash_url': 'https://v.redd.it/6lcs6x7kbpdf1/DASHPlaylist.mpd?a=1755596077%2COTA5YmM4YTMyMTM2NmMxMDlkMGFmYTI3MTUyNzlkMTQ5N2JlYTllMDdkMzI4NzllODYyNjUwMWI2YmM5YjUzNg%3D%3D&v=1&f=sd', 'duration': 166, 'fallback_url': 'https://v.redd.it/6lcs6x7kbpdf1/DASH_1080.mp4?source=fallback', '...
t3_1m3eo8z
/r/LocalLLaMA/comments/1m3eo8z/tool_for_creating_artifacts_with_different/
false
false
https://external-preview…ac99f6f533b7941d
0
{'enabled': False, 'images': [{'id': 'aHVnZHJ3N2ticGRmMZD12iNHB93MMi8LtHlPbPeDhW7skbYVBFP4SEE4kEWI', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/aHVnZHJ3N2ticGRmMZD12iNHB93MMi8LtHlPbPeDhW7skbYVBFP4SEE4kEWI.png?width=108&crop=smart&format=pjpg&auto=webp&s=0fca3ac14ba74c780773b91a080e561b02f08...
Best subreddits for AI engineers?
1
[removed]
2025-07-18T20:35:10
https://www.reddit.com/r/LocalLLaMA/comments/1m3d8b5/best_subreddits_for_ai_engineers/
AskAnAIEngineer
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3d8b5
false
null
t3_1m3d8b5
/r/LocalLLaMA/comments/1m3d8b5/best_subreddits_for_ai_engineers/
false
false
self
1
null
Open source OCR options for handwritten text, dates
7
Hi, I am working on a project where I want to extract handwritten text, dates, digits. What's important - Reliability and Accuracy. I don't care about how fast it is. I used Paddle and didn't get great results. I haven't worked too much with OCR, so anything helps!
2025-07-18T20:17:51
https://www.reddit.com/r/LocalLLaMA/comments/1m3ct76/open_source_ocr_options_for_handwritten_text_dates/
ollyollyupnfree
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3ct76
false
null
t3_1m3ct76
/r/LocalLLaMA/comments/1m3ct76/open_source_ocr_options_for_handwritten_text_dates/
false
false
self
7
null
Looking for help with terrible vLLM performance
5
I recently inherited a GPU workstation at work from a project that got shut down. It's an older Vector Lambda with 4x RTX a5000, so I decided to set it up running either one full instance of the new devstral model or some quantized versions. The problem I'm running into is I'm just getting \*terrible\* performance out ...
2025-07-18T20:03:02
https://www.reddit.com/r/LocalLLaMA/comments/1m3cfy9/looking_for_help_with_terrible_vllm_performance/
Render_Arcana
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1m3cfy9
false
null
t3_1m3cfy9
/r/LocalLLaMA/comments/1m3cfy9/looking_for_help_with_terrible_vllm_performance/
false
false
self
5
{'enabled': False, 'images': [{'id': 'MV8WQnKGiSSypEI5QKJe4g08BAeccM6KobeueLMOJdY', 'resolutions': [{'height': 60, 'url': 'https://external-preview.redd.it/MV8WQnKGiSSypEI5QKJe4g08BAeccM6KobeueLMOJdY.png?width=108&crop=smart&auto=webp&s=a08158a2ec290c8157b492f314bfb148408be1fc', 'width': 108}, {'height': 121, 'url': 'h...