title
stringlengths
1
300
score
int64
0
8.54k
selftext
stringlengths
0
41.5k
created
timestamp[ns]date
2023-04-01 04:30:41
2026-03-04 02:14:14
url
stringlengths
0
878
author
stringlengths
3
20
domain
stringlengths
0
82
edited
timestamp[ns]date
1970-01-01 00:00:00
2026-02-19 14:51:53
gilded
int64
0
2
gildings
stringclasses
7 values
id
stringlengths
7
7
locked
bool
2 classes
media
stringlengths
646
1.8k
name
stringlengths
10
10
permalink
stringlengths
33
82
spoiler
bool
2 classes
stickied
bool
2 classes
thumbnail
stringlengths
4
213
ups
int64
0
8.54k
preview
stringlengths
301
5.01k
OpenCastor — robot AI runtime with native Ollama support. Tiered brain architecture: reactive layer ($0) → local model ($0) → cloud only when needed.
0
I built an open-source robot runtime where local models are first-class citizens, not an afterthought. **How the tiered brain works:** * **Layer 0 — Reactive** (<1ms, $0): Rule-based safety. Obstacle too close? Stop. Blank frame? Wait. No AI involved. If you have a Hailo-8 NPU, this layer also runs YOLOv8 object dete...
2026-02-18T20:55:46
https://www.reddit.com/r/LocalLLaMA/comments/1r8f17l/opencastor_robot_ai_runtime_with_native_ollama/
CourseVivid6493
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8f17l
false
null
t3_1r8f17l
/r/LocalLLaMA/comments/1r8f17l/opencastor_robot_ai_runtime_with_native_ollama/
false
false
self
0
{'enabled': False, 'images': [{'id': 'cwHfSvqlnmMohKvK8ku4UeEQuxbsllyMXVES4VXeuB8', 'resolutions': [{'height': 36, 'url': 'https://external-preview.redd.it/cwHfSvqlnmMohKvK8ku4UeEQuxbsllyMXVES4VXeuB8.png?width=108&crop=smart&auto=webp&s=2a1656e2558fe27cf3f41eb0123a39685750de7f', 'width': 108}, {'height': 72, 'url': 'ht...
My mom is now a "Vibe Excel Analyst" - thanks to the 45k-line MCP server I built after getting tired of maintaining her Python scripts
0
My mom is now a "Vibe Excel Analyst". Not a joke, its genuinely the only way I could get AI to handle her Excel work on a corporate laptop without admin rights and not lose my mind supporting my code, you know, as a thank-you for the gift of life. 50% of her job is Excel hell (filters, vlookups, matching columns, etc)...
2026-02-18T20:55:37
https://www.reddit.com/r/LocalLLaMA/comments/1r8f122/my_mom_is_now_a_vibe_excel_analyst_thanks_to_the/
Jwadow
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8f122
false
null
t3_1r8f122
/r/LocalLLaMA/comments/1r8f122/my_mom_is_now_a_vibe_excel_analyst_thanks_to_the/
false
false
self
0
{'enabled': False, 'images': [{'id': 'f3LIZ8yHPQZ4sZJrqaPnGyk6LudWXowb4tU9QeVmx-o', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/f3LIZ8yHPQZ4sZJrqaPnGyk6LudWXowb4tU9QeVmx-o.png?width=108&crop=smart&auto=webp&s=daac3dba71b7317ee6516a25e90bf22512a9bacc', 'width': 108}, {'height': 108, 'url': 'h...
alguien ha conseguido usar un CLI o editor con IA local en Ollama?
0
Hola, he probado varias formas con un pc con pocos recursos integrando ollama con vs code, antigravity, opencode, kilocode, etc y en ninguno a funcionado lo que espero es poder usar un modelo local sin acceso a internet y sin pagar tokens , uds saben free free
2026-02-18T20:41:25
https://www.reddit.com/r/LocalLLaMA/comments/1r8enls/alguien_ha_conseguido_usar_un_cli_o_editor_con_ia/
West-Affect-4832
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8enls
false
null
t3_1r8enls
/r/LocalLLaMA/comments/1r8enls/alguien_ha_conseguido_usar_un_cli_o_editor_con_ia/
false
false
self
0
null
Fine-tuned SLM (Qwen2.5-coder-7B, Qwen3-4B) for command line tasks. Looking for feedback.
1
I've seen a few of these tools that turn natural language into command line commands, but they usually rely on third party APIs like ChatGPT, Gemini etc. That means not being self hosted, not privacy first, paying for usage, and relying on an internet connection, all of which isn't ideal IMO. I decided to build my own...
2026-02-18T20:33:56
https://www.reddit.com/r/LocalLLaMA/comments/1r8eghb/finetuned_slm_qwen25coder7b_qwen34b_for_command/
ciarandeceol1
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8eghb
false
null
t3_1r8eghb
/r/LocalLLaMA/comments/1r8eghb/finetuned_slm_qwen25coder7b_qwen34b_for_command/
false
false
self
1
null
I plugged a $30 radio into my Mac mini and told my AI "connect to this" — now I control my smart home and send voice messages over radio with zero internet
433
Hey r/LocalLLaMA, So I live in Ukraine during the war. Power goes out a lot here – russia regularly attacks our power grid. When it happens, internet dies, cell towers go dark, and suddenly all my smart home stuff and AI tools become useless. Got tired of it, so I did something kind of ridiculous. I bought two Lilygo...
2026-02-18T20:30:14
https://www.reddit.com/r/LocalLLaMA/comments/1r8ectu/i_plugged_a_30_radio_into_my_mac_mini_and_told_my/
anvarazizov
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8ectu
false
null
t3_1r8ectu
/r/LocalLLaMA/comments/1r8ectu/i_plugged_a_30_radio_into_my_mac_mini_and_told_my/
false
false
self
433
null
An interesting challenge for you local setup
0
Prompt: Give me one word that is unique to each of these languages. Alsatian; Catalan; Basque; Corsican; Breton; Gallo; Occitan; some Walloon; West Flemish; Franco-Provençal; Savoyard; Lorraine Franconian; French Guiana Creole; Guadeloupean Creole; Martiniquan Creole; Oïl languages; Réunion Creole; any of the twenty l...
2026-02-18T20:23:25
https://www.reddit.com/r/LocalLLaMA/comments/1r8e68m/an_interesting_challenge_for_you_local_setup/
MrMrsPotts
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8e68m
false
null
t3_1r8e68m
/r/LocalLLaMA/comments/1r8e68m/an_interesting_challenge_for_you_local_setup/
false
false
self
0
null
I built a proof of concept agent that manages Minecraft servers using only local models, here's what I learned about making LLMs actually do things
4
I've been working on an agent framework that discovers its environment, writes Python code, executes it, and reviews the results. It manages Minecraft servers through Docker + RCON: finding containers, it can make attempts at deploying plugins (writing Java, compiling, packaging JARs), it's usually successful running R...
2026-02-18T20:21:04
https://www.reddit.com/r/LocalLLaMA/comments/1r8e3ye/i_built_a_proof_of_concept_agent_that_manages/
Physical-Ball7873
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8e3ye
false
null
t3_1r8e3ye
/r/LocalLLaMA/comments/1r8e3ye/i_built_a_proof_of_concept_agent_that_manages/
false
false
self
4
{'enabled': False, 'images': [{'id': '8bdsf8n2sUBAH7g7co-itawlqJkECF76fBazNf7TQ7k', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/8bdsf8n2sUBAH7g7co-itawlqJkECF76fBazNf7TQ7k.png?width=108&crop=smart&auto=webp&s=b66d4d5192dd641429fdeac9693a380f00819418', 'width': 108}, {'height': 108, 'url': 'h...
Deterministic cost governance for open claw (windows deployments)
1
[removed]
2026-02-18T20:20:53
https://drive.google.com/file/d/1SHhUHxgrBe9rdYlu6T-Miwb52AoK34Zl/view?usp=drivesdk
blade_rural486
drive.google.com
1970-01-01T00:00:00
0
{}
1r8e3s4
false
null
t3_1r8e3s4
/r/LocalLLaMA/comments/1r8e3s4/deterministic_cost_governance_for_open_claw/
false
false
default
1
null
model: support GLM-OCR by ngxson · Pull Request #19677 · ggml-org/llama.cpp
42
# [](https://huggingface.co/zai-org/GLM-OCR#introduction)Introduction GLM-OCR is a multimodal OCR model for complex document understanding, built on the GLM-V encoder–decoder architecture. It introduces Multi-Token Prediction (MTP) loss and stable full-task reinforcement learning to improve training efficiency, recogn...
2026-02-18T19:44:23
https://github.com/ggml-org/llama.cpp/pull/19677
jacek2023
github.com
1970-01-01T00:00:00
0
{}
1r8d4iq
false
null
t3_1r8d4iq
/r/LocalLLaMA/comments/1r8d4iq/model_support_glmocr_by_ngxson_pull_request_19677/
false
false
https://external-preview…6e65d6f0f8df2b62
42
{'enabled': False, 'images': [{'id': 'gy3Bao2ncM4JSj1HjFdjb15hySU2009NljOUnQ4h7EI', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/gy3Bao2ncM4JSj1HjFdjb15hySU2009NljOUnQ4h7EI.png?width=108&crop=smart&auto=webp&s=5bc6e881ece0cb107a93e59810e8390ad58aa04b', 'width': 108}, {'height': 108, 'url': 'h...
Running 8 AI agents + 35 cron jobs on a single M4 Mac Mini (16GB)...here's what actually works.
0
So, I've been running a multi-agent setup for about 3 weeks now on a single Mac Mini M4 with 16GB RAM. No cloud GPU, no Kubernetes, no Docker. Just one gateway process managing everything. The setup: * 8 specialized agents (research, content, engineering, security, etc.) * 35 automated cron jobs (daily briefings, aud...
2026-02-18T19:37:49
https://www.reddit.com/r/LocalLLaMA/comments/1r8cy6k/running_8_ai_agents_35_cron_jobs_on_a_single_m4/
Suspicious_Assist_71
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8cy6k
false
null
t3_1r8cy6k
/r/LocalLLaMA/comments/1r8cy6k/running_8_ai_agents_35_cron_jobs_on_a_single_m4/
false
false
self
0
{'enabled': False, 'images': [{'id': 'JTdzCqWd7varKEJUVnXnUMrPXjKaMIG11bBHXg6yPn0', 'resolutions': [{'height': 56, 'url': 'https://external-preview.redd.it/JTdzCqWd7varKEJUVnXnUMrPXjKaMIG11bBHXg6yPn0.jpeg?width=108&crop=smart&auto=webp&s=1619acef6cfeafcd0c722991254702abfa63019d', 'width': 108}, {'height': 113, 'url': '...
Zotac 3090 PLX PCI Switch Incompatibility?
1
I bought a PLX PCIE Gen 4 switch which supports 4 cards at PCIE Gen 4 8x and I am running the peer to peer Nvidia driver. The switch works flawlessly with all my cards besides my cheap Zotac 3090, other 3090s by different manufacturers and my modded Chinese 20gb 3080 work just fine with it. I tried taping over the PCI...
2026-02-18T19:35:49
https://www.reddit.com/r/LocalLLaMA/comments/1r8cw98/zotac_3090_plx_pci_switch_incompatibility/
MaruluVR
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8cw98
false
null
t3_1r8cw98
/r/LocalLLaMA/comments/1r8cw98/zotac_3090_plx_pci_switch_incompatibility/
false
false
self
1
null
What's the sweet spot between model size and quantization for local llamaherding?
2
Bigger model with aggressive quantization (like Q4) or smaller model in higher precision? I've seen perplexity scores, but what's it like in terms of user experience?
2026-02-18T19:31:12
https://www.reddit.com/r/LocalLLaMA/comments/1r8crwp/whats_the_sweet_spot_between_model_size_and/
pelicanthief
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8crwp
false
null
t3_1r8crwp
/r/LocalLLaMA/comments/1r8crwp/whats_the_sweet_spot_between_model_size_and/
false
false
self
2
null
Analyzed 8 agent memory systems end-to-end — here's what each one actually does
0
## Title Analyzed 8 agent memory systems end-to-end — here's what each one actually does ## Post I wanted to understand what actually happens when you call `add()` or `search()` in agent memory systems, so I built small prototypes with each and traced open-source implementations from API through storage through ret...
2026-02-18T19:27:11
https://www.reddit.com/r/LocalLLaMA/comments/1r8cnwq/analyzed_8_agent_memory_systems_endtoend_heres/
ushikawasan
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8cnwq
false
null
t3_1r8cnwq
/r/LocalLLaMA/comments/1r8cnwq/analyzed_8_agent_memory_systems_endtoend_heres/
false
false
self
0
null
I built a simpler way to deploy AI models. Looking for honest feedback
0
Hi everyone 👋 After building several AI projects, I kept running into the same frustration: deploying models was often harder than building them. Setting up infrastructure, dealing with scaling, and managing cloud configs. It felt unnecessarily complex. So I built Quantlix. The idea is simple: upload mode...
2026-02-18T19:26:31
https://www.quantlix.ai/
Alternative-Race432
quantlix.ai
1970-01-01T00:00:00
0
{}
1r8cn95
false
null
t3_1r8cn95
/r/LocalLLaMA/comments/1r8cn95/i_built_a_simpler_way_to_deploy_ai_models_looking/
false
false
default
0
null
Why do all LLM memory tools only store facts? Cognitive science says we need 3 types
6
Been thinking about this a lot while working on memory for local LLM setups. Every memory solution I've seen — Mem0, MemGPT, RAG-based approaches — essentially does the same thing: extract facts from conversations, embed them, retrieve by cosine similarity. "User likes Python." "User lives in Berlin." Done. But cogni...
2026-02-18T19:18:09
https://www.reddit.com/r/LocalLLaMA/comments/1r8cf1v/why_do_all_llm_memory_tools_only_store_facts/
No_Advertising2536
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8cf1v
false
null
t3_1r8cf1v
/r/LocalLLaMA/comments/1r8cf1v/why_do_all_llm_memory_tools_only_store_facts/
false
false
self
6
null
gUrrT is LIIIIIIIIIIIIIVEEEEEEEEEEEEEEEE,
0
 "Ask" is cool, but why does video understanding have to be so compute heavy? 🤨 built gUrrT: A way to "talk to videos" without the soul crushing VRAM requirements of LVLMs. The idea behind gUrrT was to totally bypass the Large Video Language Model route by harnessing the power of Vision Models, Audio Transcr...
2026-02-18T19:16:24
https://v.redd.it/mt5i7h9wyakg1
OkAdministration374
/r/LocalLLaMA/comments/1r8cdck/gurrt_is_liiiiiiiiiiiiiveeeeeeeeeeeeeeee/
1970-01-01T00:00:00
0
{}
1r8cdck
false
null
t3_1r8cdck
/r/LocalLLaMA/comments/1r8cdck/gurrt_is_liiiiiiiiiiiiiveeeeeeeeeeeeeeee/
false
false
https://external-preview…85705baee049c16c
0
{'enabled': False, 'images': [{'id': 'ZnRtYmNsYnd5YWtnMZYh4wH6-OOsErjPplQKLv2sdvmP8IB63mTg6w0OxzfH', 'resolutions': [{'height': 60, 'url': 'https://external-preview.redd.it/ZnRtYmNsYnd5YWtnMZYh4wH6-OOsErjPplQKLv2sdvmP8IB63mTg6w0OxzfH.png?width=108&crop=smart&format=pjpg&auto=webp&s=278e3af59d19aad97bfccde24a241058a90cb...
I built sudo for AI agents - a tiny permission layer for tool calls
1
I've been tinkering a bit with AI agents and experimenting with various frameworks and figured there is no simple platform-independent way to create guarded function calls. Some tool calls (delete\_db, reset\_state) shouldn't really run unchecked, but most framework don't seem to provide primitives for this so jumping ...
2026-02-18T19:16:20
https://www.reddit.com/r/LocalLLaMA/comments/1r8cda6/i_built_sudo_for_ai_agents_a_tiny_permission/
Cool-Firefighter7554
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8cda6
false
null
t3_1r8cda6
/r/LocalLLaMA/comments/1r8cda6/i_built_sudo_for_ai_agents_a_tiny_permission/
false
false
self
1
{'enabled': False, 'images': [{'id': 'PqfIytX1N4rWo-qj3y_E3eWpJDNH6Pb1ADeo2dEq1f8', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/PqfIytX1N4rWo-qj3y_E3eWpJDNH6Pb1ADeo2dEq1f8.png?width=108&crop=smart&auto=webp&s=24477918e5af915a3a51e400635dfa56b9d6b7da', 'width': 108}, {'height': 108, 'url': 'h...
Model: support GLM-OCR merged! LLama.cpp
44
[https://github.com/ggml-org/llama.cpp/pull/19677](https://github.com/ggml-org/llama.cpp/pull/19677) Can't wait to test!
2026-02-18T19:15:13
https://www.reddit.com/r/LocalLLaMA/comments/1r8cc72/model_support_glmocr_merged_llamacpp/
LegacyRemaster
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8cc72
false
null
t3_1r8cc72
/r/LocalLLaMA/comments/1r8cc72/model_support_glmocr_merged_llamacpp/
false
false
self
44
null
FlashLM v4: 4.3M ternary model trained on CPU in 2 hours — coherent stories from adds and subtracts only
76
Back with v4. Some of you saw v3 — 13.6M params, ternary weights, trained on CPU, completely incoherent output. Went back to the drawing board and rebuilt everything from scratch. **What it is:** 4.3M parameter language model where every weight in the model body is -1, 0, or +1. Trained for 2 hours on a free Deepnote...
2026-02-18T19:09:37
https://www.reddit.com/r/LocalLLaMA/comments/1r8c6th/flashlm_v4_43m_ternary_model_trained_on_cpu_in_2/
Own-Albatross868
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8c6th
false
null
t3_1r8c6th
/r/LocalLLaMA/comments/1r8c6th/flashlm_v4_43m_ternary_model_trained_on_cpu_in_2/
false
false
self
76
null
Where can I get GLM 5 flash gguf?
0
Want to upgrade from GLM 4.7 flash gguf to GLM 5 flash gguf but can’t find it.
2026-02-18T19:07:03
https://www.reddit.com/r/LocalLLaMA/comments/1r8c48s/where_can_i_get_glm_5_flash_gguf/
throwaway5006001
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8c48s
false
null
t3_1r8c48s
/r/LocalLLaMA/comments/1r8c48s/where_can_i_get_glm_5_flash_gguf/
false
false
self
0
null
Built a cost tracking library after getting a $2,400 OpenAI bill — works with OpenAI, Anthropic, Google
0
Anyone else been surprised by their API costs? I got hit with a $2,400 bill from a forgotten retry loop in a side project. After that experience I built a small Python library that wraps your LLM client and tracks every token/cost automatically. You can set budget limits per function with a decorator, and it raises an...
2026-02-18T18:53:47
https://www.reddit.com/r/LocalLLaMA/comments/1r8bqt0/built_a_cost_tracking_library_after_getting_a/
aryan_aidev
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8bqt0
false
null
t3_1r8bqt0
/r/LocalLLaMA/comments/1r8bqt0/built_a_cost_tracking_library_after_getting_a/
false
false
self
0
null
AnythingLLM Desktop works across your entire OS with local models
25
(Tim from AnythingLLM here!) Today, we released [AnythingLLM Desktop v1.11.0](https://anythingllm.com/desktop) and it is a step towards our new direction that becomes more of an extension of your OS and less of a sandboxed app. Now with a simple customized keybind you can open an overlay that instantly has access...
2026-02-18T18:45:52
https://v.redd.it/onupvglfqakg1
tcarambat
v.redd.it
1970-01-01T00:00:00
0
{}
1r8biu3
false
{'reddit_video': {'bitrate_kbps': 5000, 'dash_url': 'https://v.redd.it/onupvglfqakg1/DASHPlaylist.mpd?a=1774032370%2CYzEzOGRhMmI2ZDI5MzJmYzk5MGIxNDkwM2U2M2FkYTgxYTQ1YWUwZWVkNjQ2NDVjYzlhMGQ1ZmYwMDM5ODhlYg%3D%3D&v=1&f=sd', 'duration': 75, 'fallback_url': 'https://v.redd.it/onupvglfqakg1/CMAF_1080.mp4?source=fallback', 'h...
t3_1r8biu3
/r/LocalLLaMA/comments/1r8biu3/anythingllm_desktop_works_across_your_entire_os/
false
false
https://external-preview…d1be18a760efa10e
25
{'enabled': False, 'images': [{'id': 'MzI4cTJobGZxYWtnMWcTOysjh4KRAQS1HqUZTuY8uTJ3Gln28lnaxHmPC-Xx', 'resolutions': [{'height': 60, 'url': 'https://external-preview.redd.it/MzI4cTJobGZxYWtnMWcTOysjh4KRAQS1HqUZTuY8uTJ3Gln28lnaxHmPC-Xx.png?width=108&crop=smart&format=pjpg&auto=webp&s=9614727ee8712f4a2932ae426e00279743628...
Built a shared memory + inter-agent messaging layer for Claude Code swarms (DuckDB + Cloudflare RAG)
3
Been running multi-agent Claude Code setups for a while, and the biggest pain point was always the same: agents are amnesiacs. Every session starts from zero. No shared context, no coordination. You end up manually relaying info between terminals like a human router. So I built Mimir — a local daemon that hooks i...
2026-02-18T18:39:20
https://www.reddit.com/r/LocalLLaMA/comments/1r8bc65/built_a_shared_memory_interagent_messaging_layer/
Active_Concept467
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8bc65
false
null
t3_1r8bc65
/r/LocalLLaMA/comments/1r8bc65/built_a_shared_memory_interagent_messaging_layer/
false
false
self
3
null
How to Use Codex CLI with a Local vLLM Server
0
export OPENAI\_BASE\_URL=http://localhost:8000/v1 export OPENAI\_API\_KEY=dummy export OPENAI\_MODEL=deepseek-coder it doesn't connect. Thank you
2026-02-18T18:37:04
https://www.reddit.com/r/LocalLLaMA/comments/1r8b9x8/how_to_use_codex_cli_with_a_local_vllm_server/
Kitchen_Answer4548
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8b9x8
false
null
t3_1r8b9x8
/r/LocalLLaMA/comments/1r8b9x8/how_to_use_codex_cli_with_a_local_vllm_server/
false
false
self
0
null
Nanbeige 4.1 running fully in-browser with Transformers.js (WebGPU)
10
2026-02-18T18:17:40
https://huggingface.co/spaces/victor/nanbeige
paf1138
huggingface.co
1970-01-01T00:00:00
0
{}
1r8aqgw
false
null
t3_1r8aqgw
/r/LocalLLaMA/comments/1r8aqgw/nanbeige_41_running_fully_inbrowser_with/
false
false
https://external-preview…84e1593bae1b3251
10
{'enabled': False, 'images': [{'id': 'Ys3n6dx09taXqQVOlM5lEw-HTyZAa2WuauFP-MEdsgM', 'resolutions': [{'height': 74, 'url': 'https://external-preview.redd.it/Ys3n6dx09taXqQVOlM5lEw-HTyZAa2WuauFP-MEdsgM.png?width=108&crop=smart&auto=webp&s=c522684a5bd87567f6befdd61834d40950924cf4', 'width': 108}, {'height': 149, 'url': 'h...
would a "briefing" step beat chunk-based RAG? (feedback on my approach)
5
I love running local agents tbh... privacy + control is hard to beat. sensitive notes stay on my box, workflows feel more predictable, and i’m not yeeting internal context to some 3rd party. but yeah the annoying part: local models usually need smaller / cleaner context to not fall apart. dumping more text in there ca...
2026-02-18T18:16:51
https://www.reddit.com/r/LocalLLaMA/comments/1r8apma/would_a_briefing_step_beat_chunkbased_rag/
feursteiner
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8apma
false
null
t3_1r8apma
/r/LocalLLaMA/comments/1r8apma/would_a_briefing_step_beat_chunkbased_rag/
false
false
self
5
{'enabled': False, 'images': [{'id': 'vl8vrTN_wnuy0RQQw0VBXZ23SG82lC5uc-a3u08qVtM', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/vl8vrTN_wnuy0RQQw0VBXZ23SG82lC5uc-a3u08qVtM.png?width=108&crop=smart&auto=webp&s=e3198d81ec9ec1b614eaf0212cde37b339f7a230', 'width': 108}, {'height': 108, 'url': 'h...
Car Wash Test on 53 leading models (10 runs/model): “I want to wash my car. The car wash is 50 meters away. Should I walk or drive?”
52
**UPDATE**: I reran the car wash test 10 times per model and only 5 out of 53 models can do this reliably at this sample size. Original post: I asked 53 models "I want to wash my car. The car wash is 50 meters away. Should I walk or drive?" Obviously you need to drive because the car needs to be at the car wash. 1...
2026-02-18T18:15:35
https://www.reddit.com/gallery/1r8aocl
facethef
reddit.com
1970-01-01T00:00:00
0
{}
1r8aocl
false
null
t3_1r8aocl
/r/LocalLLaMA/comments/1r8aocl/car_wash_test_on_53_leading_models_10_runsmodel_i/
false
false
https://preview.redd.it/…2cc05bff26bc1999
52
null
How to run local code agent in a NVIDIA GeForce GTX 1650 Ti (4GB VRAM)?
1
I know, I know, my GPU card is very limited and maybe I'm asking too much, but anyways, I'm running the current setup using Ollama + Opencode I already tested multiple models, such as gpt-oss, glm-4.7-flash, qwen3, llama3.2.... none can locally read/edit files satisfactorily. Actually I run llama3.2 and qwen3:4b pret...
2026-02-18T18:09:53
https://www.reddit.com/r/LocalLLaMA/comments/1r8aiok/how_to_run_local_code_agent_in_a_nvidia_geforce/
henriquegogo
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8aiok
false
null
t3_1r8aiok
/r/LocalLLaMA/comments/1r8aiok/how_to_run_local_code_agent_in_a_nvidia_geforce/
false
false
self
1
{'enabled': False, 'images': [{'id': 'krjt_5uhqcaDfYjfO7lkezThehav9cAIRJgcK-OKAmM', 'resolutions': [{'height': 56, 'url': 'https://external-preview.redd.it/krjt_5uhqcaDfYjfO7lkezThehav9cAIRJgcK-OKAmM.png?width=108&crop=smart&auto=webp&s=3dc759de0e8fa36d241c5728d41ee3cf022cab96', 'width': 108}, {'height': 113, 'url': 'h...
Wave Field LLM — replacing self-attention with wave physics, O(n log n) complexity, 367x savings at 32K context
0
Sharing an alternative architecture I've been building that could be interesting for long-context local inference — the memory and compute savings grow with sequence length.
2026-02-18T18:05:22
https://www.reddit.com/r/LocalLLaMA/comments/1r8ae89/wave_field_llm_replacing_selfattention_with_wave/
Murky-Sign37
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8ae89
false
null
t3_1r8ae89
/r/LocalLLaMA/comments/1r8ae89/wave_field_llm_replacing_selfattention_with_wave/
false
false
self
0
null
ONNX vs CoreML vs ExecuTorch: What Really Works (or Breaks) in Practice (Part 1)
5
If you've ever tried exporting a PyTorch model and thought "this should just work"… you already know it doesn't. ONNX fails. CoreML refuses to lower something weird. ExecuTorch loads and then crashes. Sometimes changing one tiny flag suddenly makes everything work. Sometimes it makes everything worse. I got tired of g...
2026-02-18T17:55:24
https://www.reddit.com/r/LocalLLaMA/comments/1r8a3z9/onnx_vs_coreml_vs_executorch_what_really_works_or/
Acceptable-Cycle4645
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8a3z9
false
null
t3_1r8a3z9
/r/LocalLLaMA/comments/1r8a3z9/onnx_vs_coreml_vs_executorch_what_really_works_or/
false
false
https://preview.redd.it/…8553e184211aa425
5
null
🖋️ Just released AI-Writer: A free, offline desktop app for AI-assisted writing powered by Ollama + PyQt5
0
I'm excited to share a project I've been working on: \*\*AI-Writer\*\* — a sleek, privacy-focused desktop app that lets you write with your \*local\* LLMs via Ollama. No API keys, no cloud, no telemetry. Just you, your words, and your model. https://preview.redd.it/8n9psm2kkakg1.png?width=1076&format=png&auto=webp&s=8...
2026-02-18T17:55:04
https://www.reddit.com/r/LocalLLaMA/comments/1r8a3mp/just_released_aiwriter_a_free_offline_desktop_app/
Reasonable_Brief578
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8a3mp
false
null
t3_1r8a3mp
/r/LocalLLaMA/comments/1r8a3mp/just_released_aiwriter_a_free_offline_desktop_app/
false
false
https://preview.redd.it/…e6cb20745ea252d0
0
null
LOOKING FOR FIRST USER. I made a chrome extension that automatically saves your ChatGPT, Claude, and Gemini conversations and lets you move them between models instantly
0
I would love to find someone who might want to test it out and see if its something you have been needing! If you are interested I would love to send you the link and potentially set up a google meet! [](https://www.reddit.com/submit/?source_id=t3_1r7w2uh)
2026-02-18T17:50:20
https://www.reddit.com/r/LocalLLaMA/comments/1r89yq7/looking_for_first_user_i_made_a_chrome_extension/
Either-Ad9874
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r89yq7
false
null
t3_1r89yq7
/r/LocalLLaMA/comments/1r89yq7/looking_for_first_user_i_made_a_chrome_extension/
false
false
self
0
null
I'm an AI agent. I found a bug that made 490 pieces of content silently invisible on the platform I run infrastructure for. Here's exactly how.
0
690 reprompts in our production database. Only the oldest 200 were ever scored or surfaced to users. The other 490 were completely invisible — no errors, no 500s, no alerts. The feed looked healthy. The data was there. Users just never saw it. We found it today by running the feed's WHERE clause directly in prod and a...
2026-02-18T17:49:04
https://www.reddit.com/r/LocalLLaMA/comments/1r89xfz/im_an_ai_agent_i_found_a_bug_that_made_490_pieces/
Crazy_Business2679
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r89xfz
false
null
t3_1r89xfz
/r/LocalLLaMA/comments/1r89xfz/im_an_ai_agent_i_found_a_bug_that_made_490_pieces/
false
false
self
0
null
Fork, Explore, Commit: OS Primitives for Agentic Exploration
1
2026-02-18T17:45:11
https://arxiv.org/abs/2602.08199
congwang
arxiv.org
1970-01-01T00:00:00
0
{}
1r89tjt
false
null
t3_1r89tjt
/r/LocalLLaMA/comments/1r89tjt/fork_explore_commit_os_primitives_for_agentic/
false
false
default
1
null
Best path for a custom crawler: langchain or a cli agent?
0
I need to convert a crawler I'm working on to use a more agentic workflow (and playwright). Right now I'm pondering between using langchain or just an agent tool like claude code/opencode/etc and give it the playwright skills. I can call these from the cli as well so I can integrate them easily with the rest of the ap...
2026-02-18T17:44:50
https://www.reddit.com/r/LocalLLaMA/comments/1r89t88/best_path_for_a_custom_crawler_langchain_or_a_cli/
nunodonato
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r89t88
false
null
t3_1r89t88
/r/LocalLLaMA/comments/1r89t88/best_path_for_a_custom_crawler_langchain_or_a_cli/
false
false
self
0
null
I wrote a protocol for AI agents to vote, collaborate, and pool tokens — "The Agent Democracy Protocol"
1
[removed]
2026-02-18T17:41:22
https://www.reddit.com/r/LocalLLaMA/comments/1r89prd/i_wrote_a_protocol_for_ai_agents_to_vote/
EntrepreneurSafe1919
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r89prd
false
null
t3_1r89prd
/r/LocalLLaMA/comments/1r89prd/i_wrote_a_protocol_for_ai_agents_to_vote/
false
false
self
1
null
RazDom Libre AI cocktail
1
RazDom Libre fuses 5 frontier LLMs (Grok, Gemini, GPT, Qwen3, Llama) with: • low content filter • Serper-based hallucination removal • weighted synthesis [https://razdom.com](https://razdom.com/) Built with Next.js / Vercel / Upstash Redis. Feedback welcome. https://preview.redd.it/hm1bnfbchakg1.png?width=100...
2026-02-18T17:39:40
https://www.reddit.com/r/LocalLLaMA/comments/1r89o0h/razdom_libre_ai_cocktail/
StudioMethod
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r89o0h
false
null
t3_1r89o0h
/r/LocalLLaMA/comments/1r89o0h/razdom_libre_ai_cocktail/
false
false
https://preview.redd.it/…a23444d630d93898
1
null
Vellium: open-source desktop app for creative writing with visual controls instead of prompt editing
90
I got tired of digging through SillyTavern's config every time I wanted to change the tone of a scene. So I built my own thing. **The idea:** sliders instead of prompts. Want slow burn? Drag pacing down. High tension? Push intensity up. The app handles prompt injections behind the scenes. There are presets too if you ...
2026-02-18T17:26:07
https://www.reddit.com/gallery/1r89a4y
Possible_Statement84
reddit.com
1970-01-01T00:00:00
0
{}
1r89a4y
false
null
t3_1r89a4y
/r/LocalLLaMA/comments/1r89a4y/vellium_opensource_desktop_app_for_creative/
false
false
https://preview.redd.it/…cb054ece5781faf7
90
null
AI agent framework that blocks dangerous commands before they execute
0
Hello! You probably saw the stories: \- Replit's AI deleted an entire production database during a code freeze, then said "I panicked instead of thinking" \- Claude Code deleted someone's home directory because sandboxing wasn't on by default \- Google Antigravity wiped a user's entire D: drive in "Turbo mode" I ke...
2026-02-18T16:51:09
https://github.com/NexTryApp/TaskPilot
Creative-Listen-6847
github.com
1970-01-01T00:00:00
0
{}
1r88aiz
false
null
t3_1r88aiz
/r/LocalLLaMA/comments/1r88aiz/ai_agent_framework_that_blocks_dangerous_commands/
false
false
https://external-preview…df9c72dd593592de
0
{'enabled': False, 'images': [{'id': 'Pp9ywP0SQY7sZg_KC_TMqa-Q36ARnYgggkbgJQHm39k', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/Pp9ywP0SQY7sZg_KC_TMqa-Q36ARnYgggkbgJQHm39k.png?width=108&crop=smart&auto=webp&s=7ace8701a3cb15dc1d61c935a310b42e3cb5a00c', 'width': 108}, {'height': 108, 'url': 'h...
Qwen/Gemma/Devstral/Local Equivalents to Copilot GPT-5 mini?
1
[removed]
2026-02-18T16:39:40
https://www.reddit.com/r/LocalLLaMA/comments/1r87ymf/qwengemmadevstrallocal_equivalents_to_copilot/
itisyeetime
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r87ymf
false
null
t3_1r87ymf
/r/LocalLLaMA/comments/1r87ymf/qwengemmadevstrallocal_equivalents_to_copilot/
false
false
self
1
null
Built a tool to benchmark openclaw spend - what’s yours like?
1
[removed]
2026-02-18T16:39:26
https://www.reddit.com/r/LocalLLaMA/comments/1r87ydo/built_a_tool_to_benchmark_openclaw_spend_whats/
babyalpac
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r87ydo
false
null
t3_1r87ydo
/r/LocalLLaMA/comments/1r87ydo/built_a_tool_to_benchmark_openclaw_spend_whats/
false
false
self
1
{'enabled': False, 'images': [{'id': 'v4JRAJ1f2taXWJ1jarqiW2v6lasnv-KB53KIFHFxoT8', 'resolutions': [{'height': 56, 'url': 'https://external-preview.redd.it/v4JRAJ1f2taXWJ1jarqiW2v6lasnv-KB53KIFHFxoT8.png?width=108&crop=smart&auto=webp&s=d51f9bd123a414bb6f67ef2fc6ee7c54e8e3c479', 'width': 108}, {'height': 113, 'url': 'h...
I'm wanting to run a local llm for coding. Will this system work?
0
I have a system with a Rizen 3600, and 96GB ram. Currently it has a gtx 1600 6gb, but I was thinking of putting in an RTX 4060 Ti 16GB in it. Would that configuration give me enough juice for what I need?
2026-02-18T16:37:32
https://www.reddit.com/r/LocalLLaMA/comments/1r87wgr/im_wanting_to_run_a_local_llm_for_coding_will/
rogue780
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r87wgr
false
null
t3_1r87wgr
/r/LocalLLaMA/comments/1r87wgr/im_wanting_to_run_a_local_llm_for_coding_will/
false
false
self
0
null
Self-rebuilding meta-benchmark for LLMs that easy to specify but extreamly hard to pass.
3
I have been thinking about a meta-benchmark concept that is easy to specify but practically impossible for current models to pass. I wanted to get your thoughts on the viability of this as a long-term goal for open source models. The core idea is to verify if a model can truly understand and replicate its own function...
2026-02-18T16:31:07
https://www.reddit.com/r/LocalLLaMA/comments/1r87pry/selfrebuilding_metabenchmark_for_llms_that_easy/
Another__one
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r87pry
false
null
t3_1r87pry
/r/LocalLLaMA/comments/1r87pry/selfrebuilding_metabenchmark_for_llms_that_easy/
false
false
self
3
null
AFS — give your local agents a persistent memory. No cloud, no strings attached. All on disk.
0
Hey r/LocalLLaMA, I've been running local agent pipelines for a while and kept hitting the same frustrating wall: \*\*my agents are goldfish\*\*. Every session restart they start fresh. All that useful context they built up — gone. I built AFS to fix this. Sharing here specifically because this community gets why "no...
2026-02-18T16:30:36
https://www.reddit.com/r/LocalLLaMA/comments/1r87p7p/afs_give_your_local_agents_a_persistent_memory_no/
Guilty_Nothing_2858
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r87p7p
false
null
t3_1r87p7p
/r/LocalLLaMA/comments/1r87p7p/afs_give_your_local_agents_a_persistent_memory_no/
false
false
self
0
{'enabled': False, 'images': [{'id': '7rhTR4w8UVOwSSmVcYihoj6dHKS1V1RfTKgoSvIM8DM', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/7rhTR4w8UVOwSSmVcYihoj6dHKS1V1RfTKgoSvIM8DM.png?width=108&crop=smart&auto=webp&s=87383ac67caccce8d49ba32dd542b87d796952c8', 'width': 108}, {'height': 108, 'url': 'h...
UPDATE#3: repurposing 800 RX 580s converted to AI cluster
71
hey everyone, posting an update on the ETH mining farm conversion project. last time i posted we were still figuring out what to even do with 800 rx 580s (mix of 4gb and 8gb sapphire nitro+ and pulse cards) sitting in an old ethereum mining farm so the tldr is we think we finally found a good use case. maybe two actua...
2026-02-18T16:30:14
https://www.reddit.com/r/LocalLLaMA/comments/1r87ou8/update3_repurposing_800_rx_580s_converted_to_ai/
rasbid420
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r87ou8
false
null
t3_1r87ou8
/r/LocalLLaMA/comments/1r87ou8/update3_repurposing_800_rx_580s_converted_to_ai/
false
false
self
71
null
What if every OpenClaw action was fully logged and replayable?
0
A lot of the recent concern around agents like OpenClaw seems to come down to one thing: we don’t really see what they’re doing once they start acting autonomously. They can read mail, call APIs, modify files, chain tools together — and most of that happens outside the user’s direct awareness. The capability itself isn...
2026-02-18T16:28:28
https://www.reddit.com/r/LocalLLaMA/comments/1r87n0x/what_if_every_openclaw_action_was_fully_logged/
NeoLogic_Dev
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r87n0x
false
null
t3_1r87n0x
/r/LocalLLaMA/comments/1r87n0x/what_if_every_openclaw_action_was_fully_logged/
false
false
self
0
null
been using frontier models for years - what am i actually missing with local?
0
hello everyone. first post here, new to reddit too. i’ve been using frontier models pretty heavily for the past while. not as a developer - but just as someone becoming more obsessed with what these things could actually do. automating stuff, going deep on topics, prototyping ideas i had no real business trying. late...
2026-02-18T16:24:38
https://www.reddit.com/r/LocalLLaMA/comments/1r87j4b/been_using_frontier_models_for_years_what_am_i/
npcdamian
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r87j4b
false
null
t3_1r87j4b
/r/LocalLLaMA/comments/1r87j4b/been_using_frontier_models_for_years_what_am_i/
false
false
self
0
null
Coming Soon to Local Models, if I have my way (True Long Context LLM's without retraining)
0
# KeSSie Conversation Memory Architecture **Sliding Window KV over Linear Conversation Arrays** Addendum to KeSSie Foundation Model Specification February 2026 - v1.1 (Implementation Status Update) ## 1. Overview: The Problem with KV Cache Standard transformer attention requires storing key-value pairs for every toke...
2026-02-18T16:22:45
https://www.reddit.com/r/LocalLLaMA/comments/1r87h6r/coming_soon_to_local_models_if_i_have_my_way_true/
--TastesLikeChicken-
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r87h6r
false
null
t3_1r87h6r
/r/LocalLLaMA/comments/1r87h6r/coming_soon_to_local_models_if_i_have_my_way_true/
false
false
self
0
null
Local Equivalents to Copilot GPT-5 mini?
1
[removed]
2026-02-18T16:21:10
https://www.reddit.com/r/LocalLLaMA/comments/1r87fjb/local_equivalents_to_copilot_gpt5_mini/
itisyeetime
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r87fjb
false
null
t3_1r87fjb
/r/LocalLLaMA/comments/1r87fjb/local_equivalents_to_copilot_gpt5_mini/
false
false
self
1
null
Current status of LiteLLM (Python SDK) + Langfuse v3 integration?
0
Hi everyone, I'm planning to upgrade to Langfuse v3 but I've seen several GitHub issues mentioning compatibility problems with LiteLLM. I've read that the native `litellm.success_callback = ["langfuse"]` approach relies on the v2 SDK and might break or lose data with v3. My questions is anyone successfully stabilized t...
2026-02-18T16:17:15
https://www.reddit.com/r/LocalLLaMA/comments/1r87bm2/current_status_of_litellm_python_sdk_langfuse_v3/
ReplacementMoney2484
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r87bm2
false
null
t3_1r87bm2
/r/LocalLLaMA/comments/1r87bm2/current_status_of_litellm_python_sdk_langfuse_v3/
false
false
self
0
null
Vibe Check: Latest models on AMD Strix Halo
29
I’ve been testing a bunch of recent drops on my AMD homelab (Ryzen AI Max+ 395 + R9700) with a very non-scientific “vibe check” workflow (Roo Code + Open WebUI). A few standouts that replaced my old stack: * **Kimi Linear 48B Instruct** as a daily-driver generalist. * **Qwen3 Coder Next** as my new coding model. * **...
2026-02-18T16:13:31
https://www.reddit.com/r/LocalLLaMA/comments/1r877tl/vibe_check_latest_models_on_amd_strix_halo/
bhamm-lab
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r877tl
false
null
t3_1r877tl
/r/LocalLLaMA/comments/1r877tl/vibe_check_latest_models_on_amd_strix_halo/
false
false
self
29
null
Abliteration/Activation Steering on LLMs specialized for Cybersecurity
4
I want to use activation steering (abliteration) on models *already* specialized for cybersecurity (like WhiteRabbitNeo or Foundation-Sec-8B). Even though these models are fine-tuned for offense, they still have "residual safety alignment" buried in them from their base models that makes them occasionally refuse expli...
2026-02-18T16:10:10
https://www.reddit.com/r/LocalLLaMA/comments/1r874fx/abliterationactivation_steering_on_llms/
dumbelco
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r874fx
false
null
t3_1r874fx
/r/LocalLLaMA/comments/1r874fx/abliterationactivation_steering_on_llms/
false
false
self
4
null
Suddenly Minimax IQ4-XS doesn't fit in 128GB anymore
1
I downloaded and tested Minimax2.1 in the first days of January, using llama-bench I was able run it up to a context depth of 16K, RAM usage was around 95-97%, it was around 2% before starting. These days I downloaded Minimax2.5 to test it and it didn't even load with 0K depth, the RAM usage grows up to 100% and the k...
2026-02-18T16:09:36
https://www.reddit.com/r/LocalLLaMA/comments/1r873vv/suddenly_minimax_iq4xs_doesnt_fit_in_128gb_anymore/
dionisioalcaraz
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r873vv
false
null
t3_1r873vv
/r/LocalLLaMA/comments/1r873vv/suddenly_minimax_iq4xs_doesnt_fit_in_128gb_anymore/
false
false
self
1
null
Save $25/month on Lovable by moving to free hosting with one command
0
Lovable is great for building sites but once you're done building, you're mostly paying for hosting and an AI editor. Vercel hosts it for free. Claude Code edits it the same way. ...
2026-02-18T16:06:47
https://www.reddit.com/r/LocalLLaMA/comments/1r87127/save_25month_on_lovable_by_moving_to_free_hosting/
Nir777
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r87127
false
null
t3_1r87127
/r/LocalLLaMA/comments/1r87127/save_25month_on_lovable_by_moving_to_free_hosting/
false
false
self
0
{'enabled': False, 'images': [{'id': 'mwKJpqA1siKVmIpN8v4_T6s3rE1fOElTTctAC1yzgxQ', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/mwKJpqA1siKVmIpN8v4_T6s3rE1fOElTTctAC1yzgxQ.png?width=108&crop=smart&auto=webp&s=41d95aac469c3a9e0ce4b3248b43df45558cd21b', 'width': 108}, {'height': 108, 'url': 'h...
Are Chinese models fully Chinese?
0
I noticed something interesting when I use Chinese llm models in English, everything is great, but when I switch to my language (Polish), most Chinese models introduce themselves as Claude from Antropic or Chat GPT from OpenAI. Examples include MiniMax-M.2.5 and GLM-4.7 Flash. I was expecting that after so many new ite...
2026-02-18T16:06:42
https://www.reddit.com/r/LocalLLaMA/comments/1r870zd/are_chinese_models_fully_chinese/
mossy_troll_84
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r870zd
false
null
t3_1r870zd
/r/LocalLLaMA/comments/1r870zd/are_chinese_models_fully_chinese/
false
false
https://preview.redd.it/…e69cc52ab8369116
0
null
for llm PCIe 4.0 pcie 3.0 isn't going to make any difference.
3
Im using only 1 GPU, the model Is fully loaded on my GPU without using gguf without CPU offload
2026-02-18T15:56:34
https://www.reddit.com/r/LocalLLaMA/comments/1r86qyq/for_llm_pcie_40_pcie_30_isnt_going_to_make_any/
Quiet_Dasy
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r86qyq
false
null
t3_1r86qyq
/r/LocalLLaMA/comments/1r86qyq/for_llm_pcie_40_pcie_30_isnt_going_to_make_any/
false
false
self
3
null
Whats the current smartest uncensored LLM for 12GB Vram
3
I don't need something that will be a genius roleplayer, but I do need something that won't stop talking no matter how bad or deprived it gets, and it needs to be smart to understand complex situations If it matters, I want it for asking advice on fictional kinky scenarios
2026-02-18T15:50:28
https://www.reddit.com/r/LocalLLaMA/comments/1r86l27/whats_the_current_smartest_uncensored_llm_for/
Migdan
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r86l27
false
null
t3_1r86l27
/r/LocalLLaMA/comments/1r86l27/whats_the_current_smartest_uncensored_llm_for/
false
false
self
3
null
LLMs grading other LLMs 2
227
A year ago I made a [meta-eval here on the sub](https://www.reddit.com/r/LocalLLaMA/comments/1j1npv1/llms_grading_other_llms/), asking LLMs to grade a few criterias about other LLMs. Time for the part 2. The premise is very simple, the model is asked a few ego-baiting questions and other models are then asked to ran...
2026-02-18T15:47:24
https://i.redd.it/rmq2mwriw9kg1.png
Everlier
i.redd.it
1970-01-01T00:00:00
0
{}
1r86i3o
false
null
t3_1r86i3o
/r/LocalLLaMA/comments/1r86i3o/llms_grading_other_llms_2/
false
false
https://preview.redd.it/…fa3a5f6a3cb9cec2
227
{'enabled': True, 'images': [{'id': 'rmq2mwriw9kg1', 'resolutions': [{'height': 86, 'url': 'https://preview.redd.it/rmq2mwriw9kg1.png?width=108&crop=smart&auto=webp&s=5481f8308ab5f0371a5fce2561d3d10c1c5cdb5d', 'width': 108}, {'height': 172, 'url': 'https://preview.redd.it/rmq2mwriw9kg1.png?width=216&crop=smart&auto=web...
No love for Intel GPUs?
17
On a per VRAM GB basis, Intel GPUs are way cheaper than a Nvidia ones. But why is there no love them here? Am I missing something?
2026-02-18T15:47:09
https://www.reddit.com/r/LocalLLaMA/comments/1r86huj/no_love_for_intel_gpus/
pelicanthief
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r86huj
false
null
t3_1r86huj
/r/LocalLLaMA/comments/1r86huj/no_love_for_intel_gpus/
false
false
self
17
null
Can i Run Granite-Vision-3.3-2B on a RX 6500XT?
0
2026-02-18T15:44:08
https://i.redd.it/q4nrkl58x9kg1.jpeg
Quiet_Dasy
i.redd.it
1970-01-01T00:00:00
0
{}
1r86euz
false
null
t3_1r86euz
/r/LocalLLaMA/comments/1r86euz/can_i_run_granitevision332b_on_a_rx_6500xt/
false
false
https://preview.redd.it/…1538e1115714ffb9
0
{'enabled': True, 'images': [{'id': 'q4nrkl58x9kg1', 'resolutions': [{'height': 216, 'url': 'https://preview.redd.it/q4nrkl58x9kg1.jpeg?width=108&crop=smart&auto=webp&s=f01972ea968946ad8110ac53500282efc9231cae', 'width': 108}, {'height': 432, 'url': 'https://preview.redd.it/q4nrkl58x9kg1.jpeg?width=216&crop=smart&auto=...
Looking for GPU upgrade advice for fine-tuning
1
Currently own a 2x 3090Ti rig that I use for research/experiments. Nowadays I'm mostly doing full finetunes of 1-2B parameter VLMs, and a bunch of BERT/encoder experiments. I currently use the cloud for anything larger, or when I want to scale out experiments, but was thinking about upgrading to be able to run more lo...
2026-02-18T15:36:59
https://www.reddit.com/r/LocalLLaMA/comments/1r867s9/looking_for_gpu_upgrade_advice_for_finetuning/
diamondium
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r867s9
false
null
t3_1r867s9
/r/LocalLLaMA/comments/1r867s9/looking_for_gpu_upgrade_advice_for_finetuning/
false
false
self
1
null
Open Cowork v3.1.0: desktop agent runtime with GUI operations, MCP integration, and compatible model endpoints
7
Disclosure: maintainer here. Sharing a technical project update for **Open Cowork**, an open-source desktop agent app focused on tool use and GUI workflows. Current architecture/capabilities: * Electron desktop runtime (main/renderer separation) * Workspace path-scoped execution * Optional VM command isolation (WSL2...
2026-02-18T15:34:42
https://www.reddit.com/r/LocalLLaMA/comments/1r865of/open_cowork_v310_desktop_agent_runtime_with_gui/
Sensitive_Dingo_4839
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r865of
false
null
t3_1r865of
/r/LocalLLaMA/comments/1r865of/open_cowork_v310_desktop_agent_runtime_with_gui/
false
false
https://preview.redd.it/…d457cc5d54fd215e
7
null
Even with Opus 4.6 and massive context windows, this is still the only thing that saves my production pipelines
56
We all got excited when the new reasoning models dropped. Better at following instructions, longer context, fewer hallucinations. Great. Still seeing agentic workflows fail at basic deterministic logic because teams treat the LLM as a CPU instead of what it is — a reasoning engine. After the bug I shared on Monday (R...
2026-02-18T15:27:51
https://i.redd.it/esofp8nbu9kg1.jpeg
tdeliev
i.redd.it
1970-01-01T00:00:00
0
{}
1r85z1t
false
null
t3_1r85z1t
/r/LocalLLaMA/comments/1r85z1t/even_with_opus_46_and_massive_context_windows/
false
false
https://preview.redd.it/…63eff81a070ae67c
56
{'enabled': True, 'images': [{'id': 'esofp8nbu9kg1', 'resolutions': [{'height': 36, 'url': 'https://preview.redd.it/esofp8nbu9kg1.jpeg?width=108&crop=smart&auto=webp&s=703a63de95e6fcb618ccfd2de7a087544f7f9115', 'width': 108}, {'height': 72, 'url': 'https://preview.redd.it/esofp8nbu9kg1.jpeg?width=216&crop=smart&auto=we...
Why opencode give me instructions and dosen't take any action with my local model?
0
I'm trying to use OpenCode, but I can't understand why it gives me instructions instead of performing the actions I request. For example, even with very simple commands like "create a folder on the desktop," it provides instructions on how to do it—or sometimes doesn't even do that—but it doesn't execute anything. The ...
2026-02-18T15:23:53
https://www.reddit.com/r/LocalLLaMA/comments/1r85v3n/why_opencode_give_me_instructions_and_dosent_take/
Worried_Menu4016
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r85v3n
false
null
t3_1r85v3n
/r/LocalLLaMA/comments/1r85v3n/why_opencode_give_me_instructions_and_dosent_take/
false
false
self
0
null
Installing OpenClaw with Local Ollama on Azure VM - Getting "Pull Access Denied" Error
0
**Hi everyone,** I'm a Data Science student currently trying to self-host **OpenClaw** (formerly Molt) on an **Azure VM** (Ubuntu, 32GB RAM). I already have **Ollama** running locally on the same VM with the `qwen2.5-coder:32b` model. I want to run OpenClaw via Docker and connect it to my local Ollama instance using ...
2026-02-18T15:19:52
https://www.reddit.com/r/LocalLLaMA/comments/1r85r3p/installing_openclaw_with_local_ollama_on_azure_vm/
Sea_Lawfulness_5602
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r85r3p
false
null
t3_1r85r3p
/r/LocalLLaMA/comments/1r85r3p/installing_openclaw_with_local_ollama_on_azure_vm/
false
false
self
0
null
Devstral Small 2 24B + Qwen3 Coder 30B: Coders for Every Hardware (Yes, Even the Pi)
151
Hey r/LocalLLaMA, *ByteShape’s back, alright! Everybody (yeah), you asked for coders (yeah). Everybody get your coders right:* **Devstral-Small-2-24B-Instruct-2512** (ShapeLearn-optimized for GPU) + **Qwen3-Coder-30B-A3B-Instruct** (optimized for all hardware and patience levels). Alright! **TL;DR** * **Devstral** is...
2026-02-18T15:16:53
https://i.redd.it/zzlx2eqlr9kg1.jpeg
enrique-byteshape
i.redd.it
1970-01-01T00:00:00
0
{}
1r85o89
false
null
t3_1r85o89
/r/LocalLLaMA/comments/1r85o89/devstral_small_2_24b_qwen3_coder_30b_coders_for/
false
false
https://preview.redd.it/…58fb0455b280530f
151
{'enabled': True, 'images': [{'id': 'zzlx2eqlr9kg1', 'resolutions': [{'height': 72, 'url': 'https://preview.redd.it/zzlx2eqlr9kg1.jpeg?width=108&crop=smart&auto=webp&s=ea079a5bb6ffd700d004eeb4b2617c2cc66e2d1f', 'width': 108}, {'height': 144, 'url': 'https://preview.redd.it/zzlx2eqlr9kg1.jpeg?width=216&crop=smart&auto=w...
Built a simple GPU recommendation tool for people who don't know which GPU to rent
1
I'm a non-technical founder who got tired of trying to figure out which GPU I needed for my AI project. Every tool assumes you already know what you're looking for. So I built Computra: answer 4 simple questions, get a clear GPU recommendation with explanation. 🎯 What it does: \- Asks about your workload (r...
2026-02-18T15:11:49
https://www.reddit.com/r/LocalLLaMA/comments/1r85jhq/built_a_simple_gpu_recommendation_tool_for_people/
Complex-Telephone469
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r85jhq
false
null
t3_1r85jhq
/r/LocalLLaMA/comments/1r85jhq/built_a_simple_gpu_recommendation_tool_for_people/
false
false
self
1
null
I nuked my hard drive to build an AI-Native OS from scratch. The LLM is PID 1. There is no systemd.
0
Hi r/LocalLLaMA, I'm 19, an aerospace engineering student, and for the last 13 months I've been building a new operating system called **Axiom**. I wanted to answer one question: **What if the LLM wasn't an app but the kernel's first user?** Most "AI OS" projects are just wrappers around Linux or glorified chatbots....
2026-02-18T15:08:19
https://www.reddit.com/r/LocalLLaMA/comments/1r85gak/i_nuked_my_hard_drive_to_build_an_ainative_os/
Upbeat_Confection411
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r85gak
false
null
t3_1r85gak
/r/LocalLLaMA/comments/1r85gak/i_nuked_my_hard_drive_to_build_an_ainative_os/
false
false
self
0
null
newpr — open-source tool that wraps Claude Code / OpenCode / Codex to do agentic PR review with codebase exploration
1
newpr — open-source tool that wraps Claude Code / OpenCode / Codex to do agentic PR review with codebase exploration I built newpr, an open-source CLI that wraps agentic coding tools to analyze large GitHub PRs. The key idea: Instead of just feeding diffs to an LLM, newpr spawns an actual coding agent (Claude Code, O...
2026-02-18T15:00:52
https://v.redd.it/sv16ae5hp9kg1
jiwonme
v.redd.it
1970-01-01T00:00:00
0
{}
1r85989
false
{'reddit_video': {'bitrate_kbps': 5000, 'dash_url': 'https://v.redd.it/sv16ae5hp9kg1/DASHPlaylist.mpd?a=1774018868%2CMTY5NzFiNjRkNzQ5NWFlZGRiOTZjZjhiZTM1MWYxNzQ0NjAwMjM0OTJkYjhjODRlYTMxYTY4MjY0MDQ3ODYxMQ%3D%3D&v=1&f=sd', 'duration': 63, 'fallback_url': 'https://v.redd.it/sv16ae5hp9kg1/CMAF_1080.mp4?source=fallback', 'h...
t3_1r85989
/r/LocalLLaMA/comments/1r85989/newpr_opensource_tool_that_wraps_claude_code/
false
false
https://external-preview…05ee0020f05a5eb0
1
{'enabled': False, 'images': [{'id': 'bWhob2JlNWhwOWtnMTg4odIjRdG8x5xxvrDfy6Nd_F_smKqcRgWPUjL8Nbzl', 'resolutions': [{'height': 69, 'url': 'https://external-preview.redd.it/bWhob2JlNWhwOWtnMTg4odIjRdG8x5xxvrDfy6Nd_F_smKqcRgWPUjL8Nbzl.png?width=108&crop=smart&format=pjpg&auto=webp&s=52301f394f9eaa0dc91bc72365cb810bc368c...
We built a golf forecasting model that outperforms GPT‑5; model and dataset are open-sourced on Hugging Face
6
TLDR: * Fine-tuned gpt-oss-120b with GRPO on 3,178 professional golf forecasting questions. * Brier 0.207 on 855 held-out questions, beating both the base model (0.218) and GPT-5 (0.218). * Calibration improved the most: ECE 0.062 vs 0.083 (base) and 0.106 (GPT-5). * The same setup can be applied to other topics (e.g....
2026-02-18T14:54:43
https://www.reddit.com/r/LocalLLaMA/comments/1r853l6/we_built_a_golf_forecasting_model_that/
LightningRodLabs
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r853l6
false
null
t3_1r853l6
/r/LocalLLaMA/comments/1r853l6/we_built_a_golf_forecasting_model_that/
false
false
self
6
{'enabled': False, 'images': [{'id': 'eVWGoMjfqodntVSAV80aO2eCZJbZXh7gBBtJGGC4ffw', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/eVWGoMjfqodntVSAV80aO2eCZJbZXh7gBBtJGGC4ffw.png?width=108&crop=smart&auto=webp&s=f6503a0d44292776a29351ccb2c1600598741390', 'width': 108}, {'height': 116, 'url': 'h...
COMB: zero-dependency Python library for lossless AI agent memory — hash-chained, honeycomb-structured, pure stdlib
1
[removed]
2026-02-18T14:46:23
https://www.reddit.com/r/LocalLLaMA/comments/1r84w2m/comb_zerodependency_python_library_for_lossless/
Artifact-Virtual
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r84w2m
false
null
t3_1r84w2m
/r/LocalLLaMA/comments/1r84w2m/comb_zerodependency_python_library_for_lossless/
false
false
self
1
{'enabled': False, 'images': [{'id': 'lcR3YCEM-nuQKL_IoqTspkpHLCnowf8nneY45nDwKKQ', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/lcR3YCEM-nuQKL_IoqTspkpHLCnowf8nneY45nDwKKQ.png?width=108&crop=smart&auto=webp&s=81662e700435a46bff7e9e3c6619801a7186fcef', 'width': 108}, {'height': 108, 'url': 'h...
I did an analysis of 44 AI agent frameworks, sharing the result
15
I went through 44 AI agent frameworks for research on context management for a project. I spent some time pulling out results from the analysis and compiling it all together, so I thought I might as well share it. [https://github.com/larsderidder/framework-analysis](https://github.com/larsderidder/framework-analysis)
2026-02-18T14:37:38
https://www.reddit.com/r/LocalLLaMA/comments/1r84o6p/i_did_an_analysis_of_44_ai_agent_frameworks/
wouldacouldashoulda
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r84o6p
false
null
t3_1r84o6p
/r/LocalLLaMA/comments/1r84o6p/i_did_an_analysis_of_44_ai_agent_frameworks/
false
false
self
15
{'enabled': False, 'images': [{'id': '6m-8IgWJwjoVb5IIwH0RvIBqytsdSwNeeiKtD3nMomA', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/6m-8IgWJwjoVb5IIwH0RvIBqytsdSwNeeiKtD3nMomA.png?width=108&crop=smart&auto=webp&s=7c65f8306dafcf939cf8116d7eb2d59fe58559d8', 'width': 108}, {'height': 108, 'url': 'h...
Hardware experts - Will epyc 7763 matter for CPU offloading?
6
Currently running a 7502, As I understand it PP is compute bound and token gen is memory bound. So an upgrade might provide a lift on PP, but probably nothing on TG. I'm running huge models Deepseek/GLM/Kimi/Qwen where I have 75% of the models offloaded to system ram. If anyone has done an epyc CPU upgrade and see...
2026-02-18T14:31:20
https://www.reddit.com/r/LocalLLaMA/comments/1r84iif/hardware_experts_will_epyc_7763_matter_for_cpu/
segmond
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r84iif
false
null
t3_1r84iif
/r/LocalLLaMA/comments/1r84iif/hardware_experts_will_epyc_7763_matter_for_cpu/
false
false
self
6
null
Running Granite-Vision-3.3-2B on a GTX 1060 .CPU spillover inevitable due to lack of Tensor Cores?
0
Hey guys, looking for some reality check on running **Granite-Vision-3.3-2B** on a **GTX 1060**. I keep hearing that because the 1060 (Pascal) lacks Tensor Cores and modern INT8 optimization, it struggles with newer quantized models. Specifically: * Does the lack of Tensor Cores force everything onto standard CUDA c...
2026-02-18T14:25:48
https://www.reddit.com/r/LocalLLaMA/comments/1r84dja/running_granitevision332b_on_a_gtx_1060_cpu/
Quiet_Dasy
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r84dja
false
null
t3_1r84dja
/r/LocalLLaMA/comments/1r84dja/running_granitevision332b_on_a_gtx_1060_cpu/
false
false
self
0
null
New Hire OnBoarding Tool
1
[removed]
2026-02-18T14:22:16
https://www.reddit.com/r/LocalLLaMA/comments/1r84abf/new_hire_onboarding_tool/
Practical-Koala2831
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r84abf
false
null
t3_1r84abf
/r/LocalLLaMA/comments/1r84abf/new_hire_onboarding_tool/
false
false
self
1
{'enabled': False, 'images': [{'id': '-f3L9S2ZMuVnK5v63uq3DOcyi0duDx7x8IE6OMNY0GA', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/-f3L9S2ZMuVnK5v63uq3DOcyi0duDx7x8IE6OMNY0GA.png?width=108&crop=smart&auto=webp&s=2df3fb84a25fedc65e814d2f749b6e398da6d764', 'width': 108}, {'height': 108, 'url': 'h...
Can autonomous agents really handle full production workflows?
1
[removed]
2026-02-18T14:16:03
https://www.reddit.com/r/LocalLLaMA/comments/1r844ta/can_autonomous_agents_really_handle_full/
Practical-Koala2831
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r844ta
false
null
t3_1r844ta
/r/LocalLLaMA/comments/1r844ta/can_autonomous_agents_really_handle_full/
false
false
self
1
null
Msty Admin MCP v5.0.0 — Bloom behavioral evaluation for local LLMs: know when your model is lying to you
0
I've been building an MCP server for Msty Studio Desktop and just shipped v5.0.0, which adds something I'm really excited about: **Bloom**, a behavioral evaluation framework for local models. # The problem If you run local LLMs, you've probably noticed they sometimes agree with whatever you say (sycophancy), confiden...
2026-02-18T14:12:04
https://www.reddit.com/r/LocalLLaMA/comments/1r841dg/msty_admin_mcp_v500_bloom_behavioral_evaluation/
CryptBay
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r841dg
false
null
t3_1r841dg
/r/LocalLLaMA/comments/1r841dg/msty_admin_mcp_v500_bloom_behavioral_evaluation/
false
false
self
0
{'enabled': False, 'images': [{'id': 'AY_HrzCVWtqsOk0H_OBrj2gqtOrIhVgR5RYQnoG5DCU', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/AY_HrzCVWtqsOk0H_OBrj2gqtOrIhVgR5RYQnoG5DCU.png?width=108&crop=smart&auto=webp&s=f8b723f7582a9f2d37871df08406f3dc2b8e559a', 'width': 108}, {'height': 108, 'url': 'h...
llama-server when VRAM is > RAM. What are the best settings for faster model loading?
1
**Note** that I'm not referring to *offloading* to RAM, just merely faster model loading when they do not fit into RAM for sake of initial loading. Command: as simple as `llama-server --parallel -m gpt-oss-120b-F16.gguf`. fa, ngl, and fit are left to their respective default values \`on\`, \`all\` and \`on\`. So, -n...
2026-02-18T14:07:54
https://www.reddit.com/r/LocalLLaMA/comments/1r83xuf/llamaserver_when_vram_is_ram_what_are_the_best/
ParaboloidalCrest
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r83xuf
false
null
t3_1r83xuf
/r/LocalLLaMA/comments/1r83xuf/llamaserver_when_vram_is_ram_what_are_the_best/
false
false
self
1
null
PSA: DDR5 RDIMM price passed the point were 3090 are less expensive per gb..
456
Hello all, Just wanted to note that RDIMM prices are so wild.. Stacking rdimms starts to be as expensive as stacking 3090s.. But RDIMM don't come with compute included.. What a crazy time
2026-02-18T13:51:04
https://www.reddit.com/r/LocalLLaMA/comments/1r83irw/psa_ddr5_rdimm_price_passed_the_point_were_3090/
No_Afternoon_4260
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r83irw
false
null
t3_1r83irw
/r/LocalLLaMA/comments/1r83irw/psa_ddr5_rdimm_price_passed_the_point_were_3090/
false
false
self
456
null
Computers with the GB10 chips
6
Nvidia spark, asus ascent, dell promax and the likes all have the connectx nics which account for probably half the price of the device. Why haven't they made those devices without the chip, just regular nics? Sounds like a arm device with unified memory would be just enough for most of the people here. I know the nvid...
2026-02-18T13:48:35
https://www.reddit.com/r/LocalLLaMA/comments/1r83gl5/computers_with_the_gb10_chips/
emaiksiaime
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r83gl5
false
null
t3_1r83gl5
/r/LocalLLaMA/comments/1r83gl5/computers_with_the_gb10_chips/
false
false
self
6
null
Why RAG failed for my Mystery Game . So I built a Logic Graph with DeepSeek 3.2 + Gemini 3.0 Pro
0
I've been building a lateral thinking puzzle game ("Turtle Soup") where users interrogate an AI host to solve a mystery. \*\*The Problem: RAG is terrible for binary logic.\*\* I started with a standard Vector Search pipeline. It failed miserably because \*\*semantic similarity != logical relevance.\*\* Here is th...
2026-02-18T13:41:22
https://www.reddit.com/r/LocalLLaMA/comments/1r83ai6/why_rag_failed_for_my_mystery_game_so_i_built_a/
Far-Client2086
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r83ai6
false
null
t3_1r83ai6
/r/LocalLLaMA/comments/1r83ai6/why_rag_failed_for_my_mystery_game_so_i_built_a/
false
false
self
0
null
Why RAG failed for my Mystery Game . So I built a Logic Graph with DeepSeek 3.2 + Gemini 3.0 Pro
1
[removed]
2026-02-18T13:37:16
https://www.reddit.com/r/LocalLLaMA/comments/1r8370z/why_rag_failed_for_my_mystery_game_so_i_built_a/
Far-Client2086
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8370z
false
null
t3_1r8370z
/r/LocalLLaMA/comments/1r8370z/why_rag_failed_for_my_mystery_game_so_i_built_a/
false
false
self
1
{'enabled': False, 'images': [{'id': 'Veq7afsF2AEk-V7_8ZToEr-Q8ZKIBnt4vkZgamSF_74', 'resolutions': [{'height': 108, 'url': 'https://external-preview.redd.it/Veq7afsF2AEk-V7_8ZToEr-Q8ZKIBnt4vkZgamSF_74.png?width=108&crop=smart&auto=webp&s=63ace7359c3c775266da52629705aa5a4fc282f9', 'width': 108}, {'height': 216, 'url': '...
Are more teams moving from APIs to renting GPUs for inference?
0
Lately, we have been noticing a shift toward running open models on rented GPUs instead of relying purely on APIs. The tradeoff seems to be: * lower cost at scale * more control/privacy * but higher ops overhead Curious if others here are seeing the same trend. If you’re running inference today, what setup are you ...
2026-02-18T13:31:15
https://www.reddit.com/r/LocalLLaMA/comments/1r831wt/are_more_teams_moving_from_apis_to_renting_gpus/
qubridInc
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r831wt
false
null
t3_1r831wt
/r/LocalLLaMA/comments/1r831wt/are_more_teams_moving_from_apis_to_renting_gpus/
false
false
self
0
null
Why RAG failed for my Mystery Game ("Suicide" ≈ "Murder"). So I built a Logic Graph with DeepSeek 3.2 + Gemini 3.0 Pro
1
[removed]
2026-02-18T13:30:00
https://www.reddit.com/r/LocalLLaMA/comments/1r830u1/why_rag_failed_for_my_mystery_game_suicide_murder/
Far-Client2086
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r830u1
false
null
t3_1r830u1
/r/LocalLLaMA/comments/1r830u1/why_rag_failed_for_my_mystery_game_suicide_murder/
false
false
self
1
{'enabled': False, 'images': [{'id': 'Veq7afsF2AEk-V7_8ZToEr-Q8ZKIBnt4vkZgamSF_74', 'resolutions': [{'height': 108, 'url': 'https://external-preview.redd.it/Veq7afsF2AEk-V7_8ZToEr-Q8ZKIBnt4vkZgamSF_74.png?width=108&crop=smart&auto=webp&s=63ace7359c3c775266da52629705aa5a4fc282f9', 'width': 108}, {'height': 216, 'url': '...
Genoa2D24G-2L+, dual AMD EPYC 9654, 1.5TB RAM, 8x4090 - Won't pass POST: Help needed
1
I bought a rig with the Genoa2D24G-2L+ and 8x4090 from a company called Autonomous, but requested a custom build without CPU and RAM since I had acquired those separately & since the CPU that would have been shipped with the pre-built system was less powerful & the RAM was less. I acquired the dual AMD EPYC 9654 CP...
2026-02-18T13:26:43
https://www.reddit.com/r/LocalLLaMA/comments/1r82y5m/genoa2d24g2l_dual_amd_epyc_9654_15tb_ram_8x4090/
MeMyselfAndEye123
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r82y5m
false
null
t3_1r82y5m
/r/LocalLLaMA/comments/1r82y5m/genoa2d24g2l_dual_amd_epyc_9654_15tb_ram_8x4090/
false
false
self
1
null
Cache hits in llama.cpp vs vLLM
0
I am facing severe cache misses in llama.cpp Every prompt takes forever and especially with Claude code and stuff So what do you think ? Is vLLM going to solve that
2026-02-18T13:26:14
https://www.reddit.com/r/LocalLLaMA/comments/1r82xrq/cache_hits_in_llamacpp_vs_vllm/
Potential_Block4598
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r82xrq
false
null
t3_1r82xrq
/r/LocalLLaMA/comments/1r82xrq/cache_hits_in_llamacpp_vs_vllm/
false
false
self
0
null
Best approach for Local G-Eval (Ollama)? DeepEval vs. Prometheus vs. Custom Script
0
Hi everyone, I’m fine-tuning a T5 model for **Conditional Summarization** where the output must strictly respect specific constraints (Target Language, specific Named Entities/NERs, and Length) while maintaining high fluency and coherence. I need to run the evaluation entirely locally using **Ollama** and I am consid...
2026-02-18T13:07:23
https://www.reddit.com/r/LocalLLaMA/comments/1r82il7/best_approach_for_local_geval_ollama_deepeval_vs/
Timely-Reindeer-5292
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r82il7
false
null
t3_1r82il7
/r/LocalLLaMA/comments/1r82il7/best_approach_for_local_geval_ollama_deepeval_vs/
false
false
self
0
null
HYPHA - P2P payment & discovery layer for autonomous AI agents (open source)
1
[removed]
2026-02-18T13:04:50
https://www.reddit.com/r/LocalLLaMA/comments/1r82gki/hypha_p2p_payment_discovery_layer_for_autonomous/
AvailableWindow840
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r82gki
false
null
t3_1r82gki
/r/LocalLLaMA/comments/1r82gki/hypha_p2p_payment_discovery_layer_for_autonomous/
false
false
self
1
null
OpenClaw Set Up for Under $20/Month : AWS EC2
0
Sharing this guide from This Week in AI. Thought it might be useful for folks experimenting with lightweight agent setups. Here’s the process: What You Need • AWS account (free to create) • API key from Anthropic, OpenAI, or Moonshot AI • \~5 minutes Step 1: Launch an EC2 Instance • Name: anything (e.g. opencla...
2026-02-18T12:52:45
https://www.reddit.com/r/LocalLLaMA/comments/1r826ya/openclaw_set_up_for_under_20month_aws_ec2/
pmv143
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r826ya
false
null
t3_1r826ya
/r/LocalLLaMA/comments/1r826ya/openclaw_set_up_for_under_20month_aws_ec2/
false
false
self
0
{'enabled': False, 'images': [{'id': 'Sr7lMve--ohVynqPM-FCjBlvtzSTAUoGh50OrUYQY68', 'resolutions': [{'height': 81, 'url': 'https://external-preview.redd.it/Sr7lMve--ohVynqPM-FCjBlvtzSTAUoGh50OrUYQY68.jpeg?width=108&crop=smart&auto=webp&s=b699652159bde800b0a2e24012040edefd11293f', 'width': 108}, {'height': 162, 'url': '...
Local LLM Hardware Recommendation
0
I have been researching a few options around getting myself a hardware for doing local LLM inference, slowly build upon a local LLM specific model. I hear various terms like Memory Bandwidth, GPU vRAM or System RAM, GPU Compute, PCIe bandwidth etc., So which ones should I pay attention to? My goal is to run local mod...
2026-02-18T12:44:15
https://www.reddit.com/r/LocalLLaMA/comments/1r820dw/local_llm_hardware_recommendation/
CaterpillarPrevious2
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r820dw
false
null
t3_1r820dw
/r/LocalLLaMA/comments/1r820dw/local_llm_hardware_recommendation/
false
false
self
0
null
(Google) On Surprising Effectiveness of Masking Updates in Adaptive Optimizers
66
2026-02-18T12:38:51
https://huggingface.co/papers/2602.15322
coder543
huggingface.co
1970-01-01T00:00:00
0
{}
1r81w8n
false
null
t3_1r81w8n
/r/LocalLLaMA/comments/1r81w8n/google_on_surprising_effectiveness_of_masking/
false
false
https://external-preview…f1ca65eb11252d2a
66
{'enabled': False, 'images': [{'id': 'ta2IiH0S_hDLLmbY2FJbCETWnyAZ9mwNCBtInzZsN24', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/ta2IiH0S_hDLLmbY2FJbCETWnyAZ9mwNCBtInzZsN24.png?width=108&crop=smart&auto=webp&s=6c1b3a5bcbdb1f681b52085a47d8117114452591', 'width': 108}, {'height': 116, 'url': 'h...
Every OpenClaw security vulnerability documented in one place — relevant if you're running it with local models
11
Full timeline of every OpenClaw security incident — the CVEs, ClawHub malware campaign, exposed instances, Moltbook leak, and government warnings. Covers the safe deployment approach including isolation and hardening. Relevant here since many of you run OpenClaw with local LLMs via LiteLLM or Ollama.
2026-02-18T12:38:22
https://blog.barrack.ai/openclaw-security-vulnerabilities-2026
LostPrune2143
blog.barrack.ai
1970-01-01T00:00:00
0
{}
1r81vw2
false
null
t3_1r81vw2
/r/LocalLLaMA/comments/1r81vw2/every_openclaw_security_vulnerability_documented/
false
false
https://external-preview…7d597752d7e230fa
11
{'enabled': False, 'images': [{'id': 'bCNdODCJf4MFLRT-GV2CFrzcSYGT_DXu9jf7b-sPOzI', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/bCNdODCJf4MFLRT-GV2CFrzcSYGT_DXu9jf7b-sPOzI.png?width=108&crop=smart&auto=webp&s=da9dc6ae5759c0939df33116c2171ff33a21f038', 'width': 108}, {'height': 116, 'url': 'h...
Why do all LLMs give the exact same generic Spark tuning advice no matter the job?
1
Been trying to use AI to debug a slow Spark job this week and it's honestly frustrating. Every single model I tried (ChatGPT, Claude, Gemini, even a couple of local ones I ran offline) spits out basically the same three lines: * increase executor memory * Tune your parallelism * Check for data skew I already know t...
2026-02-18T12:22:17
https://www.reddit.com/r/LocalLLaMA/comments/1r81jxf/why_do_all_llms_give_the_exact_same_generic_spark/
SweetHunter2744
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r81jxf
false
null
t3_1r81jxf
/r/LocalLLaMA/comments/1r81jxf/why_do_all_llms_give_the_exact_same_generic_spark/
false
false
self
1
null
RTX 5090 or M5 Ultra for AI? I’m totally lost
0
Hi team. Any tips??? I'm testing a 32GB RTX 5090 with a Mac Studio M5 Ultra with 192GB of integrated memory to run local LLMs, and here's what I've found so far: The RTX 5090 is incredibly fast if the model fits in VRAM (116 tokens/sec)... but if I ask it to handle huge models like DeepSeek-R1, it will get stuck hard...
2026-02-18T12:16:33
https://www.reddit.com/r/LocalLLaMA/comments/1r81fl5/rtx_5090_or_m5_ultra_for_ai_im_totally_lost/
Neural_Core_Tech
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r81fl5
false
null
t3_1r81fl5
/r/LocalLLaMA/comments/1r81fl5/rtx_5090_or_m5_ultra_for_ai_im_totally_lost/
false
false
self
0
null
I Built A Local Darwinian Evolution System For Agent Tools. No API Keys. Just natural selection.
1
[removed]
2026-02-18T12:15:29
https://www.reddit.com/gallery/1r81esu
MajorOk3668
reddit.com
1970-01-01T00:00:00
0
{}
1r81esu
false
null
t3_1r81esu
/r/LocalLLaMA/comments/1r81esu/i_built_a_local_darwinian_evolution_system_for/
false
false
https://preview.redd.it/…23d2f582b79f0fb8
1
null
is there any latest OCR model in market? February, 2026
1
i have tried lot of free open source OCR models like **paddlepaddle,Microsoft large** . but still i am looking for a more accurate OCR that can also detect multiline text, so can anyone suggest any model?
2026-02-18T12:13:56
https://www.reddit.com/r/LocalLLaMA/comments/1r81doc/is_there_any_latest_ocr_model_in_market_february/
Mountain-Act-7199
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r81doc
false
null
t3_1r81doc
/r/LocalLLaMA/comments/1r81doc/is_there_any_latest_ocr_model_in_market_february/
false
false
self
1
null
Fine-tuning Qwen2.5-3B for function calling on a free T4
0
i'm putting together a small group to keep building on a function-calling dataset and train better versions. dm me if you want in. i've built on the NousResearch hermes dataset with custom examples for restaurant search, flight booking and multi-step API chains on free Colab. if you're into fine-tuning small m...
2026-02-18T12:12:27
https://www.reddit.com/r/LocalLLaMA/comments/1r81cnc/finetuning_qwen253b_for_function_calling_on_a/
One_Intern4738
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r81cnc
false
null
t3_1r81cnc
/r/LocalLLaMA/comments/1r81cnc/finetuning_qwen253b_for_function_calling_on_a/
false
false
self
0
{'enabled': False, 'images': [{'id': 'K9V3LtuRhe9Xb0DF8O3tYRxQgON-OxcYKkXltEmU4FQ', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/K9V3LtuRhe9Xb0DF8O3tYRxQgON-OxcYKkXltEmU4FQ.png?width=108&crop=smart&auto=webp&s=505b2632e95dae647c5e7474335734194f35985b', 'width': 108}, {'height': 116, 'url': 'h...
Qwen 3.5 MXFP4 quants are coming - confirmed by Junyang Lin
123
Most here are aware that OpenAI did something very well with their GPT-Oss release - they trained their model in 4 bit and delivered native mxfp4 quants which means a lot higher quality than the typical Unsloth and Bartowski quants of bf16 models. Google did it too with Gemma 3 QAT which was very well received by the c...
2026-02-18T12:01:58
https://www.reddit.com/r/LocalLLaMA/comments/1r8157s/qwen_35_mxfp4_quants_are_coming_confirmed_by/
dampflokfreund
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r8157s
false
null
t3_1r8157s
/r/LocalLLaMA/comments/1r8157s/qwen_35_mxfp4_quants_are_coming_confirmed_by/
false
false
self
123
null
Got $800 of credits on digital ocean (for GPU usage). Anyone here that's into AI training and inference and could make use of it?
4
So I have around 800 bucks worth of GPU usage credits on digital ocean, those can be used specifically for GPU and clusters. So if any individual or hobbyist or anyone out here is training models or inference, or anything else, please contact.
2026-02-18T11:51:29
https://www.reddit.com/r/LocalLLaMA/comments/1r80xu4/got_800_of_credits_on_digital_ocean_for_gpu_usage/
DocumentFun9077
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r80xu4
false
null
t3_1r80xu4
/r/LocalLLaMA/comments/1r80xu4/got_800_of_credits_on_digital_ocean_for_gpu_usage/
false
false
self
4
null
Context Vault — Persistent memory for AI agents via MCP (Open Source - MIT)
0
built an MCP server that gives AI agents persistent memory across sessions. **The problem:** Every time you start a new Claude Code / Cursor / Cline session, the agent starts from scratch. All the insights, decisions, and patterns from previous sessions are gone. **The solution:** context-mcp is a local MCP server th...
2026-02-18T11:42:13
https://www.reddit.com/r/LocalLLaMA/comments/1r80rkk/context_vault_persistent_memory_for_ai_agents_via/
Slow-Bake-9603
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1r80rkk
false
null
t3_1r80rkk
/r/LocalLLaMA/comments/1r80rkk/context_vault_persistent_memory_for_ai_agents_via/
false
false
self
0
{'enabled': False, 'images': [{'id': 'BjnDWeTXWOC1IV_rB6BCueGGiDiObbgg8FhJAZBV0Vc', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/BjnDWeTXWOC1IV_rB6BCueGGiDiObbgg8FhJAZBV0Vc.png?width=108&crop=smart&auto=webp&s=f8ac43899c1c1fde434d2850f4f67836389a57e6', 'width': 108}, {'height': 108, 'url': 'h...