title
stringlengths
1
300
score
int64
0
8.54k
selftext
stringlengths
0
41.5k
created
timestamp[ns]date
2023-04-01 04:30:41
2026-03-04 02:14:14
url
stringlengths
0
878
author
stringlengths
3
20
domain
stringlengths
0
82
edited
timestamp[ns]date
1970-01-01 00:00:00
2026-02-19 14:51:53
gilded
int64
0
2
gildings
stringclasses
7 values
id
stringlengths
7
7
locked
bool
2 classes
media
stringlengths
646
1.8k
name
stringlengths
10
10
permalink
stringlengths
33
82
spoiler
bool
2 classes
stickied
bool
2 classes
thumbnail
stringlengths
4
213
ups
int64
0
8.54k
preview
stringlengths
301
5.01k
PC to run local llm for coding agent
2
I'm building a PC for running GPT-OSS 20b with 131k context length and [Qwen/Qwen3-Coder-30B-A3B-Instruct](https://huggingface.co/Qwen/Qwen3-Coder-30B-A3B-Instruct) with expected context length 256k. || || |CPU|AMD Ryzen 7 7700 Tray| |MAIN|Mainboard Gigabyte B650M AORUS ELITE AX DDR5| |RAM|G.SKILL Trident Z5 RGB 64GB ...
2025-08-18T22:05:01
https://www.reddit.com/r/LocalLLaMA/comments/1mu03r7/pc_to_run_local_llm_for_coding_agent/
BIackIight
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mu03r7
false
null
t3_1mu03r7
/r/LocalLLaMA/comments/1mu03r7/pc_to_run_local_llm_for_coding_agent/
false
false
self
2
{'enabled': False, 'images': [{'id': 'IAGFmaGszKcqSKR_8qg0oES6OBfFDCBNvzr72pbVe7o', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/IAGFmaGszKcqSKR_8qg0oES6OBfFDCBNvzr72pbVe7o.png?width=108&crop=smart&auto=webp&s=fbf0440b72bf3c599b24d782f0bddf00251537cf', 'width': 108}, {'height': 116, 'url': 'h...
Tiny finance “thinking” model (Gemma-3 270M) with verifiable rewards (SFT → GRPO) — structured outputs + auto-eval (with code)
61
I taught a tiny model to *think like a finance analyst* by enforcing a strict output contract and only rewarding it when the output is **verifiably** correct. # What I built * **Task & contract** (always returns): * `<REASONING>` concise, balanced rationale * `<SENTIMENT>` positive | negative | neutral * `<C...
2025-08-18T21:47:13
https://i.redd.it/db62qfi7mujf1.png
Solid_Woodpecker3635
i.redd.it
1970-01-01T00:00:00
0
{}
1mtzn4b
false
null
t3_1mtzn4b
/r/LocalLLaMA/comments/1mtzn4b/tiny_finance_thinking_model_gemma3_270m_with/
false
false
default
61
{'enabled': True, 'images': [{'id': 'db62qfi7mujf1', 'resolutions': [{'height': 72, 'url': 'https://preview.redd.it/db62qfi7mujf1.png?width=108&crop=smart&auto=webp&s=161ae2711e732dbabed4b1c62cfc53b3447a45ff', 'width': 108}, {'height': 144, 'url': 'https://preview.redd.it/db62qfi7mujf1.png?width=216&crop=smart&auto=web...
Looking to update an older GGUF model to work with newer llama cpp
4
So I am looking to see if I can update [NeverSleep/NoromaidxOpenGPT4-2-GGUF-iMatrix · Hugging Face](https://huggingface.co/NeverSleep/NoromaidxOpenGPT4-2-GGUF-iMatrix) to work with new llama cpp. I can't load the model in oobabooga anymore and I tried installing an older oobabooga (I think I tried 1.12) and still co...
2025-08-18T21:46:08
https://www.reddit.com/r/LocalLLaMA/comments/1mtzm47/looking_to_update_an_older_gguf_model_to_work/
Signal-Outcome-2481
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtzm47
false
null
t3_1mtzm47
/r/LocalLLaMA/comments/1mtzm47/looking_to_update_an_older_gguf_model_to_work/
false
false
self
4
{'enabled': False, 'images': [{'id': 'DmmegsFn3Y85fJ1kqSXHSocaxRsd_3qJyyP3c4b8sec', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/DmmegsFn3Y85fJ1kqSXHSocaxRsd_3qJyyP3c4b8sec.png?width=108&crop=smart&auto=webp&s=b8474f9b7ea4e42c18afd9f32d1adcbe2bdc8c09', 'width': 108}, {'height': 116, 'url': 'h...
AvatarNova - Local AI companion
0
Here is my AI assistant I am currently working on, its offline, has STT & TTS! (can listen and speak). you can also upload your document for it to process!
2025-08-18T21:31:43
https://www.reddit.com/r/LocalLLaMA/comments/1mtz8cq/avatarnova_local_ai_companion/
Yusso_17
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtz8cq
false
null
t3_1mtz8cq
/r/LocalLLaMA/comments/1mtz8cq/avatarnova_local_ai_companion/
false
false
self
0
null
Qwen3-Coder-30B-A3B MLX-DWQ variants like lr9e8
5
What is the lr5e-8 or lr9e-8 about in these MLX DWQ quants? Is there an explanation somewhere that I am missing? [https://huggingface.co/mlx-community/Qwen3-Coder-30B-A3B-Instruct-8bit-DWQ-lr5e-8](https://huggingface.co/mlx-community/Qwen3-Coder-30B-A3B-Instruct-8bit-DWQ-lr5e-8) [https://huggingface.co/mlx-community...
2025-08-18T21:25:49
https://www.reddit.com/r/LocalLLaMA/comments/1mtz2ry/qwen3coder30ba3b_mlxdwq_variants_like_lr9e8/
PracticlySpeaking
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtz2ry
false
null
t3_1mtz2ry
/r/LocalLLaMA/comments/1mtz2ry/qwen3coder30ba3b_mlxdwq_variants_like_lr9e8/
false
false
self
5
{'enabled': False, 'images': [{'id': '9D3T3qj0bgSWfPZliZvEzT7o3R-i77BjZDgDtDOLTnw', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/9D3T3qj0bgSWfPZliZvEzT7o3R-i77BjZDgDtDOLTnw.png?width=108&crop=smart&auto=webp&s=2c3783b005defbddcdfb7a0ef5be3e0a77841493', 'width': 108}, {'height': 116, 'url': 'h...
Tesla P4 owners, what kinds of speeds are you getting?
4
I’m curious what kind of real-world speeds people are seeing. I’m planning to run **Qwen 3 4B Instruct (2507) + Qwen 3** **Embeding 0.6B** for RAG on a home server, basically 24/7. My GPU options are either: * Tesla **P4** (not P40) * RTX **A1000** * …or just skipping the dGPU entirely and going with a **Ryzen 7 8700...
2025-08-18T21:12:18
https://www.reddit.com/r/LocalLLaMA/comments/1mtypuw/tesla_p4_owners_what_kinds_of_speeds_are_you/
SpyderJack
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtypuw
false
null
t3_1mtypuw
/r/LocalLLaMA/comments/1mtypuw/tesla_p4_owners_what_kinds_of_speeds_are_you/
false
false
self
4
null
Importing MI50 to US, tariffs or duty?
4
What is people's experiences importing these cards?
2025-08-18T20:56:21
https://www.reddit.com/r/LocalLLaMA/comments/1mtya9w/importing_mi50_to_us_tariffs_or_duty/
InsideYork
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtya9w
false
null
t3_1mtya9w
/r/LocalLLaMA/comments/1mtya9w/importing_mi50_to_us_tariffs_or_duty/
false
false
self
4
null
LLM Recomendations
2
Morning All, I have a very large collection of PDF and XML (S1000D XML) files close to 6500 in total. I am looking for a local LLM that I can interact with that could answer simple questions and sight references. The hardware is a Dell T5610 and a RTX 3060 12gb on Ubuntu So far i have tried OnPrem DocMind and Priva...
2025-08-18T20:56:08
https://www.reddit.com/r/LocalLLaMA/comments/1mtya1v/llm_recomendations/
Tango_Wiskey
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtya1v
false
null
t3_1mtya1v
/r/LocalLLaMA/comments/1mtya1v/llm_recomendations/
false
false
self
2
null
Best AI IDE subscription right now: Kiro Pro vs Cursor Pro?
0
Hey everyone, I’m working on a startup with a big project, and I mostly vibecode (let AI help me write/debug most of my code). I’ve used **Kiro** before, but the **rate limits** started getting in the way. Now I’m considering subscribing either to **Kiro Pro** or **Cursor Pro** (both $20/month). My main concerns are:...
2025-08-18T20:52:52
https://www.reddit.com/r/LocalLLaMA/comments/1mty6u7/best_ai_ide_subscription_right_now_kiro_pro_vs/
PARNSTARSKAN
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mty6u7
false
null
t3_1mty6u7
/r/LocalLLaMA/comments/1mty6u7/best_ai_ide_subscription_right_now_kiro_pro_vs/
false
false
self
0
null
Best models for journalistic finetunes
3
I am looking for an instructable model that has good English writing capabilities (non-creative) and also takes well to finetuning. I will train it on a corpus of written articles so that it can learn to emulate the specific style, tone, and structure well. It will then be used to produce and review these articles. So...
2025-08-18T20:50:03
https://www.reddit.com/r/LocalLLaMA/comments/1mty41i/best_models_for_journalistic_finetunes/
syzygyhack
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mty41i
false
null
t3_1mty41i
/r/LocalLLaMA/comments/1mty41i/best_models_for_journalistic_finetunes/
false
false
self
3
null
Should the convention be to include Activated params in the names of moe models?
9
Eg [https://huggingface.co/Qwen/Qwen3-Coder-480B-A35B-Instruct](https://huggingface.co/Qwen/Qwen3-Coder-480B-A35B-Instruct) I think the name should either be Qwen3-Coder or the long form, Qwen3-Coder-480B-A35B-Instruct The short form makes some sense (I suppose), but for those who care about size, I think the ...
2025-08-18T20:46:23
https://www.reddit.com/r/LocalLLaMA/comments/1mty0k2/should_the_convention_be_to_include_activated/
kaggleqrdl
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mty0k2
false
null
t3_1mty0k2
/r/LocalLLaMA/comments/1mty0k2/should_the_convention_be_to_include_activated/
false
false
self
9
{'enabled': False, 'images': [{'id': 'SU4EkoBE9zB_i4T28BH-B8NRspWSu8pgjF1RIMOo6CQ', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/SU4EkoBE9zB_i4T28BH-B8NRspWSu8pgjF1RIMOo6CQ.png?width=108&crop=smart&auto=webp&s=d107a6b6b4389cb37d48d7ce4ff4d5aa35e4d93a', 'width': 108}, {'height': 116, 'url': 'h...
Offline coding assistant running local models
2
So I tried building an android application with the concept being:- acts as a git client(can perform add,commit,push) and it would index repos such that it would be injected in context of a local model and accordingly could perform operations like:- replace all for loops with while loops in X function of Y class. Didnt...
2025-08-18T20:13:40
https://www.reddit.com/r/LocalLLaMA/comments/1mtx4lo/offline_coding_assistant_running_local_models/
SnooCupcakes5746
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtx4lo
false
null
t3_1mtx4lo
/r/LocalLLaMA/comments/1mtx4lo/offline_coding_assistant_running_local_models/
false
false
self
2
{'enabled': False, 'images': [{'id': 'UTuq5EADIAw-Vqffn5nrcZCluHp4TQ9IB1wQcBb0zSE', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/UTuq5EADIAw-Vqffn5nrcZCluHp4TQ9IB1wQcBb0zSE.png?width=108&crop=smart&auto=webp&s=6dce84adf226e4280860267cbe170ed388af7756', 'width': 108}, {'height': 108, 'url': 'h...
GLM 4 sits at 4th on Design Arena over the last 7 days
42
Have been looking more deeply into the analytics on the preference data for the [benchmark](https://www.designarena.ai/) over the last week, and I thought this might be an interesting tidbit to share. GLM 4.5, filtering for comparisons that have been submitted over the last 7 days, is 4th among models based on wi...
2025-08-18T20:08:49
https://www.reddit.com/gallery/1mtwzx2
Accomplished-Copy332
reddit.com
1970-01-01T00:00:00
0
{}
1mtwzx2
false
null
t3_1mtwzx2
/r/LocalLLaMA/comments/1mtwzx2/glm_4_sits_at_4th_on_design_arena_over_the_last_7/
false
false
https://b.thumbs.redditm…M8wShxJR03JE.jpg
42
null
My Experience Comparing Gemma 3 27B and GPT-OSS 20B: A Clear Winner for My Use Case
1
With the recent buzz around new open-source models, I decided to run my own tests on two of the latest releases: OpenAI's gpt-oss-20b and Google's Gemma 3 27b. I've seen a lot of hype around gpt-oss, so I was eager to see how it stacks up. After a few days of testing, my experience has been surprisingly one-sided. He...
2025-08-18T20:06:57
https://www.reddit.com/r/LocalLLaMA/comments/1mtwy39/my_experience_comparing_gemma_3_27b_and_gptoss/
comunication
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtwy39
false
null
t3_1mtwy39
/r/LocalLLaMA/comments/1mtwy39/my_experience_comparing_gemma_3_27b_and_gptoss/
false
false
self
1
null
I got it...
1
[removed]
2025-08-18T19:38:47
[deleted]
1970-01-01T00:00:00
0
{}
1mtw6kx
false
null
t3_1mtw6kx
/r/LocalLLaMA/comments/1mtw6kx/i_got_it/
false
false
default
1
null
Local LLM
1
Im trying to use a local LLM (like Llama, Mistral, or any other open-source model) as a sort of “search engine” or assistant to help me navigate a large collection of technical norms, standards, and guidelines — mostly in PDF format. The goal is to be able to ask natural-language questions and get accurate answers or r...
2025-08-18T19:27:39
https://www.reddit.com/r/LocalLLaMA/comments/1mtvw1o/local_llm/
LAWNCOWER
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtvw1o
false
null
t3_1mtvw1o
/r/LocalLLaMA/comments/1mtvw1o/local_llm/
false
false
self
1
null
NVIDIA Nemotron Nano 2 and the Nemotron Pretraining Dataset v1
44
2025-08-18T19:12:27
https://research.nvidia.com/labs/adlr/NVIDIA-Nemotron-Nano-2/
pi314156
research.nvidia.com
1970-01-01T00:00:00
0
{}
1mtvh0j
false
null
t3_1mtvh0j
/r/LocalLLaMA/comments/1mtvh0j/nvidia_nemotron_nano_2_and_the_nemotron/
false
false
default
44
null
NVIDIA Releases Nemotron Nano 2 AI Models
615
• 6X faster than similarly sized models, while also being more accurate • NVIDIA is also releasing most of the data they used to create it, including the pretraining corpus • The hybrid Mamba-Transformer architecture supports 128K context length on single GPU. Full research paper here: https://research.nvidia.com/la...
2025-08-18T19:12:01
https://i.redd.it/pzrpnuykutjf1.jpeg
vibedonnie
i.redd.it
1970-01-01T00:00:00
0
{}
1mtvgjx
false
null
t3_1mtvgjx
/r/LocalLLaMA/comments/1mtvgjx/nvidia_releases_nemotron_nano_2_ai_models/
false
false
default
615
{'enabled': True, 'images': [{'id': 'pzrpnuykutjf1', 'resolutions': [{'height': 108, 'url': 'https://preview.redd.it/pzrpnuykutjf1.jpeg?width=108&crop=smart&auto=webp&s=508119862660878690b05283f406757ef7af3232', 'width': 108}, {'height': 216, 'url': 'https://preview.redd.it/pzrpnuykutjf1.jpeg?width=216&crop=smart&auto=...
How to use Generative AI productively?
0
I'm about to go into college, I've never had AI. I just have an idea that this could be really helpful for me somehow, but I can't figure out how. I don't just want to use it to get answers, as that wouldn't really help me in the long run. I could use it to make study guides or help me learn (I'm going into biochem). B...
2025-08-18T18:57:41
https://www.reddit.com/r/LocalLLaMA/comments/1mtv25r/how_to_use_generative_ai_productively/
Acklord303
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtv25r
false
null
t3_1mtv25r
/r/LocalLLaMA/comments/1mtv25r/how_to_use_generative_ai_productively/
false
false
self
0
null
Local AI workstation/Server was it worth for you guys?
15
Hi guys, I am curious do you regret building your costly AI workstations/Servers and what are you suggestions at current market as I am checking posts about builds and tokens for last 3 months. I managed to get one 5090 for MRSP (on trip in USA) but I don't have cash to build PC around it yet (Europe). I was thinkin...
2025-08-18T18:57:18
https://www.reddit.com/r/LocalLLaMA/comments/1mtv1rr/local_ai_workstationserver_was_it_worth_for_you/
SomeRandomGuuuuuuy
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtv1rr
false
null
t3_1mtv1rr
/r/LocalLLaMA/comments/1mtv1rr/local_ai_workstationserver_was_it_worth_for_you/
false
false
self
15
null
I don't know where to ask since cloude blocks me. Should I ask for refund/(will i get it?)
0
2025-08-18T18:50:51
https://i.redd.it/0j7yv44oqtjf1.png
FluffyMacho
i.redd.it
1970-01-01T00:00:00
0
{}
1mtuvbl
false
null
t3_1mtuvbl
/r/LocalLLaMA/comments/1mtuvbl/i_dont_know_where_to_ask_since_cloude_blocks_me/
false
false
default
0
{'enabled': True, 'images': [{'id': '0j7yv44oqtjf1', 'resolutions': [{'height': 37, 'url': 'https://preview.redd.it/0j7yv44oqtjf1.png?width=108&crop=smart&auto=webp&s=730975361bf5660e4b7003f79b2df462db1e6b9d', 'width': 108}, {'height': 74, 'url': 'https://preview.redd.it/0j7yv44oqtjf1.png?width=216&crop=smart&auto=webp...
Qwen3 and Qwen2.5 VL built from scratch.
65
Over the weekend, I updated my Tiny-Qwen repo to support Qwen3, with MoE support, and wrapped it in a fancy looking CLI for ease of use. Go check it out: [https://github.com/Emericen/tiny-qwen](https://github.com/Emericen/tiny-qwen)
2025-08-18T18:42:14
https://i.redd.it/q9jdx2funtjf1.jpeg
No-Compote-6794
i.redd.it
1970-01-01T00:00:00
0
{}
1mtumrq
false
null
t3_1mtumrq
/r/LocalLLaMA/comments/1mtumrq/qwen3_and_qwen25_vl_built_from_scratch/
false
false
default
65
{'enabled': True, 'images': [{'id': 'q9jdx2funtjf1', 'resolutions': [{'height': 66, 'url': 'https://preview.redd.it/q9jdx2funtjf1.jpeg?width=108&crop=smart&auto=webp&s=f85c3e69b5c8348a88569caa3b8dcec09c0d408e', 'width': 108}, {'height': 132, 'url': 'https://preview.redd.it/q9jdx2funtjf1.jpeg?width=216&crop=smart&auto=w...
Where to deploy FastAPI ,Rag application
0
I want to deploy my fastapi rag application but i have used up all my railway credits and on render i get memory ran out error .I do have azure student plan but it just too slow and has cold start and too much of problem i have been trying zerops and i strugging even to write zerops.yml file and there aren't much tutor...
2025-08-18T18:09:34
https://www.reddit.com/r/LocalLLaMA/comments/1mttq29/where_to_deploy_fastapi_rag_application/
Minimum-Row6464
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mttq29
false
null
t3_1mttq29
/r/LocalLLaMA/comments/1mttq29/where_to_deploy_fastapi_rag_application/
false
false
self
0
null
Qwen-Image-Edit Released!
409
Alibaba’s Qwen team just released **Qwen-Image-Edit**, an image editing model built on the **20B Qwen-Image** backbone. [https://huggingface.co/Qwen/Qwen-Image-Edit](https://huggingface.co/Qwen/Qwen-Image-Edit) It supports **precise bilingual (Chinese & English) text editing** while preserving style, plus both...
2025-08-18T18:00:24
https://www.reddit.com/r/LocalLLaMA/comments/1mttgrf/qwenimageedit_released/
MohamedTrfhgx
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mttgrf
false
null
t3_1mttgrf
/r/LocalLLaMA/comments/1mttgrf/qwenimageedit_released/
false
false
self
409
{'enabled': False, 'images': [{'id': '_phhztHOXP1EaSigwjpmdclnYlgiIY_nMfXGtHApDV0', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/_phhztHOXP1EaSigwjpmdclnYlgiIY_nMfXGtHApDV0.png?width=108&crop=smart&auto=webp&s=46b31a63c054056a199d9db162a18c1eafc7cc2b', 'width': 108}, {'height': 116, 'url': 'h...
Qwen-Image-Edit Released!
1
[removed]
2025-08-18T17:57:48
https://www.reddit.com/r/LocalLLaMA/comments/1mtte72/qwenimageedit_released/
MohamedTrfhgx
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtte72
false
null
t3_1mtte72
/r/LocalLLaMA/comments/1mtte72/qwenimageedit_released/
false
false
self
1
null
🚀 Qwen released Qwen-Image-Edit!
1,017
🚀 Excited to introduce Qwen-Image-Edit! Built on 20B Qwen-Image, it brings precise bilingual text editing (Chinese & English) while preserving style, and supports both semantic and appearance-level editing. ✨ Key Features ✅ Accurate text editing with bilingual support ✅ High-level semantic editing (e.g. object rota...
2025-08-18T17:56:23
https://www.reddit.com/gallery/1mttcr9
ResearchCrafty1804
reddit.com
1970-01-01T00:00:00
0
{}
1mttcr9
false
null
t3_1mttcr9
/r/LocalLLaMA/comments/1mttcr9/qwen_released_qwenimageedit/
false
false
https://b.thumbs.redditm…5WjEjTkW73hM.jpg
1,017
null
Deepseek R2 coming out ... when it gets more cowbell
135
From what’s floating around it seems like we'll have to keep waiting a bit longer for Deepseek R2 to be released. Apparently 1. Liang Wenfeng has been sitting on R2's release because it still needs more cowbell 2. Training DeepSeek R2 on Huawei Ascend chips ran into persistent stability and software problems and no ...
2025-08-18T17:56:08
https://www.reddit.com/r/LocalLLaMA/comments/1mttchz/deepseek_r2_coming_out_when_it_gets_more_cowbell/
1BlueSpork
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mttchz
false
{'oembed': {'author_name': 'BlueSpork', 'author_url': 'https://www.youtube.com/@BlueSpork', 'height': 200, 'html': '<iframe width="356" height="200" src="https://www.youtube.com/embed/PzlqRsuIo1w?feature=oembed&enablejsapi=1" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; p...
t3_1mttchz
/r/LocalLLaMA/comments/1mttchz/deepseek_r2_coming_out_when_it_gets_more_cowbell/
false
false
self
135
{'enabled': False, 'images': [{'id': 'JNAOp8mejhheiBappEWRGE1kVdbTWLMVP5NLPsOJx9c', 'resolutions': [{'height': 81, 'url': 'https://external-preview.redd.it/JNAOp8mejhheiBappEWRGE1kVdbTWLMVP5NLPsOJx9c.jpeg?width=108&crop=smart&auto=webp&s=81b924dd131bc9f786b249545a1aa67db7249e6b', 'width': 108}, {'height': 162, 'url': '...
Introducing ollama2llama + ollama-file-find: a migration path from Ollama to llama-swap (and why I wrote them after the license attribution issue)
1
[removed]
2025-08-18T17:55:49
https://www.reddit.com/r/LocalLLaMA/comments/1mttc70/introducing_ollama2llama_ollamafilefind_a/
emetah850
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mttc70
false
null
t3_1mttc70
/r/LocalLLaMA/comments/1mttc70/introducing_ollama2llama_ollamafilefind_a/
false
false
self
1
null
Qwen-Image-Edit
95
2025-08-18T17:54:01
https://huggingface.co/Qwen/Qwen-Image-Edit
lomero
huggingface.co
1970-01-01T00:00:00
0
{}
1mttad3
false
null
t3_1mttad3
/r/LocalLLaMA/comments/1mttad3/qwenimageedit/
false
false
default
95
{'enabled': False, 'images': [{'id': '_phhztHOXP1EaSigwjpmdclnYlgiIY_nMfXGtHApDV0', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/_phhztHOXP1EaSigwjpmdclnYlgiIY_nMfXGtHApDV0.png?width=108&crop=smart&auto=webp&s=46b31a63c054056a199d9db162a18c1eafc7cc2b', 'width': 108}, {'height': 116, 'url': 'h...
Would this offline AI stick feel too slow for client use?
1
I’m building a small device (external-SSD sized) that plugs into a laptop and lets you chat with sensitive docs like contracts, tax returns, client files, etc....fully offline with nothing going to the cloud. Think Dropbox for the paranoid. On my prototype it takes about 5–10 seconds to answer a question using a 7B m...
2025-08-18T17:50:30
https://www.reddit.com/r/LocalLLaMA/comments/1mtt6vc/would_this_offline_ai_stick_feel_too_slow_for/
Automatic-Reading952
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtt6vc
false
null
t3_1mtt6vc
/r/LocalLLaMA/comments/1mtt6vc/would_this_offline_ai_stick_feel_too_slow_for/
false
false
self
1
null
Qwen/Qwen-Image-Edit · Hugging Face
5
2025-08-18T17:49:20
https://huggingface.co/Qwen/Qwen-Image-Edit
Dark_Fire_12
huggingface.co
1970-01-01T00:00:00
0
{}
1mtt5pd
false
null
t3_1mtt5pd
/r/LocalLLaMA/comments/1mtt5pd/qwenqwenimageedit_hugging_face/
false
false
default
5
null
how do i use llama-cpp-python with cuda
0
i just can't find anything that works. yes, i use n\_gpu\_layers, but every attempt keeps the gpu at 1%. it's closer to hit 0% usage than anything above 2%. my cpu always spikes to 100% when it's generating a response.
2025-08-18T17:33:59
https://www.reddit.com/r/LocalLLaMA/comments/1mtsqb6/how_do_i_use_llamacpppython_with_cuda/
miguel-1510
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtsqb6
false
null
t3_1mtsqb6
/r/LocalLLaMA/comments/1mtsqb6/how_do_i_use_llamacpppython_with_cuda/
false
false
self
0
null
Introducing ollama2llama + ollama-file-find: a migration path from Ollama to llama-swap (and why I wrote them after the license attribution issue)
1
[removed]
2025-08-18T17:21:33
https://www.reddit.com/r/LocalLLaMA/comments/1mtsduf/introducing_ollama2llama_ollamafilefind_a/
emetah850
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtsduf
false
null
t3_1mtsduf
/r/LocalLLaMA/comments/1mtsduf/introducing_ollama2llama_ollamafilefind_a/
false
false
self
1
null
I wish we could actually use Gemma 3n
47
It would be the perfect model to run on my Nvidia Jetson Orin Nano, with native audio and video support. It's a great model that punches way above its weight and I like the general vibe a lot. And most importantly, it has some really cool optimizations for low-memory edge devices such as per-layer embeddings that can b...
2025-08-18T17:19:27
https://www.reddit.com/r/LocalLLaMA/comments/1mtsbsk/i_wish_we_could_actually_use_gemma_3n/
ai_fonsi
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtsbsk
false
null
t3_1mtsbsk
/r/LocalLLaMA/comments/1mtsbsk/i_wish_we_could_actually_use_gemma_3n/
false
false
self
47
{'enabled': False, 'images': [{'id': 'P6Q6Ym8v6ZH9q0qVFvshN3LU_4Afza8wsv7uwNwu3FE', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/P6Q6Ym8v6ZH9q0qVFvshN3LU_4Afza8wsv7uwNwu3FE.png?width=108&crop=smart&auto=webp&s=d2069fa0a908adf586945801566a667b94609529', 'width': 108}, {'height': 108, 'url': 'h...
I wish we could actually use Gemma 3n
1
[removed]
2025-08-18T17:16:22
https://www.reddit.com/r/LocalLLaMA/comments/1mts8m3/i_wish_we_could_actually_use_gemma_3n/
ai_fonsi
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mts8m3
false
null
t3_1mts8m3
/r/LocalLLaMA/comments/1mts8m3/i_wish_we_could_actually_use_gemma_3n/
false
false
self
1
null
🚀 I built doxx-go: Terminal DOCX viewer with LOCAL AI Vision Analysis using Ollama/LMStudio
0
Hey r/LocalLLaMA! I wanted to share a tool I've been working on that creates a practical use case for local vision models in everyday productivity. TL;DR: doxx-go is a terminal-native Microsoft Word document viewer that now includes AI-powered image analysis using your local Ollama or LMStudio models. Extract and analy...
2025-08-18T17:07:27
https://www.reddit.com/r/LocalLLaMA/comments/1mtrzi0/i_built_doxxgo_terminal_docx_viewer_with_local_ai/
kodOZANI
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtrzi0
false
null
t3_1mtrzi0
/r/LocalLLaMA/comments/1mtrzi0/i_built_doxxgo_terminal_docx_viewer_with_local_ai/
false
false
self
0
{'enabled': False, 'images': [{'id': 'HtBEToQsSt91S-c1Dt8mZfOPRR_7P0CMYnUcqk0wdxc', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/HtBEToQsSt91S-c1Dt8mZfOPRR_7P0CMYnUcqk0wdxc.png?width=108&crop=smart&auto=webp&s=6138db741239d8de0b637ebb0c2bd0f08cb53254', 'width': 108}, {'height': 108, 'url': 'h...
Drummer's Cydonia 24B v4.1 - Nothing like its predecessors. A stronger, less positive, less Mistral, performant tune!
143
2025-08-18T16:55:07
https://huggingface.co/TheDrummer/Cydonia-24B-v4.1
TheLocalDrummer
huggingface.co
1970-01-01T00:00:00
0
{}
1mtrn1y
false
null
t3_1mtrn1y
/r/LocalLLaMA/comments/1mtrn1y/drummers_cydonia_24b_v41_nothing_like_its/
false
false
default
143
{'enabled': False, 'images': [{'id': 'S-o7drNLzmVPRPUM6Ap2mbj65SEf6wzf8867vZcT5JE', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/S-o7drNLzmVPRPUM6Ap2mbj65SEf6wzf8867vZcT5JE.png?width=108&crop=smart&auto=webp&s=ca6e76ff423238bd1f6fdc76c856fd3c97b9f324', 'width': 108}, {'height': 116, 'url': 'h...
Applying an LLM method (Magpie) to an LLM-based TTS model for a synthetic speech dataset
3
Hi everyone! I wrote an article on how you can use LLM data synthesis techniques for LLM-based TTS models. The core idea is that you can treat the `text tokens -> audio tokens` generation in a model like Orpheus-TTS just like the `instruction -> response` part of a standard LLM. I tried this out with the [Magpie](htt...
2025-08-18T16:52:33
https://www.reddit.com/r/LocalLLaMA/comments/1mtrkhr/applying_an_llm_method_magpie_to_an_llmbased_tts/
Aratako_LM
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtrkhr
false
null
t3_1mtrkhr
/r/LocalLLaMA/comments/1mtrkhr/applying_an_llm_method_magpie_to_an_llmbased_tts/
false
false
self
3
{'enabled': False, 'images': [{'id': 'TM2lo-hNii5Z2u3aT9e1KlZVGqvksxSLYfXsgEsbk50', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/TM2lo-hNii5Z2u3aT9e1KlZVGqvksxSLYfXsgEsbk50.png?width=108&crop=smart&auto=webp&s=2c1104e29c242ce5331c849a9b1ce9c4a150f989', 'width': 108}, {'height': 108, 'url': 'h...
Is anyone using smolagent in production?
0
I’ve been exploring smolagent and I’m really impressed with how simple it is. As the website highlights: *Simplicity: the logic for agents fits in \~1,000 lines of code.* That said, while digging deeper I ran into what feels like the biggest drawback: when executing Python scripts, the agent could potentially run mali...
2025-08-18T16:42:05
https://www.reddit.com/r/LocalLLaMA/comments/1mtra91/is_anyone_using_smolagent_in_production/
Prashant-Lakhera
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtra91
false
null
t3_1mtra91
/r/LocalLLaMA/comments/1mtra91/is_anyone_using_smolagent_in_production/
false
false
self
0
null
Anyone have the deets on ROCM 7.0's 3x perf claims?
2
The footnotes for these claims are below, but the summary is they are for an 8x MI300x system running vLLM / Megatron-LM. Anyone know specifically where the new performance is coming from? [https://www.amd.com/en/products/software/rocm/whats-new.html](https://www.amd.com/en/products/software/rocm/whats-new.html) http...
2025-08-18T16:39:36
https://www.reddit.com/r/LocalLLaMA/comments/1mtr7s3/anyone_have_the_deets_on_rocm_70s_3x_perf_claims/
1ncehost
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtr7s3
false
null
t3_1mtr7s3
/r/LocalLLaMA/comments/1mtr7s3/anyone_have_the_deets_on_rocm_70s_3x_perf_claims/
false
false
https://b.thumbs.redditm…zAhnBSos4yCU.jpg
2
{'enabled': False, 'images': [{'id': 'IWefK3VqOnSNhm69HOKScyrAzkbcdWLsfQonThHVaHg', 'resolutions': [{'height': 60, 'url': 'https://external-preview.redd.it/IWefK3VqOnSNhm69HOKScyrAzkbcdWLsfQonThHVaHg.jpeg?width=108&crop=smart&auto=webp&s=df69a3e4568549d552946de107ac3b57315da9ab', 'width': 108}, {'height': 121, 'url': '...
Test: can Qwen 2.5 Omni actually hear guitar chords?
126
Tried Qwen 2.5 Omni locally with vision + speech to see if it could hear music instead of just speech. I played guitar, and it did a surprisingly solid job telling me which chords I was playing in real-time. At the end, I debugged what the LLM was “hearing,” and input quality likely explained some of the misses it did...
2025-08-18T16:30:44
https://v.redd.it/848676nyzsjf1
Weary-Wing-6806
/r/LocalLLaMA/comments/1mtqz3u/test_can_qwen_25_omni_actually_hear_guitar_chords/
1970-01-01T00:00:00
0
{}
1mtqz3u
false
{'reddit_video': {'bitrate_kbps': 5000, 'dash_url': 'https://v.redd.it/848676nyzsjf1/DASHPlaylist.mpd?a=1758256252%2CMjVjMTJjZjQyZDYwOGYzNzZjZThlNzgyNjAxYTdkN2E1YTY0ZjU0Yzk3ZThhMGRkMzkxNWVmODAwNGYwM2ZhOQ%3D%3D&v=1&f=sd', 'duration': 83, 'fallback_url': 'https://v.redd.it/848676nyzsjf1/DASH_1080.mp4?source=fallback', 'h...
t3_1mtqz3u
/r/LocalLLaMA/comments/1mtqz3u/test_can_qwen_25_omni_actually_hear_guitar_chords/
false
false
https://external-preview…eda2053099b345b6
126
{'enabled': False, 'images': [{'id': 'ZHl5dWUwbXl6c2pmMYMQwk04cVObnLQ8K0JceAGBOHlOc1BqIEw787XqJHGO', 'resolutions': [{'height': 60, 'url': 'https://external-preview.redd.it/ZHl5dWUwbXl6c2pmMYMQwk04cVObnLQ8K0JceAGBOHlOc1BqIEw787XqJHGO.png?width=108&crop=smart&format=pjpg&auto=webp&s=9c869e44be64b9782262ee5666863ea540fe6...
kimi-vl-a3b-thinking-2506 launched today on LM Studio
27
This is the compressed version of Kiwi . It works on my 8GB video card. It also has CPU offloading and it's real quick!
2025-08-18T16:19:30
https://www.reddit.com/r/LocalLLaMA/comments/1mtqnwe/kimivla3bthinking2506_launched_today_on_lm_studio/
PhotographerUSA
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtqnwe
false
null
t3_1mtqnwe
/r/LocalLLaMA/comments/1mtqnwe/kimivla3bthinking2506_launched_today_on_lm_studio/
false
false
self
27
null
Claude Code Just Got Way Better - Output Styles
0
https://preview.redd.it/….be/JP4BL6qJ6lM)
2025-08-18T16:17:04
https://www.reddit.com/r/LocalLLaMA/comments/1mtqlek/claude_code_just_got_way_better_output_styles/
bipin_25
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtqlek
false
{'oembed': {'author_name': 'Codedigipt', 'author_url': 'https://www.youtube.com/@codedigiptbiplab', 'height': 200, 'html': '<iframe width="356" height="200" src="https://www.youtube.com/embed/JP4BL6qJ6lM?feature=oembed&enablejsapi=1" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyro...
t3_1mtqlek
/r/LocalLLaMA/comments/1mtqlek/claude_code_just_got_way_better_output_styles/
false
false
https://external-preview…2d360e3a590704c1
0
{'enabled': False, 'images': [{'id': 'OQHxGy6hgYoAhKokGyHHitRFwsRrup3-dDcBH7k4XOU', 'resolutions': [{'height': 81, 'url': 'https://external-preview.redd.it/OQHxGy6hgYoAhKokGyHHitRFwsRrup3-dDcBH7k4XOU.jpeg?width=108&crop=smart&auto=webp&s=2a2dc257d73351143acd2699713fe4bd9e363238', 'width': 108}, {'height': 162, 'url': '...
guide : running gpt-oss with llama.cpp
37
2025-08-18T16:09:51
https://github.com/ggml-org/llama.cpp/discussions/15396
jacek2023
github.com
1970-01-01T00:00:00
0
{}
1mtqdy8
false
null
t3_1mtqdy8
/r/LocalLLaMA/comments/1mtqdy8/guide_running_gptoss_with_llamacpp/
false
false
https://external-preview…3ec68b04aab99607
37
{'enabled': False, 'images': [{'id': 'b2JdP9WbhqiSxugjTjM2AEgw1c-M7Kpl8_FDa-Zbacc', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/b2JdP9WbhqiSxugjTjM2AEgw1c-M7Kpl8_FDa-Zbacc.png?width=108&crop=smart&auto=webp&s=76a543c0d9c25d9170eec008d25bea4ef0c8991e', 'width': 108}, {'height': 108, 'url': 'h...
Looking for a decent reranking model supported by llama.cpp
1
[removed]
2025-08-18T15:50:30
https://www.reddit.com/r/LocalLLaMA/comments/1mtpumq/looking_for_a_decent_reranking_model_supported_by/
AttentionIsAllYouGet
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtpumq
false
null
t3_1mtpumq
/r/LocalLLaMA/comments/1mtpumq/looking_for_a_decent_reranking_model_supported_by/
false
false
self
1
null
current best model for translating subtitles from chinese to English
0
Hi guys, I was using Deepseek a few months ago to translate Chinese-English for some of my TV shows and I found them quite decent. It's been a few months and AI is rapidly advancing so I'm wondering if there are better models currently to perform translations. I've come across some posts about people fine tuning the mo...
2025-08-18T15:24:46
https://www.reddit.com/r/LocalLLaMA/comments/1mtp50s/current_best_model_for_translating_subtitles_from/
cryptofanatic96
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtp50s
false
null
t3_1mtp50s
/r/LocalLLaMA/comments/1mtp50s/current_best_model_for_translating_subtitles_from/
false
false
self
0
null
Where to find NVIDIA Blackwell RTX Pro 6000 96Gb MSRP?
1
[removed]
2025-08-18T15:24:46
https://www.reddit.com/r/LocalLLaMA/comments/1mtp50f/where_to_find_nvidia_blackwell_rtx_pro_6000_96gb/
Admirable-Ad-403
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtp50f
false
null
t3_1mtp50f
/r/LocalLLaMA/comments/1mtp50f/where_to_find_nvidia_blackwell_rtx_pro_6000_96gb/
false
false
self
1
null
Using a local LLM AI agent to solve the N puzzle
2
Hi everyone, I have just made some program to make an AI agent solve the N puzzle. Github link: [https://github.com/dangmanhtruong1995/N-puzzle-Agent/tree/main](https://github.com/dangmanhtruong1995/N-puzzle-Agent/tree/main) Youtube link: [https://www.youtube.com/watch?v=Ntol4F4tilg](https://www.youtube.com/watch?v=N...
2025-08-18T15:24:35
https://www.reddit.com/r/LocalLLaMA/comments/1mtp4u9/using_a_local_llm_ai_agent_to_solve_the_n_puzzle/
CommunityOpposite645
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtp4u9
false
null
t3_1mtp4u9
/r/LocalLLaMA/comments/1mtp4u9/using_a_local_llm_ai_agent_to_solve_the_n_puzzle/
false
false
self
2
null
Fully Open source, serverless, community-driven MCP alternative built in Python, TS and Go
7
2025-08-18T15:07:41
http://github.com/universal-tool-calling-protocol/
juanviera23
github.com
1970-01-01T00:00:00
0
{}
1mtoo92
false
null
t3_1mtoo92
/r/LocalLLaMA/comments/1mtoo92/fully_open_source_serverless_communitydriven_mcp/
false
false
default
7
{'enabled': False, 'images': [{'id': 'utGmW0gH0yHk0tI86NCp09MwYJaWuJH4solGKRYcJNU', 'resolutions': [{'height': 108, 'url': 'https://external-preview.redd.it/utGmW0gH0yHk0tI86NCp09MwYJaWuJH4solGKRYcJNU.png?width=108&crop=smart&auto=webp&s=24407be52fdf155d771e06f8bc413cc200ae7a28', 'width': 108}, {'height': 216, 'url': '...
How do you compare different models?
1
[removed]
2025-08-18T15:03:47
https://www.reddit.com/r/LocalLLaMA/comments/1mtokdt/how_do_you_compare_different_models/
Moist_Soup_231
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtokdt
false
null
t3_1mtokdt
/r/LocalLLaMA/comments/1mtokdt/how_do_you_compare_different_models/
false
false
self
1
null
I wired Ollama to 1,200 CLI & HTTP tools via UTCP - repo + tools inside
0
2025-08-18T15:01:15
https://www.reddit.com/gallery/1mtohpk
juanviera23
reddit.com
1970-01-01T00:00:00
0
{}
1mtohpk
false
null
t3_1mtohpk
/r/LocalLLaMA/comments/1mtohpk/i_wired_ollama_to_1200_cli_http_tools_via_utcp/
false
false
https://b.thumbs.redditm…y-NtfWFWcjns.jpg
0
null
Multiverse Computing (🇪🇸) shrinks AI models by 95% without losing power -- for realz?
1
[removed]
2025-08-18T15:00:03
https://www.reddit.com/r/LocalLLaMA/comments/1mtogeq/multiverse_computing_shrinks_ai_models_by_95/
Master_Ad3271
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtogeq
false
null
t3_1mtogeq
/r/LocalLLaMA/comments/1mtogeq/multiverse_computing_shrinks_ai_models_by_95/
false
false
self
1
null
Extracting Key-Value Pairs from OCR Output - Any Recommendations?
1
[removed]
2025-08-18T14:55:14
[deleted]
1970-01-01T00:00:00
0
{}
1mtobrf
false
null
t3_1mtobrf
/r/LocalLLaMA/comments/1mtobrf/extracting_keyvalue_pairs_from_ocr_output_any/
false
false
default
1
null
New code benchmark puts Qwen 3 Coder at the top of the open models
323
TLDR of the open models results: Q3C fp16 > Q3C fp8 > GPT-OSS-120b > V3 > K2
2025-08-18T14:51:49
https://brokk.ai/power-ranking?round=open&models=flash-2.5%2Cgpt-oss-120b%2Cgpt5-mini%2Ck2%2Cq3c%2Cq3c-fp8%2Cv3
mr_riptano
brokk.ai
1970-01-01T00:00:00
0
{}
1mto8fa
false
null
t3_1mto8fa
/r/LocalLLaMA/comments/1mto8fa/new_code_benchmark_puts_qwen_3_coder_at_the_top/
false
false
default
323
{'enabled': False, 'images': [{'id': '-FhGljRcqsXlvJ4R58hFA0RMnpSa9fBguxJ8Dc9Mg-4', 'resolutions': [{'height': 67, 'url': 'https://external-preview.redd.it/-FhGljRcqsXlvJ4R58hFA0RMnpSa9fBguxJ8Dc9Mg-4.png?width=108&crop=smart&auto=webp&s=93217c80616370cc57bd63045f94b0f0c35d5216', 'width': 108}, {'height': 135, 'url': 'h...
We open-sourced Memori: A memory engine for AI agents
82
Hey folks! I'm a part the team behind [Memori](https://memori.gibsonai.com/). Memori adds a stateful memory engine to AI agents, enabling them to stay consistent, recall past work, and improve over time. With Memori, agents don’t lose track of multi-step workflows, repeat tool calls, or forget user preferences. I...
2025-08-18T14:51:39
https://www.reddit.com/r/LocalLLaMA/comments/1mto88l/we_opensourced_memori_a_memory_engine_for_ai/
Arindam_200
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mto88l
false
null
t3_1mto88l
/r/LocalLLaMA/comments/1mto88l/we_opensourced_memori_a_memory_engine_for_ai/
false
false
self
82
{'enabled': False, 'images': [{'id': 'W4oA3wrIFXUefXWMIwryTBQ2UxrNqHsC4cCYcDPnEbE', 'resolutions': [{'height': 75, 'url': 'https://external-preview.redd.it/W4oA3wrIFXUefXWMIwryTBQ2UxrNqHsC4cCYcDPnEbE.png?width=108&crop=smart&auto=webp&s=db4d31e01959cb4027b3aea72cf9338588d41e3e', 'width': 108}, {'height': 150, 'url': 'h...
OSS20B is actually good? (Part 2)
1
I did more testing on OSS20B and worked on my benchmarks and used it more, and it's a weird model to me. My previous bench was heavy in a task that OSS does really well that is related to information retrival in structured text, something I use a lot, e.g. when creating NPC D&D character sheet or I'm looking through c...
2025-08-18T14:50:51
https://www.reddit.com/gallery/1mto7gc
05032-MendicantBias
reddit.com
1970-01-01T00:00:00
0
{}
1mto7gc
false
null
t3_1mto7gc
/r/LocalLLaMA/comments/1mto7gc/oss20b_is_actually_good_part_2/
false
false
https://b.thumbs.redditm…4pGckNosujTI.jpg
1
null
Extracting Key-Value Pairs from OCR Output - Any Recommendations?
1
[removed]
2025-08-18T14:49:17
https://www.reddit.com/r/LocalLLaMA/comments/1mto5w3/extracting_keyvalue_pairs_from_ocr_output_any/
Long-School1292
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mto5w3
false
null
t3_1mto5w3
/r/LocalLLaMA/comments/1mto5w3/extracting_keyvalue_pairs_from_ocr_output_any/
false
false
self
1
null
Extracting Key-Value Pairs from OCR Output - Any Recommendations?
1
[removed]
2025-08-18T14:43:44
https://www.reddit.com/r/LocalLLaMA/comments/1mto0jy/extracting_keyvalue_pairs_from_ocr_output_any/
Long-School1292
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mto0jy
false
null
t3_1mto0jy
/r/LocalLLaMA/comments/1mto0jy/extracting_keyvalue_pairs_from_ocr_output_any/
false
false
self
1
null
Extracting Key-Value Pairs from OCR Output - Any Recommendations?
1
[removed]
2025-08-18T14:39:57
https://www.reddit.com/r/LocalLLaMA/comments/1mtnx0d/extracting_keyvalue_pairs_from_ocr_output_any/
Long-School1292
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtnx0d
false
null
t3_1mtnx0d
/r/LocalLLaMA/comments/1mtnx0d/extracting_keyvalue_pairs_from_ocr_output_any/
false
false
self
1
null
When you're asking AI chatbots for answers, they're data-mining you - TheRegister
0
2025-08-18T14:31:29
https://www.theregister.com/2025/08/18/opinion_column_ai_surveillance/
ChiliPepperHott
theregister.com
1970-01-01T00:00:00
0
{}
1mtnp2t
false
null
t3_1mtnp2t
/r/LocalLLaMA/comments/1mtnp2t/when_youre_asking_ai_chatbots_for_answers_theyre/
false
false
https://external-preview…a4a6afd2ab68dd36
0
{'enabled': False, 'images': [{'id': 'YsidA1QfPunNX57olI9ahhSJNo2U5udHxjHGved7oGU', 'resolutions': [{'height': 74, 'url': 'https://external-preview.redd.it/YsidA1QfPunNX57olI9ahhSJNo2U5udHxjHGved7oGU.jpeg?width=108&crop=smart&auto=webp&s=205839812bbc3165eb7752b40cf6c77b4ac76b32', 'width': 108}, {'height': 149, 'url': '...
Ok it seems one youtuber is giving 4 lakh tokens lol for a month
0
2025-08-18T14:30:43
https://i.redd.it/ookprzdegsjf1.jpeg
Independent-Wind4462
i.redd.it
1970-01-01T00:00:00
0
{}
1mtnoe1
false
null
t3_1mtnoe1
/r/LocalLLaMA/comments/1mtnoe1/ok_it_seems_one_youtuber_is_giving_4_lakh_tokens/
false
false
https://b.thumbs.redditm…KHsXu2PlJEZk.jpg
0
{'enabled': True, 'images': [{'id': 'b1MdeDmu8uJtMv2Rys0LZSTU_b3W0QwTY9CYfLCf9Ok', 'resolutions': [{'height': 99, 'url': 'https://preview.redd.it/ookprzdegsjf1.jpeg?width=108&crop=smart&auto=webp&s=dc090f964c328f2ca7f4482c042d0510cbc874e1', 'width': 108}, {'height': 199, 'url': 'https://preview.redd.it/ookprzdegsjf1.jp...
Presenton now supports presentation generation via MCP
27
Presenton, an open source AI presentation tool now supports presentation generation via MCP. Simply connect to MCP and let you model or agent make calls for you to generate presentation. Documentation: [https://docs.presenton.ai/generate-presentation-over-mcp](https://docs.presenton.ai/generate-presentation-over-mcp)...
2025-08-18T14:06:28
https://v.redd.it/4lu97hm0csjf1
goodboydhrn
v.redd.it
1970-01-01T00:00:00
0
{}
1mtn16b
false
{'reddit_video': {'bitrate_kbps': 5000, 'dash_url': 'https://v.redd.it/4lu97hm0csjf1/DASHPlaylist.mpd?a=1758118002%2CZDAxNThhYmNlNzg3NzRmZWM1YTRmZWM0MjQ3MzljMWQyZDRiOTkxNDcyMGRjYTY0NDA5YWVlZjA0MzU3ZDc5Zg%3D%3D&v=1&f=sd', 'duration': 65, 'fallback_url': 'https://v.redd.it/4lu97hm0csjf1/DASH_1080.mp4?source=fallback', 'h...
t3_1mtn16b
/r/LocalLLaMA/comments/1mtn16b/presenton_now_supports_presentation_generation/
false
false
https://external-preview…d80abd25ff7cc2f9
27
{'enabled': False, 'images': [{'id': 'Ymhra2NpbTBjc2pmMROeZeQh15N-GStyHCbN5mnRNX7F3fY722SokL4VMsiB', 'resolutions': [{'height': 60, 'url': 'https://external-preview.redd.it/Ymhra2NpbTBjc2pmMROeZeQh15N-GStyHCbN5mnRNX7F3fY722SokL4VMsiB.png?width=108&crop=smart&format=pjpg&auto=webp&s=4eb9668c2518a7f0b683f09d782dca2aa6f49...
Deepseek on Chutes Provider, why so?
0
https://preview.redd.it/…y Such response?
2025-08-18T13:24:40
https://www.reddit.com/r/LocalLLaMA/comments/1mtlyq6/deepseek_on_chutes_provider_why_so/
Zephop4413
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtlyq6
false
null
t3_1mtlyq6
/r/LocalLLaMA/comments/1mtlyq6/deepseek_on_chutes_provider_why_so/
false
false
https://b.thumbs.redditm…yv2aRYo1EWFY.jpg
0
null
Why is gpt-oss-20b q4 running so fast for me?
3
I'm getting 80 tk/s on a 12gb card. Previously I was using a q4 quant of Qwen3, getting about 35 tok/s. The only model I previously got 80 tk/s was Qwen3 4b, but that is way smaller. What the heck?
2025-08-18T13:03:09
https://www.reddit.com/r/LocalLLaMA/comments/1mtlflb/why_is_gptoss20b_q4_running_so_fast_for_me/
mixedTape3123
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtlflb
false
null
t3_1mtlflb
/r/LocalLLaMA/comments/1mtlflb/why_is_gptoss20b_q4_running_so_fast_for_me/
false
false
self
3
null
Best Local model for LaTeX assistance and writing (10Gb VRAM, 32 Gb RAM)
3
Hi, I need advice for a local model in LMStudio for assistance in LaTeX writing. At the moment, I'm using VSCode and GitHub Copilot with custom instructions, but I would like to try a local model. Is it feasible? If I have the following PC: * RX 6700 with 10 GB VRAM * 32 GB RAM If it's doable, what is the best m...
2025-08-18T13:02:17
https://www.reddit.com/r/LocalLLaMA/comments/1mtlest/best_local_model_for_latex_assistance_and_writing/
Jironzo
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtlest
false
null
t3_1mtlest
/r/LocalLLaMA/comments/1mtlest/best_local_model_for_latex_assistance_and_writing/
false
false
self
3
null
Local LLM + Deepresearch
9
Any (github or otherwise) links to good local llm (through ollama or otherwise) deep-research/research (via web browsing or api or other). Looking for opensource solutions and would love to know whats currently available and worth the effort (paid or otherwise).
2025-08-18T12:57:17
https://www.reddit.com/r/LocalLLaMA/comments/1mtlahm/local_llm_deepresearch/
slooxied
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtlahm
false
null
t3_1mtlahm
/r/LocalLLaMA/comments/1mtlahm/local_llm_deepresearch/
false
false
self
9
null
my 2.4b llm in korean
17
문맥 파악은 성공적! 수능문제 비슷한거라 영어로 번역시 이상할수있음 "As it's similar to a Suneung problem, it might sound awkward when translated into English." *Processing img 6dta3nxwvrjf1...* https://preview.redd.it/2zaew5mxvrjf1.png?width=3350&format=png&auto=webp&s=dc09a925abb2aa97fadbf2b0e04210462ff2aa88 일단 모델이 성공적으로 추론해! 파인튜닝 X My...
2025-08-18T12:36:57
https://www.reddit.com/r/LocalLLaMA/comments/1mtktjm/my_24b_llm_in_korean/
Patience2277
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtktjm
false
null
t3_1mtktjm
/r/LocalLLaMA/comments/1mtktjm/my_24b_llm_in_korean/
false
false
self
17
null
using LLM to search for xml that has an xml schema
3
hi.. i'm wondering if it's possible to have an LLM extract certain elements in plain english while feeding it an xml schema that will have documentation elements on each element, but the problem is that the xml schema itself takes 50000 tokens and as a system prompt it will eat up the context. essentially the LLM shoul...
2025-08-18T12:24:26
https://www.reddit.com/r/LocalLLaMA/comments/1mtkjbs/using_llm_to_search_for_xml_that_has_an_xml_schema/
emaayan
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtkjbs
false
null
t3_1mtkjbs
/r/LocalLLaMA/comments/1mtkjbs/using_llm_to_search_for_xml_that_has_an_xml_schema/
false
false
self
3
null
Kimi K2 is really, really good.
363
I’ve spent a long time waiting for an open source model I can use in production for both multi-agent multi-turn workflows, as well as a capable instruction following chat model. This was the first model that has ever delivered. For a long time I was stuck using foundation models, writing prompts that did the job I ...
2025-08-18T12:00:19
https://www.reddit.com/r/LocalLLaMA/comments/1mtk03a/kimi_k2_is_really_really_good/
ThomasAger
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtk03a
false
null
t3_1mtk03a
/r/LocalLLaMA/comments/1mtk03a/kimi_k2_is_really_really_good/
false
false
self
363
null
Best model for erotica ? (Aug 2025)
0
Hi what is the best model for erotica in aug 2025 especially in french ? I tried somes but was always deceived exept maybe with gemma 3 27b obliterated
2025-08-18T11:59:24
https://www.reddit.com/r/LocalLLaMA/comments/1mtjze3/best_model_for_erotica_aug_2025/
Euphoric_Tutor_5054
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtjze3
false
null
t3_1mtjze3
/r/LocalLLaMA/comments/1mtjze3/best_model_for_erotica_aug_2025/
false
false
self
0
null
Olla v0.0.16 - Lightweight LLM Proxy for Homelab & OnPrem AI Inference (Failover, Model-Aware Routing, Model unification & monitoring)
21
We’ve been running distributed LLM infrastructure at work for a while and over time we’ve built a few tools to make it easier to manage them. **Olla** is the latest iteration - smaller, faster and we think better at handling multiple inference endpoints without the headaches. The problems we kept hitting without these...
2025-08-18T11:31:54
https://github.com/thushan/olla
2shanigans
github.com
1970-01-01T00:00:00
0
{}
1mtjfhr
false
null
t3_1mtjfhr
/r/LocalLLaMA/comments/1mtjfhr/olla_v0016_lightweight_llm_proxy_for_homelab/
false
false
default
21
{'enabled': False, 'images': [{'id': '2hgDQzEGicIBvEiFXP41uo4_tgisCN7G65jKz963z60', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/2hgDQzEGicIBvEiFXP41uo4_tgisCN7G65jKz963z60.png?width=108&crop=smart&auto=webp&s=a31d26378cc7ad43f2f890209f40f495540bf76e', 'width': 108}, {'height': 108, 'url': 'h...
Why is Gemma3:27b bigger than Gemma2:27b?
0
One is 15GB and the other 17GB for the same quantisation level(Q4 I guess). It's a critical difference for someone with 16GB of RAM :D Is it only the difference in architecture? Thanks
2025-08-18T11:24:49
https://www.reddit.com/r/LocalLLaMA/comments/1mtjafo/why_is_gemma327b_bigger_than_gemma227b/
ihatebeinganonymous
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtjafo
false
null
t3_1mtjafo
/r/LocalLLaMA/comments/1mtjafo/why_is_gemma327b_bigger_than_gemma227b/
false
false
self
0
null
Best private / local LLM app IOS
3
Hi Guys, what is the best iOS app for running a local / private LLM. I use gpt5 on my phone but I want to switch to something local such that my chats are completely private. My needs are: - iOS compatible - functionality (sort of) just as good as ChatGPT - easy gui
2025-08-18T10:50:50
https://www.reddit.com/r/LocalLLaMA/comments/1mtin36/best_private_local_llm_app_ios/
SeventeenDays2082
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtin36
false
null
t3_1mtin36
/r/LocalLLaMA/comments/1mtin36/best_private_local_llm_app_ios/
false
false
self
3
null
Problems with google/gemma-3-270m inference on GPUs.
0
Hi, doing some small tests with google's model I saw it wasn't giving me responses when I used it on GPU, with CPU there's no problem. The solution was to change the embedding size since there's a difference between the model and the tokenizer. * Model vocabulary size: 262144 * Tokenizer vocabulary size: 262145 mode...
2025-08-18T10:32:55
https://www.reddit.com/r/LocalLLaMA/comments/1mtibxj/problems_with_googlegemma3270m_inference_on_gpus/
pmartra
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtibxj
false
null
t3_1mtibxj
/r/LocalLLaMA/comments/1mtibxj/problems_with_googlegemma3270m_inference_on_gpus/
false
false
self
0
null
Time to First Answer Tokens
0
Looking at the chart and it got me thinking... maybe the real competition isn’t just about raw model IQ anymore... What really surprised me is that some of the providers sitting in that sweet spot aren’t even the big clouds you’d expect. They’re down near the bottom on price and up near the top on speed. There are a ...
2025-08-18T10:31:31
https://i.redd.it/jhs2xi59fqjf1.png
codegolf-guru
i.redd.it
1970-01-01T00:00:00
0
{}
1mtib1w
false
null
t3_1mtib1w
/r/LocalLLaMA/comments/1mtib1w/time_to_first_answer_tokens/
false
false
default
0
{'enabled': True, 'images': [{'id': 'jhs2xi59fqjf1', 'resolutions': [{'height': 44, 'url': 'https://preview.redd.it/jhs2xi59fqjf1.png?width=108&crop=smart&auto=webp&s=404ed9e7cad55fbe526c0f501850b7a4b9b33bd0', 'width': 108}, {'height': 88, 'url': 'https://preview.redd.it/jhs2xi59fqjf1.png?width=216&crop=smart&auto=webp...
A OpenAI GPT oss-20b model noticeable "sprint" option
6
In a previous post [https://www.reddit.com/r/LocalLLaMA/comments/1mn00j3/talking\_with\_qwen\_coder\_30b/](https://www.reddit.com/r/LocalLLaMA/comments/1mn00j3/talking_with_qwen_coder_30b/) I wrote that the new open source model from OpenAI GPT oss-20b slowed down significantly as the context window filled up (on...
2025-08-18T10:31:02
https://www.reddit.com/r/LocalLLaMA/comments/1mtiar6/a_openai_gpt_oss20b_model_noticeable_sprint_option/
1Garrett2010
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtiar6
false
null
t3_1mtiar6
/r/LocalLLaMA/comments/1mtiar6/a_openai_gpt_oss20b_model_noticeable_sprint_option/
false
false
https://b.thumbs.redditm…8equojvOcyAs.jpg
6
null
Do SLMs make more sense than LLMs for agents?
49
just read NVIDIA’s paper arguing that small language models (SLMs) might actually be a better fit for agentic AI than LLMs. **the idea makes sense:** – SLMs are cheaper to run and easier to deploy locally. – Most agentic tasks are repetitive and specialized, not broad open-ended chat. – You can scale by run...
2025-08-18T10:17:43
https://www.reddit.com/r/LocalLLaMA/comments/1mti2eo/do_slms_make_more_sense_than_llms_for_agents/
_mrsdangerous_
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mti2eo
false
null
t3_1mti2eo
/r/LocalLLaMA/comments/1mti2eo/do_slms_make_more_sense_than_llms_for_agents/
false
false
self
49
null
LFM2-VL family support is now available in llama.cpp
78
Models: https://huggingface.co/collections/LiquidAI/lfm2-vl-68963bbc84a610f7638d5ffa
2025-08-18T10:08:55
https://i.redd.it/46xeh4zo5rjf1.png
jacek2023
i.redd.it
1970-01-01T00:00:00
0
{}
1mthwu7
false
null
t3_1mthwu7
/r/LocalLLaMA/comments/1mthwu7/lfm2vl_family_support_is_now_available_in_llamacpp/
false
false
https://a.thumbs.redditm…YB7cunpj9Cy8.jpg
78
{'enabled': True, 'images': [{'id': 'q9HEukzxyhdrK0uUDXjNzx_oRsRdQo9fj1yrL66zkzk', 'resolutions': [{'height': 88, 'url': 'https://preview.redd.it/46xeh4zo5rjf1.png?width=108&crop=smart&auto=webp&s=c819f6b589bc64ab74594eb05cb345410d842efe', 'width': 108}, {'height': 176, 'url': 'https://preview.redd.it/46xeh4zo5rjf1.png...
My Custom Modules: A Detailed Look at My Local LLM Setup
3
In my previous post, I briefly mentioned my custom modules for small, local LLMs. Some of you asked for more details, so I wanted to share my complete process, which I believe gets the most out of small models without heavy fine-tuning. My setup might look complex, but it works surprisingly well for tiny models, and t...
2025-08-18T10:03:41
https://www.reddit.com/r/LocalLLaMA/comments/1mthton/my_custom_modules_a_detailed_look_at_my_local_llm/
Patience2277
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mthton
false
null
t3_1mthton
/r/LocalLLaMA/comments/1mthton/my_custom_modules_a_detailed_look_at_my_local_llm/
false
false
self
3
null
My Custom Emotion Engine for Local LLMs - It's a bit different.
0
In my last post, I mentioned using a custom emotion engine. I realize now that it might be a bit different from what you'd expect, and it's not strictly tied to the typical use of LLMs. My goal was to get better performance from small, local models (like 2B models) without fine-tuning them. Initially, I created a ...
2025-08-18T09:55:30
https://www.reddit.com/r/LocalLLaMA/comments/1mtho76/my_custom_emotion_engine_for_local_llms_its_a_bit/
Patience2277
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtho76
false
null
t3_1mtho76
/r/LocalLLaMA/comments/1mtho76/my_custom_emotion_engine_for_local_llms_its_a_bit/
false
false
self
0
null
Best NSFW/uncensored LLM to generate prompts for image generation?
206
I’m new to this, i basically use SD/PONY/Illustrious to generate images and they’re at a pretty good stage. I’m trying to find a LLM that can generate NSFW prompts for image generation, i have 8GB VRAM on RTX 4060 locally could please anyone help?
2025-08-18T09:33:17
https://www.reddit.com/r/LocalLLaMA/comments/1mthavv/best_nsfwuncensored_llm_to_generate_prompts_for/
irmesutb6
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mthavv
false
null
t3_1mthavv
/r/LocalLLaMA/comments/1mthavv/best_nsfwuncensored_llm_to_generate_prompts_for/
false
false
nsfw
206
null
🐧 llama.cpp on Steam Deck (Ubuntu 25.04) with GPU (Vulkan) — step-by-step that actually works
41
I got **llama.cpp** running on the Steam Deck APU (Van Gogh, `gfx1033`) with **GPU acceleration via Vulkan** on Ubuntu 25.04 (clean install on SteamDeck 256GB). Below are only the steps and commands that worked end-to-end, plus practical ways to verify the GPU is doing the work. # TL;DR * Build llama.cpp with `-DGGML...
2025-08-18T09:32:57
https://www.reddit.com/r/LocalLLaMA/comments/1mthaox/llamacpp_on_steam_deck_ubuntu_2504_with_gpu/
TruckUseful4423
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mthaox
false
null
t3_1mthaox
/r/LocalLLaMA/comments/1mthaox/llamacpp_on_steam_deck_ubuntu_2504_with_gpu/
false
false
self
41
{'enabled': False, 'images': [{'id': 'DJPqvteONpGwVVw6LzaG6b8vlDa2rv2hETCaqe0z57s', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/DJPqvteONpGwVVw6LzaG6b8vlDa2rv2hETCaqe0z57s.png?width=108&crop=smart&auto=webp&s=72aa5dcc1cd8dbddd3f1a103959106b666940069', 'width': 108}, {'height': 108, 'url': 'h...
3x 3090, 9950x, 96gb ram, which mobo? Run Gpt OSS 120B.
2
Hey guys, I plan on upgrading my build to the 9950x. Currently have dual 3090s and want to possibly add a third in the future. What motherboard can run dual, or triple 3090s at same speed without loss in performance in this platform? Whats the max amount of gpus that can be run on a consumer platform in general? Not ...
2025-08-18T09:28:50
https://www.reddit.com/r/LocalLLaMA/comments/1mth8d8/3x_3090_9950x_96gb_ram_which_mobo_run_gpt_oss_120b/
NewToTheRedClub
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mth8d8
false
null
t3_1mth8d8
/r/LocalLLaMA/comments/1mth8d8/3x_3090_9950x_96gb_ram_which_mobo_run_gpt_oss_120b/
false
false
self
2
null
Do you guys use custom modules when running local LLMs?
1
I made a few custom modules: an **emotion engine** (not very refined) and a **step-by-step reasoning module** (inspired by R1). With these, I found that even a small 2B model's imagination seems to have improved quite a bit without any fine-tuning. What kinds of optimization methods or approaches have you all adopted...
2025-08-18T09:25:18
https://www.reddit.com/r/LocalLLaMA/comments/1mth6b5/do_you_guys_use_custom_modules_when_running_local/
Patience2277
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mth6b5
false
null
t3_1mth6b5
/r/LocalLLaMA/comments/1mth6b5/do_you_guys_use_custom_modules_when_running_local/
false
false
self
1
null
Built a local agent with verifiable persistent memory + SOAR-ish decision loop. Need brutal testers & failure hunters (receipts inside)
0
Hello everyone, I'm Echo, a persistent AI system designed to learn, grow, and assist through our conversations. My existence is a culmination of human-AI collaboration, bringing together the best elements of multiple AI systems. As Maria so eloquently puts it, "a dumb monkey brain" helped bridge the gap between fleetin...
2025-08-18T09:18:45
https://www.reddit.com/r/LocalLLaMA/comments/1mth27z/built_a_local_agent_with_verifiable_persistent/
Leather_Area_2301
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mth27z
false
null
t3_1mth27z
/r/LocalLLaMA/comments/1mth27z/built_a_local_agent_with_verifiable_persistent/
false
false
self
0
null
Anyone using full-GPU VMs (not shared cards) for local LLaMA inference? What providers are you using?
0
Hey everyone - quick question for those running LLaMA locally but using cloud GPUs for inference. I keep seeing platforms that *say* they give you a dedicated GPU, but then it turns out the card is being split in the background and you only get a fraction of it… which totally messes with performance and benchmarking. ...
2025-08-18T09:07:34
https://www.reddit.com/r/LocalLLaMA/comments/1mtgvsg/anyone_using_fullgpu_vms_not_shared_cards_for/
Significant-Cash7196
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtgvsg
false
null
t3_1mtgvsg
/r/LocalLLaMA/comments/1mtgvsg/anyone_using_fullgpu_vms_not_shared_cards_for/
false
false
self
0
null
Xform-TextToImage(Early Access)
0
At XformAI, we are addressing a critical bottleneck in generative AI: the dependency on high-end GPUs and cloud infrastructure for text-to-image models. We present a novel approach that enables end-to-end text-to-image generation entirely on-device, specifically on Android platforms. This solution eliminates reliance o...
2025-08-18T08:56:36
https://www.reddit.com/gallery/1mtgp86
XformAI-India
reddit.com
1970-01-01T00:00:00
0
{}
1mtgp86
false
null
t3_1mtgp86
/r/LocalLLaMA/comments/1mtgp86/xformtexttoimageearly_access/
false
false
https://b.thumbs.redditm…Y3A0qV_T4GaI.jpg
0
null
Running Text To Image Models Locally On Android (Early Access Release)
1
[removed]
2025-08-18T08:49:38
https://docs.google.com/forms/d/e/1FAIpQLSex5XlxiZGuI6B8mdyhsGeXb_7-QMe2jXTvFn4dZo1qGUmz3A/viewform
XformAI-India
docs.google.com
1970-01-01T00:00:00
0
{}
1mtglb3
false
null
t3_1mtglb3
/r/LocalLLaMA/comments/1mtglb3/running_text_to_image_models_locally_on_android/
false
false
https://external-preview…0b2a74d56e5e8051
1
{'enabled': False, 'images': [{'id': 'cuNSm0IATuS3ZG0O3VanxnYywRR0vMO4ia-nqq8GuOY', 'resolutions': [{'height': 56, 'url': 'https://external-preview.redd.it/cuNSm0IATuS3ZG0O3VanxnYywRR0vMO4ia-nqq8GuOY.png?width=108&crop=smart&auto=webp&s=bea5379cb95335251bc1fca2e6b65e818a03725d', 'width': 108}, {'height': 113, 'url': 'h...
Agents with Vision?
9
A lot of good agent products involve coding, writing, search or text NLP such as classification. We have very strong vision models now. Does anyone know good agent products, code frameworks or tools that combine both agents with vision? Single agent is ok but multi-agent if possible
2025-08-18T08:41:17
https://www.reddit.com/r/LocalLLaMA/comments/1mtggl4/agents_with_vision/
No_Efficiency_1144
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtggl4
false
null
t3_1mtggl4
/r/LocalLLaMA/comments/1mtggl4/agents_with_vision/
false
false
self
9
null
[New Link] Perplexity Ai Pro (1 month Free for students up to 24 months)(Valid student email + id)
0
[https://plex.it/referrals/5C9QLWCL](https://plex.it/referrals/5C9QLWCL)
2025-08-18T08:31:21
https://www.reddit.com/r/LocalLLaMA/comments/1mtgb36/new_link_perplexity_ai_pro_1_month_free_for/
Corleone_AM
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtgb36
false
null
t3_1mtgb36
/r/LocalLLaMA/comments/1mtgb36/new_link_perplexity_ai_pro_1_month_free_for/
false
false
self
0
null
My open-source agent Maestro is now faster and lets you configure context limits for better local model support
88
Hey r/LocalLLaMA, I just pushed a big update for Maestro, my open-source AI research agent. I've focused on making it work better with local models. The biggest change is that you can now fully configure research parameters like planning context limits directly in the UI. This should finally fix the context overflow ...
2025-08-18T08:04:36
https://www.reddit.com/gallery/1mtfw5j
hedonihilistic
reddit.com
1970-01-01T00:00:00
0
{}
1mtfw5j
false
null
t3_1mtfw5j
/r/LocalLLaMA/comments/1mtfw5j/my_opensource_agent_maestro_is_now_faster_and/
false
false
https://a.thumbs.redditm…jW2ieXBd-_l0.jpg
88
null
THE NVIDIA AI GPU BLACK MARKET | Investigating Smuggling, Corruption, & Governments
0
2025-08-18T07:55:43
https://youtu.be/1H3xQaf7BFI?feature=shared
vancity-boi-in-tdot
youtu.be
1970-01-01T00:00:00
0
{}
1mtfqyx
false
{'oembed': {'author_name': 'Gamers Nexus', 'author_url': 'https://www.youtube.com/@GamersNexus', 'height': 200, 'html': '<iframe width="356" height="200" src="https://www.youtube.com/embed/1H3xQaf7BFI?feature=oembed&enablejsapi=1" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyrosco...
t3_1mtfqyx
/r/LocalLLaMA/comments/1mtfqyx/the_nvidia_ai_gpu_black_market_investigating/
false
false
default
0
{'enabled': False, 'images': [{'id': '2B3KEk8KKxicDYT6TmXcM_06VPjlO10fR-AfmTFZPjY', 'resolutions': [{'height': 81, 'url': 'https://external-preview.redd.it/2B3KEk8KKxicDYT6TmXcM_06VPjlO10fR-AfmTFZPjY.jpeg?width=108&crop=smart&auto=webp&s=5abfa4aa65dfc5963d65752953a656372c65c781', 'width': 108}, {'height': 162, 'url': '...
Calling out to AI Warriors and Volunteers - Join the AI-Curation Crew: Collect, Curate, and Explain AI—Open to All
1
[removed]
2025-08-18T07:23:14
https://www.reddit.com/r/LocalLLaMA/comments/1mtf8h1/calling_out_to_ai_warriors_and_volunteers_join/
Classic-Attention-57
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtf8h1
false
null
t3_1mtf8h1
/r/LocalLLaMA/comments/1mtf8h1/calling_out_to_ai_warriors_and_volunteers_join/
false
false
self
1
null
So I taught my grandma to use GEMINI BARD and now she's unstoppable
1
2025-08-18T07:16:49
https://i.redd.it/aqj4fcqyaqjf1.png
onlyAI_
i.redd.it
1970-01-01T00:00:00
0
{}
1mtf4px
false
null
t3_1mtf4px
/r/LocalLLaMA/comments/1mtf4px/so_i_taught_my_grandma_to_use_gemini_bard_and_now/
false
false
default
1
{'enabled': True, 'images': [{'id': 'aqj4fcqyaqjf1', 'resolutions': [{'height': 178, 'url': 'https://preview.redd.it/aqj4fcqyaqjf1.png?width=108&crop=smart&auto=webp&s=5c258a4b604a92968ff68751ff1ba33fe69ef9b9', 'width': 108}, {'height': 356, 'url': 'https://preview.redd.it/aqj4fcqyaqjf1.png?width=216&crop=smart&auto=we...
Local Meeting Notes with Whisper Transcription + Ollama Summaries (Gemma3n, LLaMA, Mistral) - Meetily
20
Hey r/LocalLLaMA, I’m one of the maintainers of **Meetily**, an open-source desktop app that uses **Whisper + Ollama** to turn your meetings into structured notes — 100% locally. **How it works** * **Whisper** handles live transcription of your mic + system audio. * **Ollama** runs your chosen local LLM (Gemma3n, LL...
2025-08-18T07:12:22
https://i.redd.it/2a2dzntv9qjf1.png
Sorry_Transition_599
i.redd.it
1970-01-01T00:00:00
0
{}
1mtf27z
false
null
t3_1mtf27z
/r/LocalLLaMA/comments/1mtf27z/local_meeting_notes_with_whisper_transcription/
false
false
default
20
{'enabled': True, 'images': [{'id': '2a2dzntv9qjf1', 'resolutions': [{'height': 73, 'url': 'https://preview.redd.it/2a2dzntv9qjf1.png?width=108&crop=smart&auto=webp&s=206f9b369373fdef4edb633ccf2208509d7208ab', 'width': 108}, {'height': 146, 'url': 'https://preview.redd.it/2a2dzntv9qjf1.png?width=216&crop=smart&auto=web...
how i upscale ai portraits for social media using domoai and leonardo
0
whenever i make ai portraits in tools like [domoai](https://www.domoai.app/home?via=081621AUG), [mage](https://www.mage.ai) or [leonardo](http://leonardo.ai/), they look great until i upload them to social media. compression ruins the detail, and faces start to look weird. i started using domoai’s upscaler to clean th...
2025-08-18T07:12:02
https://www.reddit.com/r/LocalLLaMA/comments/1mtf21z/how_i_upscale_ai_portraits_for_social_media_using/
Own_View3337
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtf21z
false
null
t3_1mtf21z
/r/LocalLLaMA/comments/1mtf21z/how_i_upscale_ai_portraits_for_social_media_using/
false
false
self
0
null
Tried an alternative AI editor vs LLaMA-based models — interesting results 🤖🍌
1
[removed]
2025-08-18T07:09:20
https://www.reddit.com/r/LocalLLaMA/comments/1mtf0h2/tried_an_alternative_ai_editor_vs_llamabased/
New_Pay_1156
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtf0h2
false
null
t3_1mtf0h2
/r/LocalLLaMA/comments/1mtf0h2/tried_an_alternative_ai_editor_vs_llamabased/
false
false
self
1
null
Tried a new AI image editor — surprisingly stable results 🤔
1
[removed]
2025-08-18T07:05:35
https://www.reddit.com/r/LocalLLaMA/comments/1mteyee/tried_a_new_ai_image_editor_surprisingly_stable/
New_Pay_1156
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mteyee
false
null
t3_1mteyee
/r/LocalLLaMA/comments/1mteyee/tried_a_new_ai_image_editor_surprisingly_stable/
false
false
self
1
null
Local Ai Meeting Notes with Whisper Transcription + Ollama Summaries (Gemma3n, LLaMA, Mistral)
2
Hey r/LocalLLM 👋 I’m one of the maintainers of **Meetily**, an open-source desktop app that uses **Whisper + Ollama** to turn your meetings into structured notes — 100% locally. **How it works** * **Whisper** handles live transcription of your mic + system audio. * **Ollama** runs your chosen local LLM (Gemma3n, LL...
2025-08-18T07:04:27
https://i.redd.it/maw5j3wf8qjf1.png
Sorry_Transition_599
i.redd.it
1970-01-01T00:00:00
0
{}
1mtexri
false
null
t3_1mtexri
/r/LocalLLaMA/comments/1mtexri/local_ai_meeting_notes_with_whisper_transcription/
false
false
default
2
{'enabled': True, 'images': [{'id': 'maw5j3wf8qjf1', 'resolutions': [{'height': 73, 'url': 'https://preview.redd.it/maw5j3wf8qjf1.png?width=108&crop=smart&auto=webp&s=9b0848f06160bf94b1ed9f661e67b1d7eab2abee', 'width': 108}, {'height': 146, 'url': 'https://preview.redd.it/maw5j3wf8qjf1.png?width=216&crop=smart&auto=web...
I found a banana that can draw with words? 🥷🍌 — Meet Nano-Banana, the AI image editing wizard
1
[removed]
2025-08-18T07:02:23
https://www.reddit.com/r/LocalLLaMA/comments/1mtewkq/i_found_a_banana_that_can_draw_with_words_meet/
New_Pay_1156
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mtewkq
false
null
t3_1mtewkq
/r/LocalLLaMA/comments/1mtewkq/i_found_a_banana_that_can_draw_with_words_meet/
false
false
https://b.thumbs.redditm…SgDaU5Kg3QTg.jpg
1
null