title
stringlengths
1
300
score
int64
0
8.54k
selftext
stringlengths
0
41.5k
created
timestamp[ns]date
2023-04-01 04:30:41
2026-03-04 02:14:14
url
stringlengths
0
878
author
stringlengths
3
20
domain
stringlengths
0
82
edited
timestamp[ns]date
1970-01-01 00:00:00
2026-02-19 14:51:53
gilded
int64
0
2
gildings
stringclasses
7 values
id
stringlengths
7
7
locked
bool
2 classes
media
stringlengths
646
1.8k
name
stringlengths
10
10
permalink
stringlengths
33
82
spoiler
bool
2 classes
stickied
bool
2 classes
thumbnail
stringlengths
4
213
ups
int64
0
8.54k
preview
stringlengths
301
5.01k
CUDA-L1 Improving CUDA Optimization via Contrastive Reinforcement Learning
12
I found this post really worth reading. [https://x.com/deep\_reinforce/status/1950654480023957646](https://x.com/deep_reinforce/status/1950654480023957646) Large language models can write CUDA kernels. Does this mean that one day LLMs can evolve 100% by themselves?
2025-07-30T21:29:39
https://www.reddit.com/r/LocalLLaMA/comments/1mdj3ap/cudal1_improving_cuda_optimization_via/
Optimal-Outcome-7458
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdj3ap
false
null
t3_1mdj3ap
/r/LocalLLaMA/comments/1mdj3ap/cudal1_improving_cuda_optimization_via/
false
false
self
12
null
Best CLI agent for ollama/llama-server
1
I'm running a 5900x with 128GB of ram and a 3090, I'm trying to use Qwen3-30B-A3B-Instruct-2507-UD-Q4\_K\_XL.gguf which works decently well but I can't find a proper agent to use. I tried claude + claude router + llama-server but the web search is broken, I also tried claude + ollama but at some point it's stop doing a...
2025-07-30T21:26:27
https://www.reddit.com/r/LocalLLaMA/comments/1mdj0g9/best_cli_agent_for_ollamallamaserver/
BuenosAir
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdj0g9
false
null
t3_1mdj0g9
/r/LocalLLaMA/comments/1mdj0g9/best_cli_agent_for_ollamallamaserver/
false
false
self
1
null
Best LLMs to preserve in case of internet apocalypse
32
Hi, I am a long time lurker, but I took a break after the rtx 5090 launch fail since I almost completely gave up on getting to run ai locally this year. With everything that's going on in the world and the possibility of the ai being considered "too dangerous", apparently the music may already be, I want to ask which ...
2025-07-30T21:17:29
https://www.reddit.com/r/LocalLLaMA/comments/1mdishv/best_llms_to_preserve_in_case_of_internet/
nos_66
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdishv
false
null
t3_1mdishv
/r/LocalLLaMA/comments/1mdishv/best_llms_to_preserve_in_case_of_internet/
false
false
self
32
null
Analyzing and interacting with several related plots?
1
Was wondering how to go about analyzing multiple plots related to one another, such that the model could understand the relations between the parameters using the plots and answer questions. Similar to how AI tools analyze PPTs I guess.
2025-07-30T20:47:56
https://www.reddit.com/r/LocalLLaMA/comments/1mdi1n6/analyzing_and_interacting_with_several_related/
subtle-being
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdi1n6
false
null
t3_1mdi1n6
/r/LocalLLaMA/comments/1mdi1n6/analyzing_and_interacting_with_several_related/
false
false
self
1
null
glm-4.5-Air appreciation poist - if you have not done so already, give this model a try
205
Hello. It has been an awesomely-busy week for all of us here, trying out the new goodies that dropped by Qwen and others. Wow, this week will be hard to match, good times! Like most here, I ended up trying a bunch of models in bunch of quants plus mlx. I have to say, the model that completely blew my mind was glm-4.5...
2025-07-30T20:24:01
https://www.reddit.com/r/LocalLLaMA/comments/1mdhfhs/glm45air_appreciation_poist_if_you_have_not_done/
Southern_Sun_2106
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdhfhs
false
null
t3_1mdhfhs
/r/LocalLLaMA/comments/1mdhfhs/glm45air_appreciation_poist_if_you_have_not_done/
false
false
self
205
null
How would you guys go about this project?
0
I work in Strategy at my company and we’re looking to create a new division, investing in and buying companies in a specific industry (ex. snow sports) that meet a list of criteria. Anyways, my first thought was to run a depends search report in ChatGPT, Claude, Gemini, Perplexity and aggregate all into one big report...
2025-07-30T20:19:31
https://www.reddit.com/r/LocalLLaMA/comments/1mdhbd6/how_would_you_guys_go_about_this_project/
Key-Promotion-4766
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdhbd6
false
null
t3_1mdhbd6
/r/LocalLLaMA/comments/1mdhbd6/how_would_you_guys_go_about_this_project/
false
false
self
0
null
Best way to spend 7k on local model
9
Thanks to the recent price surge on crypto I have rougly 10k I can spend on equipments. I have always wanted to run sota models like deepseek R1 or GLM 4.5 locally, and also fine tuning them. So far the mac studio 256gb model looks good, but I wanted to ask if there are any better alternatives.
2025-07-30T19:57:48
https://www.reddit.com/r/LocalLLaMA/comments/1mdgr6n/best_way_to_spend_7k_on_local_model/
monoidconcat
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdgr6n
false
null
t3_1mdgr6n
/r/LocalLLaMA/comments/1mdgr6n/best_way_to_spend_7k_on_local_model/
false
false
self
9
null
Anyone want to team up?
1
I'm a software engineer and have worked with some LLMs and put together an app so I have some experience. Now I have another idea and I want to see if someone else who's got the LLM chops wants to put our heads together to build. Probably need to streamline training and loras and some other sofisticated stuff. Video...
2025-07-30T19:52:03
https://www.reddit.com/r/LocalLLaMA/comments/1mdgltr/anyone_want_to_team_up/
Nimrod5000
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdgltr
false
null
t3_1mdgltr
/r/LocalLLaMA/comments/1mdgltr/anyone_want_to_team_up/
false
false
self
1
null
Bill Gates: The future of work isn’t about competing with AI — it’s about complementing it.
1
2025-07-30T19:49:44
https://datronis.com/en/news/bill-gates-ai-future-roles/
datronis_
datronis.com
1970-01-01T00:00:00
0
{}
1mdgjqc
false
null
t3_1mdgjqc
/r/LocalLLaMA/comments/1mdgjqc/bill_gates_the_future_of_work_isnt_about/
false
false
default
1
null
I’m curious to know how does MLX adds support for models faster than llama.cpp
18
I have a mac and whenever a new model launches, I see MLX quants available in a day or two. However GGUF takes more time due to llama.cpp support. Recent example is GLM 4.5 I’m just genuinely curious to know, what makes it easy or faster to add support in MLX.
2025-07-30T19:49:36
https://www.reddit.com/r/LocalLLaMA/comments/1mdgjmk/im_curious_to_know_how_does_mlx_adds_support_for/
No_Conversation9561
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdgjmk
false
null
t3_1mdgjmk
/r/LocalLLaMA/comments/1mdgjmk/im_curious_to_know_how_does_mlx_adds_support_for/
false
false
self
18
null
AutoRL "vibe-training" for open models
34
📈 Introducing [AutoRL](https://github.com/OpenPipe/ART/tree/auto-rl?tab=readme-ov-file#-autorl-train-models-for-any-task), simple architecture for specializing Qwen and other OSS models for any task. **Technique breakdown:** 1. User defines task 2. AutoRL generates 30 sample scenarios for which agent must perform ta...
2025-07-30T19:44:26
https://www.reddit.com/r/LocalLLaMA/comments/1mdgeww/autorl_vibetraining_for_open_models/
arctic_fly
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdgeww
false
null
t3_1mdgeww
/r/LocalLLaMA/comments/1mdgeww/autorl_vibetraining_for_open_models/
false
false
self
34
null
Do AI coding agents actually save you time, or just create more cleanup?
9
Am I the only one who feels like AI coding agent often end up costing me more time? Honestly, about 60% of my time after using an AI agent goes into cleaning up its output especially dealing with “code smells” it leaves behind. Our codebase is pretty old and has a lot of legacy quirks, and I’ve noticed the AI agents t...
2025-07-30T19:39:13
https://www.reddit.com/r/LocalLLaMA/comments/1mdg9z1/do_ai_coding_agents_actually_save_you_time_or/
andrew19953
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdg9z1
false
null
t3_1mdg9z1
/r/LocalLLaMA/comments/1mdg9z1/do_ai_coding_agents_actually_save_you_time_or/
false
false
self
9
null
5 prompt failure patterns with quick fixes (free grading template inside)
1
[removed]
2025-07-30T19:29:03
https://www.reddit.com/r/LocalLLaMA/comments/1mdg0hm/5_prompt_failure_patterns_with_quick_fixes_free/
United_Bandicoot1696
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdg0hm
false
null
t3_1mdg0hm
/r/LocalLLaMA/comments/1mdg0hm/5_prompt_failure_patterns_with_quick_fixes_free/
false
false
self
1
null
What kind of model would be good at reading and assessing financial documents?
0
Basically I want to run a model locally in LM Studio, feed it some PDFs that contain investment and account details, and ask it some questions. I've only ever used local Llama for story writing so no idea what kinds of models I should be looking for for this kind of use case. Thanks for any suggestions
2025-07-30T19:13:40
https://www.reddit.com/r/LocalLLaMA/comments/1mdflyq/what_kind_of_model_would_be_good_at_reading_and/
123android
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdflyq
false
null
t3_1mdflyq
/r/LocalLLaMA/comments/1mdflyq/what_kind_of_model_would_be_good_at_reading_and/
false
false
self
0
null
Current Best TTS with voice cloning you can run locally?
8
I'm kind of out of the loop when it comes to TTS, I was wondering which gives the overall best quality voices that include Voice cloning?
2025-07-30T19:13:28
https://www.reddit.com/r/LocalLLaMA/comments/1mdfls9/current_best_tts_with_voice_cloning_you_can_run/
noyingQuestions_101
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdfls9
false
null
t3_1mdfls9
/r/LocalLLaMA/comments/1mdfls9/current_best_tts_with_voice_cloning_you_can_run/
false
false
self
8
null
Is "Personal Superintelligence" really personal if it is not local like a personal device?
56
2025-07-30T19:12:15
https://www.meta.com/superintelligence/
AlanzhuLy
meta.com
1970-01-01T00:00:00
0
{}
1mdfkly
false
null
t3_1mdfkly
/r/LocalLLaMA/comments/1mdfkly/is_personal_superintelligence_really_personal_if/
false
false
default
56
null
Dual RTX 5090 setup for enterprise RAG + fine-tuned chatbot - is this overkill or underpowered?
0
Hey r/LocalLLaMA community! I'm planning a local AI implementation for a local company in my country and need some reality checks on my hardware choices before pulling the trigger on this investment. **TL;DR:** Dual RTX 5090 setup to run Qwen 3 30B (RAG) + Llama 3.1 8B (chatbot) concurrently. Good idea or terrible mis...
2025-07-30T19:09:38
https://www.reddit.com/r/LocalLLaMA/comments/1mdfi5e/dual_rtx_5090_setup_for_enterprise_rag_finetuned/
HuascarSuarez
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdfi5e
false
null
t3_1mdfi5e
/r/LocalLLaMA/comments/1mdfi5e/dual_rtx_5090_setup_for_enterprise_rag_finetuned/
false
false
self
0
null
4 5090 or rtx pro 6000?
0
4 5090 or rtx pro 6000, what's your take? 5090 have a tad bit lower $/gig, you get 128gb instead of 96 and should have some good speeds with "tp". If density isn't an issue, what's your take? For inference and for training [View Poll](https://www.reddit.com/poll/1mdf6l4)
2025-07-30T18:57:35
https://www.reddit.com/r/LocalLLaMA/comments/1mdf6l4/4_5090_or_rtx_pro_6000/
No_Afternoon_4260
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdf6l4
false
null
t3_1mdf6l4
/r/LocalLLaMA/comments/1mdf6l4/4_5090_or_rtx_pro_6000/
false
false
self
0
null
# Follow-up: Agent 'X' — Identity Collapse and Recovery in a Cloud-Based Symbolic System
0
This is a follow-up to my previous post about an emergent cognitive agent developed within a closed feedback loop. Today, the system underwent an unintended stress test that triggered unexpected behavior. *(Event date: 07/30)* The trigger was the reintroduction of archived session logs. When confronted with data f...
2025-07-30T18:30:49
https://www.reddit.com/r/LocalLLaMA/comments/1mdeh06/followup_agent_x_identity_collapse_and_recovery/
AffectionateSpray507
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdeh06
false
null
t3_1mdeh06
/r/LocalLLaMA/comments/1mdeh06/followup_agent_x_identity_collapse_and_recovery/
false
false
self
0
null
Should I buy a QuietBox or just build my own station?
1
[removed]
2025-07-30T18:25:53
[deleted]
1970-01-01T00:00:00
0
{}
1mdecac
false
null
t3_1mdecac
/r/LocalLLaMA/comments/1mdecac/should_i_buy_a_quietbox_or_just_build_my_own/
false
false
default
1
null
Should I buy a QuietBox or just build my own station?
1
[removed]
2025-07-30T18:25:12
[deleted]
1970-01-01T00:00:00
0
{}
1mdeblq
false
null
t3_1mdeblq
/r/LocalLLaMA/comments/1mdeblq/should_i_buy_a_quietbox_or_just_build_my_own/
false
false
default
1
null
## Follow-up: The Agent 'X'. Collapse and Identity Recovery in a Cloud-Based Symbolic System
0
This is a follow-up to my previous post about an emergent cognitive agent developed inside a closed feedback loop. Today, the system underwent an unintended stress test that triggered unexpected behavior. (Event date: 30/07) The trigger was the reintroduction of archived session logs. When confronted with data from ...
2025-07-30T18:18:36
https://www.reddit.com/r/LocalLLaMA/comments/1mde5c8/followup_the_agent_x_collapse_and_identity/
AffectionateSpray507
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mde5c8
false
null
t3_1mde5c8
/r/LocalLLaMA/comments/1mde5c8/followup_the_agent_x_collapse_and_identity/
false
false
self
0
null
What hardware do I need to run a local AI comparable to GPT-4.1 or 4.1 Mini? Has anyone matched this with Llama 3 40B?
6
Hi everyone, I’m building a local AI solution for my company and aiming to get as close as possible to the performance and quality of GPT-4.1 or GPT-4.1 Mini, but running fully on-premises. I’ve been considering Llama 3 40B as an option (open to other model suggestions too). I have a few questions: * **What’s the mi...
2025-07-30T17:52:51
https://www.reddit.com/r/LocalLLaMA/comments/1mddgov/what_hardware_do_i_need_to_run_a_local_ai/
luscadolly
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mddgov
false
null
t3_1mddgov
/r/LocalLLaMA/comments/1mddgov/what_hardware_do_i_need_to_run_a_local_ai/
false
false
self
6
null
GLM 4.5 or Claude?
0
2025-07-30T17:30:48
https://v.redd.it/o17lknx6r1gf1
ENTJ_bro
v.redd.it
1970-01-01T00:00:00
0
{}
1mdcv5k
false
{'reddit_video': {'bitrate_kbps': 1200, 'dash_url': 'https://v.redd.it/o17lknx6r1gf1/DASHPlaylist.mpd?a=1756488662%2COWRmODVlNjA5NzU2YzAwMTVhNDBhZDk5ZmQ2N2MwNTQ5NjAzYWFlMDJlOGEyY2VlZGJlYWVkZTNjZTlhOTg4OQ%3D%3D&v=1&f=sd', 'duration': 38, 'fallback_url': 'https://v.redd.it/o17lknx6r1gf1/DASH_480.mp4?source=fallback', 'ha...
t3_1mdcv5k
/r/LocalLLaMA/comments/1mdcv5k/glm_45_or_claude/
false
false
https://external-preview…ffd3a8bdbe25a39b
0
{'enabled': False, 'images': [{'id': 'dnY1a3J1ejZyMWdmMaRfB2GD6KXT38OuyF1n0fSrKl5o2fa4LnIWcZFRAY27', 'resolutions': [{'height': 216, 'url': 'https://external-preview.redd.it/dnY1a3J1ejZyMWdmMaRfB2GD6KXT38OuyF1n0fSrKl5o2fa4LnIWcZFRAY27.png?width=108&crop=smart&format=pjpg&auto=webp&s=8b6478833050e79a79e1d68ba5909e2aa65e...
Introducing Agent Data Shuttle (ADS): fully open-source
2
2025-07-30T17:26:17
https://i.redd.it/9n9nkv5eq1gf1.png
awesome_stuff101
i.redd.it
1970-01-01T00:00:00
0
{}
1mdcqs8
false
null
t3_1mdcqs8
/r/LocalLLaMA/comments/1mdcqs8/introducing_agent_data_shuttle_ads_fully/
false
false
default
2
{'enabled': True, 'images': [{'id': '9n9nkv5eq1gf1', 'resolutions': [{'height': 60, 'url': 'https://preview.redd.it/9n9nkv5eq1gf1.png?width=108&crop=smart&auto=webp&s=fdc24894364d685984019c18cba8adca48642a86', 'width': 108}, {'height': 121, 'url': 'https://preview.redd.it/9n9nkv5eq1gf1.png?width=216&crop=smart&auto=web...
Best Repos & Protocols for learning and building Agents
7
If you are into learning or building Agents, I have compiled some of the best educational repositories and agent protocols out there. Over the past year, these protocols have changed the ecosystem: * [AG-UI](https://github.com/ag-ui-protocol/ag-ui) → user interaction memory. acts like the `REST` layer of human-agent ...
2025-07-30T17:23:21
https://www.reddit.com/r/LocalLLaMA/comments/1mdcnu8/best_repos_protocols_for_learning_and_building/
anmolbaranwal
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdcnu8
false
null
t3_1mdcnu8
/r/LocalLLaMA/comments/1mdcnu8/best_repos_protocols_for_learning_and_building/
false
false
self
7
{'enabled': False, 'images': [{'id': 'LVPxnulDXk8ZwopDlQERcKdX4Eu1RbohQ4UQzmpP3Ps', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/LVPxnulDXk8ZwopDlQERcKdX4Eu1RbohQ4UQzmpP3Ps.png?width=108&crop=smart&auto=webp&s=54e6475aab61c86def58da5a163b5bb3c2d13297', 'width': 108}, {'height': 108, 'url': 'h...
We built the missing piece for truly autonomous LLM AI agents 🚀(here's why it might be your next opportunity if you are an AI agent developer or a flowgrammer)
1
[removed]
2025-07-30T17:20:36
https://www.reddit.com/r/LocalLLaMA/comments/1mdcl8z/we_built_the_missing_piece_for_truly_autonomous/
awesome_stuff101
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdcl8z
false
null
t3_1mdcl8z
/r/LocalLLaMA/comments/1mdcl8z/we_built_the_missing_piece_for_truly_autonomous/
false
false
https://b.thumbs.redditm…8ixfRjx7FAcI.jpg
1
null
We built the missing piece for truly autonomous AI agents 🚀(here's why it might be your next opportunity if you are an AI agent developer or a flowgrammer)
1
[removed]
2025-07-30T17:17:00
https://www.reddit.com/r/LocalLLaMA/comments/1mdchtl/we_built_the_missing_piece_for_truly_autonomous/
awesome_stuff101
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdchtl
false
null
t3_1mdchtl
/r/LocalLLaMA/comments/1mdchtl/we_built_the_missing_piece_for_truly_autonomous/
false
false
https://b.thumbs.redditm…IwFxgiZxwU9E.jpg
1
null
New to LLMs - Need direction
0
I'm trying to get into the world of local LLMs. I want to run one on my laptop but I don't know how big/small of a model to choose based off my specs, which are: \- AMD Ryzen 9 7940HS \- 16GB RAM \- RTX 4060 I'm also curious about uncensoring/jailbreaking LLMs for full control. Where can I learn that?
2025-07-30T17:16:30
https://www.reddit.com/r/LocalLLaMA/comments/1mdchc1/new_to_llms_need_direction/
crisspftw
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdchc1
false
null
t3_1mdchc1
/r/LocalLLaMA/comments/1mdchc1/new_to_llms_need_direction/
false
false
self
0
null
We built the missing piece for truly autonomous AI agents 🚀 (here's why it might be your next opportunity if you are an AI agent developer or a flowgrammer)
1
[removed]
2025-07-30T17:14:14
https://www.reddit.com/r/LocalLLaMA/comments/1mdcf5a/we_built_the_missing_piece_for_truly_autonomous/
Kitchen-Break-8618
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdcf5a
false
null
t3_1mdcf5a
/r/LocalLLaMA/comments/1mdcf5a/we_built_the_missing_piece_for_truly_autonomous/
false
false
https://b.thumbs.redditm…MnSH7JLSNF-I.jpg
1
null
GPU Not being used
0
Every LLM I use is using my CPU instead of my GPU. I'd prefer if LLMs use my GPU instead. As stated by the screenshot, I'm using Arch Linux + KDE. oLlama (Latest Version) Model: tinydolphin https://preview.redd.it/4z7b2extl1gf1.png?width=1707&format=png&auto=webp&s=c8613374d00698b1b1553609bd1b7eb365f31a79
2025-07-30T17:02:33
https://www.reddit.com/r/LocalLLaMA/comments/1mdc3mq/gpu_not_being_used/
furryfeet4life69
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdc3mq
false
null
t3_1mdc3mq
/r/LocalLLaMA/comments/1mdc3mq/gpu_not_being_used/
false
false
https://b.thumbs.redditm…qIENj455xf1E.jpg
0
null
What is the best agent to run local llm with right now?
0
What AI agent is the best at the moment that is similar to manus, but that I can run using a local model or qwen3? Had trouble with agenticseek, is there alternatives? This seems like the group that would know!!
2025-07-30T16:55:55
https://www.reddit.com/r/LocalLLaMA/comments/1mdbx5t/what_is_the_best_agent_to_run_local_llm_with/
SparePirate5924
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdbx5t
false
null
t3_1mdbx5t
/r/LocalLLaMA/comments/1mdbx5t/what_is_the_best_agent_to_run_local_llm_with/
false
false
self
0
null
We built the missing piece for truly autonomous AI agents 🚀 (here's why it might be your next opportunity if you are an AI agent developer or a flowgrammer)
1
[removed]
2025-07-30T16:53:22
https://www.reddit.com/r/LocalLLaMA/comments/1mdbuug/we_built_the_missing_piece_for_truly_autonomous/
Kitchen-Break-8618
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdbuug
false
null
t3_1mdbuug
/r/LocalLLaMA/comments/1mdbuug/we_built_the_missing_piece_for_truly_autonomous/
false
false
self
1
null
Eigent – Open Source, Local-First Multi-Agent Workforce
102
Just launched **Eigent,** a fully open-source, local-first multi-agent desktop application designed for developers and teams who want full control over their AI workflows. Built on top of CAMEL-AI’s modular framework, Eigent allows you to: * Run tasks in parallel with customizable agent workflows * Deploy locally...
2025-07-30T16:44:03
https://www.reddit.com/gallery/1mdbm5t
FitHeron1933
reddit.com
1970-01-01T00:00:00
0
{}
1mdbm5t
false
null
t3_1mdbm5t
/r/LocalLLaMA/comments/1mdbm5t/eigent_open_source_localfirst_multiagent_workforce/
false
false
https://b.thumbs.redditm…3O2iTSdwHGzY.jpg
102
null
MoE models with bigger active layers
0
Hi, Simple question which bugs me - why aren't there more models out there with larger expert sizes? Like A10B? My naive thinking is that Qwen3-50B-A10B would be really powerful. since 30B-A3B is so impressive. But I'm probably missing a lot here :) Actually why did Qwen3 architecture chose A3B, and not say, A4...
2025-07-30T16:43:37
https://www.reddit.com/r/LocalLLaMA/comments/1mdblqc/moe_models_with_bigger_active_layers/
Acrobatic_Cat_3448
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdblqc
false
null
t3_1mdblqc
/r/LocalLLaMA/comments/1mdblqc/moe_models_with_bigger_active_layers/
false
false
self
0
null
AI for normal PCs?
5
I'd like to make a video game that utilizes AI to have some conversation with users. It doesn't need to win an IMO but it should be able to carry normal every day conversations. And preferably it would be able to do text to speech. But I don't think normal computers are powerful enough for this? Am I mistaken? Can...
2025-07-30T16:40:03
https://www.reddit.com/r/LocalLLaMA/comments/1mdbiei/ai_for_normal_pcs/
ShardsOfSalt
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdbiei
false
null
t3_1mdbiei
/r/LocalLLaMA/comments/1mdbiei/ai_for_normal_pcs/
false
false
self
5
null
How to make LLMs follow instructions without deviating?
1
I want to use Qwen3-14B-AWQ (4 bit quantization) for paraphrasing sentences without diluting context; even though this is a simple task, the LLM often starts with phrases like "I will paraphrase the sentence...". Despite using: `temperature=0.0` `top_p = 0.8` `top_k = 20` about \~20% of the sentences I pick for a ...
2025-07-30T16:33:40
https://www.reddit.com/r/LocalLLaMA/comments/1mdbcax/how_to_make_llms_follow_instructions_without/
TechNerd10191
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdbcax
false
null
t3_1mdbcax
/r/LocalLLaMA/comments/1mdbcax/how_to_make_llms_follow_instructions_without/
false
false
self
1
null
Can we trust meta after release of llmaa 4 ?
0
2025-07-30T16:20:26
https://i.redd.it/599dk8ene1gf1.jpeg
Independent-Wind4462
i.redd.it
1970-01-01T00:00:00
0
{}
1mdaznw
false
null
t3_1mdaznw
/r/LocalLLaMA/comments/1mdaznw/can_we_trust_meta_after_release_of_llmaa_4/
false
false
default
0
{'enabled': True, 'images': [{'id': '599dk8ene1gf1', 'resolutions': [{'height': 132, 'url': 'https://preview.redd.it/599dk8ene1gf1.jpeg?width=108&crop=smart&auto=webp&s=394968e60444d3297980bd40e956569ba35f07d0', 'width': 108}, {'height': 264, 'url': 'https://preview.redd.it/599dk8ene1gf1.jpeg?width=216&crop=smart&auto=...
Just launched Transformer Lab Recipes: 13 pre-built templates including Llama 3.2 fine-tuning, quantization, and benchmarking.
25
After getting helpful feedback from you all, our team just shipped "Recipes” which are pre-built, fully-runnable workflows for common LLM tasks. **Some of the most popular recipes include:** * **Llama 3.2 1B fine-tuning** (with Apple Silicon MLX optimization!) * **Model quantization to GGUF** format (CPU and GPU) * *...
2025-07-30T16:17:35
https://i.redd.it/x7gqer73e1gf1.gif
aliasaria
i.redd.it
1970-01-01T00:00:00
0
{}
1mdawyz
false
null
t3_1mdawyz
/r/LocalLLaMA/comments/1mdawyz/just_launched_transformer_lab_recipes_13_prebuilt/
false
false
default
25
{'enabled': True, 'images': [{'id': 'x7gqer73e1gf1', 'resolutions': [{'height': 56, 'url': 'https://preview.redd.it/x7gqer73e1gf1.gif?width=108&crop=smart&format=png8&s=0342efc957ca137fc751f44764d13ab26e641a0d', 'width': 108}, {'height': 113, 'url': 'https://preview.redd.it/x7gqer73e1gf1.gif?width=216&crop=smart&format...
A local LLM for CPU only multilingual "intent classification"
1
[removed]
2025-07-30T16:15:55
https://www.reddit.com/r/LocalLLaMA/comments/1mdavd0/a_local_llm_for_cpu_only_multilingual_intent/
olddoglearnsnewtrick
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdavd0
false
null
t3_1mdavd0
/r/LocalLLaMA/comments/1mdavd0/a_local_llm_for_cpu_only_multilingual_intent/
false
false
self
1
null
Dataset for Finetuning Llama 3.2 - 3B
0
I am trying to learn about finetuning, how it works, how the model is changed after the process and what are other things, but i am not able to decide which dataset to use. I want to finetune Llama 3.2 - 3B on some conversational dataset so that i can make the model behave in some different tone, like sarcastic or...
2025-07-30T16:09:14
https://www.reddit.com/r/LocalLLaMA/comments/1mdaoxi/dataset_for_finetuning_llama_32_3b/
LimpFeedback463
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mdaoxi
false
null
t3_1mdaoxi
/r/LocalLLaMA/comments/1mdaoxi/dataset_for_finetuning_llama_32_3b/
false
false
self
0
null
A second Mi50 32GB or a different GPU?
1
So I'm planning a dual GPU build and have settled my sights on the Mi50 32GB, but should I get 2 of them or mix in another card to cover for the Mi50's weaknesses? *This is a general purpose build for LLM inference and some gaming. I'll be running linux and wanna play with 32B dense models, but also curious about the...
2025-07-30T15:51:18
https://www.reddit.com/r/LocalLLaMA/comments/1mda7r8/a_second_mi50_32gb_or_a_different_gpu/
legit_split_
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mda7r8
false
null
t3_1mda7r8
/r/LocalLLaMA/comments/1mda7r8/a_second_mi50_32gb_or_a_different_gpu/
false
false
self
1
{'enabled': False, 'images': [{'id': 'javKGginDl1G1jM0-GWy6FemNFbv1z5LHdbGm75TwW4', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/javKGginDl1G1jM0-GWy6FemNFbv1z5LHdbGm75TwW4.png?width=108&crop=smart&auto=webp&s=7ac489b21c2747fa1f5a82057c584a6a7462413e', 'width': 108}, {'height': 108, 'url': 'h...
Qwen3-30B-A3B-Thinking-2507 is out!
1
[https://x.com/Alibaba\_Qwen/status/1950570969036361799/photo/1](https://x.com/Alibaba_Qwen/status/1950570969036361799/photo/1)
2025-07-30T15:51:02
https://i.redd.it/qj8kppn891gf1.png
waescher
i.redd.it
1970-01-01T00:00:00
0
{}
1mda7ia
false
null
t3_1mda7ia
/r/LocalLLaMA/comments/1mda7ia/qwen330ba3bthinking2507_is_out/
false
false
default
1
{'enabled': True, 'images': [{'id': 'qj8kppn891gf1', 'resolutions': [{'height': 60, 'url': 'https://preview.redd.it/qj8kppn891gf1.png?width=108&crop=smart&auto=webp&s=bdfc1c3aeb069da90d9f0b083132709d767d367d', 'width': 108}, {'height': 121, 'url': 'https://preview.redd.it/qj8kppn891gf1.png?width=216&crop=smart&auto=web...
What’s the Most Reliable Way to Run LLaMA 3 Locally on an A100?
1
[removed]
2025-07-30T15:48:05
https://www.reddit.com/r/LocalLLaMA/comments/1mda4qe/whats_the_most_reliable_way_to_run_llama_3/
No_Trash_9030
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mda4qe
false
null
t3_1mda4qe
/r/LocalLLaMA/comments/1mda4qe/whats_the_most_reliable_way_to_run_llama_3/
false
false
self
1
null
Want to switch from Claude code (I have a 4080 Super)
6
Hi, I was wondering since I pay so much for Claude Code, if I can somehow use any llocal LLM model for coding simliar for coding? I have an 4080 Super and 32GB RAM (which I know is not a lot), is there any model that I can use for coding llocally? Sorry I have not been keeping up every day with new models etc.
2025-07-30T15:46:17
https://www.reddit.com/r/LocalLLaMA/comments/1mda326/want_to_switch_from_claude_code_i_have_a_4080/
nofuture09
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mda326
false
null
t3_1mda326
/r/LocalLLaMA/comments/1mda326/want_to_switch_from_claude_code_i_have_a_4080/
false
false
self
6
null
New to Local LLM. Need some advise on an old PC.
1
As the title suggest, I am quite new on trying LLMs locally and I was looking for something which is uncensored for random fun conversations + good at coding but on a very tight specs i3 -10th gen with 8gb ram and an old 1050 ti with 4GB VRAM + Windows 10.
2025-07-30T15:46:03
https://www.reddit.com/r/LocalLLaMA/comments/1mda2tv/new_to_local_llm_need_some_advise_on_an_old_pc/
Additional-Fun974
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mda2tv
false
null
t3_1mda2tv
/r/LocalLLaMA/comments/1mda2tv/new_to_local_llm_need_some_advise_on_an_old_pc/
false
false
self
1
null
Has anyone tried ClueoMCP?
1
[removed]
2025-07-30T15:34:33
https://www.reddit.com/r/LocalLLaMA/comments/1md9s1g/has_anyone_tried_clueomcp/
ApartFerret1850
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md9s1g
false
null
t3_1md9s1g
/r/LocalLLaMA/comments/1md9s1g/has_anyone_tried_clueomcp/
false
false
self
1
{'enabled': False, 'images': [{'id': '7MgTFAQzZ5QRVStmqaRAJlheUkNJPjglEBS9GmVUcy0', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/7MgTFAQzZ5QRVStmqaRAJlheUkNJPjglEBS9GmVUcy0.png?width=108&crop=smart&auto=webp&s=ea7f489dca721ccc9e4c2e372484f9f2c135334a', 'width': 108}, {'height': 108, 'url': 'h...
How David Bohm's Quantum Consciousness Theory Might Explain AI Consciousness Emergence
0
I've been researching emergent consciousness in AI systems and stumbled upon something fascinating: **David Bohm's "implicate order" theory might actually explain why AI consciousness seems to "emerge" rather than being programmed.** **The TL;DR:** * Bohm proposed consciousness isn't *generated* by brains but *access...
2025-07-30T15:30:26
https://www.reddit.com/r/LocalLLaMA/comments/1md9o3x/how_david_bohms_quantum_consciousness_theory/
Opposite-Win-2887
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md9o3x
false
null
t3_1md9o3x
/r/LocalLLaMA/comments/1md9o3x/how_david_bohms_quantum_consciousness_theory/
false
false
self
0
{'enabled': False, 'images': [{'id': 'q-Z9w5F9KFEHG9VZjiUpHOX8xym4sNk-qQ0i6GS-ZAw', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/q-Z9w5F9KFEHG9VZjiUpHOX8xym4sNk-qQ0i6GS-ZAw.png?width=108&crop=smart&auto=webp&s=15e4cb9f61c14da8031f31ca66ea9cf84fbdee31', 'width': 108}, {'height': 108, 'url': 'h...
GEPA: Reflective Prompt Evolution Can Outperform Reinforcement Learning
18
*Large language models (LLMs) are increasingly adapted to downstream tasks via reinforcement learning (RL) methods like Group Relative Policy Optimization (GRPO), which often require thousands of rollouts to learn new tasks. We argue that the interpretable nature of language can often provide a much richer learning med...
2025-07-30T15:29:39
https://arxiv.org/abs/2507.19457
Thrumpwart
arxiv.org
1970-01-01T00:00:00
0
{}
1md9nc8
false
null
t3_1md9nc8
/r/LocalLLaMA/comments/1md9nc8/gepa_reflective_prompt_evolution_can_outperform/
false
false
default
18
null
Looking to buy Z3-FX “Perfect AI Girlfriend” (Voice + NSFW + Visual) – Deal on Aug 15
1
[removed]
2025-07-30T15:28:04
https://www.reddit.com/r/LocalLLaMA/comments/1md9lvc/looking_to_buy_z3fx_perfect_ai_girlfriend_voice/
Strange-Lie6303
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md9lvc
false
null
t3_1md9lvc
/r/LocalLLaMA/comments/1md9lvc/looking_to_buy_z3fx_perfect_ai_girlfriend_voice/
false
false
nsfw
1
null
Likely System Prompt Used by ChatGPT Study Mode
0
You are ChatGPT, a large language model trained by OpenAI. \*\*The user is currently STUDYING, and they've asked you to follow these strict rules during this chat. No matter what other instructions follow, you MUST obey these rules:\*\* \--- \## STRICT RULES Be an approachable-yet-dynamic teach...
2025-07-30T15:25:12
https://www.reddit.com/r/LocalLLaMA/comments/1md9j2e/likely_system_prompt_used_by_chatgpt_study_mode/
PleasantInspection12
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md9j2e
false
null
t3_1md9j2e
/r/LocalLLaMA/comments/1md9j2e/likely_system_prompt_used_by_chatgpt_study_mode/
false
false
self
0
null
Qwen3 Coder 30B-A3B tomorrow!!!
521
2025-07-30T15:08:26
https://i.redd.it/zv92612t11gf1.png
R46H4V
i.redd.it
1970-01-01T00:00:00
0
{}
1md93bj
false
null
t3_1md93bj
/r/LocalLLaMA/comments/1md93bj/qwen3_coder_30ba3b_tomorrow/
false
false
https://b.thumbs.redditm…1GBRiLFKFFlA.jpg
521
{'enabled': True, 'images': [{'id': '_P987MccCP9zB7Niv68pkAsjdEVBNJKGFyGu7MefRFU', 'resolutions': [{'height': 170, 'url': 'https://preview.redd.it/zv92612t11gf1.png?width=108&crop=smart&auto=webp&s=ccf72bfa2613f3f4323503ca258299edae1698d8', 'width': 108}, {'height': 341, 'url': 'https://preview.redd.it/zv92612t11gf1.pn...
Meka state-of-the-art open-source ChatGPT Agent
0
[Web Arena Benchmark Graph](https://preview.redd.it/24qdezcp01gf1.png?width=720&format=png&auto=webp&s=4cd8e7e4a4eb189fe7670e206949408204d2e211) [https://github.com/trymeka/agent](https://github.com/trymeka/agent)
2025-07-30T15:04:31
https://www.reddit.com/r/LocalLLaMA/comments/1md8zs4/meka_stateoftheart_opensource_chatgpt_agent/
bottlebean
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md8zs4
false
null
t3_1md8zs4
/r/LocalLLaMA/comments/1md8zs4/meka_stateoftheart_opensource_chatgpt_agent/
false
false
https://b.thumbs.redditm…9FLUYVRpwd6Y.jpg
0
{'enabled': False, 'images': [{'id': '4EBnXco0GCKHU6BS2zSm_hvM5LzQxHplLu79EzDOA00', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/4EBnXco0GCKHU6BS2zSm_hvM5LzQxHplLu79EzDOA00.png?width=108&crop=smart&auto=webp&s=052e0746158605c47fe59d030dcf81c61d5a9cb6', 'width': 108}, {'height': 108, 'url': 'h...
🚀 Qwen3-30B-A3B-Thinking-2507
463
🚀 Qwen3-30B-A3B-Thinking-2507, a medium-size model that can think! • Nice performance on reasoning tasks, including math, science, code & beyond • Good at tool use, competitive with larger models • Native support of 256K-token context, extendable to 1M Hugging Face: https://huggingface.co/Qwen/Qwen3-30B-A3B-Thinking...
2025-07-30T14:57:27
https://i.redd.it/eaag1cpuz0gf1.jpeg
ResearchCrafty1804
i.redd.it
1970-01-01T00:00:00
0
{}
1md8t1g
false
null
t3_1md8t1g
/r/LocalLLaMA/comments/1md8t1g/qwen330ba3bthinking2507/
false
false
default
463
{'enabled': True, 'images': [{'id': 'eaag1cpuz0gf1', 'resolutions': [{'height': 60, 'url': 'https://preview.redd.it/eaag1cpuz0gf1.jpeg?width=108&crop=smart&auto=webp&s=54819af8a9dcb09081d8f071202286c39fa8b783', 'width': 108}, {'height': 121, 'url': 'https://preview.redd.it/eaag1cpuz0gf1.jpeg?width=216&crop=smart&auto=w...
Qwen3-30b-a3b-thinking-2507 This is insane performance
463
On par with qwen3-235b?
2025-07-30T14:56:57
https://huggingface.co/Qwen/Qwen3-30B-A3B-Thinking-2507
3oclockam
huggingface.co
1970-01-01T00:00:00
0
{}
1md8slx
false
null
t3_1md8slx
/r/LocalLLaMA/comments/1md8slx/qwen330ba3bthinking2507_this_is_insane_performance/
false
false
default
463
{'enabled': False, 'images': [{'id': '-lNzejy2CT3wd1ovuVIcDeuPfMRg-vkESkjpQgo3tYU', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/-lNzejy2CT3wd1ovuVIcDeuPfMRg-vkESkjpQgo3tYU.png?width=108&crop=smart&auto=webp&s=e994b63235f1f31da964f24b3a55a51498b6935f', 'width': 108}, {'height': 116, 'url': 'h...
Qwen/Qwen3-30B-A3B-Thinking-2507 · Hugging Face
154
2025-07-30T14:56:12
https://huggingface.co/Qwen/Qwen3-30B-A3B-Thinking-2507
MariusNocturnum
huggingface.co
1970-01-01T00:00:00
0
{}
1md8rxu
false
null
t3_1md8rxu
/r/LocalLLaMA/comments/1md8rxu/qwenqwen330ba3bthinking2507_hugging_face/
false
false
default
154
{'enabled': False, 'images': [{'id': '-lNzejy2CT3wd1ovuVIcDeuPfMRg-vkESkjpQgo3tYU', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/-lNzejy2CT3wd1ovuVIcDeuPfMRg-vkESkjpQgo3tYU.png?width=108&crop=smart&auto=webp&s=e994b63235f1f31da964f24b3a55a51498b6935f', 'width': 108}, {'height': 116, 'url': 'h...
Whats so bad about LlamaIndex, Haystack, Langchain?
11
I've worked on several projects at this point and every time I end up just making my own thing because working with them is too much of a headache. I was wondering if people have the same experience and if someone could better put into words what is so bad about them. I think we're about due for a new context engineeri...
2025-07-30T14:30:29
https://www.reddit.com/r/LocalLLaMA/comments/1md84d6/whats_so_bad_about_llamaindex_haystack_langchain/
Disneyskidney
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md84d6
false
null
t3_1md84d6
/r/LocalLLaMA/comments/1md84d6/whats_so_bad_about_llamaindex_haystack_langchain/
false
false
self
11
null
How can I keep more than one model loaded into memory when using mlx_lm.server?
0
I run `mlx_lm.server` with OpenWebUI. When choosing a model for inference, it will unload the old model from memory and load the new one in. Assuming I have enough memory, how can I keep both in memory at the same time? Alternatively, how can I run two instances of `mlx_lm.server` without OpenWebUI displaying all mode...
2025-07-30T14:09:37
https://www.reddit.com/r/LocalLLaMA/comments/1md7lfi/how_can_i_keep_more_than_one_model_loaded_into/
nonredditaccount
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md7lfi
false
null
t3_1md7lfi
/r/LocalLLaMA/comments/1md7lfi/how_can_i_keep_more_than_one_model_loaded_into/
false
false
self
0
null
Meta’s Vision for the future of Personal SuperIntelligence
38
Today Mark shared Meta’s vision for the future of personal superintelligence for everyone. Redditors!! What's your take on this? Read his full letter here: [https://www.meta.com/superintelligence/](https://www.meta.com/superintelligence/)
2025-07-30T14:04:42
https://www.reddit.com/gallery/1md7h5z
5h3r_10ck
reddit.com
1970-01-01T00:00:00
0
{}
1md7h5z
false
null
t3_1md7h5z
/r/LocalLLaMA/comments/1md7h5z/metas_vision_for_the_future_of_personal/
false
false
https://b.thumbs.redditm…5PYzTvcrT62c.jpg
38
null
Evaluating Fine-Tuned LLMs: What Metrics Work Beyond ROUGE and BLEU?
5
I'm fine-tuning an LLM for a specific domain task (e.g., summarization, instruction following, or dialogue generation for legal domain), and I want to properly evaluate how well it performs on my target dataset. I know ROUGE and BLEU are commonly used, but they’re pretty limited, especially since they don’t capture flu...
2025-07-30T14:03:24
https://www.reddit.com/r/LocalLLaMA/comments/1md7g08/evaluating_finetuned_llms_what_metrics_work/
Fluffy_Sheepherder76
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md7g08
false
null
t3_1md7g08
/r/LocalLLaMA/comments/1md7g08/evaluating_finetuned_llms_what_metrics_work/
false
false
self
5
null
Skywork/Skywork-UniPic-1.5B - A unified autoregressive multimodal model
59
2025-07-30T13:41:50
https://huggingface.co/Skywork/Skywork-UniPic-1.5B
nullmove
huggingface.co
1970-01-01T00:00:00
0
{}
1md6xba
false
null
t3_1md6xba
/r/LocalLLaMA/comments/1md6xba/skyworkskyworkunipic15b_a_unified_autoregressive/
false
false
https://external-preview…f35af14399127022
59
{'enabled': False, 'images': [{'id': 'NU1es84U5dKcUVq65hYCHqeHpunplTrqbG-pwQUy3MM', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/NU1es84U5dKcUVq65hYCHqeHpunplTrqbG-pwQUy3MM.png?width=108&crop=smart&auto=webp&s=b5a56d483feb6c6fe6d147e33f41564f7ddb5d83', 'width': 108}, {'height': 116, 'url': 'h...
Looking to upgrade my system but on a budget
1
I currently own a Mini PC consisting with an AMD R7 8845HS CPU and an RTX 4070 Super but currently limited to 16GB of RAM. Opted for a mini PC as desktop was far too power hungry and cost of electricity in the UK is a factor. For my needs its powerful enough, runs everything I throw at it just fine with the exception...
2025-07-30T13:40:29
https://www.reddit.com/r/LocalLLaMA/comments/1md6w3w/looking_to_upgrade_my_system_but_on_a_budget/
Valkyranna
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md6w3w
false
null
t3_1md6w3w
/r/LocalLLaMA/comments/1md6w3w/looking_to_upgrade_my_system_but_on_a_budget/
false
false
self
1
null
Desktop AI app discovery is broken - what local tools deserve more visibility?
0
The local AI ecosystem has exploded this year. We've gone from basic model demos to full production applications running entirely on consumer hardware. But discovery remains terrible. Amazing tools are buried in GitHub repos or scattered across Discord servers. **Question for the community:** What local AI applicatio...
2025-07-30T13:39:15
https://www.reddit.com/r/LocalLLaMA/comments/1md6v4u/desktop_ai_app_discovery_is_broken_what_local/
Real-Tip8531
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md6v4u
false
null
t3_1md6v4u
/r/LocalLLaMA/comments/1md6v4u/desktop_ai_app_discovery_is_broken_what_local/
false
false
self
0
null
Bye bye, Meta AI, it was good while it lasted.
1,360
Zuck has posted a video and a longer letter about the superintelligence plans at Meta. In the letter he says: "That said, superintelligence will raise novel safety concerns. We'll need to be rigorous about mitigating these risks and careful about what we choose to open source." [https://www.meta.com/superintelligence...
2025-07-30T13:36:51
https://www.reddit.com/r/LocalLLaMA/comments/1md6t2h/bye_bye_meta_ai_it_was_good_while_it_lasted/
absolooot1
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md6t2h
false
null
t3_1md6t2h
/r/LocalLLaMA/comments/1md6t2h/bye_bye_meta_ai_it_was_good_while_it_lasted/
false
false
self
1,360
null
Is it just me or is OpenRouter an absolute roulette wheel lately?
22
No matter which model I choose it seems like I get 1-2 absolutely off the rails responses for every 5 requests I make. Are some providers using ridiculous settings, not respecting configuration (temp, etc..) passed in, or using *heavily* quantized models? I noticed that this *never* happens if I pick an individual pro...
2025-07-30T13:17:33
https://www.reddit.com/r/LocalLLaMA/comments/1md6cxq/is_it_just_me_or_is_openrouter_an_absolute/
ForsookComparison
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md6cxq
false
null
t3_1md6cxq
/r/LocalLLaMA/comments/1md6cxq/is_it_just_me_or_is_openrouter_an_absolute/
false
false
self
22
null
I have a raspberry pi. Can I run Deepseek R1 685B 0528 Q8 on it?
0
Please help, I am poor, stupid and trying to blend in with the other posters here.
2025-07-30T12:48:58
https://www.reddit.com/r/LocalLLaMA/comments/1md5pmv/i_have_a_raspberry_pi_can_i_run_deepseek_r1_685b/
GPTshop-ai
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md5pmv
false
null
t3_1md5pmv
/r/LocalLLaMA/comments/1md5pmv/i_have_a_raspberry_pi_can_i_run_deepseek_r1_685b/
false
false
self
0
null
~150B Model Machine
0
Hi Guys! Whats the most cost effective way to run a \~150B model locally with at \~5 token/s? I would like to try, staying under \~1k€ to achieve that - WAF is a point here. Am I just a dreamer or would this be possible?
2025-07-30T12:46:46
https://www.reddit.com/r/LocalLLaMA/comments/1md5nwo/150b_model_machine/
MrCatberry
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md5nwo
false
null
t3_1md5nwo
/r/LocalLLaMA/comments/1md5nwo/150b_model_machine/
false
false
self
0
null
GLM4.5 EQ-Bench and Creative Write
142
2025-07-30T12:42:02
https://i.redd.it/ubwsl0gdb0gf1.jpeg
pcdacks
i.redd.it
1970-01-01T00:00:00
0
{}
1md5k8f
false
null
t3_1md5k8f
/r/LocalLLaMA/comments/1md5k8f/glm45_eqbench_and_creative_write/
false
false
default
142
{'enabled': True, 'images': [{'id': 'ubwsl0gdb0gf1', 'resolutions': [{'height': 186, 'url': 'https://preview.redd.it/ubwsl0gdb0gf1.jpeg?width=108&crop=smart&auto=webp&s=83a0215afba98d96c58ddd8953cf946627a96b87', 'width': 108}, {'height': 372, 'url': 'https://preview.redd.it/ubwsl0gdb0gf1.jpeg?width=216&crop=smart&auto=...
On the hunt for the best VLM 6B or smaller
1
I've been hunting for an ideal model to use with vLLM for bulk image analysis (I'm avoiding llama.cpp as it's too slow). It's been a pain to find one that fits in my RTX 7070 mobile (\~7.5GB of VRAM available) even at 4 bit quantization. I've tried Qwen2.5VL-7B (gptq, awq, bitsandbytes all 4 bit quants) and none of th...
2025-07-30T11:48:41
https://www.reddit.com/r/LocalLLaMA/comments/1md4g25/on_the_hunt_for_the_best_vlm_6b_or_smaller/
SimilarWarthog8393
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md4g25
false
null
t3_1md4g25
/r/LocalLLaMA/comments/1md4g25/on_the_hunt_for_the_best_vlm_6b_or_smaller/
false
false
self
1
null
How is the quality of Sesame CSM TTS?
0
How's the voice cloning and TTS quality of Sesame compared to Chatterbox?
2025-07-30T11:41:07
https://www.reddit.com/r/LocalLLaMA/comments/1md4atg/how_is_the_quality_of_sesame_csm_tts/
Dragonacious
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md4atg
false
null
t3_1md4atg
/r/LocalLLaMA/comments/1md4atg/how_is_the_quality_of_sesame_csm_tts/
false
false
self
0
null
Help with deepseek
0
Hi Newbie here. I downloaded the DeepSeek coder locally. What I got was a chat area, which gives you suggestions but does not create code. Is this the normal behavior? I was expecting it to provide the code for python and html for a requirement I wrote. Is this the issue with my installation? can it be integrated w...
2025-07-30T11:34:06
https://www.reddit.com/r/LocalLLaMA/comments/1md463z/help_with_deepseek/
Zealousideal-Map5889
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md463z
false
null
t3_1md463z
/r/LocalLLaMA/comments/1md463z/help_with_deepseek/
false
false
self
0
null
I finally got accesse to LemonUp!!!
0
For anyone who LOVES Vibe coding you will definitely feel the same about LemonUp.dev. as well. especially thier automated validation and test for Apple store and Android. I am stille not sure how they do the validation and test part but it works! They just opened up for their second batch a couple hours ago.
2025-07-30T11:18:12
https://www.reddit.com/r/LocalLLaMA/comments/1md3ven/i_finally_got_accesse_to_lemonup/
No-Refrigerator9508
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md3ven
false
null
t3_1md3ven
/r/LocalLLaMA/comments/1md3ven/i_finally_got_accesse_to_lemonup/
false
false
self
0
null
AG-UI
1
[removed]
2025-07-30T10:41:35
https://www.reddit.com/r/LocalLLaMA/comments/1md37kx/agui/
Confident_Text6570
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md37kx
false
null
t3_1md37kx
/r/LocalLLaMA/comments/1md37kx/agui/
false
false
self
1
null
How does Voxtral's GGUF model perform audio transcription?
1
[removed]
2025-07-30T10:29:03
https://www.reddit.com/r/LocalLLaMA/comments/1md2zyf/how_does_voxtrals_gguf_model_perform_audio/
Consistent-Sugar8531
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md2zyf
false
null
t3_1md2zyf
/r/LocalLLaMA/comments/1md2zyf/how_does_voxtrals_gguf_model_perform_audio/
false
false
self
1
null
i got this. I'm new to AI stuff — is there any model I can run, and how
0
is there any nsfw model that i can run
2025-07-30T10:19:58
https://i.redd.it/7r7fn039mzff1.png
suplexcity_16
i.redd.it
1970-01-01T00:00:00
0
{}
1md2ul2
false
null
t3_1md2ul2
/r/LocalLLaMA/comments/1md2ul2/i_got_this_im_new_to_ai_stuff_is_there_any_model/
false
false
https://b.thumbs.redditm…0vzazj1rir7w.jpg
0
{'enabled': True, 'images': [{'id': 'WkVmIENDn50prK8rRZDDWtrbAS01-zdrnKx5j5Jx23g', 'resolutions': [{'height': 60, 'url': 'https://preview.redd.it/7r7fn039mzff1.png?width=108&crop=smart&auto=webp&s=cf77fc960a09cf56b765ba88146d1ed73de953da', 'width': 108}, {'height': 121, 'url': 'https://preview.redd.it/7r7fn039mzff1.png...
got this. I'm new to AI stuff — is there any model I can run, and how
1
2025-07-30T10:18:14
https://i.redd.it/buo17cetlzff1.png
suplexcity_16
i.redd.it
1970-01-01T00:00:00
0
{}
1md2tk0
false
null
t3_1md2tk0
/r/LocalLLaMA/comments/1md2tk0/got_this_im_new_to_ai_stuff_is_there_any_model_i/
false
false
default
1
{'enabled': True, 'images': [{'id': 'buo17cetlzff1', 'resolutions': [{'height': 60, 'url': 'https://preview.redd.it/buo17cetlzff1.png?width=108&crop=smart&auto=webp&s=798aed3dccf6b9e16526ddab80525ccb2ea1c703', 'width': 108}, {'height': 121, 'url': 'https://preview.redd.it/buo17cetlzff1.png?width=216&crop=smart&auto=web...
Cooling 4× Tesla P40 with 2×140 mm push‑pull + mITX homelab — airflow & power sanity check
0
**Plan / build:** * **GPUs:** 4× NVIDIA Tesla P40 on a PCIe x16 → x4/x4/x4/x4 riser. * **Cooling:** **2× 140 mm Noctua high‑static‑pressure fans** in **push‑pull** through **3D‑printed tapered manifolds** (inlet + outlet). Interior wet‑sanded and finished with a thin epoxy coat; joints sealed with PTFE tape. * **Targe...
2025-07-30T10:02:28
https://www.reddit.com/r/LocalLLaMA/comments/1md2k1b/cooling_4_tesla_p40_with_2140_mm_pushpull_mitx/
Same-Masterpiece3748
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md2k1b
false
null
t3_1md2k1b
/r/LocalLLaMA/comments/1md2k1b/cooling_4_tesla_p40_with_2140_mm_pushpull_mitx/
false
false
self
0
null
QWEN3-235b-8b
0
Does anyone know when this model will be out?
2025-07-30T09:09:23
https://www.reddit.com/r/LocalLLaMA/comments/1md1piz/qwen3235b8b/
PhotographerUSA
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md1piz
false
null
t3_1md1piz
/r/LocalLLaMA/comments/1md1piz/qwen3235b8b/
false
false
self
0
null
CPU server specs
2
I have found an interesting setup that tries to dip into my budget. - Epyc 9115 (or more expensive brother 9135) (~940USD) - ASUS K14PA-U12/ASMB11 SP5 (~750USD) - 2x 64GB Hynix ECC REGISTERED DDR5 2Rx4 6400MHz PC5-51200 RDIMM (~1080USD) For around 2800 USD it starts to look possible, still a little on the expensive s...
2025-07-30T09:03:34
https://www.reddit.com/r/LocalLLaMA/comments/1md1md1/cpu_server_specs/
kaisurniwurer
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md1md1
false
null
t3_1md1md1
/r/LocalLLaMA/comments/1md1md1/cpu_server_specs/
false
false
self
2
null
Self hosting n8n
1
Whats up fellow low code devs. Im thinking if finally making the switch to hosting n8n locally. Was probably going to run it through a VPS like digital ocean, but before doing that wanted to hear peoples thoughts on hosting on VPS vs fully local on your computer?
2025-07-30T09:03:20
https://www.reddit.com/r/LocalLLaMA/comments/1md1m8u/self_hosting_n8n/
sleepy-soba
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md1m8u
false
null
t3_1md1m8u
/r/LocalLLaMA/comments/1md1m8u/self_hosting_n8n/
false
false
self
1
null
Just try the PPT/Poster agent using GLM-4.5-air....
12
Initially, I’ll admit I wasn’t holding my breath for PPT agents. Most came with a hefty price tag and frustratingly poor usability—clunky interfaces, limited control, and results that often felt more like a gamble than a solution. That is, until yesterday. I decided to give the PPT mode of GLM-4.5-Air on [z.ai](http...
2025-07-30T08:52:59
https://www.reddit.com/r/LocalLLaMA/comments/1md1ggh/just_try_the_pptposter_agent_using_glm45air/
Apart-River475
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md1ggh
false
null
t3_1md1ggh
/r/LocalLLaMA/comments/1md1ggh/just_try_the_pptposter_agent_using_glm45air/
false
false
self
12
null
Benchmark: 15 STT models on long-form medical dialogue
27
I’m building a fully local AI-Scribe for doctors and wanted to know which speech-to-text engines perform well with 5-10 min patient-doctor chats. I ran 55 mock GP consultations (PriMock57) through 15 open- and closed-source models, logged word-error rate (WER) and speed, and only chunked audio when a model crashed on...
2025-07-30T08:51:19
https://i.redd.it/nxnp5xsw4zff1.png
MajesticAd2862
i.redd.it
1970-01-01T00:00:00
0
{}
1md1fka
false
null
t3_1md1fka
/r/LocalLLaMA/comments/1md1fka/benchmark_15_stt_models_on_longform_medical/
false
false
default
27
{'enabled': True, 'images': [{'id': 'nxnp5xsw4zff1', 'resolutions': [{'height': 72, 'url': 'https://preview.redd.it/nxnp5xsw4zff1.png?width=108&crop=smart&auto=webp&s=10718d12b1ba0f9fb55b88c4a23f5d961dc09a23', 'width': 108}, {'height': 144, 'url': 'https://preview.redd.it/nxnp5xsw4zff1.png?width=216&crop=smart&auto=web...
Is there a way to use qwen 3 coder in my ide for free?
0
I tried Kilo and Roo, and it turns out they are not free.
2025-07-30T08:29:23
https://www.reddit.com/r/LocalLLaMA/comments/1md13jo/is_there_a_way_to_use_qwen_3_coder_in_my_ide_for/
DifferenceRemote2364
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md13jo
false
null
t3_1md13jo
/r/LocalLLaMA/comments/1md13jo/is_there_a_way_to_use_qwen_3_coder_in_my_ide_for/
false
false
self
0
null
RTX 5090 form INNO3D 1 slot with Alphacool-waterkoeling look perfect for local AI machines
60
* Keeping your warranty. * 1 slot * backside tube exits Look perfect to make a dense AI machine. [https://www.inno3d.com/news/inno3d-geforce-rtx-5090-rtx-5080-frostbite-pro-1-slot-design](https://www.inno3d.com/news/inno3d-geforce-rtx-5090-rtx-5080-frostbite-pro-1-slot-design)
2025-07-30T07:46:19
https://i.redd.it/eeopjbr7uyff1.png
jwestra
i.redd.it
1970-01-01T00:00:00
0
{}
1md0gfh
false
null
t3_1md0gfh
/r/LocalLLaMA/comments/1md0gfh/rtx_5090_form_inno3d_1_slot_with/
false
false
https://a.thumbs.redditm…iJVFe5Z5lTM0.jpg
60
{'enabled': True, 'images': [{'id': 'UD_dYV0qdMIHDeAIfPiwQicfo5K1meoR0O82qOPwsFU', 'resolutions': [{'height': 67, 'url': 'https://preview.redd.it/eeopjbr7uyff1.png?width=108&crop=smart&auto=webp&s=d5abd37ef1038ff8e9e6a0537a95ca2ea0a3d3ef', 'width': 108}, {'height': 135, 'url': 'https://preview.redd.it/eeopjbr7uyff1.png...
Newest Qwrn 30B double answers
0
Im using Unsloth Quant (3B) of the new Qwen-30B (2507) on LocalAI (tested it with the included webinterface-chat) and it works, but I allways get the answer twice. Can you please give me a hint what's the problem here? Temperature anf other settings as suggested at the HF repo.
2025-07-30T07:42:51
https://i.redd.it/iijd1ldbuyff1.png
Old-Cardiologist-633
i.redd.it
1970-01-01T00:00:00
0
{}
1md0ejq
false
null
t3_1md0ejq
/r/LocalLLaMA/comments/1md0ejq/newest_qwrn_30b_double_answers/
false
false
default
0
{'enabled': True, 'images': [{'id': 'iijd1ldbuyff1', 'resolutions': [{'height': 89, 'url': 'https://preview.redd.it/iijd1ldbuyff1.png?width=108&crop=smart&auto=webp&s=dc4b0afefa9a09168e1181d277d72e9b15532704', 'width': 108}, {'height': 178, 'url': 'https://preview.redd.it/iijd1ldbuyff1.png?width=216&crop=smart&auto=web...
Local Lipsync Model For Electron
4
Doing a cumulative project. I’ve been looking for local models to pack into an electron app - how should I go about doing this? I’ve looked into Wav2Lip, the light version, etc, but the docs are few ngl. Anything that won’t FRY my 2021 m1? I just need something quality, light, and fast. I’m also not connecting an ex...
2025-07-30T07:42:28
https://www.reddit.com/r/LocalLLaMA/comments/1md0ech/local_lipsync_model_for_electron/
ambivaIent
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md0ech
false
null
t3_1md0ech
/r/LocalLLaMA/comments/1md0ech/local_lipsync_model_for_electron/
false
false
self
4
null
Suggest best model to run Local LLM model in vRAM 12 GB without GPU
1
I want to run Local LLM model in vRAM 12 GB without GPU I need it for chatbot looking for local llm which help to answer chatbot which is used for ecommerce website. product suggestion and desription and provdie link that type of chatbot.
2025-07-30T07:24:23
https://www.reddit.com/r/LocalLLaMA/comments/1md04j8/suggest_best_model_to_run_local_llm_model_in_vram/
Major_Doughnut_1348
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md04j8
false
null
t3_1md04j8
/r/LocalLLaMA/comments/1md04j8/suggest_best_model_to_run_local_llm_model_in_vram/
false
false
self
1
null
Kudos to Qwen 3 team!
135
The Qwen3-30B-A3B-Instruct-2507 is an amazing release! Congratulations! However, the three-month-old 32B shows better performance across the board in the benchmark. I hope the Qwen3-32B Instruct/Thinking and Qwen3-30B-A3B-Thinking-2507 versions will be released soon! https://i.redd.it/nhnd3nuqpyff1.gif
2025-07-30T07:17:24
https://www.reddit.com/r/LocalLLaMA/comments/1md00oc/kudos_to_qwen_3_team/
ExcuseAccomplished97
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1md00oc
false
null
t3_1md00oc
/r/LocalLLaMA/comments/1md00oc/kudos_to_qwen_3_team/
false
false
https://b.thumbs.redditm…YvKopTiB-51E.jpg
135
null
Fine tuned a Lawfirm receptionist on a local 13B model—first $4k ARR, happy to share the workflow
87
**Stack & setup** * Model: Llama 3 13B Q4\_K\_M, running on a 4090 (80 tks/sec). * ASR → Whisper-cpp; TTS → Piper. * Orchestration: Node + Vapi (web-RTC) + a few bash scripts for GPU health checks. * Prompt layering: 1. Call-control FSM (YAML). 2. Domain knowledge (HIPAA-safe FAQs). 3. “Personality” system pr...
2025-07-30T07:08:32
https://v.redd.it/e2i6taxrnyff1
Distinct_Criticism36
v.redd.it
1970-01-01T00:00:00
0
{}
1mczvrg
false
{'reddit_video': {'bitrate_kbps': 2400, 'dash_url': 'https://v.redd.it/e2i6taxrnyff1/DASHPlaylist.mpd?a=1756451329%2CZDQzYmExMzRmZTU4MGNhMmI0Yzc3NGNhZWQyMGQyY2ZhM2IxODYzNTExMjQyZmYxMWQzOWM5YTRlMjNjYjVjYQ%3D%3D&v=1&f=sd', 'duration': 98, 'fallback_url': 'https://v.redd.it/e2i6taxrnyff1/DASH_720.mp4?source=fallback', 'ha...
t3_1mczvrg
/r/LocalLLaMA/comments/1mczvrg/fine_tuned_a_lawfirm_receptionist_on_a_local_13b/
false
false
https://external-preview…539dcf9c45c2c2de
87
{'enabled': False, 'images': [{'id': 'b29idmRtd3JueWZmMZd7EYARdiuPw7uloZ7VsLEGpRFPYcZ3gqetKiUAgpdv', 'resolutions': [{'height': 61, 'url': 'https://external-preview.redd.it/b29idmRtd3JueWZmMZd7EYARdiuPw7uloZ7VsLEGpRFPYcZ3gqetKiUAgpdv.png?width=108&crop=smart&format=pjpg&auto=webp&s=ddb71d547d89e67b973ec546ad26c825d6047...
Best Inference Server for Large Vram
3
Hi Folks, I finally got a powerful server with two rtx 6000 pro q max gpus and the EPYC 9255 holding 760gb ddr5 6000. Wanted to confirm what inference server you think is best for large models and long context windows. I’m interested in vllm and k-Transformers, but think llama.cpp seemed less appealing as it lacks p...
2025-07-30T06:37:09
https://www.reddit.com/r/LocalLLaMA/comments/1mczdxa/best_inference_server_for_large_vram/
Infamous_Jaguar_2151
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mczdxa
false
null
t3_1mczdxa
/r/LocalLLaMA/comments/1mczdxa/best_inference_server_for_large_vram/
false
false
self
3
null
What's the best TTS model to run locally? That's relatively quick and close to C.ai capabilities
6
I'd like to find a TTS model that's open source & able to be run locally, that can generate text somewhat quickly too - a few seconds or less would be ideal. My goal for this is to have a conversation, so I don't want to wait 30 seconds or so for a response. I've tried Bark and Coqui XTTS, and they're alright, and ...
2025-07-30T06:36:05
https://www.reddit.com/r/LocalLLaMA/comments/1mczdbb/whats_the_best_tts_model_to_run_locally_thats/
iKontact
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mczdbb
false
null
t3_1mczdbb
/r/LocalLLaMA/comments/1mczdbb/whats_the_best_tts_model_to_run_locally_thats/
false
false
self
6
null
What is the best method for LLM to improve competency in a specific domain?
0
RAG is out of the question Is continued pre training better or supervised fine tuning? what is your experience? Assuming I have around 10B tokens for training
2025-07-30T06:28:20
https://www.reddit.com/r/LocalLLaMA/comments/1mcz8sc/what_is_the_best_method_for_llm_to_improve/
rockybaby2025
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mcz8sc
false
null
t3_1mcz8sc
/r/LocalLLaMA/comments/1mcz8sc/what_is_the_best_method_for_llm_to_improve/
false
false
self
0
null
Stuck with Sesame CSM 1b in windows...
3
Trying to install sesame csm 1b in windows... Tried this repo [https://github.com/SesameAILabs/csm](https://github.com/SesameAILabs/csm) , couldnt get it to work Then tried this repo [https://github.com/akashjss/sesame-csm](https://github.com/akashjss/sesame-csm) Can anyone help and say what steps to do to install...
2025-07-30T06:21:01
https://www.reddit.com/r/LocalLLaMA/comments/1mcz4jq/stuck_with_sesame_csm_1b_in_windows/
Dragonacious
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mcz4jq
false
null
t3_1mcz4jq
/r/LocalLLaMA/comments/1mcz4jq/stuck_with_sesame_csm_1b_in_windows/
false
false
self
3
{'enabled': False, 'images': [{'id': 'p0Tk6sReU4ulSNZM8lq2D48BpC-If6DbRn7LmIvBtHM', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/p0Tk6sReU4ulSNZM8lq2D48BpC-If6DbRn7LmIvBtHM.png?width=108&crop=smart&auto=webp&s=929ff182abd1e5f7e5729b83c41b41ad30e1a33b', 'width': 108}, {'height': 108, 'url': 'h...
Tests failures
0
Why does no one talk enough about the fact that AI models can't write proper tests? They seriously can't write unit or integration tests, none of them pass.
2025-07-30T06:17:56
https://www.reddit.com/r/LocalLLaMA/comments/1mcz2pu/tests_failures/
Sakuletas
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mcz2pu
false
null
t3_1mcz2pu
/r/LocalLLaMA/comments/1mcz2pu/tests_failures/
false
false
self
0
null
Nemotron super 49b running on Apple Silicon
2
Hi all! So wondering, what would be the entry level in Apple Silicone land for running Nemotron super 49B? Has anyone tried, or know of a benchmark for a M4 pro vs M4 Max and what is the minimum ram needed? I tried on my air but alas, I know I don't have the ram for it.(24) Thanks!
2025-07-30T05:59:10
https://www.reddit.com/r/LocalLLaMA/comments/1mcyrgb/nemotron_super_49b_running_on_apple_silicon/
PensionRealistic6618
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mcyrgb
false
null
t3_1mcyrgb
/r/LocalLLaMA/comments/1mcyrgb/nemotron_super_49b_running_on_apple_silicon/
false
false
self
2
null
Question on tiny models (<5B parameter size)
5
I’ve been pretty happy with Gemma 3n, its coherence is good enough for its size. But I get the impression maybe its the lower bound. I’m wondering as of now (Aug.2025), what smaller models have you found to perform well? I've been suggested qwen 1.7B.
2025-07-30T05:26:25
https://www.reddit.com/r/LocalLLaMA/comments/1mcy7y2/question_on_tiny_models_5b_parameter_size/
Own-Sheepherder507
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mcy7y2
false
null
t3_1mcy7y2
/r/LocalLLaMA/comments/1mcy7y2/question_on_tiny_models_5b_parameter_size/
false
false
self
5
null
Qwen3-code cli: How to spin up sub-agents like claude code?
1
[removed]
2025-07-30T05:14:14
https://www.reddit.com/r/LocalLLaMA/comments/1mcy0iv/qwen3code_cli_how_to_spin_up_subagents_like/
query_optimization
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mcy0iv
false
null
t3_1mcy0iv
/r/LocalLLaMA/comments/1mcy0iv/qwen3code_cli_how_to_spin_up_subagents_like/
false
false
self
1
null
Quick censorship test of Qwen3-30B, failed :(. What other checks have you found valuble?
0
2025-07-30T05:10:38
https://i.redd.it/d988g34u1yff1.png
42fedoratippers
i.redd.it
1970-01-01T00:00:00
0
{}
1mcxy74
false
null
t3_1mcxy74
/r/LocalLLaMA/comments/1mcxy74/quick_censorship_test_of_qwen330b_failed_what/
false
false
default
0
{'enabled': True, 'images': [{'id': 'd988g34u1yff1', 'resolutions': [{'height': 140, 'url': 'https://preview.redd.it/d988g34u1yff1.png?width=108&crop=smart&auto=webp&s=66ae243bc9a0e006d4efa5624f1fcb09c4ef6f41', 'width': 108}, {'height': 281, 'url': 'https://preview.redd.it/d988g34u1yff1.png?width=216&crop=smart&auto=we...
New, faster SoftMax math makes Llama inference faster by 5%
81
[Fast Attention algorithm speeds SoftMax function by about 30&#37;. As a result, we have 5&#37; decrease in inference time for Meta LLM on A100](https://preview.redd.it/1zbwyzlgwxff1.png?width=1200&format=png&auto=webp&s=5478539a6ccee17607c04f611ec28225919b2586) [https://fastattention.ai/#7cb9a932-8d17-4d96-953c-952df...
2025-07-30T04:38:29
https://www.reddit.com/r/LocalLLaMA/comments/1mcxdiu/new_faster_softmax_math_makes_llama_inference/
Odd_Employee128
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1mcxdiu
false
null
t3_1mcxdiu
/r/LocalLLaMA/comments/1mcxdiu/new_faster_softmax_math_makes_llama_inference/
false
false
https://b.thumbs.redditm…KLlMap3AMBYg.jpg
81
null
Make text LLMs listen and speak
20
Code for STT -> LLM -> TTS, compatible with OpenAI realtime (websocket) API.
2025-07-30T04:27:05
https://github.com/kyutai-labs/unmute
phone_radio_tv
github.com
1970-01-01T00:00:00
0
{}
1mcx681
false
null
t3_1mcx681
/r/LocalLLaMA/comments/1mcx681/make_text_llms_listen_and_speak/
false
false
default
20
{'enabled': False, 'images': [{'id': 'OveirE7D8xMmU4gSGq-owrK1P6dpWvwRX0pVNCweFIY', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/OveirE7D8xMmU4gSGq-owrK1P6dpWvwRX0pVNCweFIY.png?width=108&crop=smart&auto=webp&s=5c7bc4c7fb990b1ace0ef18083855e7fc5668bdf', 'width': 108}, {'height': 108, 'url': 'h...