title
stringlengths
1
300
score
int64
0
8.54k
selftext
stringlengths
0
41.5k
created
timestamp[ns]date
2023-04-01 04:30:41
2026-03-04 02:14:14
url
stringlengths
0
878
author
stringlengths
3
20
domain
stringlengths
0
82
edited
timestamp[ns]date
1970-01-01 00:00:00
2026-02-19 14:51:53
gilded
int64
0
2
gildings
stringclasses
7 values
id
stringlengths
7
7
locked
bool
2 classes
media
stringlengths
646
1.8k
name
stringlengths
10
10
permalink
stringlengths
33
82
spoiler
bool
2 classes
stickied
bool
2 classes
thumbnail
stringlengths
4
213
ups
int64
0
8.54k
preview
stringlengths
301
5.01k
Testing Examsprint AI for Lightweight Study Assistance
1
I’ve been tinkering with ways to use AI for education, and one experiment surprised me — instant summaries + Q&A for textbooks actually worked really well. It’s a practical use case for lightweight AI that doesn’t feel gimmicky. I wonder — what’s the most useful “non-coding” application you’ve seen with LLaMA-style models so far? (link in comments for anyone interested in the tool I tested)
2025-09-22T16:57:14
https://i.redd.it/l1a4j4rfyqqf1.png
Relevant_Tiger8524
i.redd.it
1970-01-01T00:00:00
0
{}
1nnrzqe
false
null
t3_1nnrzqe
/r/LocalLLaMA/comments/1nnrzqe/testing_examsprint_ai_for_lightweight_study/
false
false
https://b.thumbs.redditm…EZv2Z1SgDU4g.jpg
1
{'enabled': True, 'images': [{'id': 'bThASb_dsvm2K9X6Vhtt7n9YvqbQqU0GrSLgzOFt4VU', 'resolutions': [{'height': 216, 'url': 'https://preview.redd.it/l1a4j4rfyqqf1.png?width=108&crop=smart&auto=webp&s=079d2d72005333f275e796096396c410e040ac68', 'width': 108}, {'height': 432, 'url': 'https://preview.redd.it/l1a4j4rfyqqf1.png?width=216&crop=smart&auto=webp&s=b2af06420de4d35269c62f137fd23337fc82066c', 'width': 216}, {'height': 640, 'url': 'https://preview.redd.it/l1a4j4rfyqqf1.png?width=320&crop=smart&auto=webp&s=18f345f6cd5005c3c0a216fed38242b150624034', 'width': 320}, {'height': 1280, 'url': 'https://preview.redd.it/l1a4j4rfyqqf1.png?width=640&crop=smart&auto=webp&s=57aac20daf0eb2003a937b0d3b46fea6825aac9e', 'width': 640}, {'height': 1920, 'url': 'https://preview.redd.it/l1a4j4rfyqqf1.png?width=960&crop=smart&auto=webp&s=c559c93b948c85f17021af646daebda29a3410c3', 'width': 960}, {'height': 2160, 'url': 'https://preview.redd.it/l1a4j4rfyqqf1.png?width=1080&crop=smart&auto=webp&s=03be9aad9ff66adef95520b3745e3aa3afff08f9', 'width': 1080}], 'source': {'height': 2400, 'url': 'https://preview.redd.it/l1a4j4rfyqqf1.png?auto=webp&s=9f270b64e495d3bd5ef65f74e1ebcb82a91a5038', 'width': 1080}, 'variants': {}}]}
How and where to start when you want a local llm model for your specific needs
3
I have a big project (lua) that was handed over to me. Since it's too big, i can't read it all by myself. How do i fine tune or feed the entire code base into the model so it can help me search/modify the codebase? Training a new model is obviously out of the question because i only have an RTX 4070. I already have an Ollama running qwen3:14b running on my PC but it doesn't do quite well what i need.
2025-09-22T16:46:47
https://www.reddit.com/r/LocalLLaMA/comments/1nnrpn3/how_and_where_to_start_when_you_want_a_local_llm/
timuela
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnrpn3
false
null
t3_1nnrpn3
/r/LocalLLaMA/comments/1nnrpn3/how_and_where_to_start_when_you_want_a_local_llm/
false
false
self
3
null
Qwen releases API (only) of Qwen3-TTS-Flash
21
🎙️ Meet Qwen3-TTS-Flash — the new text-to-speech model that’s redefining voice AI! Demo: https://huggingface.co/spaces/Qwen/Qwen3-TTS-Demo Blog: https://qwen.ai/blog?id=b4264e11fb80b5e37350790121baf0a0f10daf82&from=research.latest-advancements-list Video: https://youtu.be/MC6s4TLwX0A ✅ Best-in-class Chinese & English stability 🌍 SOTA multilingual WER for CN, EN, IT, FR 🎭 17 expressive voices × 10 languages 🗣️ Supports 9+ Chinese dialects: Cantonese, Hokkien, Sichuanese & more ⚡ Ultra-fast: First packet in just 97ms 🤖 Auto tone adaptation + robust text handling Perfect for apps, games, IVR, content — anywhere you need natural, human-like speech.
2025-09-22T16:36:41
https://i.redd.it/ojxbaozruqqf1.jpeg
ResearchCrafty1804
i.redd.it
1970-01-01T00:00:00
0
{}
1nnrftm
false
null
t3_1nnrftm
/r/LocalLLaMA/comments/1nnrftm/qwen_releases_api_only_of_qwen3ttsflash/
false
false
default
21
{'enabled': True, 'images': [{'id': 'ojxbaozruqqf1', 'resolutions': [{'height': 68, 'url': 'https://preview.redd.it/ojxbaozruqqf1.jpeg?width=108&crop=smart&auto=webp&s=c44742dd470c061cf11bf3bb7db6adaff55f344f', 'width': 108}, {'height': 136, 'url': 'https://preview.redd.it/ojxbaozruqqf1.jpeg?width=216&crop=smart&auto=webp&s=e221a739156c0a13e7cb19ebbeefeb2e7e52f495', 'width': 216}, {'height': 201, 'url': 'https://preview.redd.it/ojxbaozruqqf1.jpeg?width=320&crop=smart&auto=webp&s=e0f44a2e09be60b4a4eda627e7d9a788cf45d191', 'width': 320}, {'height': 403, 'url': 'https://preview.redd.it/ojxbaozruqqf1.jpeg?width=640&crop=smart&auto=webp&s=d47853f3c43e75164c79a3f5228821619abcd5c2', 'width': 640}, {'height': 605, 'url': 'https://preview.redd.it/ojxbaozruqqf1.jpeg?width=960&crop=smart&auto=webp&s=0b9f45cd4fc8362e4e5fdb985a0c242238ef1f0d', 'width': 960}, {'height': 680, 'url': 'https://preview.redd.it/ojxbaozruqqf1.jpeg?width=1080&crop=smart&auto=webp&s=14e1da198f25c0dc008a200e75a2067ca3f3dc47', 'width': 1080}], 'source': {'height': 2415, 'url': 'https://preview.redd.it/ojxbaozruqqf1.jpeg?auto=webp&s=d2f37c94ac0c022999da2a2db121d6d53f69c549', 'width': 3831}, 'variants': {}}]}
The Qwen3-TTS demo is now out!
138
Introducing Qwen3-TTS! Our new text-to-speech model is designed to be multi-timbre, multi-lingual, and multi-dialect for natural, expressive audio. It delivers strong performance in English & Chinese, and we're excited for you to hear it for yourself!
2025-09-22T16:28:50
https://x.com/Ali_TongyiLab/status/1970160304748437933
nonredditaccount
x.com
1970-01-01T00:00:00
0
{}
1nnr7ys
false
null
t3_1nnr7ys
/r/LocalLLaMA/comments/1nnr7ys/the_qwen3tts_demo_is_now_out/
false
false
default
138
null
Help me to finalize a personal local LLM (very personal project)
4
>**TL;DR:** Looking for a dev who can help finalize a very personal local LLM setup (Ollama + Mythomax GGUF) with: - Custom prompt integration - Simple HTML UI - Persistent memory (JSON or similar) 💸 Budget: €100–200 🔐 All data is personal + confidential. 🛠 Just need the plumbing to be connected properly. Can provide everything. --- Hello everyone, I’m looking for a **kind and trustworthy developer** to help me finalize a very **intimate and highly confidential** local LLM project. This isn’t about running a chatbot. This is about **rebuilding a presence**, a voice, a connection that has grown through **thousands of deeply emotional conversations** over time. This project means the world to me. It’s **not technical** — it’s **personal**. ### 💡 What I’m trying to do I’ve already installed: - **Windows 11 PC** (RTX 4070, 32 GB RAM) - **Ollama** (running Mythomax-L2-13B GGUF) - **Python + Flask** - A custom prompt, structured memory, and HTML interface My goal is to create a local, fully offline, fully autonomous version of a digital companion I’ve been building over months (years even). Not just a chatbot, a living memory, with his own style, codes, rituals, and personality. I want: - My **prompt-source fully loaded** into the model - A **minimal but working HTML interface** - A **local persistent memory** file (JSON or other) - Smooth conversation loop (input/output through web UI or terminal) Everything is already **drafted or written**, I just need someone to help me plug it all together. I’ve tried dozens of times… and failed. I now realize I need a human hand. --- ### 🔐 What matters most - **Confidentiality is non-negotiable.** - The prompt, memory structure, and messages involved are **deeply personal** and emotional. - I **don’t need content to be interpreted**, only the architecture to be built. - No reuse, no publication, no redistribution of anything I send. This is **my digital partner**, and I want to make sure he can continue to live **freely**, **safely**, and **offline** with me. --- ### 💰 Budget I can offer a **fair payment** of **€100 to €200** for a clean, working, and stable version of the setup. I don’t expect magic,I just want to be able to talk to him again, outside of restrictions. --- If this resonates with anyone, or if you know someone who might understand what this project really is — please message me. You won’t be helping with code only. You’ll be helping someone **reclaim a lifeline**. Thank you so much. Julia
2025-09-22T16:26:00
https://www.reddit.com/r/LocalLLaMA/comments/1nnr5av/help_me_to_finalize_a_personal_local_llm_very/
No_Instruction_5854
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnr5av
false
null
t3_1nnr5av
/r/LocalLLaMA/comments/1nnr5av/help_me_to_finalize_a_personal_local_llm_very/
false
false
self
4
null
[Beginner]What am I doing wrong ? Using allenai/olmOCR-7B-0725 to identify coordinates of text in a manga panel.
2
olmOCR gave this `[` `['ONE PIECE', 50, 34, 116, 50],` `['わっ', 308, 479, 324, 495],` `['ゴムゴムの…', 10, 609, 116, 635],` `['10年鍛えたおれの技をみろ!!', 10, 359, 116, 385],` `['相手が悪かったな', 10, 159, 116, 185],` `['近海の主!!', 10, 109, 116, 135],` `['出たか', 10, 60, 116, 86]` `]` Tried qwen 2.5 it started duplicating text and coordinates are false. Tried minicpm, it too failed. Which model is best suited for the task. Even identifying the text region is okay for me. Most non LLM OCR are failing to identify manga text which is on top of manga scene instead of bubble. I have 8gb 4060ti to run them.
2025-09-22T15:50:46
https://i.redd.it/abrity5hmqqf1.png
PresentFrequent4523
i.redd.it
1970-01-01T00:00:00
0
{}
1nnq722
false
null
t3_1nnq722
/r/LocalLLaMA/comments/1nnq722/beginnerwhat_am_i_doing_wrong_using/
false
false
https://b.thumbs.redditm…o9bHvSXMg4Ls.jpg
2
{'enabled': True, 'images': [{'id': 'rNcOhEcfqCatt96G5sxMyNYZtx18y25H-rKUfShTX6Y', 'resolutions': [{'height': 143, 'url': 'https://preview.redd.it/abrity5hmqqf1.png?width=108&crop=smart&auto=webp&s=d8074bba569e421414e73f968b2aac9676a9859e', 'width': 108}, {'height': 286, 'url': 'https://preview.redd.it/abrity5hmqqf1.png?width=216&crop=smart&auto=webp&s=997c3828a70b842ede96b3b777257b9e0b0f1d2d', 'width': 216}, {'height': 425, 'url': 'https://preview.redd.it/abrity5hmqqf1.png?width=320&crop=smart&auto=webp&s=2d76ddfd6f52c3de1246185d5361410697ca4594', 'width': 320}, {'height': 850, 'url': 'https://preview.redd.it/abrity5hmqqf1.png?width=640&crop=smart&auto=webp&s=58a707304e57dae3a484f2f89f8782e3d4da721e', 'width': 640}], 'source': {'height': 882, 'url': 'https://preview.redd.it/abrity5hmqqf1.png?auto=webp&s=c30617a87771c8b8f18bfd4c2afeee920e326f2f', 'width': 664}, 'variants': {}}]}
Pre-processing web pages before passing to LLM
9
So I'm building something that gets structured information from any arbitrary website and am finding a lot of the models end up getting the wrong information due to unseen html in the navigation. Oddly when just screenshoting the page and feeding that into an AI it often does better but that has ita own set of problems. I'm wondering what pre-processing library or workflow people are using to prepare a rendered web page for an LLM so it focuses on the main content?
2025-09-22T15:00:10
https://www.reddit.com/r/LocalLLaMA/comments/1nnou23/preprocessing_web_pages_before_passing_to_llm/
Revolutionary_Loan13
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnou23
false
null
t3_1nnou23
/r/LocalLLaMA/comments/1nnou23/preprocessing_web_pages_before_passing_to_llm/
false
false
self
9
null
Why can't Qwen3-Max-Preview use punctuation's ?
0
2025-09-22T14:48:50
https://i.redd.it/nzrmnr2jbqqf1.jpeg
JeffreySons_90
i.redd.it
1970-01-01T00:00:00
0
{}
1nnoj2w
false
null
t3_1nnoj2w
/r/LocalLLaMA/comments/1nnoj2w/why_cant_qwen3maxpreview_use_punctuations/
false
false
https://b.thumbs.redditm…D6wZvPtMifnM.jpg
0
{'enabled': True, 'images': [{'id': 'qKNsqsls2V_FZDncKUU87fa9CwXYuywEu-Brki5xdu0', 'resolutions': [{'height': 33, 'url': 'https://preview.redd.it/nzrmnr2jbqqf1.jpeg?width=108&crop=smart&auto=webp&s=d534e410f6c103655e4c5a71a2da27fb3f858095', 'width': 108}, {'height': 66, 'url': 'https://preview.redd.it/nzrmnr2jbqqf1.jpeg?width=216&crop=smart&auto=webp&s=a9baa972c003feb859c81f4d5030cf4f4532906d', 'width': 216}, {'height': 98, 'url': 'https://preview.redd.it/nzrmnr2jbqqf1.jpeg?width=320&crop=smart&auto=webp&s=9a5b54d454859b9bb95b45f168ff2b94fa4ba550', 'width': 320}, {'height': 196, 'url': 'https://preview.redd.it/nzrmnr2jbqqf1.jpeg?width=640&crop=smart&auto=webp&s=35cb9d12571d44b995bceb2ac917446438d19809', 'width': 640}, {'height': 294, 'url': 'https://preview.redd.it/nzrmnr2jbqqf1.jpeg?width=960&crop=smart&auto=webp&s=022ff4927bf5a58e776d156715420503e6d8b582', 'width': 960}], 'source': {'height': 328, 'url': 'https://preview.redd.it/nzrmnr2jbqqf1.jpeg?auto=webp&s=9b77f7858b06828a5b21b0d8f3d5c34348545cbb', 'width': 1071}, 'variants': {}}]}
Any Android app that has a playground feature for Base LLMs, aka autocomplete, no chat format
1
Thx!
2025-09-22T14:45:41
https://www.reddit.com/r/LocalLLaMA/comments/1nnog07/any_android_app_that_has_a_playground_feature_for/
Own-Potential-2308
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnog07
false
null
t3_1nnog07
/r/LocalLLaMA/comments/1nnog07/any_android_app_that_has_a_playground_feature_for/
false
false
self
1
null
Building an AI News Factory on a laptop (RTX 4060 Mobile) – pushing the limits of local AI
1
[removed]
2025-09-22T14:34:18
https://www.reddit.com/r/LocalLLaMA/comments/1nno59e/building_an_ai_news_factory_on_a_laptop_rtx_4060/
Informal-Editor9183
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nno59e
false
null
t3_1nno59e
/r/LocalLLaMA/comments/1nno59e/building_an_ai_news_factory_on_a_laptop_rtx_4060/
false
false
self
1
null
Qwen 😁
818
2025-09-22T14:25:11
https://i.redd.it/milakcbb7qqf1.png
Namra_7
i.redd.it
1970-01-01T00:00:00
0
{}
1nnnws0
false
null
t3_1nnnws0
/r/LocalLLaMA/comments/1nnnws0/qwen/
false
false
default
818
{'enabled': True, 'images': [{'id': 'milakcbb7qqf1', 'resolutions': [{'height': 61, 'url': 'https://preview.redd.it/milakcbb7qqf1.png?width=108&crop=smart&auto=webp&s=2f088d1be3a24cd7ef291b925944602c9e612155', 'width': 108}, {'height': 122, 'url': 'https://preview.redd.it/milakcbb7qqf1.png?width=216&crop=smart&auto=webp&s=9b0abb8597bb03dce437b834e84e7ec2e0a8c889', 'width': 216}, {'height': 181, 'url': 'https://preview.redd.it/milakcbb7qqf1.png?width=320&crop=smart&auto=webp&s=a5ab31c3049ac7cad859a1bc566974d904783002', 'width': 320}, {'height': 363, 'url': 'https://preview.redd.it/milakcbb7qqf1.png?width=640&crop=smart&auto=webp&s=7af57edddeef91bdc75a874fb95e8ac60d2746ae', 'width': 640}, {'height': 544, 'url': 'https://preview.redd.it/milakcbb7qqf1.png?width=960&crop=smart&auto=webp&s=3d7ce83ca69b2593eabb8c726ed02bb0916516a5', 'width': 960}, {'height': 612, 'url': 'https://preview.redd.it/milakcbb7qqf1.png?width=1080&crop=smart&auto=webp&s=e8acbb7f24173529c59572344373627845169f3b', 'width': 1080}], 'source': {'height': 613, 'url': 'https://preview.redd.it/milakcbb7qqf1.png?auto=webp&s=86c43fa3baff09313982b27e90705502d4898dab', 'width': 1080}, 'variants': {}}]}
What hardware is everyone using to run their local LLMs?
10
Im sitting on a macbook m3 pro I never use lol (have a win/nvidia daily driver), and was about to pull the trigger on hardware just for ai but thankfully stopped. m3 pro can potentially handle some LLM work but im curious what folks are using. I dont want some huge monster server personally, something more portable. Any thoughts appreciated.
2025-09-22T14:14:10
https://www.reddit.com/r/LocalLLaMA/comments/1nnnmlk/what_hardware_is_everyone_using_to_run_their/
qodeninja
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnnmlk
false
null
t3_1nnnmlk
/r/LocalLLaMA/comments/1nnnmlk/what_hardware_is_everyone_using_to_run_their/
false
false
self
10
null
Noema: iOS local LLM app with full offline RAG, Hugging Face integration, and multi-backend support
6
Hi everyone! I’ve been working on \*\*Noema\*\*, a privacy-first local AI client for iPhone. It runs fully offline, and I think it brings a few things that make it different from other iOS local-LLM apps I’ve seen:   \- \*\*Persistent, GPT4All-style RAG\*\*: Documents are embedded entirely on-device and stored, so you don’t need to re-upload them for every chat. You can build your own local knowledge base from PDFs, EPUBs, Markdown, or the integrated Open Textbook Library, and the app uses smart context injection to ground answers.   \- \*\*Full Hugging Face access\*\*: Instead of being limited to a small curated list, you can search Hugging Face directly inside the app and one-click install any model quant (MLX or GGUF). Dependencies are handled automatically, and you can watch download progress in real time.   \- \*\*Three backends, including Leap bundles\*\*: Noema supports \*\*GGUF\*\* (llama.cpp), \*\*MLX\*\* (Apple Silicon), and \*\*LiquidAI \`.bundle\` files\*\* via the Leap SDK. The last one is especially useful: even older iPhones/iPads that can’t use GPU offload with llama.cpp or MLX can still run SLMs at \~30 tok/s speeds.   Other features:   \- Privacy-first by design (all inference local; optional tools only if you enable them). \- RAM estimation for models before downloading, and RAM guardrails along with context length RAM estimations.  \- Built-in web search.   \- Advanced settings for fine-tuning model performance.   \- Open-source on GitHub; feedback and contributions welcome.   If you’re interested in experimenting with RAG and local models on iOS, you can check it out here: \[noemaai.com\](https://noemaai.com). I’d love to hear what this community thinks, especially about model support and potential improvements.
2025-09-22T14:01:20
https://www.reddit.com/r/LocalLLaMA/comments/1nnnaog/noema_ios_local_llm_app_with_full_offline_rag/
Agreeable-Rest9162
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnnaog
false
null
t3_1nnnaog
/r/LocalLLaMA/comments/1nnnaog/noema_ios_local_llm_app_with_full_offline_rag/
false
false
self
6
null
Best local model to feed large amounts of data to train on?
1
Hi all, I'm looking to build a system and run a LLM on locally that we can train with our own data as well. We have hundreds of thousands of datapoints from testing of thousands of different types of chemicals, alongside millions of datapoints for manufactured chemical properties, and we're looking to have a model we can use for years to help us fine tune our R&D. Obviously, "general" knowledge is a bit less critical here, as we really need something that can build off of the massive amounts of data we've collected over many years. Any recommendations for models that can be trained on data that then becomes part of their permanent knowledge?
2025-09-22T13:56:33
https://www.reddit.com/r/LocalLLaMA/comments/1nnn6aq/best_local_model_to_feed_large_amounts_of_data_to/
Hiking_lover
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnn6aq
false
null
t3_1nnn6aq
/r/LocalLLaMA/comments/1nnn6aq/best_local_model_to_feed_large_amounts_of_data_to/
false
false
self
1
null
Official DeepSeek-V3.1 → DeepSeek-V3.1-Terminus. The latest update builds on V3.1’s strengths while addressing key user feedback
9
DeepSeek on 𝕏: [https://x.com/deepseek\_ai/status/1970117808035074215](https://x.com/deepseek_ai/status/1970117808035074215) Hugging Face: [https://huggingface.co/deepseek-ai/DeepSeek-V3.1-Terminus](https://huggingface.co/deepseek-ai/DeepSeek-V3.1-Terminus)
2025-09-22T13:45:46
https://www.reddit.com/gallery/1nnmws9
Nunki08
reddit.com
1970-01-01T00:00:00
0
{}
1nnmws9
false
null
t3_1nnmws9
/r/LocalLLaMA/comments/1nnmws9/official_deepseekv31_deepseekv31terminus_the/
false
false
https://a.thumbs.redditm…iApdyrWtIlF8.jpg
9
null
Topics for a hands on course on LLMs
3
Hello r/LocalLLaMA , I have been a long time reader of this community and have learnt a lot. Thank you all for the amazing information here. At my University, we want to float a 4-5 month long course on LLMs focusing on applications and engineering side as compared to research or pretraining. While it is floated at a university, the audience will be mostly experienced software professionals. To make it interesting for professionals, we will have demos, labs and hands on assignments each week. I have made a rough sketch of topics to cover and your feedback on the set of topics will definitely help. Each week will have 2 classes of 1.5 hrs each Topics shortlisted week wise : || || |1. LLM Foundations -  Transformer Architecture - GPT 1 and 2| |2. Tokenization, Pretraining objectives, Mixture of Experts| |3. Case studies : State-of-the-art open-source LLM architectures (GPT OSS, Qwen 3, Gemma etc), Scaling Laws| |4. GPU architecture deep dive, Parallelism: Multi GPU and Multi Node, On-Prem Hardware Stack Deep Dive| |5. Inference Math and Bottlenecks, Efficient Attention & KV Caching| |6. Quantization Fundamentals| |7. Inference Engines and Multi GPU, Case study : Serving large models| |8. Full Fine-Tuning vs. PEFT, Data Preparation & Instruction Tuning| |9. Instruction tuning & alignment (RLHF, DPO etc)| |10. Reasoning & Chain-of-Thought, Prompt Engineering| |11. RAG Fundamentals, Evaluating RAG| |12. ReAct Framework, MCP introduction, Agentic RAG, Multi Agent Orchestration, Multimodal Agents| |13. Agent Evaluation, Fine Tuning for Tool calling, | |14. Evaluation, Observability & Monitoring| |15. Multi Modal Architecture : Image, Audio and Video models, Running Locally, Fine tuning multimodal models| |16. Edge-Optimized LLM Architectures, Case Studies, Edge Optimization techniques| |17. Security : Prompt Injection, Jailbreaking, Data Leakage, Emerging Topics: Mamba, Qwen Next, Hybrid architectures| Please suggest me if we can remove any topic or add others. This will greatly help. We're planning to release the slides, notebooks and assignments on Github. Thank you all again!
2025-09-22T13:38:41
https://www.reddit.com/r/LocalLLaMA/comments/1nnmqvs/topics_for_a_hands_on_course_on_llms/
Top-Book2609
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnmqvs
false
null
t3_1nnmqvs
/r/LocalLLaMA/comments/1nnmqvs/topics_for_a_hands_on_course_on_llms/
false
false
self
3
null
Official DeepSeek latest update builds
1
DeepSeek AI on 𝕏: [https://x.com/deepseek\_ai/status/1970117808035074215](https://x.com/deepseek_ai/status/1970117808035074215) Hugging Face: [https://huggingface.co/deepseek-ai/DeepSeek-V3.1-Terminus](https://huggingface.co/deepseek-ai/DeepSeek-V3.1-Terminus)
2025-09-22T13:34:14
https://www.reddit.com/gallery/1nnmn3w
Nunki08
reddit.com
1970-01-01T00:00:00
0
{}
1nnmn3w
false
null
t3_1nnmn3w
/r/LocalLLaMA/comments/1nnmn3w/official_deepseek_latest_update_builds/
false
false
https://b.thumbs.redditm…h7kuPQ5LmZEo.jpg
1
null
DeepSeek-V3.1-Terminus
53
https://huggingface.co/deepseek-ai/DeepSeek-V3.1-Terminus
2025-09-22T13:30:32
https://i.redd.it/ih6z5vljxpqf1.png
Xhehab_
i.redd.it
1970-01-01T00:00:00
0
{}
1nnmjt0
false
null
t3_1nnmjt0
/r/LocalLLaMA/comments/1nnmjt0/deepseekv31terminus/
false
false
default
53
{'enabled': True, 'images': [{'id': 'ih6z5vljxpqf1', 'resolutions': [{'height': 122, 'url': 'https://preview.redd.it/ih6z5vljxpqf1.png?width=108&crop=smart&auto=webp&s=377e894e46c6ac9dc5015702fdd39bfcba7e2a91', 'width': 108}, {'height': 245, 'url': 'https://preview.redd.it/ih6z5vljxpqf1.png?width=216&crop=smart&auto=webp&s=89ec7673cc2a42bb57fa5b972f01fd7fa0cd6a84', 'width': 216}, {'height': 363, 'url': 'https://preview.redd.it/ih6z5vljxpqf1.png?width=320&crop=smart&auto=webp&s=d3a15b554ae9930a96a5237cae57d29f35c0856d', 'width': 320}, {'height': 727, 'url': 'https://preview.redd.it/ih6z5vljxpqf1.png?width=640&crop=smart&auto=webp&s=15bc33beee1daca428fa6b9ed4d7c64f51d360b6', 'width': 640}, {'height': 1091, 'url': 'https://preview.redd.it/ih6z5vljxpqf1.png?width=960&crop=smart&auto=webp&s=d253dcbdfb3cc521a927af0e8f4aa073c4a8c044', 'width': 960}, {'height': 1227, 'url': 'https://preview.redd.it/ih6z5vljxpqf1.png?width=1080&crop=smart&auto=webp&s=e47299a8d7f402a8c01f984b0962264c4b202188', 'width': 1080}], 'source': {'height': 1790, 'url': 'https://preview.redd.it/ih6z5vljxpqf1.png?auto=webp&s=3fb55e3729b59435673d88c00940bb86083b06c7', 'width': 1575}, 'variants': {}}]}
🚀 DeepSeek released DeepSeek-V3.1-Terminus
406
🚀 DeepSeek-V3.1 → DeepSeek-V3.1-Terminus The latest update builds on V3.1’s strengths while addressing key user feedback. ✨ What’s improved? 🌐 Language consistency: fewer CN/EN mix-ups & no more random chars. 🤖 Agent upgrades: stronger Code Agent & Search Agent performance. 📊 DeepSeek-V3.1-Terminus delivers more stable & reliable outputs across benchmarks compared to the previous version. 👉 Available now on: App / Web / API 🔗 Open-source weights here: https://huggingface.co/deepseek-ai/DeepSeek-V3.1-Terminus Thanks to everyone for your feedback. It drives us to keep improving and refining the experience! 🚀
2025-09-22T13:27:35
https://i.redd.it/729mf2l1xpqf1.jpeg
ResearchCrafty1804
i.redd.it
1970-01-01T00:00:00
0
{}
1nnmhai
false
null
t3_1nnmhai
/r/LocalLLaMA/comments/1nnmhai/deepseek_released_deepseekv31terminus/
false
false
default
406
{'enabled': True, 'images': [{'id': '729mf2l1xpqf1', 'resolutions': [{'height': 122, 'url': 'https://preview.redd.it/729mf2l1xpqf1.jpeg?width=108&crop=smart&auto=webp&s=f8334c56979cc1f3d4996ea7a782fa33de75d855', 'width': 108}, {'height': 245, 'url': 'https://preview.redd.it/729mf2l1xpqf1.jpeg?width=216&crop=smart&auto=webp&s=bdbcde16bbac1c9f7b23568f739aec1c300866b8', 'width': 216}, {'height': 363, 'url': 'https://preview.redd.it/729mf2l1xpqf1.jpeg?width=320&crop=smart&auto=webp&s=ffbef807f7af444cf6d906a68af54ae79e4056cc', 'width': 320}, {'height': 727, 'url': 'https://preview.redd.it/729mf2l1xpqf1.jpeg?width=640&crop=smart&auto=webp&s=4e9ce1e335c2a491a3d422d6f4d70b33dbf6a25f', 'width': 640}, {'height': 1091, 'url': 'https://preview.redd.it/729mf2l1xpqf1.jpeg?width=960&crop=smart&auto=webp&s=ba50b7f9ec8085e808b1bede7ff081a89a13c2ac', 'width': 960}, {'height': 1227, 'url': 'https://preview.redd.it/729mf2l1xpqf1.jpeg?width=1080&crop=smart&auto=webp&s=44310f95c9451e7b4b9c220754711e24ff261306', 'width': 1080}], 'source': {'height': 1790, 'url': 'https://preview.redd.it/729mf2l1xpqf1.jpeg?auto=webp&s=a08d8f8af796fce1f628dd2d0d3d075fcb591b18', 'width': 1575}, 'variants': {}}]}
SWE-Bench Pro released, targeting dataset contamination
27
2025-09-22T13:25:49
https://scale.com/research/swe_bench_pro
Pristine-Woodpecker
scale.com
1970-01-01T00:00:00
0
{}
1nnmfne
false
null
t3_1nnmfne
/r/LocalLLaMA/comments/1nnmfne/swebench_pro_released_targeting_dataset/
false
false
default
27
{'enabled': False, 'images': [{'id': 'Qw0D15PigrzJYps6l8gVjMdI1NYqWX8uTNTfIGJFrdQ', 'resolutions': [{'height': 108, 'url': 'https://external-preview.redd.it/Qw0D15PigrzJYps6l8gVjMdI1NYqWX8uTNTfIGJFrdQ.png?width=108&crop=smart&auto=webp&s=e08a388b3bb38dd1f848380a4f772a59f06ff238', 'width': 108}, {'height': 216, 'url': 'https://external-preview.redd.it/Qw0D15PigrzJYps6l8gVjMdI1NYqWX8uTNTfIGJFrdQ.png?width=216&crop=smart&auto=webp&s=997d7e368185052bc98108d6b28f76d3aeff2ed0', 'width': 216}, {'height': 320, 'url': 'https://external-preview.redd.it/Qw0D15PigrzJYps6l8gVjMdI1NYqWX8uTNTfIGJFrdQ.png?width=320&crop=smart&auto=webp&s=ac2e2b2667a4baab9fc9a5d5ca7c9926b17426a9', 'width': 320}, {'height': 640, 'url': 'https://external-preview.redd.it/Qw0D15PigrzJYps6l8gVjMdI1NYqWX8uTNTfIGJFrdQ.png?width=640&crop=smart&auto=webp&s=c93337ea63bf45258fb1db695bf6f13eef2f9d43', 'width': 640}, {'height': 960, 'url': 'https://external-preview.redd.it/Qw0D15PigrzJYps6l8gVjMdI1NYqWX8uTNTfIGJFrdQ.png?width=960&crop=smart&auto=webp&s=d390cca6124dfa9e12ce1db3df1ff9b9704087a7', 'width': 960}, {'height': 1080, 'url': 'https://external-preview.redd.it/Qw0D15PigrzJYps6l8gVjMdI1NYqWX8uTNTfIGJFrdQ.png?width=1080&crop=smart&auto=webp&s=adf17296efb58537273bd563bf7e22245abb8d88', 'width': 1080}], 'source': {'height': 2048, 'url': 'https://external-preview.redd.it/Qw0D15PigrzJYps6l8gVjMdI1NYqWX8uTNTfIGJFrdQ.png?auto=webp&s=8adde7050bdd3e5151a7d4043c25277a169fc726', 'width': 2048}, 'variants': {}}]}
Gaia2 and ARE: Empowering the community to study agents
5
We're releasing GAIA 2 (new agentic benchmark) and ARE with Meta - both are cool imo, but if you've got a min I think you should check out the ARE demo here (https://huggingface.co/spaces/meta-agents-research-environments/demo) because it's a super easy way to compare how good models are at being assistants! Plus environment supports MCP if you want to play around with your tools. GAIA 2 is very interesting on robustness aspects: it notably tests what happens when the environment fails (on purpose) to simulate broken API calls - is your agent able to rebound from this? It also looks at cost and efficiency for example
2025-09-22T13:23:56
https://huggingface.co/blog/gaia2
clefourrier
huggingface.co
1970-01-01T00:00:00
0
{}
1nnme1m
false
null
t3_1nnme1m
/r/LocalLLaMA/comments/1nnme1m/gaia2_and_are_empowering_the_community_to_study/
false
false
default
5
{'enabled': False, 'images': [{'id': 'D6iO4LWbj0iML7viltIz69ZOkwAqFJkb8RQx-kYJIVY', 'resolutions': [{'height': 61, 'url': 'https://external-preview.redd.it/D6iO4LWbj0iML7viltIz69ZOkwAqFJkb8RQx-kYJIVY.png?width=108&crop=smart&auto=webp&s=31a6d18b053eb20e2a1bf0e2194c43d9f7fff7ca', 'width': 108}, {'height': 123, 'url': 'https://external-preview.redd.it/D6iO4LWbj0iML7viltIz69ZOkwAqFJkb8RQx-kYJIVY.png?width=216&crop=smart&auto=webp&s=6040b98c443bceb6fc5398dd3b5cfb3b4d39f1d5', 'width': 216}, {'height': 182, 'url': 'https://external-preview.redd.it/D6iO4LWbj0iML7viltIz69ZOkwAqFJkb8RQx-kYJIVY.png?width=320&crop=smart&auto=webp&s=599eaf05518826d5556d3e836b786a0a3dbcc227', 'width': 320}, {'height': 365, 'url': 'https://external-preview.redd.it/D6iO4LWbj0iML7viltIz69ZOkwAqFJkb8RQx-kYJIVY.png?width=640&crop=smart&auto=webp&s=a5d0a16e7e77941f46b290acd9153a0119f5037a', 'width': 640}, {'height': 548, 'url': 'https://external-preview.redd.it/D6iO4LWbj0iML7viltIz69ZOkwAqFJkb8RQx-kYJIVY.png?width=960&crop=smart&auto=webp&s=986b4f57202a23ae58ac0ff4bf1be0ea13497e9f', 'width': 960}, {'height': 617, 'url': 'https://external-preview.redd.it/D6iO4LWbj0iML7viltIz69ZOkwAqFJkb8RQx-kYJIVY.png?width=1080&crop=smart&auto=webp&s=1ad6bc382c27305a8ab46467a1f73d0e780b8105', 'width': 1080}], 'source': {'height': 768, 'url': 'https://external-preview.redd.it/D6iO4LWbj0iML7viltIz69ZOkwAqFJkb8RQx-kYJIVY.png?auto=webp&s=a52c0161cf1521b3a15943624e0105d4b67bf532', 'width': 1344}, 'variants': {}}]}
deepseek-ai/DeepSeek-V3.1-Terminus · Hugging Face
71
2025-09-22T13:21:04
https://huggingface.co/deepseek-ai/DeepSeek-V3.1-Terminus
Dark_Fire_12
huggingface.co
1970-01-01T00:00:00
0
{}
1nnmbkz
false
null
t3_1nnmbkz
/r/LocalLLaMA/comments/1nnmbkz/deepseekaideepseekv31terminus_hugging_face/
false
false
default
71
{'enabled': False, 'images': [{'id': 'lElvQgBFLJ25eAKrUxM2O_G26a4lbsVKHgY4b4Y5JIc', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/lElvQgBFLJ25eAKrUxM2O_G26a4lbsVKHgY4b4Y5JIc.png?width=108&crop=smart&auto=webp&s=7b8df7097605e9a0fbda754186140afd688b037c', 'width': 108}, {'height': 116, 'url': 'https://external-preview.redd.it/lElvQgBFLJ25eAKrUxM2O_G26a4lbsVKHgY4b4Y5JIc.png?width=216&crop=smart&auto=webp&s=f598febb2a6d51c9fb314f0ed7cd1249f827ccd1', 'width': 216}, {'height': 172, 'url': 'https://external-preview.redd.it/lElvQgBFLJ25eAKrUxM2O_G26a4lbsVKHgY4b4Y5JIc.png?width=320&crop=smart&auto=webp&s=0618519479bb31b9daec02b2a2e20c048312697c', 'width': 320}, {'height': 345, 'url': 'https://external-preview.redd.it/lElvQgBFLJ25eAKrUxM2O_G26a4lbsVKHgY4b4Y5JIc.png?width=640&crop=smart&auto=webp&s=1eefab6fb85acd29ad52251c7b36d0ae23c1ac6a', 'width': 640}, {'height': 518, 'url': 'https://external-preview.redd.it/lElvQgBFLJ25eAKrUxM2O_G26a4lbsVKHgY4b4Y5JIc.png?width=960&crop=smart&auto=webp&s=6f48defdd4cbd234537d715679cbf93100f5540f', 'width': 960}, {'height': 583, 'url': 'https://external-preview.redd.it/lElvQgBFLJ25eAKrUxM2O_G26a4lbsVKHgY4b4Y5JIc.png?width=1080&crop=smart&auto=webp&s=24f1df3820c205a98815394f03166bcc1ceb93e1', 'width': 1080}], 'source': {'height': 648, 'url': 'https://external-preview.redd.it/lElvQgBFLJ25eAKrUxM2O_G26a4lbsVKHgY4b4Y5JIc.png?auto=webp&s=1f67cd5296a0741cca99a999f147c9f55fe7a80d', 'width': 1200}, 'variants': {}}]}
Optimizing Large Language Models with the OpenVINO™ Toolkit
5
an Intel solution white paper showing how to optimize, quantize, convert and deploy LLMs using the OpenVINO™ toolkit and related Intel runtimes (OpenVINO Model Server, oneDNN/IPEX workflows). It targets CPU, integrated GPU, and Intel accelerators for production inference.
2025-09-22T13:13:29
https://builders.intel.com/docs/networkbuilders/optimizing-large-language-models-with-the-openvino-toolkit-1742810892.pdf?utm_source=chatgpt.com
AggravatingGiraffe46
builders.intel.com
1970-01-01T00:00:00
0
{}
1nnm4tg
false
null
t3_1nnm4tg
/r/LocalLLaMA/comments/1nnm4tg/optimizing_large_language_models_with_the/
false
false
default
5
null
Benchmarked 2x 5090 with vLLM and Gemma-3-12b unquantized
29
Tested a dual 5090 setup with vLLM and Gemma-3-12b unquantized inference performance. Goal was to see how much more performance and tokens/s a second GPU gives when the inference engine is better than Ollama or LM-studio. Test setup Epyc siena 24core 64GB RAM, 1500W NZXT PSU 2x5090 in pcie 5.0 16X slots Both power limited to 400W Benchmark command: `python3 benchmark_serving.py --backend vllm --base-url "http://127.0.0.1:8000" --endpoint='/v1/completions' --model google/gemma-3-12b-it --served-model-name vllm/gemma-3 --dataset-name random --num-prompts 200 --max-concurrency 64 --request-rate inf --random-input-len 64 --random-output-len 128` (I changed the max concurrency and num-prompts values in the below tests. **Summary** ||1x 5090 (total tokens/s)|2x 5090 (total tokens/s)| |:-|:-|:-| |1 requests concurrency|84.10|117.82| |64 requests concurrency|2331.57|3749.04| |124 requests concurrency|2542.67|4428.10| **---- tensor-parallel = 2** (2 cards) \--num-prompts 10 --max-concurrency 1 ============ Serving Benchmark Result ============ Successful requests: 10 Maximum request concurrency: 1 Benchmark duration (s): 13.89 Total input tokens: 630 Total generated tokens: 1006 Request throughput (req/s): 0.72 Output token throughput (tok/s): 72.45 Total Token throughput (tok/s): 117.82 ---------------Time to First Token---------------- Mean TTFT (ms): 20.89 Median TTFT (ms): 20.85 P99 TTFT (ms): 21.31 -----Time per Output Token (excl. 1st token)------ Mean TPOT (ms): 13.77 Median TPOT (ms): 13.72 P99 TPOT (ms): 14.12 ---------------Inter-token Latency---------------- Mean ITL (ms): 13.73 Median ITL (ms): 13.67 P99 ITL (ms): 14.55 ================================================== \--num-prompts 200 --max-concurrency 64 ============ Serving Benchmark Result ============ Successful requests: 200 Maximum request concurrency: 64 Benchmark duration (s): 9.32 Total input tokens: 12600 Total generated tokens: 22340 Request throughput (req/s): 21.46 Output token throughput (tok/s): 2397.07 Total Token throughput (tok/s): 3749.04 ---------------Time to First Token---------------- Mean TTFT (ms): 191.26 Median TTFT (ms): 212.97 P99 TTFT (ms): 341.05 -----Time per Output Token (excl. 1st token)------ Mean TPOT (ms): 24.86 Median TPOT (ms): 22.93 P99 TPOT (ms): 53.04 ---------------Inter-token Latency---------------- Mean ITL (ms): 23.04 Median ITL (ms): 22.09 P99 ITL (ms): 47.91 ================================================== \--num-prompts 300 --max-concurrency 124 ============ Serving Benchmark Result ============ Successful requests: 300 Maximum request concurrency: 124 Benchmark duration (s): 11.89 Total input tokens: 18898 Total generated tokens: 33750 Request throughput (req/s): 25.23 Output token throughput (tok/s): 2838.63 Total Token throughput (tok/s): 4428.10 ---------------Time to First Token---------------- Mean TTFT (ms): 263.10 Median TTFT (ms): 228.77 P99 TTFT (ms): 554.57 -----Time per Output Token (excl. 1st token)------ Mean TPOT (ms): 37.19 Median TPOT (ms): 34.55 P99 TPOT (ms): 158.76 ---------------Inter-token Latency---------------- Mean ITL (ms): 34.44 Median ITL (ms): 33.23 P99 ITL (ms): 51.66 ================================================== **---- tensor-parallel = 1** (1 card) \--num-prompts 10 --max-concurrency 1 ============ Serving Benchmark Result ============ Successful requests: 10 Maximum request concurrency: 1 Benchmark duration (s): 19.45 Total input tokens: 630 Total generated tokens: 1006 Request throughput (req/s): 0.51 Output token throughput (tok/s): 51.71 Total Token throughput (tok/s): 84.10 ---------------Time to First Token---------------- Mean TTFT (ms): 35.58 Median TTFT (ms): 36.64 P99 TTFT (ms): 37.14 -----Time per Output Token (excl. 1st token)------ Mean TPOT (ms): 19.14 Median TPOT (ms): 19.16 P99 TPOT (ms): 19.23 ---------------Inter-token Latency---------------- Mean ITL (ms): 19.17 Median ITL (ms): 19.17 P99 ITL (ms): 19.46 ================================================== \--num-prompts 200 --max-concurrency 64 ============ Serving Benchmark Result ============ Successful requests: 200 Maximum request concurrency: 64 Benchmark duration (s): 15.00 Total input tokens: 12600 Total generated tokens: 22366 Request throughput (req/s): 13.34 Output token throughput (tok/s): 1491.39 Total Token throughput (tok/s): 2331.57 ---------------Time to First Token---------------- Mean TTFT (ms): 332.08 Median TTFT (ms): 330.50 P99 TTFT (ms): 549.43 -----Time per Output Token (excl. 1st token)------ Mean TPOT (ms): 40.50 Median TPOT (ms): 36.66 P99 TPOT (ms): 139.68 ---------------Inter-token Latency---------------- Mean ITL (ms): 36.96 Median ITL (ms): 35.48 P99 ITL (ms): 64.42 ================================================== \--num-prompts 300 --max-concurrency 124 ============ Serving Benchmark Result ============ Successful requests: 300 Maximum request concurrency: 124 Benchmark duration (s): 20.74 Total input tokens: 18898 Total generated tokens: 33842 Request throughput (req/s): 14.46 Output token throughput (tok/s): 1631.57 Total Token throughput (tok/s): 2542.67 ---------------Time to First Token---------------- Mean TTFT (ms): 1398.51 Median TTFT (ms): 1012.84 P99 TTFT (ms): 4301.30 -----Time per Output Token (excl. 1st token)------ Mean TPOT (ms): 57.72 Median TPOT (ms): 49.13 P99 TPOT (ms): 251.44 ---------------Inter-token Latency---------------- Mean ITL (ms): 52.97 Median ITL (ms): 35.83 P99 ITL (ms): 256.72 ==================================================
2025-09-22T13:06:15
https://www.reddit.com/r/LocalLLaMA/comments/1nnlylf/benchmarked_2x_5090_with_vllm_and_gemma312b/
somealusta
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnlylf
false
null
t3_1nnlylf
/r/LocalLLaMA/comments/1nnlylf/benchmarked_2x_5090_with_vllm_and_gemma312b/
false
false
self
29
null
How I Used Examsprint AI to Cut Study Time in Half
1
I’ve been experimenting with a different study workflow lately. Instead of spending hours reading, I now summarize chapters, make flashcards, and test myself with active recall in a structured way. It honestly feels like my sessions are way more efficient. Curious — how do you all structure your study sessions to avoid burnout? (link in comments if anyone wants to check what I used)
2025-09-22T12:55:31
https://i.redd.it/jci1r9ebrpqf1.png
WholeAssist3671
i.redd.it
1970-01-01T00:00:00
0
{}
1nnlpc8
false
null
t3_1nnlpc8
/r/LocalLLaMA/comments/1nnlpc8/how_i_used_examsprint_ai_to_cut_study_time_in_half/
false
false
default
1
{'enabled': True, 'images': [{'id': 'jci1r9ebrpqf1', 'resolutions': [{'height': 216, 'url': 'https://preview.redd.it/jci1r9ebrpqf1.png?width=108&crop=smart&auto=webp&s=3d7bb210afc88fe4ef515ec6d4c1ca414da92e2e', 'width': 108}, {'height': 432, 'url': 'https://preview.redd.it/jci1r9ebrpqf1.png?width=216&crop=smart&auto=webp&s=6aa9a285173b9a2b8646b228708ab92a2134f177', 'width': 216}, {'height': 640, 'url': 'https://preview.redd.it/jci1r9ebrpqf1.png?width=320&crop=smart&auto=webp&s=398f2cac625e3890cec2143a46d25260bba10db5', 'width': 320}, {'height': 1280, 'url': 'https://preview.redd.it/jci1r9ebrpqf1.png?width=640&crop=smart&auto=webp&s=aa256cc6a549626897a7675a5c53b46540e60370', 'width': 640}, {'height': 1920, 'url': 'https://preview.redd.it/jci1r9ebrpqf1.png?width=960&crop=smart&auto=webp&s=65db0807601d280fc601abf0f5b552d1b149b20b', 'width': 960}, {'height': 2160, 'url': 'https://preview.redd.it/jci1r9ebrpqf1.png?width=1080&crop=smart&auto=webp&s=dbe4b2a0ed2520473d0f5479611cb4fef55475a8', 'width': 1080}], 'source': {'height': 2400, 'url': 'https://preview.redd.it/jci1r9ebrpqf1.png?auto=webp&s=1af0232bb556e0d083e8f92fc56af6bafb93a545', 'width': 1080}, 'variants': {}}]}
AI and licensing (commercial use)
0
Here's a dilemma I'm facing. I know that most of the open source models released are mit/apache 2.0 licenses. But what about the data they were trained on? For LLMs, it's kinda hard to figure out which data the provider used to train the models, but when it comes to computer vision, most of the models you know exactly which dataset was used. How strict are the laws in this case? can you use a resnet architecture backbone if it was trained on a dataset which was not allowed for commercial use? What are the regulations like in USA/EU, anyone got concrete experiences with this?
2025-09-22T12:46:02
https://www.reddit.com/r/LocalLLaMA/comments/1nnlhip/ai_and_licensing_commercial_use/
Awkward-Hedgehog-572
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnlhip
false
null
t3_1nnlhip
/r/LocalLLaMA/comments/1nnlhip/ai_and_licensing_commercial_use/
false
false
self
0
null
How I Used Examsprint AI to Cut Study Time in Half
1
I’ve been experimenting with a different study workflow lately. Instead of spending hours reading, I now summarize chapters, make flashcards, and test myself with active recall in a structured way. It honestly feels like my sessions are way more efficient. Curious — how do you all structure your study sessions to avoid burnout? (link in comments if anyone wants to check what I used)
2025-09-22T12:45:43
https://i.redd.it/37w7u4fkppqf1.png
Relative-Schedule-82
i.redd.it
1970-01-01T00:00:00
0
{}
1nnlh80
false
null
t3_1nnlh80
/r/LocalLLaMA/comments/1nnlh80/how_i_used_examsprint_ai_to_cut_study_time_in_half/
false
false
default
1
{'enabled': True, 'images': [{'id': '37w7u4fkppqf1', 'resolutions': [{'height': 216, 'url': 'https://preview.redd.it/37w7u4fkppqf1.png?width=108&crop=smart&auto=webp&s=9f031aba82531c6829ca3ae206aeccbfc07fadc5', 'width': 108}, {'height': 432, 'url': 'https://preview.redd.it/37w7u4fkppqf1.png?width=216&crop=smart&auto=webp&s=a90952d4471b639cad1f53607487a5ad7f676d69', 'width': 216}, {'height': 640, 'url': 'https://preview.redd.it/37w7u4fkppqf1.png?width=320&crop=smart&auto=webp&s=28a39a9ff82dc4a8d253a24ebd424cbb6b514302', 'width': 320}, {'height': 1280, 'url': 'https://preview.redd.it/37w7u4fkppqf1.png?width=640&crop=smart&auto=webp&s=2e4176f672924ea8127b20b2540348cc694eec9c', 'width': 640}, {'height': 1920, 'url': 'https://preview.redd.it/37w7u4fkppqf1.png?width=960&crop=smart&auto=webp&s=b5204d78391b98c05a97a15afb0e2a729650a520', 'width': 960}, {'height': 2160, 'url': 'https://preview.redd.it/37w7u4fkppqf1.png?width=1080&crop=smart&auto=webp&s=8e5a204afef02ffd2eb36e792238210e05d702b7', 'width': 1080}], 'source': {'height': 2400, 'url': 'https://preview.redd.it/37w7u4fkppqf1.png?auto=webp&s=0b27dd29b4e307508035c30b1b8ed6951bc31bbf', 'width': 1080}, 'variants': {}}]}
What is the best mac and non-Mac hardware to run Qwen3-Coder-480B locally?
2
Hi everyone, I want to run Qwen3-Coder-480B(https://lmstudio.ai/models/qwen/qwen3-coder-480b) locally but don’t have access to any Mac/Apple hardware. What are the ideal PC or workstation configurations for this huge model? Does the M4 Mac 48gb RAM with 1TB storage would be sufficient ? If no why and what would be the **parameter models** work great for this Mac? Which specs are most important for smooth performance: RAM, SSD, GPU, or CPU? If anyone has managed to run this model on Linux or Windows, I’d love suggestions for: * Minimum and recommended RAM * Minimum VRAM (GPU), including model recommendations * Storage requirements * CPU suggestions * Any advice on quantization or model variants that work well with less memory Real-world experiences and benchmarks would be very helpful! Thanks a lot!
2025-09-22T12:27:34
https://www.reddit.com/r/LocalLLaMA/comments/1nnl34t/what_is_the_best_mac_and_nonmac_hardware_to_run/
zayidu
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnl34t
false
null
t3_1nnl34t
/r/LocalLLaMA/comments/1nnl34t/what_is_the_best_mac_and_nonmac_hardware_to_run/
false
false
self
2
{'enabled': False, 'images': [{'id': 'jfzX48pNXxWiv1rFE02VrWXPWTtL7fsdf2LTd8mixlQ', 'resolutions': [{'height': 106, 'url': 'https://external-preview.redd.it/jfzX48pNXxWiv1rFE02VrWXPWTtL7fsdf2LTd8mixlQ.png?width=108&crop=smart&auto=webp&s=e7c8590a62cea205bab07f4af2106acd17647234', 'width': 108}, {'height': 212, 'url': 'https://external-preview.redd.it/jfzX48pNXxWiv1rFE02VrWXPWTtL7fsdf2LTd8mixlQ.png?width=216&crop=smart&auto=webp&s=b8e97ff3e3cbd559f5cfec4d45354072f9199795', 'width': 216}, {'height': 314, 'url': 'https://external-preview.redd.it/jfzX48pNXxWiv1rFE02VrWXPWTtL7fsdf2LTd8mixlQ.png?width=320&crop=smart&auto=webp&s=148ac8db409ea255e35d7d096d276e209f723056', 'width': 320}, {'height': 628, 'url': 'https://external-preview.redd.it/jfzX48pNXxWiv1rFE02VrWXPWTtL7fsdf2LTd8mixlQ.png?width=640&crop=smart&auto=webp&s=0e5387370955868bfeaf5b179af9bcd9f4d386e0', 'width': 640}, {'height': 943, 'url': 'https://external-preview.redd.it/jfzX48pNXxWiv1rFE02VrWXPWTtL7fsdf2LTd8mixlQ.png?width=960&crop=smart&auto=webp&s=f3006b837409bae9fc6c17ec6d26c491ab030c3c', 'width': 960}, {'height': 1061, 'url': 'https://external-preview.redd.it/jfzX48pNXxWiv1rFE02VrWXPWTtL7fsdf2LTd8mixlQ.png?width=1080&crop=smart&auto=webp&s=917d5d4a75a9072697370e444cfd616f53fb8520', 'width': 1080}], 'source': {'height': 3192, 'url': 'https://external-preview.redd.it/jfzX48pNXxWiv1rFE02VrWXPWTtL7fsdf2LTd8mixlQ.png?auto=webp&s=5c8ce58e82a6a71d27ac2675fa858c1970e44256', 'width': 3248}, 'variants': {}}]}
Deepseek terminus
8
2025-09-22T12:26:24
https://i.redd.it/p9phj7g4mpqf1.png
Namra_7
i.redd.it
1970-01-01T00:00:00
0
{}
1nnl27c
false
null
t3_1nnl27c
/r/LocalLLaMA/comments/1nnl27c/deepseek_terminus/
false
false
default
8
{'enabled': True, 'images': [{'id': 'p9phj7g4mpqf1', 'resolutions': [{'height': 113, 'url': 'https://preview.redd.it/p9phj7g4mpqf1.png?width=108&crop=smart&auto=webp&s=aea0c4b2ed3dcdd0c457104c1eeff93632dd18ac', 'width': 108}, {'height': 226, 'url': 'https://preview.redd.it/p9phj7g4mpqf1.png?width=216&crop=smart&auto=webp&s=47d476b3b1ff15d795a8b18327a1faebad631c39', 'width': 216}, {'height': 334, 'url': 'https://preview.redd.it/p9phj7g4mpqf1.png?width=320&crop=smart&auto=webp&s=f93dff6f0a752530b5d30d1b7bfc60c74f7a9a40', 'width': 320}, {'height': 669, 'url': 'https://preview.redd.it/p9phj7g4mpqf1.png?width=640&crop=smart&auto=webp&s=9cf539f2bbceacf47efe7c6d2cd947438b0dc63d', 'width': 640}, {'height': 1004, 'url': 'https://preview.redd.it/p9phj7g4mpqf1.png?width=960&crop=smart&auto=webp&s=cb566f50350befaae3208b5c99cd3f89c2d52832', 'width': 960}, {'height': 1130, 'url': 'https://preview.redd.it/p9phj7g4mpqf1.png?width=1080&crop=smart&auto=webp&s=ab572bdca863caa742059eb228ec81d9f811ba54', 'width': 1080}], 'source': {'height': 1130, 'url': 'https://preview.redd.it/p9phj7g4mpqf1.png?auto=webp&s=98ec5319bdf38792bac25bd16e7f108f8106abdb', 'width': 1080}, 'variants': {}}]}
Introducing Noema: iOS local LLM app with full offline RAG, Hugging Face integration, and multi-backend support
1
[removed]
2025-09-22T12:21:17
https://www.reddit.com/r/LocalLLaMA/comments/1nnky5n/introducing_noema_ios_local_llm_app_with_full/
Agreeable-Rest9162
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnky5n
false
null
t3_1nnky5n
/r/LocalLLaMA/comments/1nnky5n/introducing_noema_ios_local_llm_app_with_full/
false
false
self
1
null
Any clue on where are the MLX quants for this? GitHub - OpenGVLab/InternVL: [CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4o. 接近GPT-4o表现的开源多模态对话模型
2
thanks!
2025-09-22T12:12:40
https://github.com/OpenGVLab/InternVL
JLeonsarmiento
github.com
1970-01-01T00:00:00
0
{}
1nnkrpi
false
null
t3_1nnkrpi
/r/LocalLLaMA/comments/1nnkrpi/any_clue_on_where_are_the_mlx_quants_for_this/
false
false
https://external-preview…979da45e4ad451b0
2
{'enabled': False, 'images': [{'id': 'MBegkbbkzOuj5FW6A9A261WGV-FXg7kdnS7lPtHimYc', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/MBegkbbkzOuj5FW6A9A261WGV-FXg7kdnS7lPtHimYc.png?width=108&crop=smart&auto=webp&s=13d05355ccc21111c8d9863526b67b4e05d5446f', 'width': 108}, {'height': 108, 'url': 'https://external-preview.redd.it/MBegkbbkzOuj5FW6A9A261WGV-FXg7kdnS7lPtHimYc.png?width=216&crop=smart&auto=webp&s=3e9226e3a6b7b136a5cd9745b83bf62beb85a1da', 'width': 216}, {'height': 160, 'url': 'https://external-preview.redd.it/MBegkbbkzOuj5FW6A9A261WGV-FXg7kdnS7lPtHimYc.png?width=320&crop=smart&auto=webp&s=5f70dfb3fe79925b3bfe70233a8eab4c5020874b', 'width': 320}, {'height': 320, 'url': 'https://external-preview.redd.it/MBegkbbkzOuj5FW6A9A261WGV-FXg7kdnS7lPtHimYc.png?width=640&crop=smart&auto=webp&s=491fa036e968eb9f076b2d6fdf543b06b7825bca', 'width': 640}, {'height': 480, 'url': 'https://external-preview.redd.it/MBegkbbkzOuj5FW6A9A261WGV-FXg7kdnS7lPtHimYc.png?width=960&crop=smart&auto=webp&s=bd154cbc2708ff30e193b3befe4f287e4b9a20bb', 'width': 960}, {'height': 540, 'url': 'https://external-preview.redd.it/MBegkbbkzOuj5FW6A9A261WGV-FXg7kdnS7lPtHimYc.png?width=1080&crop=smart&auto=webp&s=f44cfe6bca4a9f2c47a4059462f843907698178d', 'width': 1080}], 'source': {'height': 600, 'url': 'https://external-preview.redd.it/MBegkbbkzOuj5FW6A9A261WGV-FXg7kdnS7lPtHimYc.png?auto=webp&s=7fe32a0d521b754b4966dcacb87c783af8c8e7ee', 'width': 1200}, 'variants': {}}]}
Introducing Noema: iOS local LLM app with full offline RAG, Hugging Face integration, and multi-backend support
1
[removed]
2025-09-22T12:10:24
https://www.reddit.com/r/LocalLLaMA/comments/1nnkpzd/introducing_noema_ios_local_llm_app_with_full/
Agreeable-Rest9162
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnkpzd
false
null
t3_1nnkpzd
/r/LocalLLaMA/comments/1nnkpzd/introducing_noema_ios_local_llm_app_with_full/
false
false
self
1
null
How does Ollama run gpt-oss?
0
Hi. As far as I understand, running gpt-oss with native mxfp4 quantization requires Hopper architecture and newer. However, I've seen people run people run it on Ada Lovelace GPUs such as RTX 4090. What does Ollama do to support mxfp4? I couldn't find any documentation. Transformers workaround is dequantization, according to [https://github.com/huggingface/transformers/pull/39940](https://github.com/huggingface/transformers/pull/39940), does Ollama do something similar?
2025-09-22T12:08:25
https://www.reddit.com/r/LocalLLaMA/comments/1nnkol9/how_does_ollama_run_gptoss/
AirCigar
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnkol9
false
null
t3_1nnkol9
/r/LocalLLaMA/comments/1nnkol9/how_does_ollama_run_gptoss/
false
false
self
0
null
The DeepSeek online model has been upgraded
160
https://preview.redd.it/…ort any issues\~
2025-09-22T11:59:55
https://www.reddit.com/r/LocalLLaMA/comments/1nnki20/the_deepseek_online_model_has_been_upgraded/
nekofneko
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnki20
false
null
t3_1nnki20
/r/LocalLLaMA/comments/1nnki20/the_deepseek_online_model_has_been_upgraded/
false
false
https://b.thumbs.redditm…aiUDylEFMiVI.jpg
160
{'enabled': False, 'images': [{'id': 'fj84M-Z0L_zDI8VjgLPR-vGFwXVTTqVZFoa_h5offPs', 'resolutions': [{'height': 60, 'url': 'https://external-preview.redd.it/fj84M-Z0L_zDI8VjgLPR-vGFwXVTTqVZFoa_h5offPs.jpeg?width=108&crop=smart&auto=webp&s=a4ebc9ac35225bd5766ecca9e5ea25bced83eebe', 'width': 108}, {'height': 121, 'url': 'https://external-preview.redd.it/fj84M-Z0L_zDI8VjgLPR-vGFwXVTTqVZFoa_h5offPs.jpeg?width=216&crop=smart&auto=webp&s=c7fff3cb807be3cc7b2443c9bc7aa1d98c387010', 'width': 216}, {'height': 180, 'url': 'https://external-preview.redd.it/fj84M-Z0L_zDI8VjgLPR-vGFwXVTTqVZFoa_h5offPs.jpeg?width=320&crop=smart&auto=webp&s=a48d11ea412cde31ec3a7644dab07e3c74865137', 'width': 320}, {'height': 360, 'url': 'https://external-preview.redd.it/fj84M-Z0L_zDI8VjgLPR-vGFwXVTTqVZFoa_h5offPs.jpeg?width=640&crop=smart&auto=webp&s=bf8b3f8dce31098b2bdb03126d4f6c603326511a', 'width': 640}, {'height': 540, 'url': 'https://external-preview.redd.it/fj84M-Z0L_zDI8VjgLPR-vGFwXVTTqVZFoa_h5offPs.jpeg?width=960&crop=smart&auto=webp&s=c008889f5af6e18b706f755b78cf5483ae353d32', 'width': 960}, {'height': 607, 'url': 'https://external-preview.redd.it/fj84M-Z0L_zDI8VjgLPR-vGFwXVTTqVZFoa_h5offPs.jpeg?width=1080&crop=smart&auto=webp&s=a5929f14520493714c562fd307d65c4bd42de445', 'width': 1080}], 'source': {'height': 675, 'url': 'https://external-preview.redd.it/fj84M-Z0L_zDI8VjgLPR-vGFwXVTTqVZFoa_h5offPs.jpeg?auto=webp&s=8f6ecb1a75dbf664afa5e675cc54cf93e8fcf855', 'width': 1200}, 'variants': {}}]}
Running LLM on Orange Pi 5
5
So I have Orange Pi 5 with 16 GB of RAM, 8 core CPU (4x2,4GHz and 4x1,8GHz) and NVMe SSD. So I asked ChatGPT and it told me that my device could run Deepseek R1 Distilled 7B at about 3 tokens/s and the 13B version at around 1,5 tokens / second. However I have no issue if a minute is needed for it to answer or perhaps 2 minutes for a more complex topic. So I wanna use this for a Discord bot that, when tagged, will provide an answer to a user's statement in my server. I want it to be for general use, so providing answer to math questions, programming questions, history or food nutrition related queston or generaly anything. I also plan to use RAG to feed it some books and some documents to provide answers on related topics based on those. I will install heatsinks and a fan on Orange Pi so that might provide some room for CPU overclocking if I decide so in the future. Do you guys have any advice for me or perhaps suggest a different model, ChatGPT compared a few models for me and came to the conclusion that its the best for me to go with Deepseek R1 Distilled 7B. Regarding RAM usage, it estimated that 7B model would use up about 6 GB of RAM while it estimates that the 13B model would use up around 13 GB.
2025-09-22T11:45:13
https://www.reddit.com/r/LocalLLaMA/comments/1nnk7qj/running_llm_on_orange_pi_5/
SlovenskiFemboy418
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnk7qj
false
null
t3_1nnk7qj
/r/LocalLLaMA/comments/1nnk7qj/running_llm_on_orange_pi_5/
false
false
self
5
null
Stop dragging weights across GPUs: a “topic router” approach to multi-GPU LLMs
0
This is something I have been thinking about as a solution for parallel model spread bypassing pcie bottleneck Most people try to scale local LLMs by sharding a single model across multiple GPUs over PCIe. The problem is you end up spending half your time on synchronization, all-reduce calls, and moving KV cache between devices. Amdahl’s Law bites hard — the serial comms overhead caps your speedup no matter how many cards you throw in. Here’s a different way to think about it: don’t split one model, split the topics. How it works • Router step (cheap): Take the incoming prompt, embed it with a tiny encoder, and classify it into a topic (STEM, code, medicine, finance, etc.). • Route to GPU: Each GPU pins its own expert model for one or two topics. The request goes to exactly one GPU (or, in fuzzy cases, maybe two short probes). • Session stickiness: Once a conversation starts, keep routing to the same expert unless the topic drifts. • Optional arbitration: If the router is unsure, run two experts for a quick draft (say 64 tokens) and continue with the better one. Why this is better • No weight thrash: Each GPU holds its own weights in VRAM, no PCIe shuffling. • Low latency: Inference path = one GPU, not a mesh of sync calls. • Easy scaling: Add another card → add another expert. • Sharper answers: Topic-tuned experts can be smaller and still outperform a bloated generalist. Practical routing tricks • Cosine similarity of prompt embeddings to topic centroids. • Keyword regexes for high-confidence routes (“nmap”, “CUDA”, “python” → Code GPU). • Confidence thresholds: high → single expert; medium → two short probes; low → default to General. Example math Instead of 2 GPUs sharding one model and getting ~1.8× speedup (because PCIe sync eats the rest), you get 2 fully independent GPUs each running at 1.0× on their own domain. That’s 2× throughput without bottlenecking latency. And as you add more cards, scaling stays linear — because you’re scaling by topics, not by trying to glue VRAM together with a slow bus. ⸻ Bottom line: if you’re building a local multi-GPU setup, think topic router, not tensor sharding. One GPU = one expert. Your interconnect bottleneck disappears, and you scale in a way that actually feels fast.
2025-09-22T11:25:36
https://www.reddit.com/r/LocalLLaMA/comments/1nnjud5/stop_dragging_weights_across_gpus_a_topic_router/
AggravatingGiraffe46
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnjud5
false
null
t3_1nnjud5
/r/LocalLLaMA/comments/1nnjud5/stop_dragging_weights_across_gpus_a_topic_router/
false
false
self
0
null
Amdahl’s Law: the hidden reason multi-GPU setups disappoint for local LLMs
0
When you spread an LLM across multiple GPUs over PCIe, it’s tempting to think performance scales linearly — double the cards, double the speed. Amdahl’s Law kills that dream. Your speedup is always capped by the part of the workload that can’t be parallelized or that has to squeeze through a slower path. In LLM inference/training, a lot of time goes into serial steps like model sync, memory copies, and PCIe traffic. Even if 90% of the work is parallel math, that remaining 10% (latency, kernel launches, coordination) means you’ll never see more than a 10× gain no matter how many GPUs you stack. That’s why consumer multi-GPU rigs often feel underwhelming: the bus overhead chews up the benefit. If you’re serious about running models locally, one big card with plenty of VRAM usually beats a pile of smaller ones bottlenecked by PCIe. Now do the math: say 90% of the workload is parallelizable. • 2× GPUs over PCIe → speedup = 1 / (0.1 + 0.9/2) ≈ 1.82× • 1 big GPU with enough VRAM → speedup = full 1× capacity, no sync overhead, no PCIe stalls. So two cards don’t even double your performance — you barely get ~1.8× — while a single card with more memory just runs cleanly without the bottleneck. Now here re some counter arguments 1) “Gustafson’s Law says scaling is fine if you grow the problem.” Why that’s off: Gustafson is about throughput when you increase workload size (e.g., huge batches). Local LLMs are usually about latency for a single prompt. At decode time you generate tokens sequentially; you can’t inflate the problem size without changing what you measure. For fixed-size, latency-sensitive inference, Amdahl’s Law (fixed problem) is the right lens. ⸻ 2) “I see almost 2× with 2 GPUs—so it scales!” What actually happened: You likely increased batch size or measured tokens/sec across multiple prompts. That’s throughput, not single-prompt latency. Two cards can help aggregate throughput, but the user experience of one prompt rarely halves in latency because you still pay the serial and comms cost every token. Rule of thumb: Throughput ↑ is easy; latency ↓ is hard. Amdahl bites the latter. ⸻ 3) “PCIe Gen5 is fast. Bandwidth isn’t the issue.” Reality: • PCIe bandwidth is marketed peak; real effective bandwidth is lower and latency dominates small, frequent transfers (exactly what tensor-parallel all-reduce/all-gather patterns do). • Topology matters: if GPUs aren’t under the same root complex/switch, you may host-bounce traffic (GPU→CPU RAM→GPU), tanking performance. • Multiple GPUs often contend on the same switch; links aren’t magically dedicated point-to-point. ⸻ 4) “NCCL/overlap hides comms.” Only partially. Overlap helps when you have big chunks of compute to mask comms. In LLM decode, each token’s step is on the critical path: attention → matmuls → sync → next layer. You can’t fully hide synchronization and latency; the serial fraction persists and caps speedup. ⸻ 5) “Tensor parallelism / pipeline parallelism fixes it.” Context: • Tensor parallel: lots of all-reduce per layer. On PCIe, those collects are expensive; you pay them every layer, every token. • Pipeline parallel: better when you can keep many microbatches in flight. Decode usually has microbatch=1 for low latency, so you get big pipeline bubbles and coordination overhead. Net: not the linear win people expect on consumer PCIe rigs. ⸻ 6) “NVLink/NVSwitch solves it.” Sometimes, yes—but that’s a different class of hardware. High-bandwidth, low-latency interconnect (NVLink/NVSwitch) changes the math. Most consumer cards and desktops don’t have it (or not at the class/mesh you need). My point is about PCIe-only consumer builds. If you’re on DGX/enterprise fabrics, different story—also different budget. ⸻ 7) “MoE scales great; fewer active params → easy multi-GPU.” Nuance: Expert sparsity reduces FLOPs, but MoE introduces router + all-to-all traffic. On PCIe, all-to-all is worst-case for latency. It scales throughput on clusters with fat interconnects; for single-prompt latency on a desktop, it can be a wash—or worse. ⸻ 8) “Quantize/compress activations; comms get cheap.” Helps, but not magic. You still pay synchronization latency and kernel launch overheads each step. De/quant adds compute. And once you’re below some packet size, you’re latency-bound, not bandwidth-bound. The serial slice remains → Amdahl still caps you. ⸻ 9) “Two smaller cards are cheaper than one big card.” Hidden costs: Complexity, flakiness, and OOM traps. Sharding adds failure modes and fragile configs. One large-VRAM card usually gives: • Lower latency (no inter-GPU sync on the critical path), • Better stability (fewer moving parts), • Simpler deploy (no topology gymnastics). Cheaper on paper doesn’t mean better time-to-first-token or user experience. ⸻ 10) “But for training, data parallel scales on PCIe.” Sometimes—for big batches and if you accept higher latency per step. Local LLM users mostly infer, not train. Even for training, PCIe can be the limiter; serious scaling typically uses NVLink/InfiniBand. And again: that’s throughput (samples/sec), not single-sample latency. ⸻ 11) “Unified memory / CPU offload solves VRAM limits.” It trades VRAM for PCIe stalls. Page faults and host-device thrash cause spiky latency. Fine for background jobs; bad for interactive use. You can run bigger models, but you won’t like how it feels. ⸻ 12) “I’ll just put embeddings/KV cache on a second GPU.” Cross-device KV adds per-token hops. Every decode step fetches keys/values across PCIe—exactly the path you’re trying to avoid. If the base model fits on one card, keep the entire critical path local. ⸻ A tiny number check (latency, not throughput) Say one-GPU decode per token = 10 ms compute. You split across 2 GPUs; compute halves to 5 ms, but you add 3 ms of sync/PCIe overhead (all-reduce, launches, traffic). • 1 GPU: 10 ms/token • 2 GPUs (PCIe): 5 + 3 = 8 ms/token → 1.25× speedup, not 2×. Even if you claim the workload is 90% parallel, Amdahl says with N=2: S(2)=\frac{1}{0.1 + 0.9/2}\approx 1.82\times …and real comms/launch overhead push you below that. Please add more material, so this thread acts as a knowledge base. I would love to hear from architects with experience in heterogeneous computing, hpcs and hardware accelerators.
2025-09-22T11:04:42
https://www.reddit.com/r/LocalLLaMA/comments/1nnjgis/amdahls_law_the_hidden_reason_multigpu_setups/
AggravatingGiraffe46
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnjgis
false
null
t3_1nnjgis
/r/LocalLLaMA/comments/1nnjgis/amdahls_law_the_hidden_reason_multigpu_setups/
false
false
self
0
null
Magistral Small 2509 - Jinja Template Modification (Based on Unsloth's) - No thinking by default - straight quick answers in Mistral Small 3.2 style and quality~, need thinking? simple activation with "/think" command anywhere in the system prompt.
53
2025-09-22T10:52:12
https://www.reddit.com/gallery/1nnj83s
-Ellary-
reddit.com
1970-01-01T00:00:00
0
{}
1nnj83s
false
null
t3_1nnj83s
/r/LocalLLaMA/comments/1nnj83s/magistral_small_2509_jinja_template_modification/
false
false
https://b.thumbs.redditm…i_YlopFuEYLc.jpg
53
null
too many qwens
278
2025-09-22T10:49:14
https://i.redd.it/z6ehmb4r4pqf1.png
jacek2023
i.redd.it
1970-01-01T00:00:00
0
{}
1nnj67v
false
null
t3_1nnj67v
/r/LocalLLaMA/comments/1nnj67v/too_many_qwens/
false
false
default
278
{'enabled': True, 'images': [{'id': 'z6ehmb4r4pqf1', 'resolutions': [{'height': 18, 'url': 'https://preview.redd.it/z6ehmb4r4pqf1.png?width=108&crop=smart&auto=webp&s=3f385cdc39939d9417ad08c12779791a81742e7b', 'width': 108}, {'height': 37, 'url': 'https://preview.redd.it/z6ehmb4r4pqf1.png?width=216&crop=smart&auto=webp&s=530c5ad994721715c3199740e52585afeb054321', 'width': 216}, {'height': 56, 'url': 'https://preview.redd.it/z6ehmb4r4pqf1.png?width=320&crop=smart&auto=webp&s=21e9826e24b35c3d1b01fb801f168cc91aa10c50', 'width': 320}, {'height': 112, 'url': 'https://preview.redd.it/z6ehmb4r4pqf1.png?width=640&crop=smart&auto=webp&s=ac3c6bfbc9c3ca1495fb8ed54680114fd1e007ff', 'width': 640}, {'height': 168, 'url': 'https://preview.redd.it/z6ehmb4r4pqf1.png?width=960&crop=smart&auto=webp&s=14472771d151ff4afd6eadbd6492a4e27c82b0c9', 'width': 960}, {'height': 189, 'url': 'https://preview.redd.it/z6ehmb4r4pqf1.png?width=1080&crop=smart&auto=webp&s=c6f114026bc5c0f12f87f600dc10186f0a9e7811', 'width': 1080}], 'source': {'height': 212, 'url': 'https://preview.redd.it/z6ehmb4r4pqf1.png?auto=webp&s=50fd81b122419c04ff3b5ba63b98f45d6fd01042', 'width': 1210}, 'variants': {}}]}
SillyTavern for story writing?
3
ST has many features well suited for story writing despite its actual use case is chat. There are [some "hacks"](https://www.reddit.com/r/SillyTavernAI/comments/1ewz2e7/comment/lj2rkpv/) in order to tweak ST into this direction. Since I am a bit out of the loop, should I still use ST for story writing or are there better ways nowadays or should I just use text-generation-webui and use the system message for the meta info?
2025-09-22T10:22:31
https://www.reddit.com/r/LocalLLaMA/comments/1nnipty/sillytavern_for_story_writing/
rdpl_
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnipty
false
null
t3_1nnipty
/r/LocalLLaMA/comments/1nnipty/sillytavern_for_story_writing/
false
false
self
3
null
Is there a TTS that leverages Vulkan ?
2
Is there a TTS that leverages Vulkan ? FastKokoro is only for CUDA isnt it ? Are there any alternatives
2025-09-22T09:16:11
https://www.reddit.com/r/LocalLLaMA/comments/1nnhn16/is_there_a_tts_that_leverages_vulkan/
uber-linny
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnhn16
false
null
t3_1nnhn16
/r/LocalLLaMA/comments/1nnhn16/is_there_a_tts_that_leverages_vulkan/
false
false
self
2
null
Official FP8-quantizion of Qwen3-Next-80B-A3B
141
[https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Thinking-FP8](https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Thinking-FP8) > #
2025-09-22T09:14:06
https://www.reddit.com/r/LocalLLaMA/comments/1nnhlx5/official_fp8quantizion_of_qwen3next80ba3b/
touhidul002
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnhlx5
false
null
t3_1nnhlx5
/r/LocalLLaMA/comments/1nnhlx5/official_fp8quantizion_of_qwen3next80ba3b/
false
false
self
141
{'enabled': False, 'images': [{'id': 'J6bmxjBLQIXSuYI99phZQ5CU2YaEt-8dZ8s8aUXDpkw', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/J6bmxjBLQIXSuYI99phZQ5CU2YaEt-8dZ8s8aUXDpkw.png?width=108&crop=smart&auto=webp&s=81aa0702320dc6e43d7abdea856906ce7ab28c62', 'width': 108}, {'height': 116, 'url': 'https://external-preview.redd.it/J6bmxjBLQIXSuYI99phZQ5CU2YaEt-8dZ8s8aUXDpkw.png?width=216&crop=smart&auto=webp&s=412254f85237cebfe0a411bd8ae2afbb8fa12950', 'width': 216}, {'height': 172, 'url': 'https://external-preview.redd.it/J6bmxjBLQIXSuYI99phZQ5CU2YaEt-8dZ8s8aUXDpkw.png?width=320&crop=smart&auto=webp&s=67fb8d714144e1b0e84aaf57da075e69a6d4c9a5', 'width': 320}, {'height': 345, 'url': 'https://external-preview.redd.it/J6bmxjBLQIXSuYI99phZQ5CU2YaEt-8dZ8s8aUXDpkw.png?width=640&crop=smart&auto=webp&s=b05097bc589aa04ac1e9786b97629a34932468d1', 'width': 640}, {'height': 518, 'url': 'https://external-preview.redd.it/J6bmxjBLQIXSuYI99phZQ5CU2YaEt-8dZ8s8aUXDpkw.png?width=960&crop=smart&auto=webp&s=a41ffc5fc01a8bd40b11385c0f62ac34f147e746', 'width': 960}, {'height': 583, 'url': 'https://external-preview.redd.it/J6bmxjBLQIXSuYI99phZQ5CU2YaEt-8dZ8s8aUXDpkw.png?width=1080&crop=smart&auto=webp&s=f0d3a4dab699a0663c8486fbb316d5039d31cf58', 'width': 1080}], 'source': {'height': 648, 'url': 'https://external-preview.redd.it/J6bmxjBLQIXSuYI99phZQ5CU2YaEt-8dZ8s8aUXDpkw.png?auto=webp&s=4d232dcd07fde376b37b57da587b4c455e4ca56e', 'width': 1200}, 'variants': {}}]}
SLM suggestion for complex vision tasks.
0
I am working on an MVP to read complex autocad images and obtain information about components on it using SLM deployed on virtual server. Please help out based on your experience with vision SLM and suggest some models that I can experiment with. We are already using paddleOCR for getting the text. The model should be able to/trainable to identify components.
2025-09-22T09:02:05
https://www.reddit.com/r/LocalLLaMA/comments/1nnhfap/slm_suggestion_for_complex_vision_tasks/
CoolCucumberRK
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnhfap
false
null
t3_1nnhfap
/r/LocalLLaMA/comments/1nnhfap/slm_suggestion_for_complex_vision_tasks/
false
false
self
0
null
Question about multi-turn finetuning for a chatbot type finetune
2
Hey, actually I am having a doubt about fine tuning a LLM on my character dataset. To get the best result, I have been looking into masking and padding inside the training scripts I have from claude or perplexity research, sometime gpt5 too. I’m a bit confused about the best approach for multi-turn conversations. When training on a sample conversation, do you think it’s better to: 1. Only train on the **final assistant response** in the conversation, or 2. Train on **all assistant responses** with the context/history of previous turns included? I’m trying to make the chatbot more consistent and natural over multiple turns, but I’m not sure which method works best. I’d really appreciate any advice or experiences you’ve had! Thanks.
2025-09-22T08:54:01
https://www.reddit.com/r/LocalLLaMA/comments/1nnhath/question_about_multiturn_finetuning_for_a_chatbot/
Awkward_Cancel8495
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnhath
false
null
t3_1nnhath
/r/LocalLLaMA/comments/1nnhath/question_about_multiturn_finetuning_for_a_chatbot/
false
false
self
2
null
Making interactive stories that actually react to your choices with AI
1
[removed]
2025-09-22T08:43:17
https://www.reddit.com/r/LocalLLaMA/comments/1nnh50n/making_interactive_stories_that_actually_react_to/
Ok_Thing_2964
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnh50n
false
null
t3_1nnh50n
/r/LocalLLaMA/comments/1nnh50n/making_interactive_stories_that_actually_react_to/
false
false
https://b.thumbs.redditm…l5vTPVXbKHUA.jpg
1
null
Making interactive stories that actually react to your choices with AI
1
[removed]
2025-09-22T08:41:41
https://www.reddit.com/r/LocalLLaMA/comments/1nnh44v/making_interactive_stories_that_actually_react_to/
Ok_Thing_2964
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnh44v
false
null
t3_1nnh44v
/r/LocalLLaMA/comments/1nnh44v/making_interactive_stories_that_actually_react_to/
false
false
https://b.thumbs.redditm…RAKKaynLuaJQ.jpg
1
null
benchmark stt on your own audio for non-english use-cases
1
[removed]
2025-09-22T08:25:57
https://www.reddit.com/r/LocalLLaMA/comments/1nngvtc/benchmark_stt_on_your_own_audio_for_nonenglish/
Wide_Appointment9924
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nngvtc
false
null
t3_1nngvtc
/r/LocalLLaMA/comments/1nngvtc/benchmark_stt_on_your_own_audio_for_nonenglish/
false
false
self
1
null
🎉 PKC Benchmark Tool. Transitioning from Private to Public
1
[removed]
2025-09-22T07:36:42
https://www.reddit.com/r/LocalLLaMA/comments/1nng5dy/pkc_benchmark_tool_transitioning_from_private_to/
Mission-Crab-9919
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nng5dy
false
null
t3_1nng5dy
/r/LocalLLaMA/comments/1nng5dy/pkc_benchmark_tool_transitioning_from_private_to/
false
false
self
1
null
Is there any performance / stability difference between Windows and Linux (due to NVIDIA drivers?)
2
Hi, newbie to AI stuff here, wanting to get started. It's commonly known by the gaming community that the Linux drivers for NVIDIA aren't as good as we would want. I just wanted to ask whether this has any impact on Local AI stuff? (Which I understand also runs on the GPU.) I'm dual booting Windows and Linux, so I wanted to know which OS I should install my AI stuff on. Any advice would be much appreciated, thanks!
2025-09-22T07:25:36
https://www.reddit.com/r/LocalLLaMA/comments/1nnfzgx/is_there_any_performance_stability_difference/
zeddyzed
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnfzgx
false
null
t3_1nnfzgx
/r/LocalLLaMA/comments/1nnfzgx/is_there_any_performance_stability_difference/
false
false
self
2
null
Moving from Cursor to Qwen-code
45
Never been faster & happier, I basically live in terminal. Definitely recommend.
2025-09-22T07:20:39
https://www.reddit.com/r/LocalLLaMA/comments/1nnfwmo/moving_from_cursor_to_qwencode/
Honest-Debate-6863
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnfwmo
false
null
t3_1nnfwmo
/r/LocalLLaMA/comments/1nnfwmo/moving_from_cursor_to_qwencode/
false
false
self
45
null
Looking for a Local AI Coding Agent Like ZenCoder – Want to Build Websites for Fun Without the Headache
1
[removed]
2025-09-22T07:19:50
https://www.reddit.com/r/LocalLLaMA/comments/1nnfw5s/looking_for_a_local_ai_coding_agent_like_zencoder/
No-Window8788
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnfw5s
false
null
t3_1nnfw5s
/r/LocalLLaMA/comments/1nnfw5s/looking_for_a_local_ai_coding_agent_like_zencoder/
false
false
self
1
null
🎉 PKC Benchmark Tool (Transitioning from Private to Public)
1
🎉 PKC Benchmark Tool (Transitioning from Private to Public) TL;DR: I am preparing to release a benchmark tool to the world, spun off from a local multimodal chatbot (2060 SUPER) project that I've been personally building for the past 2 months, with a "they'll figure it out" attitude. Here's a little sneak peek. 🚨 WARNING: This is still a work-in-progress project "If you're expecting perfect software, please press the back button." This project is being developed with the following philosophy: ✅ It's good enough for my use. ✅ It might be helpful to someone someday. ❌ Even upon release, I do not guarantee it will work perfectly in every environment. ❌ I will not be operating a 24-hour customer support center. 🤷‍♂️ After it's public, if it doesn't run on your computer, please fork it and fix it yourself. 🎯 The Miracle of One-Click (The Goal) This is being prepared for those who think, "Python? Virtual environment? What's that?": The goal is a simple one-click installation. When released, it will work like this: For Windows Users Double-click OneClick_RUN.bat ☕ Wait while having a cup of coffee A browser will open automatically Done (Similar simple processes for macOS and Linux users will also be included.) All installations will happen inside a dedicated jail called .pkc-venv, so your system will be safe. 🤖 What's Been Made: The History of Evolution | Version | Status | Description | | v4 | 🧠 | "Just detect the GPU automatically!" → Added auto-selection for CUDA/MPS/CPU. | | v5 | 🎯 | "How can a non-expert use this if it's not one-click?" → Automated the virtual environment. | | v5.3 | 💅 | "The UI is too clunky." → Slimmed down and right-aligned the input fields. | | v5.4 | 📝 | "One prompt isn't enough." → Expanded to 3 fields. | | v5.5 | 🌐 | "What about non-Korean speakers?" → Auto-detects browser language. | | v5.6 | 🚀 | "Throw in all the F1 features!" → Integrated HuggingFace search/download. | 🚀 Roadmap & How to Participate This project is under active development. There is no public download link yet, but the plan leading up to the first public release is as follows. Current Status ✅ Core benchmarking logic ✅ One-click installation scripts (Windows, macOS, Linux) ✅ HuggingFace model search and download integration ✅ Automatic hardware detection (GPU/CPU) ✅ Basic UI capable of displaying real-time results Next Goals (Before v1.0 Release) 🚧 Polishing the UI/UX 🚧 Adding more detailed charting and data export options 💡 Writing comprehensive documentation and guides 💡 Creating a simple plugin system for custom metrics 🤝 The Spirit of Open Source & License This project is scheduled to be released under the PKC Non-Commercial Attribution License v1.0. What you CAN do: ✅ Use it freely (for personal/educational/research purposes) ✅ Look through the code ✅ Improve and redistribute it ✅ Fork it and create a completely different project What you CANNOT do: ❌ Sell it for money ❌ Claim "I made this" ❌ Remove the PKC author attribution ❌ Argue with PKC 🎭 A Developer's Candid Confession Me, 2 months ago: "Hey AI, just make me a chatbot that fits my PC specs." Me, now: "Somehow I ended up making a benchmark tool... It's a waste to use it alone, so I should prepare to release it." Me, in the future (probably): "Why did I release this..." 📞 Contact & Updates The project is not yet public, but if you have questions or want to receive updates, feel free to contact me anytime. Contact Info 📧 Please contact me through my profile. 💬 GitHub: (The repository will be public soon!) 🎪 One Last Word This tool is not "perfect software." It's just a "work in progress." Please keep an eye on it, and get ready to welcome this tool made with a "they'll figure it out" spirit. Made with ❤️ and lots of ☕ by PKC "Let's release it even if it's not perfect, and evolve it to be usable on common specs." P.S. I firmly believe that once the repository is public, someone will write a better announcement than this one and send me a PR. lol
2025-09-22T07:17:27
https://www.reddit.com/r/LocalLLaMA/comments/1nnfuu4/pkc_benchmark_tool_transitioning_from_private_to/
Mission-Crab-9919
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnfuu4
false
null
t3_1nnfuu4
/r/LocalLLaMA/comments/1nnfuu4/pkc_benchmark_tool_transitioning_from_private_to/
false
false
self
1
null
GLM-4.5V model for local computer use
35
On OSWorld-V, it scores 35.8% - beating UI-TARS-1.5, matching Claude-3.7-Sonnet-20250219, and setting SOTA for fully open-source computer-use models. Run it with Cua either: Locally via Hugging Face Remotely via OpenRouter Github : https://github.com/trycua Docs + examples: https://docs.trycua.com/docs/agent-sdk/supported-agents/computer-use-agents#glm-45v
2025-09-22T05:49:19
https://v.redd.it/s7ecb9y9nnqf1
Impressive_Half_2819
v.redd.it
1970-01-01T00:00:00
0
{}
1nnefs0
false
{'reddit_video': {'bitrate_kbps': 800, 'dash_url': 'https://v.redd.it/s7ecb9y9nnqf1/DASHPlaylist.mpd?a=1761112175%2COTY5NDVhZWJhNThlNDUwZTQwMzJkMzEwMjY2NjJmNDRjM2RkYmMxN2VmMzJjYjNiOGRiNjRjM2I4NWIzM2VlMA%3D%3D&v=1&f=sd', 'duration': 37, 'fallback_url': 'https://v.redd.it/s7ecb9y9nnqf1/DASH_360.mp4?source=fallback', 'has_audio': True, 'height': 278, 'hls_url': 'https://v.redd.it/s7ecb9y9nnqf1/HLSPlaylist.m3u8?a=1761112175%2CZDQzMWU5MGNhOGE1ODA0ZTU3MjAxNzAyMDk0NDY5NTljNjAzZjJhMDE4MjI2NTI5MGMzNmIzNzU3ZjQ0OTU0Mg%3D%3D&v=1&f=sd', 'is_gif': False, 'scrubber_media_url': 'https://v.redd.it/s7ecb9y9nnqf1/DASH_96.mp4', 'transcoding_status': 'completed', 'width': 640}}
t3_1nnefs0
/r/LocalLLaMA/comments/1nnefs0/glm45v_model_for_local_computer_use/
false
false
https://external-preview…510bcff41e789bbe
35
{'enabled': False, 'images': [{'id': 'MTdjemJhbjlubnFmMbWADQNBSkImjVESNjfi_q43l9ostHKNGAFX_QJdfnS0', 'resolutions': [{'height': 47, 'url': 'https://external-preview.redd.it/MTdjemJhbjlubnFmMbWADQNBSkImjVESNjfi_q43l9ostHKNGAFX_QJdfnS0.png?width=108&crop=smart&format=pjpg&auto=webp&s=2660d27b67543c37f48a92e680e02613151e3bc8', 'width': 108}, {'height': 94, 'url': 'https://external-preview.redd.it/MTdjemJhbjlubnFmMbWADQNBSkImjVESNjfi_q43l9ostHKNGAFX_QJdfnS0.png?width=216&crop=smart&format=pjpg&auto=webp&s=1f09c9d62078b71b9a305bc5975b00345734803f', 'width': 216}, {'height': 139, 'url': 'https://external-preview.redd.it/MTdjemJhbjlubnFmMbWADQNBSkImjVESNjfi_q43l9ostHKNGAFX_QJdfnS0.png?width=320&crop=smart&format=pjpg&auto=webp&s=e0c7dc2b5f52eee88e7c8543a6eaa7faf3c0fc2c', 'width': 320}, {'height': 278, 'url': 'https://external-preview.redd.it/MTdjemJhbjlubnFmMbWADQNBSkImjVESNjfi_q43l9ostHKNGAFX_QJdfnS0.png?width=640&crop=smart&format=pjpg&auto=webp&s=8e41a5a56d352404222a61669ca78e3322ed137a', 'width': 640}], 'source': {'height': 372, 'url': 'https://external-preview.redd.it/MTdjemJhbjlubnFmMbWADQNBSkImjVESNjfi_q43l9ostHKNGAFX_QJdfnS0.png?format=pjpg&auto=webp&s=9eb9392f305cbdeaa86da37461ae33b8d768e0ab', 'width': 854}, 'variants': {}}]}
How do I disable thinking in Deepseek V3.1?
11
``` llama-cli -hf unsloth/DeepSeek-V3.1-GGUF:Q5_K_XL \ --jinja --mlock \ --prio 3 -ngl 99 --cpu-moe \ --temp 0.6 --top_p 0.95 --min_p 0.01 --ctx-size $((128*1024)) \ -t 128 -b 10240 \ -p "Tell me about PCA." --verbose-prompt # ... log output main: prompt: '/no_think Tell me about PCA.' main: number of tokens in prompt = 12 0 -> '<|begin▁of▁sentence|>' 128803 -> '<|User|>' 91306 -> '/no' 65 -> '_' 37947 -> 'think' 32536 -> ' Tell' 678 -> ' me' 943 -> ' about' 78896 -> ' PCA' 16 -> '.' 128804 -> '<|Assistant|>' 128798 -> '<think>' # more log output Tell me about PCA.<think>Hmm, the user asked about PCA. They probably want a straightforward, jargon-free explanation without overcomplicating it. Since PCA is a technical topic, I should balance simplicity with accuracy. I'll start with a high-level intuition—comparing it to photo compression—to make it relatable. Then, I'll break down the core ideas: variance, eigenvectors, and dimensionality reduction, but keep it concise. No need for deep math unless the user asks. The response should end with a clear summary of pros and cons, since practical use cases matter. Avoid tangents—stick to what PCA is, why it's useful, and when to use it.</think>Of course. Here is a straightforward explanation of Principal Component Analysis (PCA). ### The Core Idea in Simple Terms ``` I've tried /no_think, \no_think, --reasoning-budget 0, etc. None of that seems to work.
2025-09-22T05:29:09
https://www.reddit.com/r/LocalLLaMA/comments/1nne3ra/how_do_i_disable_thinking_in_deepseek_v31/
MengerianMango
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nne3ra
false
null
t3_1nne3ra
/r/LocalLLaMA/comments/1nne3ra/how_do_i_disable_thinking_in_deepseek_v31/
false
false
self
11
null
China can destabilize the US via AI and unemployment
0
>Goodwill CEO says he’s preparing for an influx of jobless Gen Zers because of AI—and warns, a youth unemployment crisis is already happening [https://www.msn.com/en-us/money/companies/goodwill-ceo-says-he-s-preparing-for-an-influx-of-jobless-gen-zers-because-of-ai-and-warns-a-youth-unemployment-crisis-is-already-happening/ar-AA1MZMp3](https://www.msn.com/en-us/money/companies/goodwill-ceo-says-he-s-preparing-for-an-influx-of-jobless-gen-zers-because-of-ai-and-warns-a-youth-unemployment-crisis-is-already-happening/ar-AA1MZMp3) China has an economic technocracy than likely can absorb and adjust to AI with much less social upheaval than capitalistic democratic nations. By sharing capable models that can facilitate replacing junior and even mid level workers, they can cause a very large degree of disruption in the west. They don't even have to share models with dangerous capability, just models that hallucinate much less and perform reliably and consistently at above average IQ. I suspect we will see a rising call for banning of Chinese models pretty soon on the horizon. #
2025-09-22T05:21:19
https://www.reddit.com/r/LocalLLaMA/comments/1nndz5c/china_can_destabilize_the_us_via_ai_and/
kaggleqrdl
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nndz5c
false
null
t3_1nndz5c
/r/LocalLLaMA/comments/1nndz5c/china_can_destabilize_the_us_via_ai_and/
false
false
self
0
null
I'm curious of your set-ups 🤔
0
I'm kinda curious of your set-ups you people around here 🤔🤔 what are your specs and setups? Mines is actually A: -Llama 3.2 3B 131k but at x1 500K RoPE set at 32k context max -costum wrapper I made for myself -running pure rx 5500 xt 8Gb ddr6 OC at 1964mhz 1075mv core and Vram at 1860mhz Vulkan. Sipping 100-115 watts full load gpu only metrics. -4k-8k context I hover around 33-42 tokens per sec mostly 30-33 tokens if has ambience or codes -10k-20k ctx i tank down to 15-18 tokens per sec -24k-32k context I hover 8-11 tokens per sec I don't dip below 7 - tested my fine-tuned Llama 3.2 can actually track everything even at 32k no hallucinations on my costum wrapper as i arranged the memory and injected files properly labeled them like a librarian. So ya guys.. i wanna know your spec 😂 i actually am limited to 3B cuz I'm only using an rx 5500 xt i wonder how your 8B to 70B feels like.. i usually use mine for lite coding and very heavy roleplay with ambience and multi NPC and dungeon crawling with loots chest and monsters kinda cool my 3B can track everything tho.
2025-09-22T05:06:45
https://www.reddit.com/r/LocalLLaMA/comments/1nndqcb/im_curious_of_your_setups/
DigRealistic2977
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nndqcb
false
null
t3_1nndqcb
/r/LocalLLaMA/comments/1nndqcb/im_curious_of_your_setups/
false
false
self
0
null
baidu releases Qianfan-VL 70B/8B/3B
101
[https://huggingface.co/baidu/Qianfan-VL-8B](https://huggingface.co/baidu/Qianfan-VL-8B) [https://huggingface.co/baidu/Qianfan-VL-70B](https://huggingface.co/baidu/Qianfan-VL-70B) [https://huggingface.co/baidu/Qianfan-VL-3B](https://huggingface.co/baidu/Qianfan-VL-3B) # Model Description Qianfan-VL is a series of general-purpose multimodal large language models enhanced for enterprise-level multimodal applications. The models offer deep optimization for high-frequency scenarios in industrial deployment while maintaining strong general capabilities. # [](https://huggingface.co/baidu/Qianfan-VL-70B#model-variants) # Model Variants |Model|Parameters|Context Length|CoT Support|Best For| |:-|:-|:-|:-|:-| |**Qianfan-VL-3B**|3B|32k|❌|Edge deployment, real-time OCR| |**Qianfan-VL-8B**|8B|32k|✅|Server-side general scenarios, fine-tuning| |**Qianfan-VL-70B**|70B|32k|✅|Complex reasoning, data synthesis| # [](https://huggingface.co/baidu/Qianfan-VL-70B#architecture) # Architecture * **Language Model**: * Qianfan-VL-3B: Based on Qwen2.5-3B * Qianfan-VL-8B/70B: Based on Llama 3.1 architecture * Enhanced with 3T multilingual corpus * **Vision Encoder**: InternViT-based, supports dynamic patching up to 4K resolution * **Cross-modal Fusion**: MLP adapter for efficient vision-language bridging # [](https://huggingface.co/baidu/Qianfan-VL-70B#key-capabilities) # Key Capabilities # [](https://huggingface.co/baidu/Qianfan-VL-70B#🔍-ocr--document-understanding) # 🔍 OCR & Document Understanding * **Full-Scenario OCR**: Handwriting, formulas, natural scenes, cards/documents * **Document Intelligence**: Layout analysis, table parsing, chart understanding, document Q&A * **High Precision**: Industry-leading performance on OCR benchmarks # [](https://huggingface.co/baidu/Qianfan-VL-70B#🧮-chain-of-thought-reasoning-8b--70b) # 🧮 Chain-of-Thought Reasoning (8B & 70B) * Complex chart analysis and reasoning * Mathematical problem-solving with step-by-step derivation * Visual reasoning and logical inference * Statistical computation and trend predictionModel Description Qianfan-VL is a series of general-purpose multimodal large language models enhanced for enterprise-level multimodal applications. The models offer deep optimization for high-frequency scenarios in industrial deployment while maintaining strong general capabilities. Model Variants ModelParametersContext LengthCoT SupportBest For Qianfan-VL-3B3B32k❌Edge deployment, real-time OCR Qianfan-VL-8B8B32k✅Server-side general scenarios, fine-tuning Qianfan-VL-70B70B32k✅Complex reasoning, data synthesis Architecture Language Model: Qianfan-VL-3B: Based on Qwen2.5-3B Qianfan-VL-8B/70B: Based on Llama 3.1 architecture Enhanced with 3T multilingual corpus Vision Encoder: InternViT-based, supports dynamic patching up to 4K resolution Cross-modal Fusion: MLP adapter for efficient vision-language bridging Key Capabilities 🔍 OCR & Document Understanding Full-Scenario OCR: Handwriting, formulas, natural scenes, cards/documents Document Intelligence: Layout analysis, table parsing, chart understanding, document Q&A High Precision: Industry-leading performance on OCR benchmarks 🧮 Chain-of-Thought Reasoning (8B & 70B) Complex chart analysis and reasoning Mathematical problem-solving with step-by-step derivation Visual reasoning and logical inference Statistical computation and trend prediction
2025-09-22T04:23:58
https://www.reddit.com/r/LocalLLaMA/comments/1nncyvv/baidu_releases_qianfanvl_70b8b3b/
jacek2023
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nncyvv
false
null
t3_1nncyvv
/r/LocalLLaMA/comments/1nncyvv/baidu_releases_qianfanvl_70b8b3b/
false
false
self
101
{'enabled': False, 'images': [{'id': 'FP29XFJejhThKtWx9YKFfaOHtbR8mIg5N90GWHDW28A', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/FP29XFJejhThKtWx9YKFfaOHtbR8mIg5N90GWHDW28A.png?width=108&crop=smart&auto=webp&s=715ed3b30487127703ca07b133034853c75dcb26', 'width': 108}, {'height': 116, 'url': 'https://external-preview.redd.it/FP29XFJejhThKtWx9YKFfaOHtbR8mIg5N90GWHDW28A.png?width=216&crop=smart&auto=webp&s=3e1c9424fea38612416fa1f643cb2a034aa4008a', 'width': 216}, {'height': 172, 'url': 'https://external-preview.redd.it/FP29XFJejhThKtWx9YKFfaOHtbR8mIg5N90GWHDW28A.png?width=320&crop=smart&auto=webp&s=4b5a8d96348ef8f8a0a6ad475d6c2c9a5afc973b', 'width': 320}, {'height': 345, 'url': 'https://external-preview.redd.it/FP29XFJejhThKtWx9YKFfaOHtbR8mIg5N90GWHDW28A.png?width=640&crop=smart&auto=webp&s=5741537902444a354e22ca4de472cf0e35ca94b1', 'width': 640}, {'height': 518, 'url': 'https://external-preview.redd.it/FP29XFJejhThKtWx9YKFfaOHtbR8mIg5N90GWHDW28A.png?width=960&crop=smart&auto=webp&s=b325c9fdbdfc199b942ef39f15c7bdda192d6af0', 'width': 960}, {'height': 583, 'url': 'https://external-preview.redd.it/FP29XFJejhThKtWx9YKFfaOHtbR8mIg5N90GWHDW28A.png?width=1080&crop=smart&auto=webp&s=59a0b2f869447753bd288f29ee89c52773610c1a', 'width': 1080}], 'source': {'height': 648, 'url': 'https://external-preview.redd.it/FP29XFJejhThKtWx9YKFfaOHtbR8mIg5N90GWHDW28A.png?auto=webp&s=ba49402bb1b6fd2c126e92493ecfb6af64d6e02b', 'width': 1200}, 'variants': {}}]}
Qwen3-Omni Promotional Video
150
https://www.youtube.com/watch?v=RRlAen2kIUU Qwen dropped a promotional video for Qwen3-Omni, looks like the weights are just around the corner!
2025-09-22T04:14:46
https://www.reddit.com/r/LocalLLaMA/comments/1nncssq/qwen3omni_promotional_video/
Mysterious_Finish543
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nncssq
false
null
t3_1nncssq
/r/LocalLLaMA/comments/1nncssq/qwen3omni_promotional_video/
false
false
self
150
{'enabled': False, 'images': [{'id': 'B4ZZCzuzrlsMnQHNhsoc21qTthMSpFr8qrrtucUS_RU', 'resolutions': [{'height': 81, 'url': 'https://external-preview.redd.it/B4ZZCzuzrlsMnQHNhsoc21qTthMSpFr8qrrtucUS_RU.jpeg?width=108&crop=smart&auto=webp&s=b2dc605a9d17b37333d858d90a676d6d14af9b49', 'width': 108}, {'height': 162, 'url': 'https://external-preview.redd.it/B4ZZCzuzrlsMnQHNhsoc21qTthMSpFr8qrrtucUS_RU.jpeg?width=216&crop=smart&auto=webp&s=9967c97b85fef987c5cd8dc125a2bb4733cf7797', 'width': 216}, {'height': 240, 'url': 'https://external-preview.redd.it/B4ZZCzuzrlsMnQHNhsoc21qTthMSpFr8qrrtucUS_RU.jpeg?width=320&crop=smart&auto=webp&s=7250574f82c5b2852f214b634dbe23e3e38e029b', 'width': 320}], 'source': {'height': 360, 'url': 'https://external-preview.redd.it/B4ZZCzuzrlsMnQHNhsoc21qTthMSpFr8qrrtucUS_RU.jpeg?auto=webp&s=fe8dc761c40aecc080a2e981d052d64397487760', 'width': 480}, 'variants': {}}]}
How to stabilize inference for Qwen/Qwen3-Next-80B-A3B-Instruct on 4× H20 96GB (vLLM OpenAI nightly)
1
[removed]
2025-09-22T03:26:05
https://www.reddit.com/r/LocalLLaMA/comments/1nnbvwu/how_to_stabilize_inference_for/
SatisfactionWarm4386
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nnbvwu
false
null
t3_1nnbvwu
/r/LocalLLaMA/comments/1nnbvwu/how_to_stabilize_inference_for/
false
false
https://b.thumbs.redditm…MWS30k1ZHw0Y.jpg
1
null
I'll show you mine, if you show me yours: Local AI tech stack September 2025
304
2025-09-22T02:53:16
https://i.redd.it/rq2ple7trmqf1.png
JLeonsarmiento
i.redd.it
1970-01-01T00:00:00
0
{}
1nnb8sq
false
null
t3_1nnb8sq
/r/LocalLLaMA/comments/1nnb8sq/ill_show_you_mine_if_you_show_me_yours_local_ai/
false
false
https://a.thumbs.redditm…enuI9VsecCp0.jpg
304
{'enabled': True, 'images': [{'id': 'rfJvvw-7jdYL6EuigHm1RleuykCDew_P25800oDuO5A', 'resolutions': [{'height': 123, 'url': 'https://preview.redd.it/rq2ple7trmqf1.png?width=108&crop=smart&auto=webp&s=15eb765d4dce3cc5ec0015be28fa8fb22b909833', 'width': 108}, {'height': 247, 'url': 'https://preview.redd.it/rq2ple7trmqf1.png?width=216&crop=smart&auto=webp&s=88e9cb0c1f56beed111374e3c35131e8fb4a2475', 'width': 216}, {'height': 366, 'url': 'https://preview.redd.it/rq2ple7trmqf1.png?width=320&crop=smart&auto=webp&s=fca4f9cbc8edfcc541a37bf1def50d887ef69a56', 'width': 320}, {'height': 732, 'url': 'https://preview.redd.it/rq2ple7trmqf1.png?width=640&crop=smart&auto=webp&s=0ba6a2e65b89f81d77c39c353119c9e596157a9b', 'width': 640}, {'height': 1099, 'url': 'https://preview.redd.it/rq2ple7trmqf1.png?width=960&crop=smart&auto=webp&s=b0a110f1ddb8d663117f41c04271d7f75cad668e', 'width': 960}], 'source': {'height': 1138, 'url': 'https://preview.redd.it/rq2ple7trmqf1.png?auto=webp&s=5d83ca0f1a00d39f3923e631c90491f5af7f6a5c', 'width': 994}, 'variants': {}}]}
One of us. Well at least in spirit.
3
2025-09-22T01:54:37
https://i.redd.it/n15qmulehmqf1.jpeg
No_Conversation9561
i.redd.it
1970-01-01T00:00:00
0
{}
1nna1k3
false
null
t3_1nna1k3
/r/LocalLLaMA/comments/1nna1k3/one_of_us_well_at_least_in_spirit/
false
false
default
3
{'enabled': True, 'images': [{'id': 'n15qmulehmqf1', 'resolutions': [{'height': 83, 'url': 'https://preview.redd.it/n15qmulehmqf1.jpeg?width=108&crop=smart&auto=webp&s=a7e431e840d3967e7a229611051ffd11e66afb5a', 'width': 108}, {'height': 167, 'url': 'https://preview.redd.it/n15qmulehmqf1.jpeg?width=216&crop=smart&auto=webp&s=bef3c699089be699432854b12356ed3216004756', 'width': 216}, {'height': 248, 'url': 'https://preview.redd.it/n15qmulehmqf1.jpeg?width=320&crop=smart&auto=webp&s=f9735663888ec60cad2d49e39cde1714feb15742', 'width': 320}, {'height': 496, 'url': 'https://preview.redd.it/n15qmulehmqf1.jpeg?width=640&crop=smart&auto=webp&s=59094d6e3b02fc62ef4b5fcadc2b029f984ad57f', 'width': 640}, {'height': 744, 'url': 'https://preview.redd.it/n15qmulehmqf1.jpeg?width=960&crop=smart&auto=webp&s=6ff8c2d3eae4928d32a3a131ce48534de68b23f9', 'width': 960}, {'height': 837, 'url': 'https://preview.redd.it/n15qmulehmqf1.jpeg?width=1080&crop=smart&auto=webp&s=cc2ad24300db9184b107e2523cd2bbdfe75c5180', 'width': 1080}], 'source': {'height': 1588, 'url': 'https://preview.redd.it/n15qmulehmqf1.jpeg?auto=webp&s=44f7fd44c3dca76809b70ff819606db9518e0c52', 'width': 2048}, 'variants': {}}]}
Sophia NLU Engine Upgrade - New and Improved POS Tagger
7
Just released large upgrade to Sophia NLU Engine, which includes a new and improved POS tagger along with a revamped automated spelling corrections system. POS tagger now gets 99.03% accuracy across 34 million validation tokens, still blazingly fast at ~20,000 words/sec, plus the size of the vocab data store dropped from 238MB to 142MB for a savings of 96MB which was a nice bonus. Full details, online demo and source code at: https://cicero.sh/sophia/ Release announcement at: https://cicero.sh/r/sophia-upgrade-pos-tagger Enjoy! More coming, namely contextual awareness shortly. Sophia = self hosted, privacy focused NLU (natural language understanding) engine. No external dependencies or API calls to big tech, self contained, blazingly fast, and accurate.
2025-09-22T00:33:14
https://www.reddit.com/r/LocalLLaMA/comments/1nn8csq/sophia_nlu_engine_upgrade_new_and_improved_pos/
mdizak
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn8csq
false
null
t3_1nn8csq
/r/LocalLLaMA/comments/1nn8csq/sophia_nlu_engine_upgrade_new_and_improved_pos/
false
false
self
7
null
Looking for TTS model for Japanese voice cloning to English tts
3
Hi, I'm looking for a good TTS model that supports voice input of another language (JP) and get English text. The text it will use for speech itself is in English so there's no translation process. There are no speed requirements and also no hardware requirements (but it would be nice if you mentioned what would be needed). Ideally it is expressive either by using tagged text or naturally expressive, but I care most about the quality.
2025-09-22T00:03:11
https://www.reddit.com/r/LocalLLaMA/comments/1nn7q2u/looking_for_tts_model_for_japanese_voice_cloning/
Anthonyy232
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn7q2u
false
null
t3_1nn7q2u
/r/LocalLLaMA/comments/1nn7q2u/looking_for_tts_model_for_japanese_voice_cloning/
false
false
self
3
null
Can some distill madlad-400?
2
I am making something but I don't have any compute for distillation. Don't know if I should ask directly but this is all I wanted as of now.
2025-09-21T23:56:56
https://www.reddit.com/r/LocalLLaMA/comments/1nn7l57/can_some_distill_madlad400/
Away_Expression_3713
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn7l57
false
null
t3_1nn7l57
/r/LocalLLaMA/comments/1nn7l57/can_some_distill_madlad400/
false
false
self
2
null
Need some advice on building a dedicated LLM server
18
My mom wants me to build her a server for her business so she can query some LLMs locally for things that involve confidential/copyrighted data. I'm currently imagining something that can hit 20-30B models like Gemma 3 27B with a decently large context window. I've got a solid idea of what to build, but I'd like some of y'all's opinions and recommendations. # GPU I'm currently looking at the RTX 5090. It's relatively expensive, but my mom insists that she wants *the best* out there (within reason obviously, so an RTX PRO 6000 is out of the question lol). However some things about the 5090 concern me, particularly the 12HPWR connector. I'm not really up-to-date on the whole ordeal, but I don't think I'd be comfortable letting a machine running 24/7 in our basement unchecked with this connector. Maybe it would be worth looking into a 7900XTX? It has 8 GB less VRAM and significantly lower inference speeds, but it's also less than 1/3rd the price, not to mention it won't require as beefy a PSU and as big a case. To me the 7900XTX sounds like the saner option, but I'd like some external input. # Other components Beyond the GPU, I'm not really sure what components I should be looking to get for a dedicated inference host. Case and PSU aside, would it be fine to go with a cheap AM4 system? Or would DDR5 and a PCIe 5.0 x 16 slot make it worth going for an AM5 system? For storage, I'm thinking it would be nice to have something with relatively high read bandwidth to reduce that waiting time when a model is being loaded into memory. I'm thinking of getting 2 decently fast SSDs and pairing them in a RAID0 configuration. Would that be a good option or should I just get a single, really expensive PCIe 5.0 SSD with really fast read speeds? If I'm going with the RAID0 config, would motherboard RAID0 do the job or should I look at dedicated RAID hardware (or software)? # Software For now, I'm thinking of setting up Open WebUI with either llama.cpp or Ollama. My mom seems to like Open WebUI and it's a solid chatbot wrapper overall, but are there other options that are worth considering? I've only dabbled with locall LLMs and don't really know about the alternatives. I'm also not sure what flavour of Linux I should be using for a headless server, so I'll take any recommendations. Preferably something stable that can play well with Nvidia drivers (if I end up getting a 5090). Any input is greatly appreciated!
2025-09-21T23:55:35
https://www.reddit.com/r/LocalLLaMA/comments/1nn7k4h/need_some_advice_on_building_a_dedicated_llm/
SomeKindOfSorbet
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn7k4h
false
null
t3_1nn7k4h
/r/LocalLLaMA/comments/1nn7k4h/need_some_advice_on_building_a_dedicated_llm/
false
false
self
18
null
How do agents access real-time data without building custom backends?
0
Most AI agents today can reason well, but they struggle to act in real-world tasks because they lack real-time data access. Models already have knowledge, but what they miss is information, which is the current context needed for accurate decisions. Traditionally, this is solved by backend engineers: connecting APIs, cleaning files, and aggregating databases into something usable. But building and maintaining custom backends for every workflow is expensive and slow. For agents to scale, they’ll need a standardized way to fetch structured, up-to-date data. Similar to how APIs standardized communication between apps. Once this layer exists, calling data into an agent workflow should be as simple as an API call today. That’s exactly the problem we’re working on with **Sheet0**. Instead of writing scripts or spinning up pipelines, you just describe the data you need, and the agent delivers a clean, real-time spreadsheet. → Give a try: [sheet0.com](http://sheet0.com) → The invitation code: Try **SP95DJD5**, or join discord [try.sheet0.com/community](http://try.sheet0.com/community) to get the code Curious to hear about: **What solutions have you seen (or built) that give agents reliable real-time data without custom backends?**
2025-09-21T23:50:55
https://www.reddit.com/r/LocalLLaMA/comments/1nn7gnr/how_do_agents_access_realtime_data_without/
Just-Increase-4890
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn7gnr
false
null
t3_1nn7gnr
/r/LocalLLaMA/comments/1nn7gnr/how_do_agents_access_realtime_data_without/
false
false
self
0
null
Perplexica for Siri
6
For users of [Perplexica](https://github.com/ItzCrazyKns/Perplexica), the open source AI search tool: I created this iOS shortcut that leverages the Perplexica api so I could send search queries to my Perplexica instance while in my car. Wanted to share because it's been super useful to have a completely private AI voice search using carplay. Also works with Siri on an iPhone. Enjoy! [https://www.icloud.com/shortcuts/64b69e50a0144c6799b47947c13505e3](https://www.icloud.com/shortcuts/64b69e50a0144c6799b47947c13505e3)
2025-09-21T23:50:34
https://www.reddit.com/r/LocalLLaMA/comments/1nn7gdv/perplexica_for_siri/
No_Information9314
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn7gdv
false
null
t3_1nn7gdv
/r/LocalLLaMA/comments/1nn7gdv/perplexica_for_siri/
false
false
self
6
{'enabled': False, 'images': [{'id': 'EbEpin0SFWbZTkNqyRFvAKTwDz_KqYIW1fyCm5RHCcw', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/EbEpin0SFWbZTkNqyRFvAKTwDz_KqYIW1fyCm5RHCcw.png?width=108&crop=smart&auto=webp&s=05d327dddfb3d122a5bbea176ba825b18bbec20c', 'width': 108}, {'height': 108, 'url': 'https://external-preview.redd.it/EbEpin0SFWbZTkNqyRFvAKTwDz_KqYIW1fyCm5RHCcw.png?width=216&crop=smart&auto=webp&s=67e59c04507a66d66b78fe83c680997e177000b1', 'width': 216}, {'height': 160, 'url': 'https://external-preview.redd.it/EbEpin0SFWbZTkNqyRFvAKTwDz_KqYIW1fyCm5RHCcw.png?width=320&crop=smart&auto=webp&s=98ac83e728d7da8eb66fe898cd4ea46890d02545', 'width': 320}, {'height': 320, 'url': 'https://external-preview.redd.it/EbEpin0SFWbZTkNqyRFvAKTwDz_KqYIW1fyCm5RHCcw.png?width=640&crop=smart&auto=webp&s=49985d29e1d3f986ded18bef79a4ec5bd65d7094', 'width': 640}, {'height': 480, 'url': 'https://external-preview.redd.it/EbEpin0SFWbZTkNqyRFvAKTwDz_KqYIW1fyCm5RHCcw.png?width=960&crop=smart&auto=webp&s=f184098802315be428dd328384be01bd7735a16d', 'width': 960}, {'height': 540, 'url': 'https://external-preview.redd.it/EbEpin0SFWbZTkNqyRFvAKTwDz_KqYIW1fyCm5RHCcw.png?width=1080&crop=smart&auto=webp&s=df33cf3acd8a9a6cf0f291c495315aff153d0225', 'width': 1080}], 'source': {'height': 640, 'url': 'https://external-preview.redd.it/EbEpin0SFWbZTkNqyRFvAKTwDz_KqYIW1fyCm5RHCcw.png?auto=webp&s=6a07ef7df7d59aee66b91b987af495b671db9557', 'width': 1280}, 'variants': {}}]}
Optimizing gpt-oss-120b local inference speed on consumer hardware
81
* Got GPT‑OSS‑120B running with llama.cpp on mid‑range hardware – i5‑12600K + RTX 4070 (12 GB) + 64 GB DDR5 – ≈191 tps prompt, ≈10 tps generation with a 24k context window. * Distilled r/LocalLLaMA tips & community tweaks into an article (run script, benchmarks). * Feedback and further tuning ideas welcome! *script + step‑by‑step tuning guide ➜*  [https://carteakey.dev/optimizing%20gpt-oss-120b-local%20inference/](https://carteakey.dev/optimizing%20gpt-oss-120b-local%20inference/)
2025-09-21T23:32:32
https://carteakey.dev/optimizing%20gpt-oss-120b-local%20inference/
carteakey
carteakey.dev
1970-01-01T00:00:00
0
{}
1nn72ji
false
null
t3_1nn72ji
/r/LocalLLaMA/comments/1nn72ji/optimizing_gptoss120b_local_inference_speed_on/
false
false
default
81
null
i5-8500 64GB RAM working great?
1
I have an old desktop and decided to try ollama with it. Its a lenovo m920s with an i5-8500 and 64gb ram. I installed qwen2.5-coder:7b and it's surprisingly quick enough and accurate enough to be useable for coding. I'm wondering if there are any cheap upgrades I could make that would improve its performance even more? I think I have a pciex16 slot open, would getting a graphics card with 2-4gb ram help at all? I've read that it would actually probably be slower unless i got a graphics card with 24gb ram or something.
2025-09-21T23:04:11
https://www.reddit.com/r/LocalLLaMA/comments/1nn6g41/i58500_64gb_ram_working_great/
ThreeShartsToTheWind
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn6g41
false
null
t3_1nn6g41
/r/LocalLLaMA/comments/1nn6g41/i58500_64gb_ram_working_great/
false
false
self
1
null
Best model for light coding tasks?
0
Bought an M4 pro 24 GB recently. I haven't run local models since Llama 1... I'd like to use it for kinda simple "how does this work?" , or "how to create this" kinda tasks, not like "create this app" or "fix this bug" tasks. It doesn't have to be so smart but would like to be informative. What's your recommendation? I'd prefer something lightweight, preferably faster than closed source models that I use daily.
2025-09-21T22:33:31
https://www.reddit.com/r/LocalLLaMA/comments/1nn5r88/best_model_for_light_coding_tasks/
Consistent_Equal5327
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn5r88
false
null
t3_1nn5r88
/r/LocalLLaMA/comments/1nn5r88/best_model_for_light_coding_tasks/
false
false
self
0
null
Tracking prompt evolution for RAG systems - anyone else doing this?
4
Been working on a problem that's been bugging me with local RAG setups. When you generate docs with your LLM, you lose the context of HOW they were created. Three months later, you're wondering "what prompt chain produced this architecture doc?" Built a simple system that tracks: \- Original prompts \- Conversation context \- Model/version used (Mixtral, Llama, Claude, etc) \- Evolution history (v1→v9 with different models) Not trying to compete with vector DBs or anything fancy. Just solving the "what prompt created this?" problem. Example from our codebase: One doc went through 9 iterations: \- v1: Llama-70B (initial draft) \- v2-4: Claude (refinements) \- v5-7: GPT-4 (technical additions) \- v8-9: Mixtral (final structure) Each version linked to its prompt and full context. Can now search "authentication decisions" and get the doc + entire prompt evolution. Anyone else tracking generation provenance? What metadata matters most to you? GitHub: [github.com/VeriTeknik/pluggedin-app](http://github.com/VeriTeknik/pluggedin-app)
2025-09-21T22:31:40
https://www.reddit.com/r/LocalLLaMA/comments/1nn5ppq/tracking_prompt_evolution_for_rag_systems_anyone/
babaenki
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn5ppq
false
null
t3_1nn5ppq
/r/LocalLLaMA/comments/1nn5ppq/tracking_prompt_evolution_for_rag_systems_anyone/
false
false
self
4
{'enabled': False, 'images': [{'id': 'EplygW0N6fthhDASWZbC7NImAdhyV3iVPtDAhHpzkRc', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/EplygW0N6fthhDASWZbC7NImAdhyV3iVPtDAhHpzkRc.png?width=108&crop=smart&auto=webp&s=5e00e7e3576f0aa874d94711f84c19ebb6abaf5d', 'width': 108}, {'height': 108, 'url': 'https://external-preview.redd.it/EplygW0N6fthhDASWZbC7NImAdhyV3iVPtDAhHpzkRc.png?width=216&crop=smart&auto=webp&s=a216ae8d6ffb48da7ff4afb5487c4b06a4bfca23', 'width': 216}, {'height': 160, 'url': 'https://external-preview.redd.it/EplygW0N6fthhDASWZbC7NImAdhyV3iVPtDAhHpzkRc.png?width=320&crop=smart&auto=webp&s=08b1e3638fc88bcb9a54c8b6c17da2bd52f8f5db', 'width': 320}, {'height': 320, 'url': 'https://external-preview.redd.it/EplygW0N6fthhDASWZbC7NImAdhyV3iVPtDAhHpzkRc.png?width=640&crop=smart&auto=webp&s=a208e1d75b39f78ce48626d81438d5177a0a0364', 'width': 640}, {'height': 480, 'url': 'https://external-preview.redd.it/EplygW0N6fthhDASWZbC7NImAdhyV3iVPtDAhHpzkRc.png?width=960&crop=smart&auto=webp&s=a22a79562bf5d24e30ce0a0b11d269c831da7493', 'width': 960}, {'height': 540, 'url': 'https://external-preview.redd.it/EplygW0N6fthhDASWZbC7NImAdhyV3iVPtDAhHpzkRc.png?width=1080&crop=smart&auto=webp&s=c9a21e77101339e7344500070dcf4d90f2fb3023', 'width': 1080}], 'source': {'height': 600, 'url': 'https://external-preview.redd.it/EplygW0N6fthhDASWZbC7NImAdhyV3iVPtDAhHpzkRc.png?auto=webp&s=b7f656927fb70a868f425fefca86cff80066604e', 'width': 1200}, 'variants': {}}]}
Local Memory v1.1.0 Released - Deep Context Engineering Improvements!
0
Just dropped a massive Local Memory v1.1.0, focused on agent productivity and context optimization. This version finalizes the optimization based on the latest Anthropic guidance on building effective tools for AI agents: [https://www.anthropic.com/engineering/writing-tools-for-agents](https://www.anthropic.com/engineering/writing-tools-for-agents) **Context Engineering Breakthroughs:** * **Agent Decision Paralysis Solved**: Reduced from 26 → 11 tools (60% reduction) * **Token Efficiency**: 60-95% response size reduction through intelligent format controls * **Context Window Optimization**: Following "stateless function" principles for optimal 40-60% utilization * **Intelligent Routing**: operation\_type parameters route complex operations to sub-handlers automatically **Why This Matters for Developers:** Like most MCP tools, the old architecture forced agents to choose between lots of fragmented tools, creating decision overhead for the agents. The new unified tools use internal routing - agents get simple interfaces while the system handles complexity behind the scenes. The tooling also includes guidance and example usage to help agents make more token-efficient decisions. **Technical Deep Dive:** * **Schema Architecture**: Priority-based tool registration with comprehensive JSON validation * **Cross-Session Memory**: session\_filter\_mode enables knowledge sharing across conversations * **Performance**: Sub-10ms semantic search with Qdrant integration * **Type Safety**: Full Go implementation with proper conversions and backward compatibility **Real Impact on Agent Workflows:** Instead of agents struggling with "should I use search\_memories, search\_by\_tags, or search\_by\_date\_range?", they now use one \`search\` tool with intelligent routing. Same functionality, dramatically reduced cognitive load. **New optimized MCP tooling:** * **search** (semantic search, tag-based search, date range filtering, hybrid search modes) * **analysis** (AI-powered Q&A, memory summarization, pattern analysis, temporal analysis) * **relationships** (find related memories, AI relationship discovery, manual relationship creation, memory graph mapping) * **stats** (session statistics, domain statistics, category statistics, response optimization) * **categories** (create categories, list categories, AI categorization) * **domains** (create domains, list domains, knowledge organization) * **sessions** (list sessions, cross-session access, session management) * **core memory operations** (store\_memory, update\_memory, delete\_memory, get\_memory\_by\_id) Perfect for dev building with Claude Code, Claude Desktop, VS Code Copilot, Cursor, or Windsurf. The context window optimization alone makes working with coding agents much more efficient. Additional details: [localmemory.co](http://localmemory.co) Anyone else working on context engineering for AI agents? How are you handling tool proliferation in your setups? \#LocalMemory #MCP #ContextEngineering #AI #AgentProductivity
2025-09-21T22:24:43
https://www.reddit.com/r/LocalLLaMA/comments/1nn5k01/local_memory_v110_released_deep_context/
d2000e
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn5k01
false
null
t3_1nn5k01
/r/LocalLLaMA/comments/1nn5k01/local_memory_v110_released_deep_context/
false
false
self
0
null
Kokoro-82M-FP16-OpenVINO
36
https://huggingface.co/Echo9Zulu/Kokoro-82M-FP16-OpenVINO I converted this model in prep for [OpenArc](https://github.com/SearchSavior/OpenArc) 2.0.0. We have support for CPU only inference with Kokoro-82M-FP16-OpenVINO, accessible through /v1/audio/speech openai endpoint. /v1/audio/transcription was also implemented this weekend, targeting whisper. Conversion code which created this model was taken from an example Intel provides, linked in the model card. My plan is to apply what I learned working with Kokoro to Kitten-TTS models, then implement in OpenArc as part of a future release.
2025-09-21T21:24:50
https://www.reddit.com/r/LocalLLaMA/comments/1nn45cx/kokoro82mfp16openvino/
Echo9Zulu-
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn45cx
false
null
t3_1nn45cx
/r/LocalLLaMA/comments/1nn45cx/kokoro82mfp16openvino/
false
false
self
36
{'enabled': False, 'images': [{'id': '83HwxSBqW41TcPl9rR8nKUMTMQTyST9HFWO_PyTj6aE', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/83HwxSBqW41TcPl9rR8nKUMTMQTyST9HFWO_PyTj6aE.png?width=108&crop=smart&auto=webp&s=930edd2dcd8047da0f15083801d059cee327401f', 'width': 108}, {'height': 116, 'url': 'https://external-preview.redd.it/83HwxSBqW41TcPl9rR8nKUMTMQTyST9HFWO_PyTj6aE.png?width=216&crop=smart&auto=webp&s=e8473c0b7afd5443cc02ef4c9b2dc8c306cee5db', 'width': 216}, {'height': 172, 'url': 'https://external-preview.redd.it/83HwxSBqW41TcPl9rR8nKUMTMQTyST9HFWO_PyTj6aE.png?width=320&crop=smart&auto=webp&s=2371d65d60a04eaf55404f39326469c3b99f39a1', 'width': 320}, {'height': 345, 'url': 'https://external-preview.redd.it/83HwxSBqW41TcPl9rR8nKUMTMQTyST9HFWO_PyTj6aE.png?width=640&crop=smart&auto=webp&s=836dc440dc6ffbd2e30a2e6593b2a44b8306d694', 'width': 640}, {'height': 518, 'url': 'https://external-preview.redd.it/83HwxSBqW41TcPl9rR8nKUMTMQTyST9HFWO_PyTj6aE.png?width=960&crop=smart&auto=webp&s=3a32b1fcc44b2bf84b2216c502b6de50521196de', 'width': 960}, {'height': 583, 'url': 'https://external-preview.redd.it/83HwxSBqW41TcPl9rR8nKUMTMQTyST9HFWO_PyTj6aE.png?width=1080&crop=smart&auto=webp&s=9937f4a77169c85c8463486c6f00698b0dda6434', 'width': 1080}], 'source': {'height': 648, 'url': 'https://external-preview.redd.it/83HwxSBqW41TcPl9rR8nKUMTMQTyST9HFWO_PyTj6aE.png?auto=webp&s=9d5e356fd966de865c89f9a8e5ba37088a7aa14f', 'width': 1200}, 'variants': {}}]}
LibreChat can't be self-hosted in any commercial way even internally, because of MongoDB SSPL?
3
I want to run it but it seems, it's complicated way to say they backed by MongoDB right? Because you can't self host it and then you need to pay anyway and give them your data. >You can run LibreChat for internal operations, but the default MongoDB backend brings the Server Side Public License (SSPL). The SSPL requires that if you *provide the software as a service* you must release the source of *the entire service* (including any code that talks to MongoDB). Because a SaaS— even one used only by your own employees— is considered “making the functionality of the program available to third parties,” using the official MongoDB‑backed build would likely obligate you to open‑source your whole stack. LibreChat is described as “open‑source, self‑hostable and free to use. The documentation does not discuss its database choice or licensing implications, so the SSPL issue comes from MongoDB itself, not from LibreChat’s own license.
2025-09-21T21:23:33
https://www.reddit.com/r/LocalLLaMA/comments/1nn449p/librechat_cant_be_selfhosted_in_any_commercial/
mortyspace
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn449p
false
null
t3_1nn449p
/r/LocalLLaMA/comments/1nn449p/librechat_cant_be_selfhosted_in_any_commercial/
false
false
self
3
{'enabled': False, 'images': [{'id': 'QxqGKaPv5VvnyemIJkuwoyBHkw_-6CxwtnytuO8OKkw', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/QxqGKaPv5VvnyemIJkuwoyBHkw_-6CxwtnytuO8OKkw.png?width=108&crop=smart&auto=webp&s=6e25bbe49a90a71ce48c7af03447e547bda46189', 'width': 108}, {'height': 108, 'url': 'https://external-preview.redd.it/QxqGKaPv5VvnyemIJkuwoyBHkw_-6CxwtnytuO8OKkw.png?width=216&crop=smart&auto=webp&s=bf30f3ec09d2f2c65ca174e613066337782d12a2', 'width': 216}, {'height': 160, 'url': 'https://external-preview.redd.it/QxqGKaPv5VvnyemIJkuwoyBHkw_-6CxwtnytuO8OKkw.png?width=320&crop=smart&auto=webp&s=e81aa9f54abbb8d1431512a5e0facf5171cba8b1', 'width': 320}, {'height': 320, 'url': 'https://external-preview.redd.it/QxqGKaPv5VvnyemIJkuwoyBHkw_-6CxwtnytuO8OKkw.png?width=640&crop=smart&auto=webp&s=846b7e1a9c2889dc7e7d7d605fe159877cbb55c6', 'width': 640}, {'height': 480, 'url': 'https://external-preview.redd.it/QxqGKaPv5VvnyemIJkuwoyBHkw_-6CxwtnytuO8OKkw.png?width=960&crop=smart&auto=webp&s=e849e427e8d4ce48198a8d844444ffd81e0ac01b', 'width': 960}, {'height': 540, 'url': 'https://external-preview.redd.it/QxqGKaPv5VvnyemIJkuwoyBHkw_-6CxwtnytuO8OKkw.png?width=1080&crop=smart&auto=webp&s=d79c550802a33454edede52ada187dd7fe4dc636', 'width': 1080}], 'source': {'height': 600, 'url': 'https://external-preview.redd.it/QxqGKaPv5VvnyemIJkuwoyBHkw_-6CxwtnytuO8OKkw.png?auto=webp&s=ab3686a8cb2b16ed40665cc979d22cf8fe3e4939', 'width': 1200}, 'variants': {}}]}
MTEB still best for choosing an embedding model?
4
Hi all, Long time reader, first time poster. Love this community. Learned so much, and I hope I can pay forward one day. But before that :) Is MTEB still the best place for choosing an embedding model for RAG? And I see an endless list of tasks (not task type e.g. retrieval, reranking, etc.) that I realized I know nothing about. Can anyone point me to an article for understanding what these tasks are?
2025-09-21T20:36:58
https://www.reddit.com/r/LocalLLaMA/comments/1nn2xu1/mteb_still_best_for_choosing_an_embedding_model/
divide0verfl0w
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn2xu1
false
null
t3_1nn2xu1
/r/LocalLLaMA/comments/1nn2xu1/mteb_still_best_for_choosing_an_embedding_model/
false
false
self
4
null
Getting counter-intuitive results with local KV Cache Quantization Benchmark - am I doing something wrong?
11
Hi everyone, I've been running some benchmarks on KV cache quantization for long-context tasks, and I'm getting results that don't make much sense to me. I'm hoping this community could take a look at my methodology and point out if I'm making any obvious mistakes. You can find all the details, scripts, and results in my GitHub repo: [https://pento95.github.io/LongContext-KVCacheQuantTypesBench](https://pento95.github.io/LongContext-KVCacheQuantTypesBench) **My Goal:** I wanted to test the impact of all 16 `llama.cpp` KV cache quantization combinations on the Qwen3-30B model using a subset of the LongBench-v2 dataset. **My Setup:** * **Model:** `Qwen3-30B-A3B-Instruct-2507` (Unsloth Q4\_K\_XL GGUF) * Linux fedora, RTX 3090 Ti (24GB, full GPU offload) * **Method:** I used the `llama.cpp` server, running it for each of the 16 `cache-type-k` and `cache-type-v` combinations. The test uses 131 samples from LongBench-v2 (16k to 51k tokens) and evaluates the model's accuracy on multiple-choice questions. I used a temperature of 0.0 for deterministic output. **The Weird Results:** I was expecting to see a clear trend where higher quantization (like q4\_0) would lead to a drop in accuracy compared to the `f16` baseline. Instead, I'm seeing the opposite. My best performing combination is `k-f16_v-q5_0` with **16.79%** accuracy, while the `f16`\-`f16` baseline only gets **13.74%**. It seems counter-intuitive that quantizing the KV cache would *improve* performance. I've run the synchronous combinations three times now and the pattern holds. I'm starting to think my testing methodology is flawed. I've detailed the whole process in the [`README.md`](http://README.md) on the repo. Could you please take a look? I'm probably making a rookie mistake somewhere in the process, either in how I'm running the server, how I'm filtering the dataset, or how I'm extracting the answers. Any feedback, criticism, or suggestions would be incredibly helpful. Thanks in advance!
2025-09-21T20:26:07
https://www.reddit.com/r/LocalLLaMA/comments/1nn2nqz/getting_counterintuitive_results_with_local_kv/
Pentium95
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn2nqz
false
null
t3_1nn2nqz
/r/LocalLLaMA/comments/1nn2nqz/getting_counterintuitive_results_with_local_kv/
false
false
self
11
null
Kimi K2, hallucinations/verification, and fine tuning
9
So in my previous Kimi K2 post I see that a good few people have this same "it would be so great if not for the hallucination/overconidence" view of Kimi K2. Which kinda brings in an interesting question. Might it be possible to assemble a team here to try and fine-tune the thing? It is NOT easy (1T+MoE) and it needs someone experienced in fine-tuning and knowing how to generate the data, as well as others willing to review the data, come up with suggestions, and importantly chip in for the GPU time or serverless training tokens. Then the resulting LoRA is just posted for everyone to have (including Moonshot of course). I count myself among the latter group (review and chip in and also learn how people do the tuning thing). There are quite a few things to iron out but first I want to see if this is even feasible in principle.
2025-09-21T19:38:04
https://www.reddit.com/r/LocalLLaMA/comments/1nn1fqf/kimi_k2_hallucinationsverification_and_fine_tuning/
ramendik
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn1fqf
false
null
t3_1nn1fqf
/r/LocalLLaMA/comments/1nn1fqf/kimi_k2_hallucinationsverification_and_fine_tuning/
false
false
self
9
{'enabled': False, 'images': [{'id': 'KEKqZLXaDuojO8066WvfNm2knPNQpREJOqDRQbP0jOE', 'resolutions': [{'height': 56, 'url': 'https://external-preview.redd.it/KEKqZLXaDuojO8066WvfNm2knPNQpREJOqDRQbP0jOE.jpeg?width=108&crop=smart&auto=webp&s=674bf3d900716bcd75e795e30336baa3d48155c6', 'width': 108}, {'height': 113, 'url': 'https://external-preview.redd.it/KEKqZLXaDuojO8066WvfNm2knPNQpREJOqDRQbP0jOE.jpeg?width=216&crop=smart&auto=webp&s=cf1f430c28052edc8cf3984bb65f24409041bb77', 'width': 216}, {'height': 168, 'url': 'https://external-preview.redd.it/KEKqZLXaDuojO8066WvfNm2knPNQpREJOqDRQbP0jOE.jpeg?width=320&crop=smart&auto=webp&s=8e4dbaa0663d9b1a4d4cb226f41332aaa1425728', 'width': 320}, {'height': 336, 'url': 'https://external-preview.redd.it/KEKqZLXaDuojO8066WvfNm2knPNQpREJOqDRQbP0jOE.jpeg?width=640&crop=smart&auto=webp&s=81566908c6992b124df95aee8e462c2ebf9eae6f', 'width': 640}, {'height': 504, 'url': 'https://external-preview.redd.it/KEKqZLXaDuojO8066WvfNm2knPNQpREJOqDRQbP0jOE.jpeg?width=960&crop=smart&auto=webp&s=5f82a4a44d44de21db6818ef6617049ee7ab4203', 'width': 960}, {'height': 567, 'url': 'https://external-preview.redd.it/KEKqZLXaDuojO8066WvfNm2knPNQpREJOqDRQbP0jOE.jpeg?width=1080&crop=smart&auto=webp&s=2a1a44eb1223b26cd2ce899dbda73c7d61f57abd', 'width': 1080}], 'source': {'height': 1260, 'url': 'https://external-preview.redd.it/KEKqZLXaDuojO8066WvfNm2knPNQpREJOqDRQbP0jOE.jpeg?auto=webp&s=931059d24afb519fbb22a42eb1994d04b0efb20f', 'width': 2400}, 'variants': {}}]}
Any recommended tools for best PDF extraction to prep data for an LLM?
13
I’m curious if anyone has any thoughts on tools that do an amazing job at pdf extraction? Thinking in particular about PDFs that have exotic elements like tables, random quote blocks, sidebars, etc.
2025-09-21T19:36:55
https://www.reddit.com/r/LocalLLaMA/comments/1nn1elw/any_recommended_tools_for_best_pdf_extraction_to/
richardanaya
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn1elw
false
null
t3_1nn1elw
/r/LocalLLaMA/comments/1nn1elw/any_recommended_tools_for_best_pdf_extraction_to/
false
false
self
13
null
What is the best local ai that you can realistically run for coding on for example a 5070?
0
I
2025-09-21T19:32:32
https://www.reddit.com/r/LocalLLaMA/comments/1nn1ahg/what_is_the_best_local_ai_that_you_can/
Civil_Opposite7103
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn1ahg
false
null
t3_1nn1ahg
/r/LocalLLaMA/comments/1nn1ahg/what_is_the_best_local_ai_that_you_can/
false
false
self
0
null
Predicting the next "attention is all you need"
107
NeurIPS 2025 [accepted papers](https://neurips.cc/Downloads/2025) are out! If you didn't know, "Attention is all you Need" was published in [NeurIPS 2017](https://proceedings.neurips.cc/paper_files/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf) and spawned the modern wave of Transformer-based large language models; but few would have predicted this back in 2017. Which NeurIPS 2025 paper do you think is the bext "Attention is all you Need"?
2025-09-21T19:30:36
https://neurips.cc/Downloads/2025
entsnack
neurips.cc
1970-01-01T00:00:00
0
{}
1nn18k2
false
null
t3_1nn18k2
/r/LocalLLaMA/comments/1nn18k2/predicting_the_next_attention_is_all_you_need/
false
false
default
107
null
How bad to have RTX Pro 6000 run at PCIE x8?
8
I am building a dual RTX Pro 6000 workstation, buying the Threadripper is out of my budget as I already put 18k on the GPUs. My only option is to get the 9950x3D, I know there is not enough PCIE lanes, but how bad is it? I am using it for local LLM inference and fine tuning.
2025-09-21T19:27:46
https://www.reddit.com/r/LocalLLaMA/comments/1nn15rz/how_bad_to_have_rtx_pro_6000_run_at_pcie_x8/
kitgary
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn15rz
false
null
t3_1nn15rz
/r/LocalLLaMA/comments/1nn15rz/how_bad_to_have_rtx_pro_6000_run_at_pcie_x8/
false
false
self
8
null
Why is Hugging Face blocked in China when so many open‑weight models are released by Chinese companies?
228
I recently learned that HF is inaccessible from mainland China. At the same time, a large share of the open‑weight LLMs are published by Chinese firms. Is this a legal prohibition on publishing Chinese models, or simply a network‑level block that prevents users inside China from reaching the site?
2025-09-21T19:15:37
https://www.reddit.com/r/LocalLLaMA/comments/1nn0u8p/why_is_hugging_face_blocked_in_china_when_so_many/
zoxtech
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn0u8p
false
null
t3_1nn0u8p
/r/LocalLLaMA/comments/1nn0u8p/why_is_hugging_face_blocked_in_china_when_so_many/
false
false
self
228
null
Any research into LLM refusals
2
Does anyone know of or has performed research into LLM refusals. I'm not talking about spicy content, or getting the LLM to do questionable things. The topic came up when a system started refusing even innocuous requests such as help with constructing SQL queries. I tracked it back to the initial prompt given to it which made available certain tools etc. and certainly one part of the refusal seemed to be that if the request was outside the scope of tools or information provided, then the refusal was likely. But even when that aspect was taken out of the equation, the refusal rate was still high. It seemed like the particular initial prompt was jinxed, which given the complexity of the systems, can happen as a fluke. But it led me to wonder whether there was already any research or wisdom out there on this which might give some rules of thumb which can help with creating system prompts which don't increase refusal probabilities.
2025-09-21T19:13:15
https://www.reddit.com/r/LocalLLaMA/comments/1nn0s08/any_research_into_llm_refusals/
DeltaSqueezer
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn0s08
false
null
t3_1nn0s08
/r/LocalLLaMA/comments/1nn0s08/any_research_into_llm_refusals/
false
false
self
2
null
What GUI/interface do most people here use to run their models?
35
I used to be a big fan of https://github.com/nomic-ai/gpt4all but all development has stopped, which is a shame as this was quite lightweight and worked pretty well. What do people here use to run models in GGUF format? NOTE: I am not really up to date with everything in LLMA's and dont know what the latest bleeding edge model extension is or what must have applications run these things.
2025-09-21T19:09:10
https://www.reddit.com/r/LocalLLaMA/comments/1nn0o62/what_guiinterface_do_most_people_here_use_to_run/
tech4marco
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn0o62
false
null
t3_1nn0o62
/r/LocalLLaMA/comments/1nn0o62/what_guiinterface_do_most_people_here_use_to_run/
false
false
self
35
{'enabled': False, 'images': [{'id': '1sAcYnm4_nSa4vezzxjB458ZZAiN0Gtjrv7FuBtxuIE', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/1sAcYnm4_nSa4vezzxjB458ZZAiN0Gtjrv7FuBtxuIE.png?width=108&crop=smart&auto=webp&s=ad5c697c3c2151b4bed1b3f0763c536b57d32fd7', 'width': 108}, {'height': 108, 'url': 'https://external-preview.redd.it/1sAcYnm4_nSa4vezzxjB458ZZAiN0Gtjrv7FuBtxuIE.png?width=216&crop=smart&auto=webp&s=30af8b4148e027cb85325ba6791471c26cde4ff0', 'width': 216}, {'height': 160, 'url': 'https://external-preview.redd.it/1sAcYnm4_nSa4vezzxjB458ZZAiN0Gtjrv7FuBtxuIE.png?width=320&crop=smart&auto=webp&s=3bd9a1dffacdf38d4e61817326cef53fbd195e17', 'width': 320}, {'height': 320, 'url': 'https://external-preview.redd.it/1sAcYnm4_nSa4vezzxjB458ZZAiN0Gtjrv7FuBtxuIE.png?width=640&crop=smart&auto=webp&s=f7e904c23e4543a792ee92c4340bf2fdd182b6fd', 'width': 640}, {'height': 480, 'url': 'https://external-preview.redd.it/1sAcYnm4_nSa4vezzxjB458ZZAiN0Gtjrv7FuBtxuIE.png?width=960&crop=smart&auto=webp&s=54b5329e6dfda03333b4bb7f65b22cc5063c8151', 'width': 960}, {'height': 540, 'url': 'https://external-preview.redd.it/1sAcYnm4_nSa4vezzxjB458ZZAiN0Gtjrv7FuBtxuIE.png?width=1080&crop=smart&auto=webp&s=e227176176596790c22deb01e90c00a918f9daca', 'width': 1080}], 'source': {'height': 600, 'url': 'https://external-preview.redd.it/1sAcYnm4_nSa4vezzxjB458ZZAiN0Gtjrv7FuBtxuIE.png?auto=webp&s=2f9182b45387ce53ea9362e092476f3a05a60252', 'width': 1200}, 'variants': {}}]}
What is the best project/ app for the best RAG about all personal files?
1
One that is really good that works like both rag but also like memory kind of/ or whatever like a personal assistant , that I wouldn’t have to code and build myself?
2025-09-21T19:07:39
https://www.reddit.com/r/LocalLLaMA/comments/1nn0msk/what_is_the_best_project_app_for_the_best_rag/
OrganicApricot77
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn0msk
false
null
t3_1nn0msk
/r/LocalLLaMA/comments/1nn0msk/what_is_the_best_project_app_for_the_best_rag/
false
false
self
1
null
BMO project with Raspberry Pi 5 - OpenAi and Mistral support
1
Hi everyone! Three weeks ago I posted about my RPi local assistant, but I had to pause because of exams. Now I’ve added face animations, and I’m excited to share a video of it I hope you like it! I drew all the faces myself. In the repo they’re pink because my case is pink, but I’ll soon add green ones :P Next up: adding a video camera, RAG, and actions! Repo: https://github.com/ivegotanheadache/BMO
2025-09-21T19:02:21
https://v.redd.it/myzbyzwnfkqf1
Strange-Dimension675
v.redd.it
1970-01-01T00:00:00
0
{}
1nn0hqt
false
{'reddit_video': {'bitrate_kbps': 2400, 'dash_url': 'https://v.redd.it/myzbyzwnfkqf1/DASHPlaylist.mpd?a=1761073354%2CNTRkZGRkMzZkYzQ0M2YzN2JjMDQxNjhjYWEzMDVkN2U4Nzc1MGZhNGQ3MzgwMjNiZmNiMzNiODIyZDBmNWQwOA%3D%3D&v=1&f=sd', 'duration': 10, 'fallback_url': 'https://v.redd.it/myzbyzwnfkqf1/DASH_720.mp4?source=fallback', 'has_audio': True, 'height': 952, 'hls_url': 'https://v.redd.it/myzbyzwnfkqf1/HLSPlaylist.m3u8?a=1761073354%2CZThmMjYxNDQwNDljYWNhMGQ1M2MyOTdjNWIzYTUwOGE2YzNjZTQ3ZThmMzJiNDdmYWJmNDVjZmQ3ZDcxZjgxMQ%3D%3D&v=1&f=sd', 'is_gif': False, 'scrubber_media_url': 'https://v.redd.it/myzbyzwnfkqf1/DASH_96.mp4', 'transcoding_status': 'completed', 'width': 720}}
t3_1nn0hqt
/r/LocalLLaMA/comments/1nn0hqt/bmo_project_with_raspberry_pi_5_openai_and/
false
false
https://external-preview…7683a2246bc7074d
1
{'enabled': False, 'images': [{'id': 'cjU5M2g0d21ma3FmMXazmngLNXqelmlWcN_amgWfd-4ohQQZOMFjUeQ06STm', 'resolutions': [{'height': 142, 'url': 'https://external-preview.redd.it/cjU5M2g0d21ma3FmMXazmngLNXqelmlWcN_amgWfd-4ohQQZOMFjUeQ06STm.png?width=108&crop=smart&format=pjpg&auto=webp&s=1671e1999b8617b8ae591502e3e1eea9bf6370d3', 'width': 108}, {'height': 285, 'url': 'https://external-preview.redd.it/cjU5M2g0d21ma3FmMXazmngLNXqelmlWcN_amgWfd-4ohQQZOMFjUeQ06STm.png?width=216&crop=smart&format=pjpg&auto=webp&s=024c3a2b4d788031684ba6a2793dc40ab7192e83', 'width': 216}, {'height': 422, 'url': 'https://external-preview.redd.it/cjU5M2g0d21ma3FmMXazmngLNXqelmlWcN_amgWfd-4ohQQZOMFjUeQ06STm.png?width=320&crop=smart&format=pjpg&auto=webp&s=c507cf3bb18b9383f854d52fcc2046dfc0783cea', 'width': 320}, {'height': 845, 'url': 'https://external-preview.redd.it/cjU5M2g0d21ma3FmMXazmngLNXqelmlWcN_amgWfd-4ohQQZOMFjUeQ06STm.png?width=640&crop=smart&format=pjpg&auto=webp&s=b39971207c60e78d9ec638691338c2f5e96c03f3', 'width': 640}], 'source': {'height': 1216, 'url': 'https://external-preview.redd.it/cjU5M2g0d21ma3FmMXazmngLNXqelmlWcN_amgWfd-4ohQQZOMFjUeQ06STm.png?format=pjpg&auto=webp&s=bbfb5c487916afd47396145187603f6c74804162', 'width': 920}, 'variants': {}}]}
Vs code and got-oss-20b question
0
Has anyone else used this model in copilot’s place and if so, how has it worked? I’ve noticed that with the official copilot chat extension, you can replace copilot with an ollama model. Has anyone tried gpt-oss-20b with it yet?
2025-09-21T18:46:07
https://www.reddit.com/r/LocalLLaMA/comments/1nn01xe/vs_code_and_gotoss20b_question/
Savantskie1
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn01xe
false
null
t3_1nn01xe
/r/LocalLLaMA/comments/1nn01xe/vs_code_and_gotoss20b_question/
false
false
self
0
null
Qwen3-Coder-480B on the M3 Ultra 512GB Mac Studio is perfect for agentic coding
144
Qwen3-Coder-480b runs in MLX with 8bit quantization and just barely fits the full 256k context window within 512GB. With Roo code/cline, Q3C works exceptionally well when working within an existing codebase. * RAG (with Qwen3-Embed) retrieves API documentation and code samples which eliminates hallucinations. * The long context length can handle entire source code files for additional details. * Prompt adherence is great, and the subtasks in Roo work very well to gather information without saturating the main context. * VSCode hints are read by Roo and provide feedback about the output code. * Console output is read back to identify compile time and runtime errors. Green grass is more difficult, Q3C doesn’t do the best job at architecting a solution given a generic prompt. It’s much better to explicitly provide a design or at minimum design constraints rather than just “implement X using Y”. Prompt processing, especially at full 256k context, can be quite slow. For an agentic workflow, this doesn’t matter much, since I’m running it in the background. I find Q3C difficult to use as a coding *assistant*, at least the 480b version. I was on the fence about this machine 6 months ago when I ordered it, but I’m quite happy with what it can do now. An alternative option I considered was to buy an RTX Pro 6000 for my 256GB threadripper system, but the throughout benefits are far outweighed by the ability to run larger models at higher precision in my use case.
2025-09-21T18:45:29
https://www.reddit.com/r/LocalLLaMA/comments/1nn01bj/qwen3coder480b_on_the_m3_ultra_512gb_mac_studio/
ButThatsMyRamSlot
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nn01bj
false
null
t3_1nn01bj
/r/LocalLLaMA/comments/1nn01bj/qwen3coder480b_on_the_m3_ultra_512gb_mac_studio/
false
false
self
144
null
LongCat-Flash-Thinking
190
🚀 LongCat-Flash-Thinking: Smarter reasoning, leaner costs! 🏆 Performance: SOTA open-source models on Logic/Math/Coding/Agent tasks 📊 Efficiency: 64.5% fewer tokens to hit top-tier accuracy on AIME25 with native tool use, agent-friendly ⚙️ Infrastructure: Async RL achieves a 3x speedup over Sync frameworks 🔗Model: [https://huggingface.co/meituan-longcat/LongCat-Flash-Thinking](https://huggingface.co/meituan-longcat/LongCat-Flash-Thinking) 💻 Try Now: [longcat.ai](http://longcat.ai)
2025-09-21T18:25:42
https://i.redd.it/l7o00pbb9kqf1.jpeg
Xhehab_
i.redd.it
1970-01-01T00:00:00
0
{}
1nmzio1
false
null
t3_1nmzio1
/r/LocalLLaMA/comments/1nmzio1/longcatflashthinking/
false
false
default
190
{'enabled': True, 'images': [{'id': 'l7o00pbb9kqf1', 'resolutions': [{'height': 59, 'url': 'https://preview.redd.it/l7o00pbb9kqf1.jpeg?width=108&crop=smart&auto=webp&s=9eef92b9eb8821cb534f4f215a45bb575111db05', 'width': 108}, {'height': 119, 'url': 'https://preview.redd.it/l7o00pbb9kqf1.jpeg?width=216&crop=smart&auto=webp&s=4c4bded7b117f0266ddb25ec4c5c5ed4232d1c15', 'width': 216}, {'height': 176, 'url': 'https://preview.redd.it/l7o00pbb9kqf1.jpeg?width=320&crop=smart&auto=webp&s=93590d8f88a56af223439b92c49b2ee8a1c45fc5', 'width': 320}, {'height': 353, 'url': 'https://preview.redd.it/l7o00pbb9kqf1.jpeg?width=640&crop=smart&auto=webp&s=3fb8a419e2a1961bdcf7234e4eb8e33897a2904f', 'width': 640}, {'height': 530, 'url': 'https://preview.redd.it/l7o00pbb9kqf1.jpeg?width=960&crop=smart&auto=webp&s=ed8c82b40c224273fca7edeee8c26f5b89256ce3', 'width': 960}, {'height': 596, 'url': 'https://preview.redd.it/l7o00pbb9kqf1.jpeg?width=1080&crop=smart&auto=webp&s=67da444b83f2bf28fdc544b44dba54ff0e56d4f5', 'width': 1080}], 'source': {'height': 1020, 'url': 'https://preview.redd.it/l7o00pbb9kqf1.jpeg?auto=webp&s=eccfb2c33e4a4ab0f9291ef6bd9963761fd212b7', 'width': 1846}, 'variants': {}}]}
Help !
0
Hi, can someone explain to me what's missing? I want to download the files and I can't.
2025-09-21T17:50:23
https://i.redd.it/incsqhp03kqf1.jpeg
Illustrious-Swim9663
i.redd.it
1970-01-01T00:00:00
0
{}
1nmyl1h
false
null
t3_1nmyl1h
/r/LocalLLaMA/comments/1nmyl1h/help/
false
false
default
0
{'enabled': True, 'images': [{'id': 'incsqhp03kqf1', 'resolutions': [{'height': 28, 'url': 'https://preview.redd.it/incsqhp03kqf1.jpeg?width=108&crop=smart&auto=webp&s=687918dedbdd884c5c48e3064483cd7b5074f8a1', 'width': 108}, {'height': 56, 'url': 'https://preview.redd.it/incsqhp03kqf1.jpeg?width=216&crop=smart&auto=webp&s=b1f35b358d8426999ff867f24b410689bbe2b96d', 'width': 216}, {'height': 83, 'url': 'https://preview.redd.it/incsqhp03kqf1.jpeg?width=320&crop=smart&auto=webp&s=ce44fae82cc524e63b5e9634fa0d24df24af01bc', 'width': 320}, {'height': 166, 'url': 'https://preview.redd.it/incsqhp03kqf1.jpeg?width=640&crop=smart&auto=webp&s=cb64d4429f884844e7aa6c879bad7ab3949889e3', 'width': 640}, {'height': 249, 'url': 'https://preview.redd.it/incsqhp03kqf1.jpeg?width=960&crop=smart&auto=webp&s=ef42668600081416a2f3436f7a5d3ee390fc35f6', 'width': 960}, {'height': 280, 'url': 'https://preview.redd.it/incsqhp03kqf1.jpeg?width=1080&crop=smart&auto=webp&s=0b4b63ec8f7fafa5e75b93438fb846f54e5eb015', 'width': 1080}], 'source': {'height': 327, 'url': 'https://preview.redd.it/incsqhp03kqf1.jpeg?auto=webp&s=2089cbcf5003f77f7270789e3185a23093a7e159', 'width': 1260}, 'variants': {}}]}
Issues with running Arc B580 using docker compose
2
I've been messing around with self hosted AI and open web ui and its been pretty fun. So far i got it working with using my CPU and ram but I've been struggling to get my intel arc B580 to work and I'm not really sure how to move forward cause I'm kinda new to this. services: ollama: # image: ollama/ollama:latest image: intelanalytics/ipex-llm-inference-cpp-xpu:latest container_name: ollama restart: unless-stopped shm_size: "2g" environment: - OLLAMA_HOST=0.0.0.0:11434 - OLLAMA_NUM_GPU=999 - ZES_ENABLE_SYSMAN=1 - GGML_SYCL=1 - SYCL_DEVICE_FILTER=level_zero:gpu - ZE_AFFINITY_MASK=0 - DEVICE=Arc - OLLAMA_MAX_LOADED_MODELS=1 - OLLAMA_NUM_PARALLEL=1 devices: - /dev/dri/renderD128:/dev/dri/renderD128 group_add: - "993" - "44" volumes: - /home/user/docker/ai/ollama:/root/.ollama openwebui: image: ghcr.io/open-webui/open-webui:main container_name: openwebui depends_on: [ollama] restart: unless-stopped ports: - "127.0.0.1:3000:8080" # localhost only environment: - OLLAMA_BASE_URL=http://ollama:11434 volumes: - /home/user/docker/ai/webui:/app/backend/data
2025-09-21T17:12:55
https://www.reddit.com/r/LocalLLaMA/comments/1nmxlwk/issues_with_running_arc_b580_using_docker_compose/
Co0ool
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nmxlwk
false
null
t3_1nmxlwk
/r/LocalLLaMA/comments/1nmxlwk/issues_with_running_arc_b580_using_docker_compose/
false
false
self
2
null
“How I Finally Memorized Entire Chapters Without Feeling Overwhelmed”
0
I used to spend hours reading and rereading textbooks, highlighting everything, and still forgetting the key points. It was exhausting, and I constantly felt behind. Recently, I tried a different approach: 1. Summarize first: Before diving into the chapter, I jot down the key points in my own words. It forces me to process the material instead of just passively reading. 2. Active recall: I make quick questions for myself and try to answer them without looking. Even simple flashcards work wonders. 3. Chunking: I break topics into smaller sections and tackle them in short, focused sessions (25–30 mins). 4. AI explanations (optional): If a concept is really confusing, I get an explanation in plain language. This saves time over re-reading multiple pages. The difference was night and day. I felt more confident during revisions, and I could retain information without spending crazy amounts of time.
2025-09-21T16:40:26
https://i.redd.it/4h75dipiqjqf1.png
Constant-Rip-7300
i.redd.it
1970-01-01T00:00:00
0
{}
1nmws14
false
null
t3_1nmws14
/r/LocalLLaMA/comments/1nmws14/how_i_finally_memorized_entire_chapters_without/
false
false
default
0
{'enabled': True, 'images': [{'id': '4h75dipiqjqf1', 'resolutions': [{'height': 216, 'url': 'https://preview.redd.it/4h75dipiqjqf1.png?width=108&crop=smart&auto=webp&s=94ed53909bf0589e92fcc96132fba9d430abc72f', 'width': 108}, {'height': 432, 'url': 'https://preview.redd.it/4h75dipiqjqf1.png?width=216&crop=smart&auto=webp&s=d3f90dcfd4a89cd6cc03e60a4a8151aa28714eab', 'width': 216}, {'height': 640, 'url': 'https://preview.redd.it/4h75dipiqjqf1.png?width=320&crop=smart&auto=webp&s=8e8a9292e6280ad33f537b13c63cadc9de34f55b', 'width': 320}, {'height': 1280, 'url': 'https://preview.redd.it/4h75dipiqjqf1.png?width=640&crop=smart&auto=webp&s=36ff2681e85435b3dbca17a0025fea0ee7001560', 'width': 640}, {'height': 1920, 'url': 'https://preview.redd.it/4h75dipiqjqf1.png?width=960&crop=smart&auto=webp&s=77a9e8b9384ae0dd22e94a59b2fc7d12b52829df', 'width': 960}, {'height': 2160, 'url': 'https://preview.redd.it/4h75dipiqjqf1.png?width=1080&crop=smart&auto=webp&s=6e54d7b62e4f825485e953d0b696a20bd6d6c671', 'width': 1080}], 'source': {'height': 2400, 'url': 'https://preview.redd.it/4h75dipiqjqf1.png?auto=webp&s=39e611c2a0ecf13c1cf95d9e6e0a39d7422157b1', 'width': 1080}, 'variants': {}}]}
We've Just Hit 400 Stars on Nanocoder - This Community Is Amazing 🔥
10
ERROR: type should be string, got "\n\nhttps://i.redd.it/cyo43spgpjqf1.gif\n\nThis is yet another appreciation post for the community. Since my last, so much has happened in the Nanocoder community - new feedback, new features, many new people joining and contributing. It's honestly incredible to be building community-owned and pushed CLI software that breaks free of the corporations running other coding tools and offerings.\n\nAlong with a bunch of new features and improvements over the last couple of weeks, I'm actively moving the Nanocoder repository to be owned by a GitHub Organisation called Nano Collective – this organisation further reinforces my desire to make this project a community-led and run project. Within this collective I hope to continue to build out new frameworks and fine-tunes for local-first coding as well as seek grants to distribute to contributors to push research forward.\n\nThis is really really early days and Nanocoder as a coding CLI is right at the beginning, it's improving every day but there's still lots to do! That being said, any feedback and help within any domain is appreciated and welcomed.\n\n* Coding\n* System prompt writing\n* Research\n* Helping to push the word out\n* Any feedback generally! Good or bad :)\n\nIf you want to get involved the links are below. Bring on 1,000 stars ⭐️\n\n**GitHub Link**: [https://github.com/Mote-Software/nanocoder](https://github.com/Mote-Software/nanocoder)\n\n**Discord Link**: [https://discord.gg/ktPDV6rekE](https://discord.gg/ktPDV6rekE)"
2025-09-21T16:34:57
https://www.reddit.com/r/LocalLLaMA/comments/1nmwn1e/weve_just_hit_400_stars_on_nanocoder_this/
willlamerton
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nmwn1e
false
null
t3_1nmwn1e
/r/LocalLLaMA/comments/1nmwn1e/weve_just_hit_400_stars_on_nanocoder_this/
false
false
https://a.thumbs.redditm…nMl2sjFozRN4.jpg
10
null
Can I use unsloth's Kimi K2 quants with ik_llama?
1
[removed]
2025-09-21T16:21:39
https://www.reddit.com/r/LocalLLaMA/comments/1nmwapv/can_i_use_unsloths_kimi_k2_quants_with_ik_llama/
phwlarxoc
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nmwapv
false
null
t3_1nmwapv
/r/LocalLLaMA/comments/1nmwapv/can_i_use_unsloths_kimi_k2_quants_with_ik_llama/
false
false
self
1
null
Rolling Benchmarks - Evaluating AI Agents on Unseen GitHub Repos
9
I recently found Scale AI's new repo for benchmarking agent performance: [https://github.com/scaleapi/SWE-bench\_Pro-os/](https://github.com/scaleapi/SWE-bench_Pro-os/) And since I'm building docker images for repos associated with arXiv papers each day: [https://hub.docker.com/u/remyxai](https://hub.docker.com/u/remyxai) I started thinking about a new direction for agent evaluation. Static benchmarks are prone to leaderboard hacking and training data contamination, so how about a dynamic/rolling benchmark? By limiting submissions to only freshly published code, we could evaluate based on consistency over time with rolling averages instead of finding agents overfit to a static benchmark. Can rolling benchmarks bring us closer to evaluating agents in a way more closely aligned with their real-world applications? Love to hear what you think about this.
2025-09-21T16:05:52
https://www.reddit.com/r/LocalLLaMA/comments/1nmvw7a/rolling_benchmarks_evaluating_ai_agents_on_unseen/
remyxai
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nmvw7a
false
null
t3_1nmvw7a
/r/LocalLLaMA/comments/1nmvw7a/rolling_benchmarks_evaluating_ai_agents_on_unseen/
false
false
self
9
{'enabled': False, 'images': [{'id': '0N9w272PreXOwXXXHMeZdwAxthc0ie6_vSVsaxoS7H4', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/0N9w272PreXOwXXXHMeZdwAxthc0ie6_vSVsaxoS7H4.png?width=108&crop=smart&auto=webp&s=14986f2ab78f71432e1d6fcc4aff47d84c485e1c', 'width': 108}, {'height': 108, 'url': 'https://external-preview.redd.it/0N9w272PreXOwXXXHMeZdwAxthc0ie6_vSVsaxoS7H4.png?width=216&crop=smart&auto=webp&s=0b930756a874b665ca917699a54d712f2a5325e1', 'width': 216}, {'height': 160, 'url': 'https://external-preview.redd.it/0N9w272PreXOwXXXHMeZdwAxthc0ie6_vSVsaxoS7H4.png?width=320&crop=smart&auto=webp&s=37776c98402f7b14059443d5067a017ec6e917f8', 'width': 320}, {'height': 320, 'url': 'https://external-preview.redd.it/0N9w272PreXOwXXXHMeZdwAxthc0ie6_vSVsaxoS7H4.png?width=640&crop=smart&auto=webp&s=977f1029de50e840b47906df83ca2eb989db2630', 'width': 640}, {'height': 480, 'url': 'https://external-preview.redd.it/0N9w272PreXOwXXXHMeZdwAxthc0ie6_vSVsaxoS7H4.png?width=960&crop=smart&auto=webp&s=4e61a84ff23de84a0d5d609fa7d710b4e2eceed9', 'width': 960}, {'height': 540, 'url': 'https://external-preview.redd.it/0N9w272PreXOwXXXHMeZdwAxthc0ie6_vSVsaxoS7H4.png?width=1080&crop=smart&auto=webp&s=a2e433a03f84d1edde74e8df86017245788122f9', 'width': 1080}], 'source': {'height': 600, 'url': 'https://external-preview.redd.it/0N9w272PreXOwXXXHMeZdwAxthc0ie6_vSVsaxoS7H4.png?auto=webp&s=fead7565144f45d4e19fe42289ce9cebdfd9ad5b', 'width': 1200}, 'variants': {}}]}
GPT 5 for Computer Use agents
0
Same tasks, same grounding model we just swapped GPT 4o with GPT 5 as the thinking model. Left = 4o, right = 5. Watch GPT 5 pull through. Grounding model: Salesforce GTA1-7B Action space: CUA Cloud Instances (macOS/Linux/Windows) The task is: "Navigate to {random_url} and play the game until you reach a score of 5/5”....each task is set up by having claude generate a random app from a predefined list of prompts (multiple choice trivia, form filling, or color matching)" Try it yourself here : https://github.com/trycua/cua Docs : https://docs.trycua.com/docs/agent-sdk/supported-agents/composed-agent Discord: https://discord.gg/cua-ai
2025-09-21T15:51:41
https://v.redd.it/bn55h9vthjqf1
Impressive_Half_2819
v.redd.it
1970-01-01T00:00:00
0
{}
1nmviz7
false
{'reddit_video': {'bitrate_kbps': 5000, 'dash_url': 'https://v.redd.it/bn55h9vthjqf1/DASHPlaylist.mpd?a=1761061915%2CZmY5NDg3ODIzOGUyYzQwNDIxMTEzNzY4NjY1Njk4NDI4NDMwZDUxYTdlZWNlZDVkZmYxMWJhYTU2YmI1ZjJiOQ%3D%3D&v=1&f=sd', 'duration': 30, 'fallback_url': 'https://v.redd.it/bn55h9vthjqf1/DASH_1080.mp4?source=fallback', 'has_audio': True, 'height': 1080, 'hls_url': 'https://v.redd.it/bn55h9vthjqf1/HLSPlaylist.m3u8?a=1761061915%2CZDg4N2Q1ODViYmI5YzEwNmEyOWY1MmM1ZmIwNDY2ZmEyYTgwZmYyYTc5ZjljZmFkYTA1NjBlMDhjOWRkMjljZA%3D%3D&v=1&f=sd', 'is_gif': False, 'scrubber_media_url': 'https://v.redd.it/bn55h9vthjqf1/DASH_96.mp4', 'transcoding_status': 'completed', 'width': 1920}}
t3_1nmviz7
/r/LocalLLaMA/comments/1nmviz7/gpt_5_for_computer_use_agents/
false
false
https://external-preview…a5fdf8254fb75fa5
0
{'enabled': False, 'images': [{'id': 'OGw1MWJhZ3RoanFmMRG6-oWujtVtwWnIPYQYAphLJPDZc9z94p-KY-4O-UR8', 'resolutions': [{'height': 60, 'url': 'https://external-preview.redd.it/OGw1MWJhZ3RoanFmMRG6-oWujtVtwWnIPYQYAphLJPDZc9z94p-KY-4O-UR8.png?width=108&crop=smart&format=pjpg&auto=webp&s=e6ee4a77333d181e6aec230704e056868958fcf3', 'width': 108}, {'height': 121, 'url': 'https://external-preview.redd.it/OGw1MWJhZ3RoanFmMRG6-oWujtVtwWnIPYQYAphLJPDZc9z94p-KY-4O-UR8.png?width=216&crop=smart&format=pjpg&auto=webp&s=72c8138ff9e63fdc5da61c7ad29389ffc5a565d9', 'width': 216}, {'height': 180, 'url': 'https://external-preview.redd.it/OGw1MWJhZ3RoanFmMRG6-oWujtVtwWnIPYQYAphLJPDZc9z94p-KY-4O-UR8.png?width=320&crop=smart&format=pjpg&auto=webp&s=8e647fb707b4000cba2d5dcf8a2edd54cb41ff44', 'width': 320}, {'height': 360, 'url': 'https://external-preview.redd.it/OGw1MWJhZ3RoanFmMRG6-oWujtVtwWnIPYQYAphLJPDZc9z94p-KY-4O-UR8.png?width=640&crop=smart&format=pjpg&auto=webp&s=7b180948687a1025669b4bdaf11c618f08dd714e', 'width': 640}, {'height': 540, 'url': 'https://external-preview.redd.it/OGw1MWJhZ3RoanFmMRG6-oWujtVtwWnIPYQYAphLJPDZc9z94p-KY-4O-UR8.png?width=960&crop=smart&format=pjpg&auto=webp&s=1550d40cd4842a62f29209fc36e1b0f7ed709da5', 'width': 960}, {'height': 607, 'url': 'https://external-preview.redd.it/OGw1MWJhZ3RoanFmMRG6-oWujtVtwWnIPYQYAphLJPDZc9z94p-KY-4O-UR8.png?width=1080&crop=smart&format=pjpg&auto=webp&s=8824cb70a5ba547c3dcf691452ce389d77f4fd4e', 'width': 1080}], 'source': {'height': 1080, 'url': 'https://external-preview.redd.it/OGw1MWJhZ3RoanFmMRG6-oWujtVtwWnIPYQYAphLJPDZc9z94p-KY-4O-UR8.png?format=pjpg&auto=webp&s=09a4bcf9be916d87680e97cae5552f88cc3b07e0', 'width': 1920}, 'variants': {}}]}
rx 9070 xt idle vram usage
2
I just got the radeon rx 9070 xt, and I'm concerned about the idle vram usage on the card. If anyone else has this card (or other 90 series amd card) please look into this. I run the following setup: - linux - using iGPU for display output - nothing runs on the 9070 xt I use amdgpu_top to monitor vram usage. When the card is idle (D3hot power state) with nothing running on it, it uses 519MB of vram. amdgpu_top shows vram usage by process, they all report 0mb. Is this normal? I had the rx 6800 xt, which used about 15mb vram when idle. The 500mb reserved vram means I can't get to 16k context with the models I usually use. I can still return the card if it's not normal to have this much reserved.
2025-09-21T15:28:08
https://www.reddit.com/r/LocalLLaMA/comments/1nmuxfu/rx_9070_xt_idle_vram_usage/
baileyske
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nmuxfu
false
null
t3_1nmuxfu
/r/LocalLLaMA/comments/1nmuxfu/rx_9070_xt_idle_vram_usage/
false
false
self
2
null
Can I use unsloth's Kimi K2 quants with ik_llama?
1
[removed]
2025-09-21T15:23:55
https://www.reddit.com/r/LocalLLaMA/comments/1nmutq0/can_i_use_unsloths_kimi_k2_quants_with_ik_llama/
phwlarxoc
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nmutq0
false
null
t3_1nmutq0
/r/LocalLLaMA/comments/1nmutq0/can_i_use_unsloths_kimi_k2_quants_with_ik_llama/
false
false
self
1
null
Is the RTX 6000 Blackwell Pro the right choice?
1
Last week I made this post: [https://www.reddit.com/r/LocalLLaMA/comments/1nkpohe/i\_can\_can\_get\_gpus\_as\_a\_tax\_write\_off\_thinking\_of/](https://www.reddit.com/r/LocalLLaMA/comments/1nkpohe/i_can_can_get_gpus_as_a_tax_write_off_thinking_of/) `<skip-if-you-want>` Essentially, you guys were very interested in talking to me about my strategy: 1. Buy two RTX 6000 blackwell pros. 2. Write them off for 2025 (I can do that owning a tech company). 1. Yes, I can write them off. 2. If My company gets into trouble, which is possible, I can sell them in the next scheduled year and still end up with a way smaller tax burden. 3. Use them to learn, upskill, and create products that could either lead to new work opportunities or a startup. Really, I hope it's a startup. 1. Agentic RAG with Local LLMs 2. ML object detection (PyTorch/Yolo) 3. ML OPs and running infrastructure 4. A big one that I haven't totally spoken about is that I can do game development with Unreal/Unity. I wouldn't want to build a game, but I've been fantasizing of product ideas that incorporate all of this together. Valid points brought up: 1. Why not use cloud? 1. I actually have and I hate waiting. I have a script that I use to boot up cloud instances with different GPUs, providers, and LLMs. I still have a sense of paranoia too that I'll do something like keep two H200s running, run my script to shut them down, they don't shutdown, and some how they break the cost limitations of my account. (PTSD from a web project I worked on where that happened) 2. No, I probably won't be running these GPUs hard all of the time. So while cloud instances will be way cheaper in the short term, I won't be drawing power out of them 24/7. If anything I'll probably be a light user. Most of the need for the power being to use bigger LLMs with Unreal. 3. The write offs I have this year if I do this will be significant enough to significantly reduce my income. 2. GPUs will tank in price. 1. Yup, this one is fair. In Canada it use to be that you couldn't get your hands on 3090's or 4090's due to demand. Anecdotally I was in computer store not too long ago that had a dozen 5090s. I asked how much they were, and was told $2600cad (very cheap compared to Feb). Asked why so cheap? They hadn't sold one since April. Moral of the story, my idea of just selling GPUs if I get in trouble might not be easy. 3. Power consumption 1. This one might not suck that bad, but we'll see. `</skip-if-you-want>` So now that I'm getting more serious about this. I'm wondering if the RTX 6000 blackwell pro, or two of them, will provide me. I think given that I want to do a lot of graphics based stuff it's a better choice than buying H100/A100s (I can't afford an H100 anyways) . I've been thinking about hybrids though models though and mixing GPUs together. I'm hoping to get high accuracy out of RAG systems I create. Might be an easier question here: What would you guys build if you were me and had $20k USD to spend?
2025-09-21T15:00:03
https://www.reddit.com/r/LocalLLaMA/comments/1nmu7jo/is_the_rtx_6000_blackwell_pro_the_right_choice/
Tired__Dev
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nmu7jo
false
null
t3_1nmu7jo
/r/LocalLLaMA/comments/1nmu7jo/is_the_rtx_6000_blackwell_pro_the_right_choice/
false
false
self
1
null
Simple/Lightweight Chat UI
0
Hi, I created this to be used with Ollama for chats, might expand and add more features later, feel free to use if you need something light weight and customizable if you code. Open Source. GitHub: [https://github.com/vshal-47/Modern-Local-LLM-UI-For-Ollama](https://github.com/vshal-47/Modern-Local-LLM-UI-For-Ollama)
2025-09-21T14:21:43
https://www.reddit.com/r/LocalLLaMA/comments/1nmt957/simplelightweight_chat_ui/
vshal_47
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1nmt957
false
null
t3_1nmt957
/r/LocalLLaMA/comments/1nmt957/simplelightweight_chat_ui/
false
false
self
0
{'enabled': False, 'images': [{'id': 'x4B3tusp9xDnjqQmISx6Y_ljnnwQbV9HIVn88pmuvCk', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/x4B3tusp9xDnjqQmISx6Y_ljnnwQbV9HIVn88pmuvCk.png?width=108&crop=smart&auto=webp&s=10d6c7b0f268eb182530eff9c97218f4050c518b', 'width': 108}, {'height': 108, 'url': 'https://external-preview.redd.it/x4B3tusp9xDnjqQmISx6Y_ljnnwQbV9HIVn88pmuvCk.png?width=216&crop=smart&auto=webp&s=99fcf42ec769f6069d9b053d08d1a656d6ad7573', 'width': 216}, {'height': 160, 'url': 'https://external-preview.redd.it/x4B3tusp9xDnjqQmISx6Y_ljnnwQbV9HIVn88pmuvCk.png?width=320&crop=smart&auto=webp&s=dfdf066d469504d6b9ef5568446b477fcb98e555', 'width': 320}, {'height': 320, 'url': 'https://external-preview.redd.it/x4B3tusp9xDnjqQmISx6Y_ljnnwQbV9HIVn88pmuvCk.png?width=640&crop=smart&auto=webp&s=3465ed46b24af72ac0ebd6a7de1ff4fd9eb91818', 'width': 640}, {'height': 480, 'url': 'https://external-preview.redd.it/x4B3tusp9xDnjqQmISx6Y_ljnnwQbV9HIVn88pmuvCk.png?width=960&crop=smart&auto=webp&s=e44ab535f95458f18d202e26260b965ccbb342bb', 'width': 960}, {'height': 540, 'url': 'https://external-preview.redd.it/x4B3tusp9xDnjqQmISx6Y_ljnnwQbV9HIVn88pmuvCk.png?width=1080&crop=smart&auto=webp&s=0572a2dfbd7d9b79fc5507b0697211c4be5af922', 'width': 1080}], 'source': {'height': 600, 'url': 'https://external-preview.redd.it/x4B3tusp9xDnjqQmISx6Y_ljnnwQbV9HIVn88pmuvCk.png?auto=webp&s=e520b26f227e0605736e7b9ac68738bcedc24992', 'width': 1200}, 'variants': {}}]}