-
Anyone else tracking datacenter GPU prices on eBay?
Posted by ttkciar@reddit | LocalLLaMA | View on Reddit | 36 comments
-
"Cheap" 24GB GPU options for fine-tuning?
Posted by deus119@reddit | LocalLLaMA | View on Reddit | 17 comments
-
I am solving AI Math Hallucinations with Hissab
Posted by prenx4x@reddit | LocalLLaMA | View on Reddit | 2 comments
-
Self-hosting LLaMA: What are your biggest pain points?
Posted by Sriyakee@reddit | LocalLLaMA | View on Reddit | 79 comments
-
Repurposing 800 x RX 580s for LLM inference - 4 months later - learnings
Posted by rasbid420@reddit | LocalLLaMA | View on Reddit | 56 comments
-
Use llama.cpp to run a model with the combined power of a networked cluster of GPUs.
Posted by farkinga@reddit | LocalLLaMA | View on Reddit | 2 comments
-
mistralai/Magistral-Small-2506
Posted by yoracale@reddit | LocalLLaMA | View on Reddit | 151 comments
-
Run Deepseek locally on a 24g GPU: Quantizing on our Giga Computing 6980P Xeon
Posted by atape_1@reddit | LocalLLaMA | View on Reddit | 36 comments
-
Best non-Chinese open models?
Posted by ProbaDude@reddit | LocalLLaMA | View on Reddit | 25 comments
-
Built an adaptive text classifier that learns continuously - no retraining needed for new classes
Posted by asankhs@reddit | LocalLLaMA | View on Reddit | 1 comments
-
Local AI setup 1x5090, 5x3090
Posted by Emergency_Fuel_2988@reddit | LocalLLaMA | View on Reddit | 34 comments
-
Current best uncensored model?
Posted by Accomplished-Feed568@reddit | LocalLLaMA | View on Reddit | 93 comments
-
Ohh. 🤔 Okay ‼️ But what if we look at AMD Mi100 instinct,⁉️🙄 I can get it for $1000.
Posted by sub_RedditTor@reddit | LocalLLaMA | View on Reddit | 9 comments
-
Qwen3 for Apple Neural Engine
Posted by Competitive-Bake4602@reddit | LocalLLaMA | View on Reddit | 22 comments
-
AMD Radeon AI PRO R9700 GPU Offers 4x More TOPS & 2x More AI Performance Than Radeon PRO W7800
Posted by _SYSTEM_ADMIN_MOD_@reddit | LocalLLaMA | View on Reddit | 22 comments
-
How much VRAM do you have and what's your daily-driver model?
Posted by EmPips@reddit | LocalLLaMA | View on Reddit | 173 comments
-
Jan got an upgrade: New design, switched from Electron to Tauri, custom assistants, and 100+ fixes - it's faster & more stable now
Posted by eck72@reddit | LocalLLaMA | View on Reddit | 161 comments
-
Intel's OpenVINO 2025.2 Brings Support For New Models, GenAI Improvements
Posted by FastDecode1@reddit | LocalLLaMA | View on Reddit | 1 comments
-
Planning to build AI PC does my Build make sense?
Posted by germaniiifelisarta@reddit | LocalLLaMA | View on Reddit | 7 comments
-
Who's the voice Narrator in this video??
Posted by mikemaina@reddit | LocalLLaMA | View on Reddit | 3 comments
-
Fine-tuning LLMs with Just One Command Using IdeaWeaver
Posted by Prashant-Lakhera@reddit | LocalLLaMA | View on Reddit | 0 comments
-
Jan-nano, a 4B model that can outperform 671B on MCP
Posted by Kooky-Somewhere-2883@reddit | LocalLLaMA | View on Reddit | 471 comments
-
Sam Altman says Meta offered OpenAI staff $100 million bonuses, as Mark Zuckerberg ramps up AI poaching efforts
Posted by choose_a_guest@reddit | LocalLLaMA | View on Reddit | 82 comments
-
Running DeepSeek locally using ONNX Runtime
Posted by DangerousGood4561@reddit | LocalLLaMA | View on Reddit | 7 comments
-
Exploring the limitations of LLMs-as-a-Judge
Posted by TelloLeEngineer@reddit | LocalLLaMA | View on Reddit | 1 comments
-
Oops
Posted by Own-Potential-2308@reddit | LocalLLaMA | View on Reddit | 48 comments
-
How run Open Source?
Posted by Easy_Marsupial_5833@reddit | LocalLLaMA | View on Reddit | 16 comments
-
We built this project to increase LLM throughput by 3x. Now it has been adopted by IBM in their LLM serving stack!
Posted by Nice-Comfortable-650@reddit | LocalLLaMA | View on Reddit | 55 comments
-
Kyutai's STT with semantic VAD now opensource
Posted by phhusson@reddit | LocalLLaMA | View on Reddit | 25 comments
-
MiniMax-M1 - a MiniMaxAI Collection
Posted by Dark_Fire_12@reddit | LocalLLaMA | View on Reddit | 42 comments
-
Linkedin Scraper / Automation / Data
Posted by Success-Dependent@reddit | LocalLLaMA | View on Reddit | 0 comments
-
Qwen releases official MLX quants for Qwen3 models in 4 quantization levels: 4bit, 6bit, 8bit, and BF16
Posted by ResearchCrafty1804@reddit | LocalLLaMA | View on Reddit | 44 comments
-
ICONN 1 is now out!
Posted by Enderchef@reddit | LocalLLaMA | View on Reddit | 159 comments
-
AMD Lemonade Server Update: Ubuntu, llama.cpp, Vulkan, webapp, and more!
Posted by jfowers_amd@reddit | LocalLLaMA | View on Reddit | 21 comments
-
96GB VRAM plus 256GB/512GB Fast RAM
Posted by SteveRD1@reddit | LocalLLaMA | View on Reddit | 11 comments
-
Smallest basic ai model for working
Posted by Future_Tonight_6626@reddit | LocalLLaMA | View on Reddit | 6 comments
-
Am I using lightrag + llama.cpp wrong?
Posted by Devonance@reddit | LocalLLaMA | View on Reddit | 1 comments
-
GMK X2(AMD Max+ 395 w/128GB) first impressions.
Posted by fallingdowndizzyvr@reddit | LocalLLaMA | View on Reddit | 67 comments
-
[Setup discussion] AMD RX 7900 XTX workstation for local LLMs — Linux or Windows as host OS?
Posted by ElkanRoelen@reddit | LocalLLaMA | View on Reddit | 18 comments
-
Dual CPU Penalty?
Posted by jsconiers@reddit | LocalLLaMA | View on Reddit | 20 comments
-
Performance expectations question (Devstral)
Posted by _-Carnage@reddit | LocalLLaMA | View on Reddit | 8 comments
-
New 24B finetune: Impish_Magic_24B
Posted by Sicarius_The_First@reddit | LocalLLaMA | View on Reddit | 15 comments
-
Local AI for a small/median accounting firm - € Buget of 10k-25k
Posted by AFruitShopOwner@reddit | LocalLLaMA | View on Reddit | 137 comments
-
How do you make your LLM apps secure?
Posted by kk17702@reddit | LocalLLaMA | View on Reddit | 8 comments
-
Dual RTX 6000, Blackwell and Ada Lovelace, with thermal imagery
Posted by Thalesian@reddit | LocalLLaMA | View on Reddit | 19 comments
-
Gemini models (yes, even the recent 2.5 ones) hallucinate crazily on video inputs
Posted by Infrared12@reddit | LocalLLaMA | View on Reddit | 2 comments
-
Best models to run with 8GB VRAM, 16GB RAM
Posted by Qxz3@reddit | LocalLLaMA | View on Reddit | 36 comments
-
[DEAL] On-demand B200 GPUs for $1.49/hr at DeepInfra (promo ends June 30)
Posted by temirulan@reddit | LocalLLaMA | View on Reddit | 4 comments
-
RTX 6000 PRO Blackwell Max Q? Non Max Q?
Posted by Opening_Progress6820@reddit | LocalLLaMA | View on Reddit | 20 comments
-
Computer-Use on Windows Sandbox
Posted by Impressive_Half_2819@reddit | LocalLLaMA | View on Reddit | 19 comments