TheaterFire

Login with Reddit

Currently browsing tags:

  • LocalLLaMA
  • Anyone else tracking datacenter GPU prices on eBay?

    Posted by ttkciar@reddit | LocalLLaMA | View on Reddit | 36 comments

  • "Cheap" 24GB GPU options for fine-tuning?

    Posted by deus119@reddit | LocalLLaMA | View on Reddit | 17 comments

  • I am solving AI Math Hallucinations with Hissab

    Posted by prenx4x@reddit | LocalLLaMA | View on Reddit | 2 comments

  • Self-hosting LLaMA: What are your biggest pain points?

    Posted by Sriyakee@reddit | LocalLLaMA | View on Reddit | 79 comments

  • Repurposing 800 x RX 580s for LLM inference - 4 months later - learnings

    Posted by rasbid420@reddit | LocalLLaMA | View on Reddit | 56 comments

  • Use llama.cpp to run a model with the combined power of a networked cluster of GPUs.

    Posted by farkinga@reddit | LocalLLaMA | View on Reddit | 2 comments

  • mistralai/Magistral-Small-2506

    Posted by yoracale@reddit | LocalLLaMA | View on Reddit | 151 comments

  • Run Deepseek locally on a 24g GPU: Quantizing on our Giga Computing 6980P Xeon

    Posted by atape_1@reddit | LocalLLaMA | View on Reddit | 36 comments

  • Best non-Chinese open models?

    Posted by ProbaDude@reddit | LocalLLaMA | View on Reddit | 25 comments

  • Built an adaptive text classifier that learns continuously - no retraining needed for new classes

    Posted by asankhs@reddit | LocalLLaMA | View on Reddit | 1 comments

  • Local AI setup 1x5090, 5x3090

    Posted by Emergency_Fuel_2988@reddit | LocalLLaMA | View on Reddit | 34 comments

  • Current best uncensored model?

    Posted by Accomplished-Feed568@reddit | LocalLLaMA | View on Reddit | 93 comments

  • Ohh. 🤔 Okay ‼️ But what if we look at AMD Mi100 instinct,⁉️🙄 I can get it for $1000.

    Posted by sub_RedditTor@reddit | LocalLLaMA | View on Reddit | 9 comments

  • Qwen3 for Apple Neural Engine

    Posted by Competitive-Bake4602@reddit | LocalLLaMA | View on Reddit | 22 comments

  • AMD Radeon AI PRO R9700 GPU Offers 4x More TOPS & 2x More AI Performance Than Radeon PRO W7800

    Posted by _SYSTEM_ADMIN_MOD_@reddit | LocalLLaMA | View on Reddit | 22 comments

  • How much VRAM do you have and what's your daily-driver model?

    Posted by EmPips@reddit | LocalLLaMA | View on Reddit | 173 comments

  • Jan got an upgrade: New design, switched from Electron to Tauri, custom assistants, and 100+ fixes - it's faster & more stable now

    Posted by eck72@reddit | LocalLLaMA | View on Reddit | 161 comments

  • Intel's OpenVINO 2025.2 Brings Support For New Models, GenAI Improvements

    Posted by FastDecode1@reddit | LocalLLaMA | View on Reddit | 1 comments

  • Planning to build AI PC does my Build make sense?

    Posted by germaniiifelisarta@reddit | LocalLLaMA | View on Reddit | 7 comments

  • Who's the voice Narrator in this video??

    Posted by mikemaina@reddit | LocalLLaMA | View on Reddit | 3 comments

  • Fine-tuning LLMs with Just One Command Using IdeaWeaver

    Posted by Prashant-Lakhera@reddit | LocalLLaMA | View on Reddit | 0 comments

  • Jan-nano, a 4B model that can outperform 671B on MCP

    Posted by Kooky-Somewhere-2883@reddit | LocalLLaMA | View on Reddit | 471 comments

  • Sam Altman says Meta offered OpenAI staff $100 million bonuses, as Mark Zuckerberg ramps up AI poaching efforts

    Posted by choose_a_guest@reddit | LocalLLaMA | View on Reddit | 82 comments

  • Running DeepSeek locally using ONNX Runtime

    Posted by DangerousGood4561@reddit | LocalLLaMA | View on Reddit | 7 comments

  • Exploring the limitations of LLMs-as-a-Judge

    Posted by TelloLeEngineer@reddit | LocalLLaMA | View on Reddit | 1 comments

  • Oops

    Posted by Own-Potential-2308@reddit | LocalLLaMA | View on Reddit | 48 comments

  • How run Open Source?

    Posted by Easy_Marsupial_5833@reddit | LocalLLaMA | View on Reddit | 16 comments

  • We built this project to increase LLM throughput by 3x. Now it has been adopted by IBM in their LLM serving stack!

    Posted by Nice-Comfortable-650@reddit | LocalLLaMA | View on Reddit | 55 comments

  • Kyutai's STT with semantic VAD now opensource

    Posted by phhusson@reddit | LocalLLaMA | View on Reddit | 25 comments

  • MiniMax-M1 - a MiniMaxAI Collection

    Posted by Dark_Fire_12@reddit | LocalLLaMA | View on Reddit | 42 comments

  • Linkedin Scraper / Automation / Data

    Posted by Success-Dependent@reddit | LocalLLaMA | View on Reddit | 0 comments

  • Qwen releases official MLX quants for Qwen3 models in 4 quantization levels: 4bit, 6bit, 8bit, and BF16

    Posted by ResearchCrafty1804@reddit | LocalLLaMA | View on Reddit | 44 comments

  • ICONN 1 is now out!

    Posted by Enderchef@reddit | LocalLLaMA | View on Reddit | 159 comments

  • AMD Lemonade Server Update: Ubuntu, llama.cpp, Vulkan, webapp, and more!

    Posted by jfowers_amd@reddit | LocalLLaMA | View on Reddit | 21 comments

  • 96GB VRAM plus 256GB/512GB Fast RAM

    Posted by SteveRD1@reddit | LocalLLaMA | View on Reddit | 11 comments

  • Smallest basic ai model for working

    Posted by Future_Tonight_6626@reddit | LocalLLaMA | View on Reddit | 6 comments

  • Am I using lightrag + llama.cpp wrong?

    Posted by Devonance@reddit | LocalLLaMA | View on Reddit | 1 comments

  • GMK X2(AMD Max+ 395 w/128GB) first impressions.

    Posted by fallingdowndizzyvr@reddit | LocalLLaMA | View on Reddit | 67 comments

  • [Setup discussion] AMD RX 7900 XTX workstation for local LLMs — Linux or Windows as host OS?

    Posted by ElkanRoelen@reddit | LocalLLaMA | View on Reddit | 18 comments

  • Dual CPU Penalty?

    Posted by jsconiers@reddit | LocalLLaMA | View on Reddit | 20 comments

  • Performance expectations question (Devstral)

    Posted by _-Carnage@reddit | LocalLLaMA | View on Reddit | 8 comments

  • New 24B finetune: Impish_Magic_24B

    Posted by Sicarius_The_First@reddit | LocalLLaMA | View on Reddit | 15 comments

  • Local AI for a small/median accounting firm - € Buget of 10k-25k

    Posted by AFruitShopOwner@reddit | LocalLLaMA | View on Reddit | 137 comments

  • How do you make your LLM apps secure?

    Posted by kk17702@reddit | LocalLLaMA | View on Reddit | 8 comments

  • Dual RTX 6000, Blackwell and Ada Lovelace, with thermal imagery

    Posted by Thalesian@reddit | LocalLLaMA | View on Reddit | 19 comments

  • Gemini models (yes, even the recent 2.5 ones) hallucinate crazily on video inputs

    Posted by Infrared12@reddit | LocalLLaMA | View on Reddit | 2 comments

  • Best models to run with 8GB VRAM, 16GB RAM

    Posted by Qxz3@reddit | LocalLLaMA | View on Reddit | 36 comments

  • [DEAL] On-demand B200 GPUs for $1.49/hr at DeepInfra (promo ends June 30)

    Posted by temirulan@reddit | LocalLLaMA | View on Reddit | 4 comments

  • RTX 6000 PRO Blackwell Max Q? Non Max Q?

    Posted by Opening_Progress6820@reddit | LocalLLaMA | View on Reddit | 20 comments

  • Computer-Use on Windows Sandbox

    Posted by Impressive_Half_2819@reddit | LocalLLaMA | View on Reddit | 19 comments

Next