TheaterFire

Login with Reddit

Currently browsing tags:

  • LocalLLaMA
  • Best Local VLMs - November 2025

    Posted by rm-rf-rm@reddit | LocalLLaMA | View on Reddit | 36 comments

  • Gaming Laptop for LLM and SD

    Posted by Appropriate-Quit1714@reddit | LocalLLaMA | View on Reddit | 0 comments

  • What's the best machine I can get for $10k?

    Posted by TWUC@reddit | LocalLLaMA | View on Reddit | 56 comments

  • Benchmarks and evals

    Posted by selund1@reddit | LocalLLaMA | View on Reddit | 9 comments

  • Looking for open source 10B model that is comparable to gpt4o-mini

    Posted by bohemianLife1@reddit | LocalLLaMA | View on Reddit | 35 comments

  • Local AI for small biz owner

    Posted by binyang@reddit | LocalLLaMA | View on Reddit | 20 comments

  • unsloth/Qwen3-Next-80B-A3B-Instruct-GGUF · Hugging Face

    Posted by WhaleFactory@reddit | LocalLLaMA | View on Reddit | 100 comments

  • Setup with Nvidia 6000 Pro

    Posted by Appropriate-Quit1714@reddit | LocalLLaMA | View on Reddit | 7 comments

  • Qwen3 Next imatrix GGUFs up!

    Posted by noneabove1182@reddit | LocalLLaMA | View on Reddit | 13 comments

  • When will AGI arrive?

    Posted by ObjectSmooth8899@reddit | LocalLLaMA | View on Reddit | 24 comments

  • A Tribute to MetaAI and Stability AI - 2 Giants Who Brought us so Much Joy... And, 2025 is the Year they Die... So Sad!😢

    Posted by Iory1998@reddit | LocalLLaMA | View on Reddit | 50 comments

  • When, if ever, do you think we will have an open-source Gemini 3 Pro?

    Posted by Unusual_Guidance2095@reddit | LocalLLaMA | View on Reddit | 23 comments

  • Open source nano banana 🍌 pro alternative?

    Posted by PumpkinNarrow6339@reddit | LocalLLaMA | View on Reddit | 6 comments

  • Yet another reason to stick with local models

    Posted by nekofneko@reddit | LocalLLaMA | View on Reddit | 20 comments

  • Thoughts on how to structure and implement RAG for genealogical datasets?

    Posted by Own_Attention_3392@reddit | LocalLLaMA | View on Reddit | 3 comments

  • Deepseek Unchained?

    Posted by muffinnmannn@reddit | LocalLLaMA | View on Reddit | 7 comments

  • Qwen3-Next-80B-A3B vs gpt-oss-120b

    Posted by bfroemel@reddit | LocalLLaMA | View on Reddit | 42 comments

  • Ask me to run models

    Posted by monoidconcat@reddit | LocalLLaMA | View on Reddit | 135 comments

  • CPU-only LLM performance - t/s with llama.cpp

    Posted by pmttyji@reddit | LocalLLaMA | View on Reddit | 64 comments

  • Try the new Z-Image-Turbo 6B (Runs on 8GB VRAM)!

    Posted by KvAk_AKPlaysYT@reddit | LocalLLaMA | View on Reddit | 26 comments

  • 3090 48GB

    Posted by tronathan@reddit | LocalLLaMA | View on Reddit | 65 comments

  • Optimising NVIDIA’s DGX Spark (Grace + Blackwell) – 1.5× PyTorch speedup with custom build

    Posted by guigsss@reddit | LocalLLaMA | View on Reddit | 22 comments

  • Perfecto!!

    Posted by Icy_Resolution8390@reddit | LocalLLaMA | View on Reddit | 15 comments

  • Grüße aus Shenzhen: Wir haben ein NAS-Gehäuse entwickelt, das DeepSeek R1 70B lokal packt (20 t/s mit interner 4090). Feedback erwünscht!

    Posted by Maleficent_Cap9844@reddit | LocalLLaMA | View on Reddit | 23 comments

  • Gemma3 27 heretic, lower divergence than mlabonne/gemma3

    Posted by coder3101@reddit | LocalLLaMA | View on Reddit | 12 comments

  • RAG from Scratch is now live on GitHub

    Posted by purellmagents@reddit | LocalLLaMA | View on Reddit | 3 comments

  • Best Open Source LLM for Langraph Agent

    Posted by geekyrahulvk@reddit | LocalLLaMA | View on Reddit | 7 comments

  • Who is Elara?

    Posted by itsjustmarky@reddit | LocalLLaMA | View on Reddit | 40 comments

  • The official vLLM support for the Ryzen AI Max+ 395 is here! (the whole AI 300 series, ie gfx1150 and gfx1151)

    Posted by waiting_for_zban@reddit | LocalLLaMA | View on Reddit | 2 comments

  • (Partly) Open Video Overview – Generate narrated videos from text with AI (requires Gemini API)

    Posted by arbayi@reddit | LocalLLaMA | View on Reddit | 0 comments

  • Compared actual usage costs for Chinese AI models. Token efficiency changes everything.

    Posted by YormeSachi@reddit | LocalLLaMA | View on Reddit | 41 comments

  • Claude code can now connect directly to llama.cpp server

    Posted by tarruda@reddit | LocalLLaMA | View on Reddit | 15 comments

  • I built an AI research platform and just open sourced it.

    Posted by CodingWithSatyam@reddit | LocalLLaMA | View on Reddit | 14 comments

  • 2 x DGX Spark! Give me your non-inference workloads

    Posted by entsnack@reddit | LocalLLaMA | View on Reddit | 130 comments

  • MI50 price hike, are they moving inventory at that price?

    Posted by emaiksiaime@reddit | LocalLLaMA | View on Reddit | 25 comments

  • Recommendation for Production Hardware for inference and fine tuning.

    Posted by Whyme-__-@reddit | LocalLLaMA | View on Reddit | 1 comments

  • Will I have any problems pairing a 3090 with a 5060 Ti 16GB?

    Posted by PhantomWolf83@reddit | LocalLLaMA | View on Reddit | 5 comments

  • Handy - a simple, open-source offline speech-to-text app written in Rust using whisper.cpp

    Posted by sipjca@reddit | LocalLLaMA | View on Reddit | 41 comments

  • Where did the Epstein emails dataset go

    Posted by egomarker@reddit | LocalLLaMA | View on Reddit | 82 comments

  • It turns out WDDM driver mode is making our RAM - GPU transfer extremely slower compared to TCC or MCDM mode. Anyone has figured out the bypass NVIDIA software level restrictions?

    Posted by CeFurkan@reddit | LocalLLaMA | View on Reddit | 52 comments

  • You can now do FP8 reinforcement learning locally! (<5GB VRAM)

    Posted by danielhanchen@reddit | LocalLLaMA | View on Reddit | 80 comments

  • Manufactured 4090 48gb AMA

    Posted by koalfied-coder@reddit | LocalLLaMA | View on Reddit | 82 comments

  • Inference-time drift reduces repetition collapse in frozen Llama-3.1-8B (repo + reproducible script)

    Posted by chazc2@reddit | LocalLLaMA | View on Reddit | 2 comments

  • Anthropic just showed how to make AI agents work on long projects without falling apart

    Posted by purealgo@reddit | LocalLLaMA | View on Reddit | 81 comments

  • AMD 395+ and NVIDIA GPU

    Posted by EntropyNegotiator@reddit | LocalLLaMA | View on Reddit | 11 comments

  • Best Models for 16GB VRAM

    Posted by LinuxIsFree@reddit | LocalLLaMA | View on Reddit | 36 comments

  • Has this happened with anyone!?

    Posted by Hamilton-Io@reddit | LocalLLaMA | View on Reddit | 1 comments

  • Call for all teachers

    Posted by GnobarEl@reddit | LocalLLaMA | View on Reddit | 3 comments

  • Current local models that work well as coding agents

    Posted by yehiaserag@reddit | LocalLLaMA | View on Reddit | 7 comments

  • Hardcore RAG & AI Search resources

    Posted by LilDemonApparel@reddit | LocalLLaMA | View on Reddit | 10 comments

Next