TheaterFire

Login with Reddit

Currently browsing tags:

  • LocalLLaMA
  • Sparse Adaptive Attention “MoE”, a potential performance breakthrough for LLMs?

    Posted by kaggleqrdl@reddit | LocalLLaMA | View on Reddit | 8 comments

  • What are your real life/WORK use cases with LOCAL LLMs

    Posted by Adventurous-Gold6413@reddit | LocalLLaMA | View on Reddit | 11 comments

  • Best current dense, nonthinking models in the 8b-14b range?

    Posted by Priceless_Pennies@reddit | LocalLLaMA | View on Reddit | 20 comments

  • Using my Mac Mini M4 as an LLM server—Looking for recommendations

    Posted by cockpit_dandruff@reddit | LocalLLaMA | View on Reddit | 7 comments

  • L16 Prompt Drift: GPT-2 vs LLaMA-3-8B — IDENTICAL DRIFT

    Posted by Mysterious_Doubt_341@reddit | LocalLLaMA | View on Reddit | 0 comments

  • Qwen3 Max Thinking this week

    Posted by ResearchCrafty1804@reddit | LocalLLaMA | View on Reddit | 33 comments

  • Getting llm on low end phone

    Posted by hemtai_lover@reddit | LocalLLaMA | View on Reddit | 2 comments

  • Hebrew_Nemo: a state-of-the-art Hebrew large language model

    Posted by Sicarius_The_First@reddit | LocalLLaMA | View on Reddit | 4 comments

  • Need advice on building a GPU-based render/Al compute setup: Unsure about hardware direction

    Posted by One_Abroad_5937@reddit | LocalLLaMA | View on Reddit | 0 comments

  • Choosing the right model

    Posted by Bowdenzug@reddit | LocalLLaMA | View on Reddit | 4 comments

  • AMD MI50 32GB/Vega20 GPU Passthrough Guide for Proxmox

    Posted by Panda24z@reddit | LocalLLaMA | View on Reddit | 41 comments

  • Open-source web crawler with markdown output?

    Posted by mxdamp@reddit | LocalLLaMA | View on Reddit | 8 comments

  • Granite 4.0 Nano Language Models

    Posted by ApprehensiveAd3629@reddit | LocalLLaMA | View on Reddit | 68 comments

  • Reliable source for used 3090 ?

    Posted by vdiallonort@reddit | LocalLLaMA | View on Reddit | 14 comments

  • Improving RAG Results with OpenWebUI - Looking for Advice on Custom Pipelines & Better Embeddings

    Posted by b5761@reddit | LocalLLaMA | View on Reddit | 8 comments

  • Best Hardware Setup to Run DeepSeek-V3 670B Locally on $40K–$80K?

    Posted by PrevelantInsanity@reddit | LocalLLaMA | View on Reddit | 65 comments

  • GPT-OSS Safeguard coming soon

    Posted by Independent-Ruin-376@reddit | LocalLLaMA | View on Reddit | 12 comments

  • Speculation or rumors on Gemma 4?

    Posted by RobotRobotWhatDoUSee@reddit | LocalLLaMA | View on Reddit | 0 comments

  • How much does 1T tokens cost? How much did all these amazing people spent on OpenAI tokens?

    Posted by aospan@reddit | LocalLLaMA | View on Reddit | 44 comments

  • Uncensored AI for scientific research

    Posted by PrintCreepy8982@reddit | LocalLLaMA | View on Reddit | 14 comments

  • L16 Prompt Drift Experiment — Live Colab (GPT-2)

    Posted by Mysterious_Doubt_341@reddit | LocalLLaMA | View on Reddit | 0 comments

  • How is the new Grok AI girlfriend animation implemented?

    Posted by EvilKY45@reddit | LocalLLaMA | View on Reddit | 42 comments

  • how the ai models are editing the code snippets???

    Posted by lavangamm@reddit | LocalLLaMA | View on Reddit | 3 comments

  • Best chat interface currently (Aug 2025)

    Posted by cmdr-William-Riker@reddit | LocalLLaMA | View on Reddit | 27 comments

  • Minimax-M2 cracks top 10 overall LLMs (production LLM performance gap shrinking: 7 points from GPT-5 in Artificial Analysis benchmark)

    Posted by medi6@reddit | LocalLLaMA | View on Reddit | 23 comments

  • Poker Tournament for LLMs

    Posted by undoing8@reddit | LocalLLaMA | View on Reddit | 28 comments

  • Serve 100 Large AI Models on a single GPU with low impact to time to first token.

    Posted by SetZealousideal5006@reddit | LocalLLaMA | View on Reddit | 6 comments

  • IBM releases Granite-4.0 Nano (300M & 1B), along with a local browser demo showing how the models can programmatically interact with websites and call tools/browser APIs on your behalf.

    Posted by xenovatech@reddit | LocalLLaMA | View on Reddit | 26 comments

  • What is the best open source TTS model with multi language support?

    Posted by Anxietrap@reddit | LocalLLaMA | View on Reddit | 22 comments

  • MiniMaxAI/MiniMax-M2 · Hugging Face

    Posted by Dark_Fire_12@reddit | LocalLLaMA | View on Reddit | 52 comments

  • Local Hosting Question

    Posted by Media_Express@reddit | LocalLLaMA | View on Reddit | 15 comments

  • Llama3.3:70b vs GPT-OSS:20b for PHP Code Generation

    Posted by AppledogHu@reddit | LocalLLaMA | View on Reddit | 13 comments

  • Phone LLM's benchmarks?

    Posted by ctrl-brk@reddit | LocalLLaMA | View on Reddit | 34 comments

  • Experience with the new model MiniMax M2 and some cost saving tips

    Posted by thalacque@reddit | LocalLLaMA | View on Reddit | 26 comments

  • AMA Announcement: Liquid AI, the team behind Liquid Foundational Models, LEAP and Apollo (Thu, Oct 30 • 10 AM – 1 PM PDT)

    Posted by LiquidAI_Team@reddit | LocalLLaMA | View on Reddit | 5 comments

  • Anyone running local LLM coding setups on 24GB VRAM laptops? Looking for real-world experiences

    Posted by AmazinglyNatural6545@reddit | LocalLLaMA | View on Reddit | 20 comments

  • MiniMax-M2 llama.cpp

    Posted by butlan@reddit | LocalLLaMA | View on Reddit | 11 comments

  • Just dropped Kani TTS English - a 400M TTS model that's 5x faster than realtime on RTX 4080

    Posted by ylankgz@reddit | LocalLLaMA | View on Reddit | 37 comments

  • Book suggestion with chapters to be 'must read' to get most from the book

    Posted by bad_detectiv3@reddit | LocalLLaMA | View on Reddit | 5 comments

  • Bad news: DGX Spark may have only half the performance claimed.

    Posted by Dr_Karminski@reddit | LocalLLaMA | View on Reddit | 253 comments

  • First impressions and thoughts on the GTR9 Pro (Beelink's 395)

    Posted by kmouratidis@reddit | LocalLLaMA | View on Reddit | 28 comments

  • tokens per second on a NASA computer

    Posted by Pro-editor-1105@reddit | LocalLLaMA | View on Reddit | 9 comments

  • GLM-4.6 on fresh SWE-bench–style tasks collected in September 2025

    Posted by CuriousPlatypus1881@reddit | LocalLLaMA | View on Reddit | 25 comments

  • Completed 8xAMD MI50 - 256GB VRAM + 256GB RAM rig for $3k

    Posted by MLDataScientist@reddit | LocalLLaMA | View on Reddit | 221 comments

  • MiniMax M2 Llama.cpp support

    Posted by ilintar@reddit | LocalLLaMA | View on Reddit | 14 comments

  • Local coding models limit

    Posted by Blues520@reddit | LocalLLaMA | View on Reddit | 14 comments

  • OSS alternative to Open WebUI - ChatGPT-like UI, API and CLI

    Posted by mythz@reddit | LocalLLaMA | View on Reddit | 93 comments

  • ⚠ NOT ALLOWED TO EVEN WONDER ⚠

    Posted by researchAmericanAI@reddit | LocalLLaMA | View on Reddit | 1 comments

  • PipesHub - Open Source Enterprise Search Platform(Generative-AI Powered)

    Posted by Effective-Ad2060@reddit | LocalLLaMA | View on Reddit | 8 comments

  • ⚠ AUTOMATIC ⚠

    Posted by researchAmericanAI@reddit | LocalLLaMA | View on Reddit | 0 comments

Next