TheaterFire

Login with Reddit

Currently browsing tags:

  • LocalLLaMA
  • Kimi-K2 0905, DeepSeek V3.1, Qwen3-Next-80B-A3B, Grok 4, and others on fresh SWE-bench–style tasks collected in August 2025

    Posted by CuriousPlatypus1881@reddit | LocalLLaMA | View on Reddit | 30 comments

  • once China is able to produce its own GPU for datacenters (which they are forced to due to both import and export bans by both China and USA), there will be less reason to release their models open weight?

    Posted by balianone@reddit | LocalLLaMA | View on Reddit | 106 comments

  • What’s the training cost for models like Qwen3 coder 30b and is the code for training it is open source or close source?

    Posted by NoFudge4700@reddit | LocalLLaMA | View on Reddit | 3 comments

  • China bans its biggest tech companies from acquiring Nvidia chips, says report — Beijing claims its homegrown AI processors now match H20 and RTX Pro 6000D

    Posted by balianone@reddit | LocalLLaMA | View on Reddit | 228 comments

  • Thread for CPU-only LLM performance comparison

    Posted by MLDataScientist@reddit | LocalLLaMA | View on Reddit | 46 comments

  • Magistral Small 2509 has been released

    Posted by jacek2023@reddit | LocalLLaMA | View on Reddit | 113 comments

  • LLM shows signs of over cautious, which has very serious consequences

    Posted by Striking-Warning9533@reddit | LocalLLaMA | View on Reddit | 4 comments

  • Granite 4 release today? Collection updated with 8 private repos.

    Posted by ironwroth@reddit | LocalLLaMA | View on Reddit | 43 comments

  • Qwen3 Coder Plus

    Posted by Dependent_Factor_204@reddit | LocalLLaMA | View on Reddit | 7 comments

  • local llm for macbook air?

    Posted by thebrokebuilder@reddit | LocalLLaMA | View on Reddit | 9 comments

  • DeepSeek-R1 on Nature: How Pure Reinforcement Learning Unlocks LLM Reasoning

    Posted by First_Ground_9849@reddit | LocalLLaMA | View on Reddit | 9 comments

  • Big AI pushes the "we need to beat China" narrative cuz they want fat government contracts and zero democratic oversight. It's an old trick. Fear sells.

    Posted by katxwoods@reddit | LocalLLaMA | View on Reddit | 73 comments

  • Every SOTA on its own data

    Posted by Cheryl_Apple@reddit | LocalLLaMA | View on Reddit | 5 comments

  • Has there been any alt-right equivalent of ChatGPT?

    Posted by Alarmed-Profile5736@reddit | LocalLLaMA | View on Reddit | 2 comments

  • Qwen3-235B-A22B-2507 Released!

    Posted by pseudoreddituser@reddit | LocalLLaMA | View on Reddit | 265 comments

  • The Qwen of Pain.

    Posted by -Ellary-@reddit | LocalLLaMA | View on Reddit | 95 comments

  • What's the smallest model you've gotten to work with OpenCode?

    Posted by RecordFront4405@reddit | LocalLLaMA | View on Reddit | 3 comments

  • Help: looking for a fast web search option with chat

    Posted by My_Unbiased_Opinion@reddit | LocalLLaMA | View on Reddit | 0 comments

  • What is the best local LLM to ask questions about homework, physics, biology, math, and school stuff?

    Posted by Indohorneado@reddit | LocalLLaMA | View on Reddit | 7 comments

  • I bought a modded 4090 48GB in Shenzhen. This is my story.

    Posted by king_priam_of_Troy@reddit | LocalLLaMA | View on Reddit | 344 comments

  • Made a pre-flight check for RAG projects - thoughts?

    Posted by me_z@reddit | LocalLLaMA | View on Reddit | 0 comments

  • I've noticed in this sub corporate tools pose as personal projects

    Posted by kuhunaxeyive@reddit | LocalLLaMA | View on Reddit | 61 comments

  • Local LLM laptop budget 2.5-5k

    Posted by 0800otto@reddit | LocalLLaMA | View on Reddit | 63 comments

  • Qwen released Qwen3-Next-80B-A3B — the FUTURE of efficient LLMs is here!

    Posted by ResearchCrafty1804@reddit | LocalLLaMA | View on Reddit | 217 comments

  • M1 Ultra Mac Studio vs AMD Ryzen AI Max 395+ for local AI?

    Posted by doweig@reddit | LocalLLaMA | View on Reddit | 37 comments

  • Can I use Cursor Agent (or similar) with a local LLM setup (8B / 13B)?

    Posted by BudgetPurple3002@reddit | LocalLLaMA | View on Reddit | 8 comments

  • Google's paper, SLED, seems to improve factuality with (all? Most?) LLMs at only a 4% speed penalty

    Posted by laser_man6@reddit | LocalLLaMA | View on Reddit | 0 comments

  • GPU advice for running local coding LLMs

    Posted by mak3rdad@reddit | LocalLLaMA | View on Reddit | 12 comments

  • I built a local AI agent that turns my messy computer into a private, searchable memory

    Posted by AlanzhuLy@reddit | LocalLLaMA | View on Reddit | 23 comments

  • Is anyone able to successfully run Qwen 30B Coder BF16?

    Posted by TokenRingAI@reddit | LocalLLaMA | View on Reddit | 20 comments

  • I evaluated several small and SOTA LLMs on Python code generation

    Posted by spacespacespapce@reddit | LocalLLaMA | View on Reddit | 1 comments

  • How to make a small LLM from scratch?

    Posted by Charming_Barber_3317@reddit | LocalLLaMA | View on Reddit | 20 comments

  • How are some of you running 6x gpu's?

    Posted by eat_those_lemons@reddit | LocalLLaMA | View on Reddit | 69 comments

  • When AI admits it doesn't know: the surprising behavior of Magistral-small-2509 in a soccer quiz contest

    Posted by Different_File6723@reddit | LocalLLaMA | View on Reddit | 1 comments

  • FULL Orchids.app System Prompt and Tools

    Posted by Independent-Box-898@reddit | LocalLLaMA | View on Reddit | 2 comments

  • Is it possible for different brand GPUs to work together?

    Posted by WizardlyBump17@reddit | LocalLLaMA | View on Reddit | 5 comments

  • How is llama.cpp or other implementations handle tokenization without tiktoken?

    Posted by EricHermosis@reddit | LocalLLaMA | View on Reddit | 11 comments

  • 500,000 public datasets on Hugging Face

    Posted by clem59480@reddit | LocalLLaMA | View on Reddit | 11 comments

  • OpenAI usage breakdown released

    Posted by LeatherRub7248@reddit | LocalLLaMA | View on Reddit | 27 comments

  • Want to set up my own AI thing for RPing (Story Driven)...

    Posted by Sufficient-Grape5366@reddit | LocalLLaMA | View on Reddit | 12 comments

  • Qwen3 coder Plus vs Grok Code Fast which is the best free model?

    Posted by Level-Dig-4807@reddit | LocalLLaMA | View on Reddit | 4 comments

  • Running LLMS on RAM ?

    Posted by Electronic_Image1665@reddit | LocalLLaMA | View on Reddit | 8 comments

  • Best sub 14b llm for long text summaries?

    Posted by GreenTreeAndBlueSky@reddit | LocalLLaMA | View on Reddit | 16 comments

  • SvelteKit-based WebUI by allozaur · Pull Request #14839 · ggml-org/llama.cpp

    Posted by jacek2023@reddit | LocalLLaMA | View on Reddit | 9 comments

  • A Quick Look At The AMD Instinct MI355X With ROCm 7.0

    Posted by Balance-@reddit | LocalLLaMA | View on Reddit | 4 comments

  • Qwen3 Next 80b is gimped. Back to Gemma 3

    Posted by meshreplacer@reddit | LocalLLaMA | View on Reddit | 8 comments

  • Best document parser

    Posted by aiwtl@reddit | LocalLLaMA | View on Reddit | 17 comments

  • Not a developer, want to make a simple web app, which LLM should I use?

    Posted by 1BlueSpork@reddit | LocalLLaMA | View on Reddit | 61 comments

  • LACT "indirect undervolt & OC" method beats `nvidia-smi -pl 400` on 3090TI FE.

    Posted by VoidAlchemy@reddit | LocalLLaMA | View on Reddit | 7 comments

  • How I'm using Claude/ChatGPT + voice to replace my entire multi-monitor setup

    Posted by Smooth-Loquat-4954@reddit | LocalLLaMA | View on Reddit | 9 comments

Next