TheaterFire

Login with Reddit

Currently browsing tags:

  • LocalLLaMA
  • How can I show log probs for a demo

    Posted by SQLGene@reddit | LocalLLaMA | View on Reddit | 0 comments

  • ​The White House just launched "The Genesis Mission": A Manhattan Project-style initiative for AI

    Posted by iamnottheabyss@reddit | LocalLLaMA | View on Reddit | 85 comments

  • Cheapest $/vRAM GPU right now? Is it a good time?

    Posted by Roy3838@reddit | LocalLLaMA | View on Reddit | 79 comments

  • New Open-source text-to-image model from Alibaba is just below Seedream 4, Coming today or tomorrow!

    Posted by abdouhlili@reddit | LocalLLaMA | View on Reddit | 11 comments

  • 10k Hardware for LLM

    Posted by Appropriate-Quit1714@reddit | LocalLLaMA | View on Reddit | 34 comments

  • Correct a dangerous racial bias in an LLM through targeted pruning

    Posted by pmartra@reddit | LocalLLaMA | View on Reddit | 68 comments

  • JanV1-Q8 still cant answer some basic of questions

    Posted by choxxolatee@reddit | LocalLLaMA | View on Reddit | 18 comments

  • BPE tokenizer in Rust - would love feedback from the community

    Posted by farhan-dev@reddit | LocalLLaMA | View on Reddit | 8 comments

  • OpenAI-GPT-OSS-120B scores on livecodebench

    Posted by Used-Negotiation-741@reddit | LocalLLaMA | View on Reddit | 6 comments

  • Trying to build a "Jarvis" that never phones home - on-device AI with full access to your digital life (free beta, roast us)

    Posted by ipav9@reddit | LocalLLaMA | View on Reddit | 68 comments

  • I built an open-source Memory API because setting up vector DBs for every AI project was annoying

    Posted by Eastern-Height2451@reddit | LocalLLaMA | View on Reddit | 5 comments

  • Is there any local LLM that comes close to GPT-4 in reasoning and capabilities? Hardware suggestion?

    Posted by ExtensionAd182@reddit | LocalLLaMA | View on Reddit | 46 comments

  • [Pre-release] Wavefront AI, the fully open-source AI middleware built over FloAI for Enterprises

    Posted by Traditional-Let-856@reddit | LocalLLaMA | View on Reddit | 0 comments

  • ClearCut – open-source tool that forces you to think before AI answers

    Posted by ComplexCanary1860@reddit | LocalLLaMA | View on Reddit | 2 comments

  • You can now do FP8 reinforcement learning locally! (<5GB VRAM)

    Posted by danielhanchen@reddit | LocalLLaMA | View on Reddit | 61 comments

  • TikTok · SOLAYai

    Posted by SOLAYAi@reddit | LocalLLaMA | View on Reddit | 1 comments

  • What are these supposed no branding 3090s?

    Posted by aeroumbria@reddit | LocalLLaMA | View on Reddit | 19 comments

  • Is the llama.cpp webui in danger from the recent npm attack?

    Posted by shroddy@reddit | LocalLLaMA | View on Reddit | 11 comments

  • I made a free playground for comparing 10+ OCR models side-by-side

    Posted by Emc2fma@reddit | LocalLLaMA | View on Reddit | 83 comments

  • I'm looking for an Uncensored LLM to produce extremely spicy prompts - What would you recommend?

    Posted by Whipit@reddit | LocalLLaMA | View on Reddit | 14 comments

  • Calling a Finetune/LoRA Wizard: Need Dataset Tips for RP Model

    Posted by AmpedHorizon@reddit | LocalLLaMA | View on Reddit | 17 comments

  • GPUs - what to do?

    Posted by engineeringstoned@reddit | LocalLLaMA | View on Reddit | 4 comments

  • Hidden causes of LLM latency, its not just the model size

    Posted by emmettvance@reddit | LocalLLaMA | View on Reddit | 2 comments

  • tencent/HunyuanOCR-1B

    Posted by nullmove@reddit | LocalLLaMA | View on Reddit | 21 comments

  • LM Studio running very slow compared to Ollama

    Posted by EaZyRecipeZ@reddit | LocalLLaMA | View on Reddit | 10 comments

  • Can application layer improve local model output quality?

    Posted by ayechat@reddit | LocalLLaMA | View on Reddit | 16 comments

  • Planning Multi-RTX 5060 Ti Local LLM Workstation (TRX40 / 32–64GB VRAM)

    Posted by Special-Art-9369@reddit | LocalLLaMA | View on Reddit | 24 comments

  • Polish is the most effective language for prompting AI, study reveals

    Posted by tengo_harambe@reddit | LocalLLaMA | View on Reddit | 202 comments

  • Feedback | Local LLM Build 2x RTX Pro 4000

    Posted by sebakirs@reddit | LocalLLaMA | View on Reddit | 21 comments

  • Spiralers vs Engineers vs Researchers … The Real Divide in AI

    Posted by prime_architect@reddit | LocalLLaMA | View on Reddit | 4 comments

  • [Beta Testing] Built infrastructure to prevent LLM drift, need testers !! (10 mins)

    Posted by Awkward_Article5427@reddit | LocalLLaMA | View on Reddit | 2 comments

  • Validating a visual orchestration tool for local LLMs (concept feedback wanted)

    Posted by HarjjotSinghh@reddit | LocalLLaMA | View on Reddit | 6 comments

  • Why talking to AI assistants sucks: a project that's finally fixing the interruption problem.

    Posted by Parking_Cricket_9194@reddit | LocalLLaMA | View on Reddit | 3 comments

  • That's why local models are better

    Posted by Illustrious-Swim9663@reddit | LocalLLaMA | View on Reddit | 220 comments

  • I built an AI research platform and just open sourced it.

    Posted by CodingWithSatyam@reddit | LocalLLaMA | View on Reddit | 11 comments

  • LLaDA2.0 (103B/16B) has been released

    Posted by jacek2023@reddit | LocalLLaMA | View on Reddit | 72 comments

  • Low restricted local model

    Posted by daibo_taibo@reddit | LocalLLaMA | View on Reddit | 5 comments

  • Looking for a solution for cleaning up/summarizing long voice transcriptions on M1 Max 32GB.

    Posted by Bulky_Jellyfish_2616@reddit | LocalLLaMA | View on Reddit | 8 comments

  • An explainer blog on attention, KV-caching, continuous batching

    Posted by unofficialmerve@reddit | LocalLLaMA | View on Reddit | 1 comments

  • Is Bert-Nebulon Alpha the new GLM model?

    Posted by Cool-Chemical-5629@reddit | LocalLLaMA | View on Reddit | 18 comments

  • Built a self-hosted form builder where you can chat to create forms (open source)

    Posted by Careful_Patience_815@reddit | LocalLLaMA | View on Reddit | 2 comments

  • Testing call handoff logic to humans best approach?

    Posted by AugustusCaesar00@reddit | LocalLLaMA | View on Reddit | 0 comments

  • Heretic: Fully automatic censorship removal for language models

    Posted by -p-e-w-@reddit | LocalLLaMA | View on Reddit | 296 comments

  • Does gpt-oss:20b’s thinking output cause more confusion than help in multi-step tasks?

    Posted by Prestigious_Peak_773@reddit | LocalLLaMA | View on Reddit | 7 comments

  • 4070 Super (12gb) vs 5070ti (16gb)

    Posted by rabbany05@reddit | LocalLLaMA | View on Reddit | 11 comments

  • I tested a few local hosted coding models with VSCode / cline so that you don't have to

    Posted by DrMicrobit@reddit | LocalLLaMA | View on Reddit | 18 comments

  • Opus 4.5 claims 1st place on fresh SWE-bench-like problems in October [SWE-rebench]

    Posted by Fabulous_Pollution10@reddit | LocalLLaMA | View on Reddit | 3 comments

  • I made a free site with file tools + a local AI chat that connects to Ollama

    Posted by opal-emporium@reddit | LocalLLaMA | View on Reddit | 8 comments

  • Can local llm's teach complex subjects? (Such as 3D modeling?)

    Posted by No_Strawberry_8719@reddit | LocalLLaMA | View on Reddit | 3 comments

  • Building agents using SMLs

    Posted by shoeshineboy_99@reddit | LocalLLaMA | View on Reddit | 1 comments

Next