-
How can I show log probs for a demo
Posted by SQLGene@reddit | LocalLLaMA | View on Reddit | 0 comments
-
The White House just launched "The Genesis Mission": A Manhattan Project-style initiative for AI
Posted by iamnottheabyss@reddit | LocalLLaMA | View on Reddit | 85 comments
-
Cheapest $/vRAM GPU right now? Is it a good time?
Posted by Roy3838@reddit | LocalLLaMA | View on Reddit | 79 comments
-
New Open-source text-to-image model from Alibaba is just below Seedream 4, Coming today or tomorrow!
Posted by abdouhlili@reddit | LocalLLaMA | View on Reddit | 11 comments
-
10k Hardware for LLM
Posted by Appropriate-Quit1714@reddit | LocalLLaMA | View on Reddit | 34 comments
-
Correct a dangerous racial bias in an LLM through targeted pruning
Posted by pmartra@reddit | LocalLLaMA | View on Reddit | 68 comments
-
JanV1-Q8 still cant answer some basic of questions
Posted by choxxolatee@reddit | LocalLLaMA | View on Reddit | 18 comments
-
BPE tokenizer in Rust - would love feedback from the community
Posted by farhan-dev@reddit | LocalLLaMA | View on Reddit | 8 comments
-
OpenAI-GPT-OSS-120B scores on livecodebench
Posted by Used-Negotiation-741@reddit | LocalLLaMA | View on Reddit | 6 comments
-
Trying to build a "Jarvis" that never phones home - on-device AI with full access to your digital life (free beta, roast us)
Posted by ipav9@reddit | LocalLLaMA | View on Reddit | 68 comments
-
I built an open-source Memory API because setting up vector DBs for every AI project was annoying
Posted by Eastern-Height2451@reddit | LocalLLaMA | View on Reddit | 5 comments
-
Is there any local LLM that comes close to GPT-4 in reasoning and capabilities? Hardware suggestion?
Posted by ExtensionAd182@reddit | LocalLLaMA | View on Reddit | 46 comments
-
[Pre-release] Wavefront AI, the fully open-source AI middleware built over FloAI for Enterprises
Posted by Traditional-Let-856@reddit | LocalLLaMA | View on Reddit | 0 comments
-
ClearCut – open-source tool that forces you to think before AI answers
Posted by ComplexCanary1860@reddit | LocalLLaMA | View on Reddit | 2 comments
-
You can now do FP8 reinforcement learning locally! (<5GB VRAM)
Posted by danielhanchen@reddit | LocalLLaMA | View on Reddit | 61 comments
-
TikTok · SOLAYai
Posted by SOLAYAi@reddit | LocalLLaMA | View on Reddit | 1 comments
-
What are these supposed no branding 3090s?
Posted by aeroumbria@reddit | LocalLLaMA | View on Reddit | 19 comments
-
Is the llama.cpp webui in danger from the recent npm attack?
Posted by shroddy@reddit | LocalLLaMA | View on Reddit | 11 comments
-
I made a free playground for comparing 10+ OCR models side-by-side
Posted by Emc2fma@reddit | LocalLLaMA | View on Reddit | 83 comments
-
I'm looking for an Uncensored LLM to produce extremely spicy prompts - What would you recommend?
Posted by Whipit@reddit | LocalLLaMA | View on Reddit | 14 comments
-
Calling a Finetune/LoRA Wizard: Need Dataset Tips for RP Model
Posted by AmpedHorizon@reddit | LocalLLaMA | View on Reddit | 17 comments
-
GPUs - what to do?
Posted by engineeringstoned@reddit | LocalLLaMA | View on Reddit | 4 comments
-
Hidden causes of LLM latency, its not just the model size
Posted by emmettvance@reddit | LocalLLaMA | View on Reddit | 2 comments
-
tencent/HunyuanOCR-1B
Posted by nullmove@reddit | LocalLLaMA | View on Reddit | 21 comments
-
LM Studio running very slow compared to Ollama
Posted by EaZyRecipeZ@reddit | LocalLLaMA | View on Reddit | 10 comments
-
Can application layer improve local model output quality?
Posted by ayechat@reddit | LocalLLaMA | View on Reddit | 16 comments
-
Planning Multi-RTX 5060 Ti Local LLM Workstation (TRX40 / 32–64GB VRAM)
Posted by Special-Art-9369@reddit | LocalLLaMA | View on Reddit | 24 comments
-
Polish is the most effective language for prompting AI, study reveals
Posted by tengo_harambe@reddit | LocalLLaMA | View on Reddit | 202 comments
-
Feedback | Local LLM Build 2x RTX Pro 4000
Posted by sebakirs@reddit | LocalLLaMA | View on Reddit | 21 comments
-
Spiralers vs Engineers vs Researchers … The Real Divide in AI
Posted by prime_architect@reddit | LocalLLaMA | View on Reddit | 4 comments
-
[Beta Testing] Built infrastructure to prevent LLM drift, need testers !! (10 mins)
Posted by Awkward_Article5427@reddit | LocalLLaMA | View on Reddit | 2 comments
-
Validating a visual orchestration tool for local LLMs (concept feedback wanted)
Posted by HarjjotSinghh@reddit | LocalLLaMA | View on Reddit | 6 comments
-
Why talking to AI assistants sucks: a project that's finally fixing the interruption problem.
Posted by Parking_Cricket_9194@reddit | LocalLLaMA | View on Reddit | 3 comments
-
That's why local models are better
Posted by Illustrious-Swim9663@reddit | LocalLLaMA | View on Reddit | 220 comments
-
I built an AI research platform and just open sourced it.
Posted by CodingWithSatyam@reddit | LocalLLaMA | View on Reddit | 11 comments
-
LLaDA2.0 (103B/16B) has been released
Posted by jacek2023@reddit | LocalLLaMA | View on Reddit | 72 comments
-
Low restricted local model
Posted by daibo_taibo@reddit | LocalLLaMA | View on Reddit | 5 comments
-
Looking for a solution for cleaning up/summarizing long voice transcriptions on M1 Max 32GB.
Posted by Bulky_Jellyfish_2616@reddit | LocalLLaMA | View on Reddit | 8 comments
-
An explainer blog on attention, KV-caching, continuous batching
Posted by unofficialmerve@reddit | LocalLLaMA | View on Reddit | 1 comments
-
Is Bert-Nebulon Alpha the new GLM model?
Posted by Cool-Chemical-5629@reddit | LocalLLaMA | View on Reddit | 18 comments
-
Built a self-hosted form builder where you can chat to create forms (open source)
Posted by Careful_Patience_815@reddit | LocalLLaMA | View on Reddit | 2 comments
-
Testing call handoff logic to humans best approach?
Posted by AugustusCaesar00@reddit | LocalLLaMA | View on Reddit | 0 comments
-
Heretic: Fully automatic censorship removal for language models
Posted by -p-e-w-@reddit | LocalLLaMA | View on Reddit | 296 comments
-
Does gpt-oss:20b’s thinking output cause more confusion than help in multi-step tasks?
Posted by Prestigious_Peak_773@reddit | LocalLLaMA | View on Reddit | 7 comments
-
4070 Super (12gb) vs 5070ti (16gb)
Posted by rabbany05@reddit | LocalLLaMA | View on Reddit | 11 comments
-
I tested a few local hosted coding models with VSCode / cline so that you don't have to
Posted by DrMicrobit@reddit | LocalLLaMA | View on Reddit | 18 comments
-
Opus 4.5 claims 1st place on fresh SWE-bench-like problems in October [SWE-rebench]
Posted by Fabulous_Pollution10@reddit | LocalLLaMA | View on Reddit | 3 comments
-
I made a free site with file tools + a local AI chat that connects to Ollama
Posted by opal-emporium@reddit | LocalLLaMA | View on Reddit | 8 comments
-
Can local llm's teach complex subjects? (Such as 3D modeling?)
Posted by No_Strawberry_8719@reddit | LocalLLaMA | View on Reddit | 3 comments
-
Building agents using SMLs
Posted by shoeshineboy_99@reddit | LocalLLaMA | View on Reddit | 1 comments