-
Besides coding and chatting, how do you use LLMs?
Posted by 330d@reddit | LocalLLaMA | View on Reddit | 170 comments
-
So, that new sampling paradigm seems to be real after all
Posted by Ok_Landscape_6819@reddit | LocalLLaMA | View on Reddit | 14 comments
-
AutoRAG Huggingface Space Release : Optimize RAG & evaluation dataset generation without installation.
Posted by jeffrey-0711@reddit | LocalLLaMA | View on Reddit | 6 comments
-
New model | Llama-3.1-nemotron-70b-instruct
Posted by redjojovic@reddit | LocalLLaMA | View on Reddit | 56 comments
-
Reverse Engineering o1 Architecture (With a little help from our friend Claude)
Posted by TechnoTherapist@reddit | LocalLLaMA | View on Reddit | 55 comments
-
DARKEST Planet 16.5B - Unusually strong non AI creative model, with "regen" randomness.
Posted by Dangerous_Fix_5526@reddit | LocalLLaMA | View on Reddit | 41 comments
-
OpenAI releases new open-source agent orchestration library: Swarm
Posted by umarmnaq@reddit | LocalLLaMA | View on Reddit | 10 comments
-
Just curious what those of you with 3090s are running for roleplay right now?
Posted by delicatemicdrop@reddit | LocalLLaMA | View on Reddit | 10 comments
-
How difficult would it be to have a text-to-speech setup like Elevenlabs at home?
Posted by iaseth@reddit | LocalLLaMA | View on Reddit | 41 comments
-
Thread of the "Reflection 70B" debacle by it's creator
Posted by rambat1994@reddit | LocalLLaMA | View on Reddit | 9 comments
-
LLM training bug fixes - Gradient accumulation was wrong
Posted by danielhanchen@reddit | LocalLLaMA | View on Reddit | 25 comments
-
Grok-2 and Grok-2-mini on Aider's Code Editing Benchmark
Posted by CheekyBastard55@reddit | LocalLLaMA | View on Reddit | 26 comments
-
I know it's not for everyone, but I think this model needs more spotlight.
Posted by cgs019283@reddit | LocalLLaMA | View on Reddit | 14 comments
-
Rig confirmation
Posted by Old_Scratch3771@reddit | LocalLLaMA | View on Reddit | 4 comments
-
Is Claude from Anthropic the best AI Code Assist in the market?
Posted by AMGraduate564@reddit | LocalLLaMA | View on Reddit | 70 comments
-
Reflection 70b
Posted by novexion@reddit | LocalLLaMA | View on Reddit | 6 comments
-
Why did Meta AI give a $100 million model for free?
Posted by Worth-Card9034@reddit | LocalLLaMA | View on Reddit | 274 comments
-
LoLCATS - a hazyresearch Collection (of Linearized Llama 3.1 models 8B, 70B, and405B)
Posted by DeepWisdomGuy@reddit | LocalLLaMA | View on Reddit | 14 comments
-
Playing AI-Generated CS:GO on a Single RTX 3090 in real time
Posted by Icy-Corgi4757@reddit | LocalLLaMA | View on Reddit | 86 comments
-
Chain-of-Thought Reasoning Without Prompting [paper by Google]
Posted by DreamGenAI@reddit | LocalLLaMA | View on Reddit | 71 comments
-
Best TTS you tried?
Posted by -Krag@reddit | LocalLLaMA | View on Reddit | 15 comments
-
Supermicro H12 + AMD EPYC + 4x 3090 build: Cooling concerns and PCIe 4.0 cable worries
Posted by Super_Spot3712@reddit | LocalLLaMA | View on Reddit | 1 comments
-
Triton for Windows (unofficial, tested to work in ComfyUI)
Posted by rerri@reddit | LocalLLaMA | View on Reddit | 13 comments
-
Windows program for RAG using local pdf files, magazines, and technical documents using gguf models
Posted by cmdrmcgarrett@reddit | LocalLLaMA | View on Reddit | 5 comments
-
Fine-tuning with small batch sizes and gradient accumulation poorly perform if you use Transformers (TRL)!
Posted by TheKaitchup@reddit | LocalLLaMA | View on Reddit | 36 comments
-
LLMs as a way to browse the web
Posted by mayodoctur@reddit | LocalLLaMA | View on Reddit | 16 comments
-
I built something and I think it's awesome.
Posted by superabhidash@reddit | LocalLLaMA | View on Reddit | 6 comments
-
Handwritten formula to latex format
Posted by DataScientia@reddit | LocalLLaMA | View on Reddit | 10 comments
-
Self-hosted LLM for scientific papers
Posted by Lost_Albatross_5673@reddit | LocalLLaMA | View on Reddit | 1 comments
-
Built my first AI + Video processing Workstation - 3x 4090
Posted by Special-Wolverine@reddit | LocalLLaMA | View on Reddit | 221 comments
-
Serving 3.3 Million Context for Llama-3-8B on a single GPU
Posted by Van_Chopiszt@reddit | LocalLLaMA | View on Reddit | 10 comments
-
How to import gguf models to ollama ? Begginer question
Posted by kleju_@reddit | LocalLLaMA | View on Reddit | 1 comments
-
Has anyone seen AI agents working in production at scale?
Posted by madredditscientist@reddit | LocalLLaMA | View on Reddit | 68 comments
-
ARM or AMD cpus for LLM / AI on Linux
Posted by Elegant_Fold_7809@reddit | LocalLLaMA | View on Reddit | 11 comments
-
Gemma2 9B is certainly something o_O
Posted by LMTMFA@reddit | LocalLLaMA | View on Reddit | 2 comments
-
Llama-3 is actually an Octopus. Which animal do LLMs identify with?
Posted by cpldcpu@reddit | LocalLLaMA | View on Reddit | 1 comments
-
How to make custom tokenizers?
Posted by RandiyOrtonu@reddit | LocalLLaMA | View on Reddit | 0 comments
-
DeepSeek silently released their DeepSeek-Coder-V2-Instruct-0724, which ranks #2 on Aider LLM Leaderboard, and it beats DeepSeek V2.5 according to the leaderboard
Posted by XMasterrrr@reddit | LocalLLaMA | View on Reddit | 41 comments
-
Hardware costs to run 90B llama at home?
Posted by Sea-Replacement7541@reddit | LocalLLaMA | View on Reddit | 144 comments
-
Technical question can you mask/hide parts of the KV cash for a request.
Posted by Noxusequal@reddit | LocalLLaMA | View on Reddit | 3 comments
-
LLM observability
Posted by papipapi419@reddit | LocalLLaMA | View on Reddit | 12 comments
-
Higher capacity regular DDR5 timeline? 64GBx2 96GBx2?
Posted by capybooya@reddit | LocalLLaMA | View on Reddit | 26 comments
-
Tell the LLM to "Repeat the Question". An Interesting Observation.
Posted by taste_my_bun@reddit | LocalLLaMA | View on Reddit | 87 comments
-
Llama Image Tagger: A project I made to help me sort thousands of images
Posted by Eisenstein@reddit | LocalLLaMA | View on Reddit | 18 comments
-
Special German Question TTS
Posted by Fun_Librarian_7699@reddit | LocalLLaMA | View on Reddit | 1 comments
-
Any such thing as a pre-setup physical AI server you can buy (for consumers)?
Posted by meeplemop159@reddit | LocalLLaMA | View on Reddit | 50 comments
-
Llama 3.1 + Flux + Hailuo AI
Posted by mso96@reddit | LocalLLaMA | View on Reddit | 8 comments
-
Best LLM for OCR invoices and make JSON and calculate values?
Posted by Difficult-Bluejay-52@reddit | LocalLLaMA | View on Reddit | 5 comments
-
Local TTS LLM in your browser?
Posted by DinUXasourus@reddit | LocalLLaMA | View on Reddit | 6 comments
-
Chain-of-Thought reasoning on the next token prediction level?
Posted by Marha01@reddit | LocalLLaMA | View on Reddit | 2 comments