-
Best LLM for coding
Posted by imedmactavish@reddit | LocalLLaMA | View on Reddit | 21 comments
-
New series of models for creative writing like no other RP models (3.8B, 8B, 12B, 70B) - ArliAI-RPMax-v1.1 Series
Posted by nero10579@reddit | LocalLLaMA | View on Reddit | 55 comments
-
Open Interpreter refunds all hardware orders for 01 Light AI device, makes it a phone app instead. App launches TODAY!
Posted by Porespellar@reddit | LocalLLaMA | View on Reddit | 3 comments
-
405B P40 performance?
Posted by My_Unbiased_Opinion@reddit | LocalLLaMA | View on Reddit | 18 comments
-
Have You Encountered Any Surprising Uses of AI in the Last Two Years?
Posted by moveitfast@reddit | LocalLLaMA | View on Reddit | 2 comments
-
CONFIRMED: REFLECTION 70B'S OFFICIAL API IS SONNET 3.5
Posted by TGSCrust@reddit | LocalLLaMA | View on Reddit | 310 comments
-
Need feedback on my AI project
Posted by Strangooo266@reddit | LocalLLaMA | View on Reddit | 5 comments
-
Battle of the cheap GPUs - Lllama 3.1 8B GGUF vs EXL2 on P102-100, M40, P100, CMP 100-210, Titan V
Posted by MachineZer0@reddit | LocalLLaMA | View on Reddit | 67 comments
-
Out of the loop on this whole "Reflection" thing? You're not alone. Here's the best summary I could come up.
Posted by Decaf_GT@reddit | LocalLLaMA | View on Reddit | 43 comments
-
I made an 822 page Google Doc with lots of arguments and citations to defend AI
Posted by Which-Tomato-8646@reddit | LocalLLaMA | View on Reddit | 3 comments
-
AMD announces unified UDNA GPU architecture — bringing RDNA and CDNA together to take on Nvidia's CUDA ecosystem
Posted by kkchangisin@reddit | LocalLLaMA | View on Reddit | 69 comments
-
Memory bandwidth values (STREAM TRIAD benchmark results) for most Epyc Genoa CPUs (single and dual configurations)
Posted by fairydreaming@reddit | LocalLLaMA | View on Reddit | 10 comments
-
Word Game Bench - evaluating language models on word puzzle games
Posted by starzmustdie@reddit | LocalLLaMA | View on Reddit | 3 comments
-
best 7b coding LLM? (java)
Posted by GreedyWorking1499@reddit | LocalLLaMA | View on Reddit | 12 comments
-
Deepsilicon runs neural nets with 5x less RAM and ~20x faster. They are building SW and custom silicon for it
Posted by hamada0001@reddit | LocalLLaMA | View on Reddit | 12 comments
-
40GB VRAM Powerhouse: Testing Performance with Dual GPUs (7900 XTX & 7800 XT) and Ryzen 7 7700X - Need Benchmark Ideas!
Posted by djdeniro@reddit | LocalLLaMA | View on Reddit | 21 comments
-
Which models for summarising texts
Posted by mrwrob@reddit | LocalLLaMA | View on Reddit | 1 comments
-
DeepSeek silently released their DeepSeek-Coder-V2-Instruct-0724, which ranks #2 on Aider LLM Leaderboard, and it beats DeepSeek V2.5 according to the leaderboard
Posted by XMasterrrr@reddit | LocalLLaMA | View on Reddit | 26 comments
-
All of this drama has diverted our attention from a truly important open weights release: DeepSeek-V2.5
Posted by paf1138@reddit | LocalLLaMA | View on Reddit | 135 comments
-
Can I do this?
Posted by Fit-Construction-869@reddit | LocalLLaMA | View on Reddit | 19 comments
-
Poor results mistery solved. Reflection 70B was infected by COVID.
Posted by Wrong_User_Logged@reddit | LocalLLaMA | View on Reddit | 91 comments
-
Mini-Omni: Language Models Can Hear, Talk While Thinking in Streaming (finetuned Qwen2-0.5B)
Posted by Balance-@reddit | LocalLLaMA | View on Reddit | 19 comments
-
Multi-GPU Utilization issues, recompiling Llama.cpp and benchmarks
Posted by desexmachina@reddit | LocalLLaMA | View on Reddit | 7 comments
-
Current best ~70B RP model?
Posted by Master-Meal-77@reddit | LocalLLaMA | View on Reddit | 26 comments
-
Neo-AI assistant, can interact directly with Linux.
Posted by TomatoWasabi@reddit | LocalLLaMA | View on Reddit | 23 comments
-
Looking to build a workstation for AI interference on LLaMa 70b have 2x4090
Posted by DangerousPathh@reddit | LocalLLaMA | View on Reddit | 28 comments
-
Fashionistas are coming for your GPUs!
Posted by DeltaSqueezer@reddit | LocalLLaMA | View on Reddit | 2 comments
-
KoboldCpp v1.74 - adds XTC (Exclude Top Choices) sampler for creative writing
Posted by HadesThrowaway@reddit | LocalLLaMA | View on Reddit | 63 comments
-
Smh: Reflection was too good to be true - reference article
Posted by notsoserious408@reddit | LocalLLaMA | View on Reddit | 16 comments
-
Looking for a llm which is fine tuned specifically for Java
Posted by anonymous_abc99@reddit | LocalLLaMA | View on Reddit | 2 comments
-
For Gemma-2-2b, You can Extend the Windows Context to 32K+
Posted by Iory1998@reddit | LocalLLaMA | View on Reddit | 13 comments
-
Drummer's Theia 21B v2 - Rocinante's big sister! An upscaled NeMo finetune with a focus on RP and storytelling.
Posted by TheLocalDrummer@reddit | LocalLLaMA | View on Reddit | 20 comments
-
Extraordinary claims require extraordinary evidence, something Reflection 70B clearly lacks
Posted by stephen-leo@reddit | LocalLLaMA | View on Reddit | 38 comments
-
To RAG or not to RAG
Posted by madway99@reddit | LocalLLaMA | View on Reddit | 2 comments
-
llama 3.1 doesn't let me push code to git
Posted by Matthew_heartful@reddit | LocalLLaMA | View on Reddit | 2 comments
-
LM Studio alternatives?
Posted by stfz@reddit | LocalLLaMA | View on Reddit | 46 comments
-
🐺🐦⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with **17** different instruct templates
Posted by WolframRavenwolf@reddit | LocalLLaMA | View on Reddit | 39 comments
-
For what purpose do you use local LLMs?
Posted by mrscript_lt@reddit | LocalLLaMA | View on Reddit | 11 comments
-
Need help on solving a life or death vision LLM problem for a nonprofit
Posted by poli-cya@reddit | LocalLLaMA | View on Reddit | 17 comments
-
POV : The anthropic employee under NDA that see all the API requests from a guy called « matt.schumer.freeaccounttrial27 »
Posted by LinkSea8324@reddit | LocalLLaMA | View on Reddit | 23 comments
-
Reliable ways to get structured output from llms
Posted by amit13k@reddit | LocalLLaMA | View on Reddit | 6 comments
-
Not impressed with Reflection Llama 3.1 70B IQ2_S quant.
Posted by LaughterOnWater@reddit | LocalLLaMA | View on Reddit | 8 comments
-
Msty - Free Local + Remote AI Chat App (w/ support for Ollama/HF) has just hit its 1.0 release!
Posted by Decaf_GT@reddit | LocalLLaMA | View on Reddit | 42 comments
-
Idea: create a test that specifically knows of all the erratas in all the tests
Posted by uniformly@reddit | LocalLLaMA | View on Reddit | 0 comments
-
Looking for Llama.cpp Alternative to Run Recent Vision Language Models on Apple Silicon
Posted by chibop1@reddit | LocalLLaMA | View on Reddit | 8 comments
-
AMD probably killing top consumer GPUs for 8000 series (and is short sighted if so)
Posted by 1ncehost@reddit | LocalLLaMA | View on Reddit | 61 comments
-
Im really confused right now...
Posted by noblex33@reddit | LocalLLaMA | View on Reddit | 62 comments
-
Announcing: Magnum 123B
Posted by lucyknada@reddit | LocalLLaMA | View on Reddit | 83 comments
-
CUDA on WSL 2?
Posted by lehsunMartins@reddit | LocalLLaMA | View on Reddit | 11 comments
-
Windows or Linux for LLMs and AI?
Posted by RelationshipNeat6468@reddit | LocalLLaMA | View on Reddit | 151 comments