-
Using LlaMa to analyze scientific texts - I am failing
Posted by RollLikeRick@reddit | LocalLLaMA | View on Reddit | 7 comments
-
DeepSeek-R1-Lite Preview Version Officially Released
Posted by nekofneko@reddit | LocalLLaMA | View on Reddit | 104 comments
-
M4 Max 128GB running Qwen 72B Q4 MLX at 11tokens/second.
Posted by tony__Y@reddit | LocalLLaMA | View on Reddit | 22 comments
-
Opencanvas - An open source alternative to OpenAI's canvas
Posted by umarmnaq@reddit | LocalLLaMA | View on Reddit | 6 comments
-
Mac Users: New Mistral Large MLX Quants for Apple Silicon (MLX)
Posted by thezachlandes@reddit | LocalLLaMA | View on Reddit | 12 comments
-
CrisperWhisper ranks #2 on Open ASR Leaderboard
Posted by vaibhavs10@reddit | LocalLLaMA | View on Reddit | 10 comments
-
Is the 4070 Ti Super Worth It Over the 4070 Super?
Posted by Paulus_Audet@reddit | LocalLLaMA | View on Reddit | 7 comments
-
Curb Your Inference: AICI for rewriting context in real time, constrained generation, backtracking KV-cache
Posted by tucnak@reddit | LocalLLaMA | View on Reddit | 13 comments
-
LLM Visualization
Posted by mcnuuk@reddit | LocalLLaMA | View on Reddit | 3 comments
-
Here the R1-Lite-Preview from DeepSeek AI showed its power... WTF!! This is amazing!!
Posted by Inspireyd@reddit | LocalLLaMA | View on Reddit | 9 comments
-
I Created an AI Research Assistant that actually DOES research! Feed it ANY topic, it searches the web, scrapes content, saves sources, and gives you a full research document + summary. Uses Ollama (FREE) - Just ask a question and let it work! No API costs, open source, runs locally!
Posted by CuriousAustralianBoy@reddit | LocalLLaMA | View on Reddit | 129 comments
-
Best NSFW model?
Posted by poet3991@reddit | LocalLLaMA | View on Reddit | 10 comments
-
Deepseek announces R1 lite , beats o1 preview on some benchmarks
Posted by TheLogiqueViper@reddit | LocalLLaMA | View on Reddit | 62 comments
-
Is the 4070 Ti Super Worth It Over the 4070 Super?
Posted by Paulus_Audet@reddit | LocalLLaMA | View on Reddit | 2 comments
-
New way to make LLM's Safer: Enhancing LLM Safety with Precision Knowledge Editing (PKE)
Posted by lial4415@reddit | LocalLLaMA | View on Reddit | 1 comments
-
What is currently a good and easy way to run local llms against an entire code base?
Posted by Particular_Paper7789@reddit | LocalLLaMA | View on Reddit | 2 comments
-
Manhattan style project race to AGI recommended to Congress by U.S congressional commission
Posted by Status-Beginning9804@reddit | LocalLLaMA | View on Reddit | 117 comments
-
NaturalLM - A Natural Sounding LLM
Posted by random-tomato@reddit | LocalLLaMA | View on Reddit | 7 comments
-
Notes on Chatgpt Search: Better than Perplexity?
Posted by SunilKumarDash@reddit | LocalLLaMA | View on Reddit | 30 comments
-
AI File Organizer Update: Now with Dry Run Mode and Llama 3.2 as Default Model
Posted by unseenmarscai@reddit | LocalLLaMA | View on Reddit | 50 comments
-
why is o1 not ranked #1 on lmsys
Posted by Ok-Engineering5104@reddit | LocalLLaMA | View on Reddit | 4 comments
-
Best Use Cases for Small LLMs
Posted by XhoniShollaj@reddit | LocalLLaMA | View on Reddit | 25 comments
-
Managed to get r1-lite to think for 207 seconds.
Posted by laser_man6@reddit | LocalLLaMA | View on Reddit | 5 comments
-
CogVideoX 5B - Open weights Text to Video AI model (less than 10GB VRAM to run) | Tsinghua KEG (THUDM)
Posted by Nunki08@reddit | LocalLLaMA | View on Reddit | 83 comments
-
Request: Someone with an M4 Macbook Pro Max 64GB
Posted by NEEDMOREVRAM@reddit | LocalLLaMA | View on Reddit | 18 comments
-
My open-source and free Apple Intelligence Writing Tools for Windows app now has instant website summaries! You can use almost any local LLM with it, with its OpenAI-compatible API support :D
Posted by TechExpert2910@reddit | LocalLLaMA | View on Reddit | 8 comments
-
How far would AMD Threadripper 3600 (24 core, 48 threads) and 256 GB of memory get me for running local LLMs?
Posted by x3derr8orig@reddit | LocalLLaMA | View on Reddit | 65 comments
-
Nvidia's LLaMA-Mesh weights have been released on HF. Inference code is available.
Posted by FullOf_Bad_Ideas@reddit | LocalLLaMA | View on Reddit | 31 comments
-
Old server, new life?
Posted by Far_Let_5678@reddit | LocalLLaMA | View on Reddit | 0 comments
-
AMD Strix Halo rumored to have APU with 7600 XT performance & 96 GB of shared VRAM
Posted by 1ncehost@reddit | LocalLLaMA | View on Reddit | 60 comments
-
Implementing reasoning in LLMs through Neural Cellular Automata (NCA) ? (imagining each pixel/cell as a 256-float embedded token)
Posted by ryunuck@reddit | LocalLLaMA | View on Reddit | 13 comments
-
Is this setup good enough to run LLaMA 70B at 8-bit quantization?
Posted by matt23458798@reddit | LocalLLaMA | View on Reddit | 6 comments
-
NTK-Aware Scaled RoPE allows LLaMA models to have extended (8k+) context size without any fine-tuning and minimal perplexity degradation.
Posted by bloc97@reddit | LocalLLaMA | View on Reddit | 8 comments
-
Another how much ram should i get on an M4 Max MBP thread
Posted by flying_unicorn@reddit | LocalLLaMA | View on Reddit | 42 comments
-
How can I train a model on both text and numeric data?
Posted by boringblobking@reddit | LocalLLaMA | View on Reddit | 3 comments
-
Best local LLM Setup for IntelliJ / coding assistance?
Posted by Lost_Fox__@reddit | LocalLLaMA | View on Reddit | 6 comments
-
Is this how chain of thought model works? đź’€
Posted by vinam_7@reddit | LocalLLaMA | View on Reddit | 46 comments
-
OpenAI API Compatible access to Google Gemini and Embedding Models
Posted by theeashman@reddit | LocalLLaMA | View on Reddit | 12 comments
-
Wanting to try out L3-70b-euryale on my computer, but don't know what version to choose
Posted by EEEEEEEEEEEEEEEE_Man@reddit | LocalLLaMA | View on Reddit | 1 comments
-
LLM hardware acceleration—on a Raspberry Pi (Top-end AMD GPU using a low cost Pi as it's base computer)
Posted by Colecoman1982@reddit | LocalLLaMA | View on Reddit | 25 comments
-
Something weird is happening with LLMs and chess
Posted by paranoidray@reddit | LocalLLaMA | View on Reddit | 80 comments
-
Does 2x Dual-Channel improve performance on models?
Posted by Fusseldieb@reddit | LocalLLaMA | View on Reddit | 20 comments
-
Tool for web scraping with LLMs?
Posted by arbayi@reddit | LocalLLaMA | View on Reddit | 9 comments
-
Vision (for bank account statements): is it better to OCR an account statement and have the LLM analyze markdown/json to get the info you need OR have the vision model extract the info you need?
Posted by dirtyring@reddit | LocalLLaMA | View on Reddit | 2 comments
-
Has anyone tested GenAI vs Advanced NER?
Posted by dalhaze@reddit | LocalLLaMA | View on Reddit | 9 comments
-
This seems pretty hype...
Posted by clduab11@reddit | LocalLLaMA | View on Reddit | 21 comments
-
Can anyone share their qwen 2.5 setup for a 4090 please?
Posted by firemeaway@reddit | LocalLLaMA | View on Reddit | 35 comments
-
Getting the info you need from bank account statement PDFs: what's the best way?
Posted by dirtyring@reddit | LocalLLaMA | View on Reddit | 0 comments
-
Can somebody please do a benchmark of the new Mistral-Large-2411?
Posted by CH1997H@reddit | LocalLLaMA | View on Reddit | 20 comments
-
Gemma2 2B IT is the most impressive small model I ever seen.
Posted by Discordpeople@reddit | LocalLLaMA | View on Reddit | 157 comments