What kind of model or harness would be the best for teaching stuff to you from documents
Posted by Trovebloxian@reddit | LocalLLaMA | View on Reddit | 10 comments
Going through university right now, and we have massive 100 page pdfs/ppts with soo much fluff that its annoying to go through. until now ive been using chatgpt for it, but realized that the output tokens are HEAVILY limited, and loses a LOT of information. rightnow im just using the 35b model locally and the qwen3.5plus model for larger docs. what can i do to make this more accurate/detailed, ie better. (telling it to be more detailed and not skip over anything didnt help xD)
Ok-Ad-8976@reddit
notebook lm
Trovebloxian@reddit (OP)
Local alternative? I have too much stuff to parse
dodox4@reddit
The information loss is a context window problem; once the doc exceeds what the model can hold at once it starts dropping or summarising. Qwen 3.5 Plus has a large context so that's a reasonable call for bigger docs. For the PPTs specifically, converting to plain text first (strip formatting, remove slide structure) before feeding helps a lot; models handle dense structured slides poorly.
If any of your course material comes as actual textbooks in EPUB format rather than lecture slides, calice.app might be worth a look. We don't do RAG; under the hood it's a heterogeneous mixture-of-experts that routes between retrieval specialists (near-perfect recall) and reasoning specialists (for sota reasoning), so you get both without naive full-context inference costs. LLM sits split-screen alongside the reader and highlights the exact passages it references. Not self-hosted, and won't help with PPTs or diagram-heavy PDFs, but for dense text-based books it works well.
Trovebloxian@reddit (OP)
Its not even i formation loss like that, it just doesnt go into detail, like when i ask it to do 1 topic per prompt it does it fine, they just heavily limit output tokens per message
nothrowaway@reddit
what exactly is your prompt, is it very specific or generic? You need to break down the learning material just as if you were studying it manually. "1 topic per prompt" is very generic. "Summarize chapter 1" will be poor output. Ask about key topics, high yield facts & concepts, testable facts, common pitfalls, etc. If you have access to old test question, you could see what topics are typically emphasized as well. Try asking ChatGPT to generate a prompt for you on how to ask the question.
Trovebloxian@reddit (OP)
I ask it to provide me a neatly organized version, highly detailed, without skipping any of the material. Format it for bionic reading.
This but more detailed prompt i even give it a topic list
dodox4@reddit
The short answers are a ChatGPT specific problem right? OpenAI discourages long outputs during RL finetuning to lower inference cost. For local Qwen, not sure there is a fix, they just don't match sota performance. If open models don't do the trick, Claude (Anthropic) handles PPTs natively and is less aggressive about output length. No other ideas for local models though...
Trovebloxian@reddit (OP)
currently ive been using qwen cloud and its working pretty good, now im after the QoL features, like showing graphs, flowcharts, better explaination
wombweed@reddit
I use ai to learn new things, personally I like notebooklm for generating quizzes, but not aware of any self hosted alternatives for that specific task. Quick question though. Forgive me if this is naive, I didn’t go to college so maybe I am missing something. But like, isn’t this exactly what you’re already paying for as a student? Aren’t the lectures basically the whole point, helping you retain the course material instead through interactive learning instead of just reading the entire textbook?
Trovebloxian@reddit (OP)
Yes but where i am from (india) they make us take irrelevant stuff like electrical engineering when my focus is on software engineering. And honestly most of the lectue PPTs and PDFs they provide is formated like trash