Self-hosted LLM for scientific papers

Posted by Lost_Albatross_5673@reddit | LocalLLaMA | View on Reddit | 2 comments

Hi everyone,

I am new to self-hosted LLMs but so far it's been an exciting journey. My main use case for LLMs is understanding and grabbing key conceptualisations from scientific papers. So far I've used and worked mostly with ChatGPTs 4o model. I have a specific prompt that gives a key summary of the main arguments, research design, supporting data and data analysis. It works really well with ChatGPTs 4o model, but when I give the same prompt to a self-hosted Gemma/Ilama 3.1, I end up with a very high level set of bullet points.

Any further exploratory work or questions are either met with high level answers or statements that the model cannot access the document.

I haven't trained the model, but I assumed that it was already trained on a set of data. Any advice on what I should do to improve the models performance? I am running the model on my MacBook using AnythingLLM. I tried Docker and can switch easily, but I am guessing the issue is that I haven't trained the model yet?