Hosting a private LLM for a client. Does this setup make sense?

Posted by nullReferenceError@reddit | LocalLLaMA | View on Reddit | 32 comments

I’m working with a client who wants to use AI to analyze sensitive business data, so public LLMs like OpenAI or Anthropic are off the table due to privacy concerns. I’ve used AI in projects before, but this is my first time hosting an LLM myself.

The initial use case is pretty straightforward: they want to upload CSVs and have the AI analyze the data. In the future, they may want to fine-tune a model on their own datasets.

Here’s my current plan. Would love any feedback or gotchas I might be missing:

Eventually I’ll build out a backend to handle CSV uploads and prompt construction, but for now I’m just aiming to get the chat UI talking to the model.

Anyone done something similar or have tips on optimizing this setup?