How Fast can I run models.

Posted by feelin-lonely-1254@reddit | LocalLLaMA | View on Reddit | 3 comments

I'm running image processing with gemma 3 27b and getting structured outputs as response, but my present pipeline is awfully slow (I use huggingface for the most part and lmformatenforcer), it processes a batch of 32 images in 5-10 minutes when I get a response of atmax 256 tokens per image. Now this is running on 4 A100 40 gig chips.

This seems awfully slow and suboptimal. Can people share some codebooks and benchmark times for image processing, and should I shift to sglang? I cannot use the latest version of VLLM in my uni's compute cluster.