llama 3.1 70B is absolutely awful at tool usage

Posted by fireKido@reddit | LocalLLaMA | View on Reddit | 41 comments

Hi guys,

I am experimenting with a langgraph multi-agent model, and I tested it with GPT-4o, everything works well and the results are pretty impressive.

after that, I tested it with ollama and llama3.1:70b-instruct-q8_0, however, the results are absolutely disappointing, it's not capable of correctly structuring a single tool call, ignoring completely the info I give them, forgetting parameters for the function calls, and other similar silly mistakes

my question is, is this your experience as well? I'm afraid I am doing something wrong because generally, I read positive stuff about llama3...