Tool calling with llm.chat #12557
Closed
alexanderbrodko
announced in
Q&A
Replies: 1 comment
-
My bad. I do not need tokenize when I use
In fact, the model is Qwen2.5-Coder-Instruct-0.5B |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
According to this example
https://github.com/vllm-project/vllm/blob/27b78c73cad00f5c7bb3b2431f02dc680f7034bc/examples/offline_inference/chat_with_tools.py
I create some model:
Do inference:
Then I ask the model about temperature in San Francisco and it fails.
Any suggestions?
Beta Was this translation helpful? Give feedback.
All reactions