-
Notifications
You must be signed in to change notification settings - Fork 221
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
suggest adding support for the local large model Qwen. #137
Comments
In fact, you can initialize the client directly for the local large model |
Running a local model with vllm is outside the scope of what I can support with this repo since it isn’t a workflow I do myself very much. One suggestion would be to try running it on FastChat instead of vllm directly? I believe it’s a bit simpler and maybe easier to integrate. I would be open to add some simple instructions on how to run that and very minimal modifications to the repo code to support the use case. Let me know what you think. |
Thank you for your quick response! I know how to use vllm for deployment. The issue I am encountering is an error when using research assistant agent . By calling the Qwen local model to build the agent, which likely requires modifications in the tools section. I hope it can support calling the lcoal Qwen large model to implement the function call feature. |
Did you try running vllm with the The tool calling is pretty fundamental to the research assistant and a fairly commodity thing at this point, so I don’t think I’ll be making any changes to make it work with an LLM API that doesn’t support it. I would guess the other built in agents or any custom agent you build without tool calling on the repo would work fine. |
Hopefully it can be included in this project.
The text was updated successfully, but these errors were encountered: