Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

suggest adding support for the local large model Qwen. #137

Open
lixy0420 opened this issue Jan 16, 2025 · 5 comments
Open

suggest adding support for the local large model Qwen. #137

lixy0420 opened this issue Jan 16, 2025 · 5 comments

Comments

@lixy0420
Copy link

Hopefully it can be included in this project.

@reatang
Copy link
Contributor

reatang commented Jan 16, 2025

In fact, you can initialize the client directly for the local large model

@lixy0420
Copy link
Author

In fact, you can initialize the client directly for the local large model

The locally deployed Qwen model with vllm will throw an error if used directly, as it does not support bind_tools, and there are still some details that need to be modified.
eb073b0e-1da9-4cf0-af98-b50043d51333

@JoshuaC215
Copy link
Owner

Running a local model with vllm is outside the scope of what I can support with this repo since it isn’t a workflow I do myself very much.

One suggestion would be to try running it on FastChat instead of vllm directly? I believe it’s a bit simpler and maybe easier to integrate.

I would be open to add some simple instructions on how to run that and very minimal modifications to the repo code to support the use case.

Let me know what you think.

@lixy0420
Copy link
Author

Thank you for your quick response! I know how to use vllm for deployment. The issue I am encountering is an error when using research assistant agent . By calling the Qwen local model to build the agent, which likely requires modifications in the tools section. I hope it can support calling the lcoal Qwen large model to implement the function call feature.

@JoshuaC215
Copy link
Owner

Did you try running vllm with the —enable-auto-tool-choice and —tool-call-parser flags mentioned in the error message you posted?

The tool calling is pretty fundamental to the research assistant and a fairly commodity thing at this point, so I don’t think I’ll be making any changes to make it work with an LLM API that doesn’t support it. I would guess the other built in agents or any custom agent you build without tool calling on the repo would work fine.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants