I use LoRA adapter, but the following message appears: "llama_lora_adapter_set: flash_attn is not compatible with LoRA" #11097
Unanswered
WangHao311
asked this question in
Q&A
Replies: 1 comment 2 replies
-
I am not sure why FA is not allowed with LoRA. It seems to me that they are both compatible. cc @ngxson |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
When I use llama-server to start the service, I use the "-fa" parameter. When I start the log and call the interface "/v1/chat/completions", the prompt message "llama_lora_adapter_set: flash_attn is not compatible with LoRA appears". Is there a problem? How can I modify it?
Beta Was this translation helpful? Give feedback.
All reactions