Replies: 3 comments 1 reply
-
we meet the same promblem |
Beta Was this translation helpful? Give feedback.
0 replies
-
Same any news? :) |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I met this error during the call LLM with the Mixtral8x7B fine-tuned LoRA Merged Model. It worked well in torch 2.1.2 and vlllm 0.3.3 but torch 2,3 and vlllm 0.4.2 met this valueError. What is missing here? Should I use the different quantization method like awq or gptq than bitsandbytes?
Beta Was this translation helpful? Give feedback.
All reactions