-
-
Notifications
You must be signed in to change notification settings - Fork 5.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: vLLM LoRA Crash when using Dynamic Loading #11702
Comments
Is this a local path? |
@jeejeelee Yes, it's the file path in the Kubernetes Persistent Volume Claim (PVC) that we used to store the base model. |
Have you tried using absolute path? |
@jeejeelee Here is the path of our lora file, and it is a absolute path
|
@haitwang-cloud I'm guessing that the error logs you pasted are from the container rebooting after the lora adapter caused it to crash, and the real root cause of the crash is in the previous container's logs. You should be able to get those with the I've opened a PR that will attempt to load the lora adapters eagerly so that you get an error response back from |
@joerunde Really appreciate your comments & code fix for this issue, I will have a try after your fix. Also, I'm experiencing issues with the vLLM pod keep crashing in my case. Unfortunately, I'm unable to retrieve the previous container logs from terminal & Dynatrace, and I don't have enough time to capture the init logs.
|
Your current environment
Model Input Dumps
No response
🐛 Describe the bug
We encountered a 500 error while testing the dynamic loading of LoRA with vLLM.
Steps to Reproduce:
Load LoRA dynamically:
The LoRA was loaded correctly as confirmed:
Attempt to use the loaded LoRA:
Observed Error:
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: