-
-
Notifications
You must be signed in to change notification settings - Fork 5.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: After successfully loading the LoRA module with load_lora_adapter, the result returned by v1/models does not include this LoRA module. #11761
Comments
Maybe you forget export the related env var , see:https://docs.vllm.ai/en/latest/usage/lora.html#dynamically-serving-lora-adapters |
I have set this environment variable; if it is not set, the LoRA module will not be successfully added. |
run: {
"object": "list",
"data": [
{
"id": "b",
"object": "model",
"created": 1736177354,
"owned_by": "vllm",
"root": "/llm_models/BaseModel/llama/Llama-2-7b-chat-hf",
"parent": null,
"max_model_len": 4096,
"permission": [
{
"id": "modelperm-e126eb667b80402d8acd7542ecd1ab0d",
"object": "model_permission",
"created": 1736177354,
"allow_create_engine": false,
"allow_sampling": true,
"allow_logprobs": true,
"allow_search_indices": false,
"allow_view": true,
"allow_fine_tuning": false,
"organization": "*",
"group": null,
"is_blocking": false
}
]
},
{
"id": "ase",
"object": "model",
"created": 1736177354,
"owned_by": "vllm",
"root": "/llm_models/BaseModel/llama/Llama-2-7b-chat-hf",
"parent": null,
"max_model_len": 4096,
"permission": [
{
"id": "modelperm-e9689675ae254695bd35e8b88abc25f9",
"object": "model_permission",
"created": 1736177354,
"allow_create_engine": false,
"allow_sampling": true,
"allow_logprobs": true,
"allow_search_indices": false,
"allow_view": true,
"allow_fine_tuning": false,
"organization": "*",
"group": null,
"is_blocking": false
}
]
},
{
"id": "sql-lora",
"object": "model",
"created": 1736177354,
"owned_by": "vllm",
"root": "/lora_model/llama-2-7b-sql-lora-test-yard1",
"parent": "b",
"max_model_len": null,
"permission": [
{
"id": "modelperm-93e1bf8266a345e9b2eadcf173c05284",
"object": "model_permission",
"created": 1736177354,
"allow_create_engine": false,
"allow_sampling": true,
"allow_logprobs": true,
"allow_search_indices": false,
"allow_view": true,
"allow_fine_tuning": false,
"organization": "*",
"group": null,
"is_blocking": false
}
]
}
]
} |
jq is used to format the output JSON, but the result I'm getting still doesn't include the added LoRA module. |
I mean I haven't been able to reproduce your issue. If your output differs from mine, you might need to upgrade to the latest version of vLLM. |
Which version are you using? I am currently using version 0.6.6. |
I was built based on the latest main branch. IIUC, This issue should be resolved by #11094 |
I am getting this issue as well. I am using the docker image. Here's my docker compose setup:
The meta-llama/Meta-Llama-3-8B-Instruct appears in the model list but never any LORAs I add. |
Your current environment
Model Input Dumps
No response
🐛 Describe the bug
After successfully loading the LoRA module with load_lora_adapter, the result returned by v1/models does not include this LoRA module.
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: