-
-
Notifications
You must be signed in to change notification settings - Fork 6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
support download Lora Model from ModelScope and download private mode… #10762
base: main
Are you sure you want to change the base?
Conversation
…l from ModelScope
👋 Hi! Thank you for contributing to the vLLM project. Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can do one of these:
🚀 |
/ready |
IMHO, downloading LoRA weights from remote servers should not be handled internally by vllm, as this would increase the risk of service crashes and impact performance. What are your thoughts on this? cc @simon-mo @youkaichao @DarkLight1337 |
Indeed, if the download path is incorrect or some internet errors occured, it will crash the server(eg. the lora from the internet is lack of some files). However, we can add additional try-except or add download instruction when starting vllm or interacting with /upload_lora . Actually, vllm now support downloading LoRA weight from huggingface, and it doesn't matter. |
@AlphaINF I know, so I want to hear their thoughts and suggestions. |
Not sure about dynamic LoRA adapters (I don't have context on this), but since we currently support downloading models from both HF and ModelScope, it makes sense to enable this for LoRA as well. |
@jeejeelee Through some days trial, your concern is right! |
Have you tried using filelock? |
That's a great idea, I will try to add it and test for some days! |
Currently, vllm can't loading Lora models from modelscope.
Beside, If a model from modelscope is private, it can't load from modelscope, have to login manually before starting vllm.
this pull request fix two problems below.
If you want to using private models, you can set MODELSCOPE_ACCESS_TOKEN as environmental variables