Mounting s3 hosted model files using s3fs is causing startup issues #765
Labels
area/lora
kind/bug
Something isn't working
triage/accepted
Indicates an issue or PR is ready to be actively worked on.
I wanted to download models from self managed s3 bucket. I got it working by downloading the files to local storage. But somehow referencing the models using their s3 path is making the vllm crash cause it interprets the s3 link as a huggingface model repo (https://aibrix.readthedocs.io/latest/features/lora-dynamic-loading.html#model-registry).
No problem I thought I can just mount the s3 bucket using s3fs. Which I can I do no problem but then when the vllm engine is trying to start it tries to pull in the model weights locally but they are not there yet cause they have to be pulled from s3 using s3fs which introduces latency over native storage. I added a stacktrace from inside my pod.
The text was updated successfully, but these errors were encountered: