You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
You can also load the model from a local path by passing a file path in modelURL in your values.yaml.
More specifically, the helm chart will try mounting a PersistentVolume to /data in the vLLM pods. Therefore, if you have your model stored in a persistent volume (for example, in the folder my_model/), you can let vLLM load it by setting modelURL: "/data/my_model" in your values.yaml.
Currently, we haven't supported the S3 folders yet. Adding an init container to download from S3 to local persistent volume when launching the vLLM pods should be a solution.
Can you create a separate issue to request this feature?
What is the process for fine-tuned models?
Just like I mentioned above, you can create a persistent volume that contains your fine-tuned models.
I see that we need hugging face token to download the model
The text was updated successfully, but these errors were encountered: