-
-
Notifications
You must be signed in to change notification settings - Fork 6.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[New Model]: Request for LLaVA-Video-7B-Qwen2 Model Implementation #13190
Comments
@Noctis-SC Seems that this model has same architecture with llava-onevision, you can convert this model to hf format llava-onevision model using this modified script: convert_llava_onevision_weights_to_hf.py I will try to open a PR in |
Hey, thank you for your reply. I've tried to follow your script to convert the model. It stuck on the "Single forward pass" step for a couple of hours is it how it is supposed to be?
|
@Noctis-SC I have converted and pushed it to hf_hub. Can you check this model repo? (https://huggingface.co/Isotr0py/LLaVA-Video-7B-Qwen2-hf) |
Thank you. I run this model and it seems it works fine with vllm. These days I will test this model. But for now there is no questions from my side. Thanks again for quick responses. |
The model to consider.
Hello, I would like to ask for this model (LLaVA-Video-7B-Qwen2) conversion to hf-weights.
I’ve been testing llava-hf/LLaVA-NeXT-Video-7B-32K-hf (vicuna base) and LLaVA-Video-7B-Qwen2 (qwen2 base), and I’ve noticed that LLaVA-Video-7B-Qwen2 LLM outperforms LLaVA-NeXT-Video-7B-32K-hf in terms of video understanding (video description). Given the performance difference, I would like to request an implementation of a new model LLaVA-Video-7B-Qwen2.
Thanks!
The closest model vllm already supports.
No response
What's your difficulty of supporting the model you want?
No response
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: