We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add the new Multi-Modal model of mistral AI: pixtral-12b:
https://huggingface.co/mistral-community/pixtral-12b-240910
It supports image encoder, can it also be added to the image generator API as an alternative to Stable Diffusion?
The text was updated successfully, but these errors were encountered:
Since yesterday vllm has internVL2 support. :-)
vllm-project/vllm/releases/tag/v0.6.1
Sorry, something went wrong.
I guess that would work already with llama.cpp GGUF models if/when is getting supported in there ( see also ggerganov/llama.cpp#9440 ).
I'd change the focus of this one to be more generic and add support for multimodal with vLLM, examples:
https://github.com/vllm-project/vllm/blob/main/examples/offline_inference_pixtral.py https://github.com/vllm-project/vllm/blob/main/examples/offline_inference_vision_language_multi_image.py
No branches or pull requests
Add the new Multi-Modal model of mistral AI: pixtral-12b:
https://huggingface.co/mistral-community/pixtral-12b-240910
It supports image encoder, can it also be added to the image generator API as an alternative to Stable Diffusion?
The text was updated successfully, but these errors were encountered: