We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
根据教程在window11 RTX 3060和macos M1下均无法实现本地部署,请问有更详细的本地部署教程吗?
The text was updated successfully, but these errors were encountered:
+1, vllm在windows系统下会报错,还有没有别的替代方案?
Sorry, something went wrong.
我直接 clone vllm 项目,然后起 vllm server 是运行起来了,无奈 M1 Pro 推理太慢了...
#64
我们实际发现ollama运行qwen2-vl模型会有明显的模型性能损失,而且ollama对bf16的支持也有限;vLLM的话可能更推荐使用CUDA+Linux的环境,如果有对应的推理问题,也可以去ollama和vLLM的仓库提issues。
No branches or pull requests
根据教程在window11 RTX 3060和macos M1下均无法实现本地部署,请问有更详细的本地部署教程吗?
The text was updated successfully, but these errors were encountered: