Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Expected System Specifications and Hosting Qwen2-VL-7B-Instruct on EC2 #664

Open
shrinesabu2000 opened this issue Jan 22, 2025 · 0 comments

Comments

@shrinesabu2000
Copy link

I am planning to host the Qwen2-VL-7B-Instruct model as a server on an EC2 instance. I would like to know the recommended system specifications for running this model efficiently. Additionally, I am exploring whether using vLLMs is the best approach for deploying this model in a production environment. I plan to use this model primarily for OCR tasks and multimodal inference, processing both images and text in real-time.

Also, would you recommend vLLMs for deploying Qwen2-VL-7B-Instruct as a server for inference?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant