You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Jan 4, 2026. It is now read-only.
Hello, I want to run Qwen/Qwen2-VL-72B-Instruct-AWQ on my local computer, currently, I have 2xrtx 3090 but it has trouble OOM. Then I see in your vision.py 's options there --max-memory option to offload on CPU. Can you please implement it also for Qwen/Qwen2-VL-72B-Instruct-AWQ