We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
MiniCPM-V 2.6跑demo,本身机器显存8G,使用cuda提示显存不够,目前使用cpu跑,贼慢十几个小时才出结果,请问如何类似于QW2-VL调整输入可以满足显存在8G以内的设备上跑。 接着又尝试了MiniCPM-V 2.6 int4也很慢。 目前尝试了MiniCPM-V 2.0得45分钟才出结果。
请问如何类似于QW2-VL调整输入可以满足显存在8G以内的设备上跑?
No response
- OS:ubuntu20.04 - Python:3.10 - Transformers:4.40.0 - PyTorch:2.1.2+cu121 - CUDA (`python -c 'import torch; print(torch.version.cuda)'`):12.1
The text was updated successfully, but these errors were encountered:
你好,cpu跑速度可能确实会慢很多,可以尝试llamacpp和ollama,速度可能有所上升
Sorry, something went wrong.
No branches or pull requests
是否已有关于该错误的issue或讨论? | Is there an existing issue / discussion for this?
该问题是否在FAQ中有解答? | Is there an existing answer for this in FAQ?
当前行为 | Current Behavior
MiniCPM-V 2.6跑demo,本身机器显存8G,使用cuda提示显存不够,目前使用cpu跑,贼慢十几个小时才出结果,请问如何类似于QW2-VL调整输入可以满足显存在8G以内的设备上跑。
接着又尝试了MiniCPM-V 2.6 int4也很慢。
目前尝试了MiniCPM-V 2.0得45分钟才出结果。
期望行为 | Expected Behavior
请问如何类似于QW2-VL调整输入可以满足显存在8G以内的设备上跑?
复现方法 | Steps To Reproduce
No response
运行环境 | Environment
备注 | Anything else?
No response
The text was updated successfully, but these errors were encountered: