-
Notifications
You must be signed in to change notification settings - Fork 894
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
t2v CogVideoX1.5-5B OOM #540
Comments
更新到最新的diffusers main分支 |
已经更新到最新的diffusers main,还是存在OOM |
(cogvideo) root@autodl-container-cd46119efa-b92bcf86: Successfully built diffusers WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager, possibly rendering your system unusable.It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv. Use the --root-user-action option if you know what you are doing and want to suppress this warning. |
diffusers 0.32.0.dev0 |
Was anyone able to solve this issue? |
Hope there could be a solution. Thank you. |
Thank you for your hard work, zRzRzRzRzRzRzR. The cogvideo1.5 t2v worked, so I'm sharing the method. A 1360x768 resolution video was generated in my environment.
I haven't tried inference/cli_demo.py yet, but I think it will be easier to modify CogVideoX1.5-5B/transformer/config.json and use cli_demo.py. |
System Info / 系統信息
CUDA12.4
diffusers 0.32.0.dev0 (使用pi p install -e . 安装的最新的)
A100 40GB VRAM
运行CogVideoX1.5-5B-I2V进行I2V正常生成
运行CogVideoX1.5-5B进行T2V,总是OOM
Information / 问题信息
Reproduction / 复现过程
python inference/cli_demo.py --prompt="Two kittens lick each other's fur" --generate_type="t2v"
torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 5.67 GiB. GPU 0 has a total capacity of 39.38 GiB of which 4.50 GiB is free. Process 168833 has 34.88 GiB memory in use. Of the allocated memory 31.49 GiB is allocated by PyTorch, and 2.89 GiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
Expected behavior / 期待表现
直接运行的官方的nference/cli_demo.py,无修改
里面是开启的
pipe.enable_sequential_cpu_offload() enabled.
依旧会OOM
The text was updated successfully, but these errors were encountered: