Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

May I ask if pruning of qwen2-72b and llama 3.1-70b is supported? #445

Open
lingq1 opened this issue Dec 6, 2024 · 2 comments
Open

May I ask if pruning of qwen2-72b and llama 3.1-70b is supported? #445

lingq1 opened this issue Dec 6, 2024 · 2 comments

Comments

@lingq1
Copy link

lingq1 commented Dec 6, 2024

Is pruning of large models similar to 70B supported?

@VainF
Copy link
Owner

VainF commented Dec 6, 2024

Hi @lingq1, I think it should work as long as we have sufficient GPU RAM. If not, we can run the pruning on CPU.

@lingq1
Copy link
Author

lingq1 commented Dec 9, 2024

Hi @lingq1, I think it should work as long as we have sufficient GPU RAM. If not, we can run the pruning on CPU.

ths

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants