Skip to content

pruning-distillation guidance or docs for Multimodal model #12975

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
xduzhangjiayu opened this issue Apr 11, 2025 · 2 comments
Open

pruning-distillation guidance or docs for Multimodal model #12975

xduzhangjiayu opened this issue Apr 11, 2025 · 2 comments
Assignees

Comments

@xduzhangjiayu
Copy link

Hi,
Is this project has some guidance or docs about how to pruning-distillation a multimodal model like qwen2.5vl-7B Qwen2.5-Omni-7B? Currently I only find the docs for Llama.
Looking forward to the reply, thanks!

@sharathts
Copy link
Collaborator

Hi @xduzhangjiayu - we currently support only pruning and distillation of LLMs. The same can be applied to compressing the LLM part of VLMs as they tend to be the bottleneck.

@xduzhangjiayu
Copy link
Author

xduzhangjiayu commented Apr 16, 2025

@sharathts
Hi,
Thanks for the reply, one more question is, how can I transfer the multimodal model (like qwen2vl) type from huggingface model to .nemo format, is there any script in this repo? Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants