You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I was wondering if llama.cpp currently supports multimodal models (like LLaVA) on Android or other mobile devices. I know that llama.cpp has added support for LLaVA, but has anyone successfully run it on mobile?
Some questions I have:
Is there official support or any experimental builds for multimodal inference on Android?
What are the main challenges in running LLaVA on mobile? (Performance, quantization, etc.)
Are there any optimized versions or alternative solutions to make it work efficiently on mobile CPUs/NPUs?
If anyone has tested it or has insights, I’d love to hear your thoughts!
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
-
Hi everyone,
I was wondering if llama.cpp currently supports multimodal models (like LLaVA) on Android or other mobile devices. I know that llama.cpp has added support for LLaVA, but has anyone successfully run it on mobile?
Some questions I have:
If anyone has tested it or has insights, I’d love to hear your thoughts!
Thanks in advance!
Beta Was this translation helpful? Give feedback.
All reactions