Replies: 2 comments
-
Do you know if llama.cpp can load MiniCPM models? |
Beta Was this translation helpful? Give feedback.
0 replies
-
Yes, they do. Supported models are: Inference with llama.cpp MiniCPM-V 2.6 can run with llama.cpp now! See our fork of llama.cpp for more detail. This implementation supports smooth inference of 16~18 token/s on iPad (test environment:iPad Pro + M4). |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Is there a way to add support or somehow be able to load MiniCPM model like this one here?
https://huggingface.co/openbmb/MiniCPM-V-2_6-gguf
Thank you in advance!
Beta Was this translation helpful? Give feedback.
All reactions