Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Llama-3.2 11B Vision Support #1209

Open
yukiarimo opened this issue Nov 11, 2024 · 3 comments
Open

Llama-3.2 11B Vision Support #1209

yukiarimo opened this issue Nov 11, 2024 · 3 comments

Comments

@yukiarimo
Copy link

I would like to run Llama-3.2 11B Vision in KoboldCPP. Ollama recently added the support, so I guess it is possible to pull it here :)

Also, there is support needed for GGUF conversion!

@LostRuins
Copy link
Owner

Ollama does it in golang, their implementation is not compatible.

KoboldCpp's upstream is llama.cpp, so you'd have to wait until ggerganov#9643 is implemented (looking very unlikely since nobody is interested in implementing it)

Alternatively, you can use MiniCPM v2, which already has support here and works well enough.

@yukiarimo
Copy link
Author

Oh, I see. That’s sad! By the way, in KoboldCPP, you can send plain raw text to the model. Do you know how to do that using Ollama without using any template? And is a model card mandatory?

@LostRuins
Copy link
Owner

I don't think it's possible in Ollama (most backends nowadays only allow chat completions). That said, I have not extensively tried their software and would naturally be biased against it.

In KoboldCpp, no model card is necessary and you are able to use whatever prompt format or instruct format you like. In Ollama, I believe they restrict it based on whatever you have downloaded.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants