-
Notifications
You must be signed in to change notification settings - Fork 371
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Any plans to bring llava support? #340
Comments
I have a prototype working. I will see if I can clean and finish the project this weekend. |
Hello! Are there any news? |
Looking forward to the change! Just discussed how this will help significantly to solve a use-case for our internal application. Do you have it available in your fork already? I'd like to pull it into mine and play around with it. |
I would like to help too. This is very promising
Чт, 7 дек. 2023 г. в 22:01, Philipp Bauer ***@***.***>:
… Looking forward to the change! Just discussed how this will help
significantly to solve a use-case for our internal application. Do you have
it available in your fork already? I'd like to pull it into mine and play
around with it.
—
Reply to this email directly, view it on GitHub
<#340 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AEKNDOSY2R4WHTFQILPOCQTYIHK5DAVCNFSM6AAAAABACXJGWWVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTQNBVGUYDINRWHA>
.
You are receiving this because you authored the thread.Message ID:
***@***.***>
|
I will try to finish the library a little more this weekend before make this public. Until know, only tested on osx-arm64. |
Cool, thank you! I'm on osx-arm64 and can do some testing (have my colleague do it perhaps too) after implementing it in our app. |
Is there any update on this? One Use case is OCR. Thanks, |
I have a branch in my fork with part of the changes. Build binaries, include runtime, etc. Those are things that I don't have done before. I first make the development with manually build binaries. With the first versions of January, it worked. With this prompt: What is unusual about this image? and this picture: This is the output: But since PR #445 it crashes on llama.cpp. I'm trying to identify the root cause and get this working again. |
Thanks for the update @SignalRT Happy to test this when you have it working. |
@SignalRT Btw, where is the code repo with the working branch you mentioned above? |
I will work this weekend to try to publish my work. The work will be in my branch: https://github.com/SignalRT/LLamaSharp/tree/MultiModal until PR. |
@SignalRT For now I switched to plan B, OllamaSharp, but I'm happy to hear that I will be able to switch back soon. |
@SignalRT please put the code you have (LLava) into your branch that we can help you finalizing it (better calling it LLava instead of LLavaSharp). |
@IntptrMax, I have looked at your example. It is a very good attempt! I have noticed a few bugs with marshaling the cpp output, for example,
should be changed to
because cpp returns a pointer to the structure and if you do not do this, then you will get some random problems...
where Also, there is a problem with the context size. If the number of tokens in the image embedding is higher, then the context size (n_ctx), then the program will crash in the function |
@IntptrMax, I have quickly corrected your example and it seems that if we use llava_image_embed_make_with_filename with the above correction and a higher context size (4096), then it works: |
@zsogitbe Thanks a lot! I have get the same problem when evaluating several images, that's a good idea to solve it. |
I need a minimum context size of = image embedding size (2880 tokens) + batch size (512). In your example 2880+512 = 3392! The image embedding size depends on the model! |
PR with first draft: #555 |
The first PR to build llava binaries #556 |
I have tried to add llava to llamaSharp and it can work, but still have to improve. My demo is https://github.com/IntptrMax/LLamaSharp/tree/add_llava |
It works IntptrMax, I have tested it, but there is a memory leak. In my trial 1.8GB GPU memory is not freed. Try to find how to free the GPU memory because 1.8 GB is too much. |
Hi! Are there any plans to support llava as well? As i see, it is merged to llama.cpp about month ago and gives possibility to work with image recognition aswell.
The text was updated successfully, but these errors were encountered: