Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support for another models (ollama models) #68

Open
testing0mon21 opened this issue May 27, 2024 · 3 comments
Open

Support for another models (ollama models) #68

testing0mon21 opened this issue May 27, 2024 · 3 comments

Comments

@testing0mon21
Copy link

testing0mon21 commented May 27, 2024

@b4rtaz Hey, thank you for your wonderful work. Could you please offer some details about how to add supported model? For example, how to to convert some ollama models like command+r or starcoder or llama3 70b to ddlama

https://ollama.com/library/command-r-plus
https://ollama.com/library/llama3:70b
https://ollama.com/library/starcoder2

@b4rtaz
Copy link
Owner

b4rtaz commented May 27, 2024

Hello @testing0mon21,

So from your list only llama3 is supported now.

To convert Llama 3 you have 2 options, you can do it by using Meta files and convert them by using convert-llama.py script, here is the tutorial. The second option is download .safetensor weights from Huggingface and convert it by using convert-hf.py.

@testing0mon21
Copy link
Author

testing0mon21 commented May 27, 2024

Did I understand correctly, for other architectures it will be difficult to implement the same thing that you implemented with llama? @b4rtaz

@b4rtaz
Copy link
Owner

b4rtaz commented May 27, 2024

I think this depends on a specyfic architecture. Some architectures will be easy, some not. Adding new architecture is always non-zero effort. Currently DL supports: llama, mixtral and grok1.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants