Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add support for Llama3.3 #2661

Open
jorgeantonio21 opened this issue Dec 6, 2024 · 4 comments
Open

Add support for Llama3.3 #2661

jorgeantonio21 opened this issue Dec 6, 2024 · 4 comments

Comments

@jorgeantonio21
Copy link
Contributor

jorgeantonio21 commented Dec 6, 2024

https://huggingface.co/meta-llama/Llama-3.3-70B-Instruct

@zackangelo
Copy link
Contributor

There were no architectural changes afaik, it should already work with the existing example, just change the repo URL.

@theHausdorffMetric
Copy link

I added the respective repo url to the llama example and created a pull request:
#2677

@LaurentMazare
Copy link
Collaborator

Were you able to try it out? I don't think it would fit in the memory of a single gpu so this would be better suited for llama_multiprocess that can use multiple gpus. Would be great if someone can give it a spin there and check that it works.

@zackangelo
Copy link
Contributor

Ah, you're right @LaurentMazare.

I don't believe the multiprocess example has been updated to create the new rope tensors that were introduced in Llama 3.1. But it should work otherwise.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants