You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
It looks like the latest Megatron-LM already supports Mixture-of-Experts -- I'd be happy to see that supported (for Mixtral)!
I can also help contribute but don't really have too much experience in rebasing Megatron-LLM to upstream..
The text was updated successfully, but these errors were encountered:
xingyaoww
changed the title
Any plans to rebase the codebase to most recent Megatron-LM for MoE
Any plans to rebase the codebase to most recent Megatron-LM for MoE?
Mar 7, 2024
It looks like the latest Megatron-LM already supports Mixture-of-Experts -- I'd be happy to see that supported (for Mixtral)!
I can also help contribute but don't really have too much experience in rebasing Megatron-LLM to upstream..
The text was updated successfully, but these errors were encountered: