Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Would you support glm4-chat-1m #65

Open
choyakawa opened this issue Aug 29, 2024 · 1 comment
Open

Would you support glm4-chat-1m #65

choyakawa opened this issue Aug 29, 2024 · 1 comment

Comments

@choyakawa
Copy link

I have some concerns about this. Based on my experience, GGUF with llama.cpp seems to work differently from transformers, whereas GGML with chatglm.cpp behaves the same as transformers. I haven't yet identified the exact differences. Therefore, an optimization for long-context handling with transformers would be very helpful.

@qiyuxinlin
Copy link
Contributor

Thank you for following our work! we will take some time to evaluate how to incorporate glm4-chat-1m into our framework

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants