Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Feature - Add support for GGUF spec to read necessary details directly from the model #1

Open
crimson-knight opened this issue Apr 15, 2024 · 1 comment
Assignees
Labels
enhancement New feature or request help wanted Extra attention is needed

Comments

@crimson-knight
Copy link
Owner

Right now Llamero requires that you add the prompt template details yourself.

Not a big deal, but those details are already present in the model. So, using the GGUF spec, we should add support to read the initial bytes from the file to get the necessary information for the chat template.

This would be an excellent convenience.

GGUF Spec for reference: https://github.com/ggerganov/ggml/blob/master/docs/gguf.md

@crimson-knight crimson-knight self-assigned this Apr 15, 2024
@crimson-knight crimson-knight added enhancement New feature or request help wanted Extra attention is needed labels Apr 15, 2024
@crimson-knight
Copy link
Owner Author

This is partially implemented allowing fetching of certain special tokens for the chat prompt.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request help wanted Extra attention is needed
Projects
None yet
Development

No branches or pull requests

1 participant