Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

LLamaBatch Automatically Grow Capacity #443

Merged
merged 1 commit into from
Jan 20, 2024

Conversation

martindevans
Copy link
Member

  • Removed some unused eval methods.
  • Added a DecodeAsync overload which runs the work in a task
  • Replaced some NativeHandle usage in BatchedDecoding with higher level equivalents.
  • Made the LLamaBatch grow when token capacity is exceeded, removing the need to manage token capacity externally.

 - Added a `DecodeAsync` overload which runs the work in a task
 - Replaced some `NativeHandle` usage in `BatchedDecoding` with higher level equivalents.
 - Made the `LLamaBatch` grow when token capacity is exceeded, removing the need to manage token capacity externally.
@martindevans martindevans merged commit 250c20b into SciSharp:master Jan 20, 2024
3 checks passed
@martindevans martindevans deleted the llama_batch_self_grow branch January 20, 2024 14:43
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant