You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on May 12, 2023. It is now read-only.
I want to use this to run a chat session so that future generate calls have the previous context. But I don't want to re-tokenize and re-run the old tokens again right? Is there some easy way to do this?
The text was updated successfully, but these errors were encountered:
@redthing1, Yes you can use interactive mode in the generate function, it behaves the same as the main example of llama.cpp, I will add a CLI soon to see how it works.
I will make it more convenient soon!
I want to use this to run a chat session so that future generate calls have the previous context. But I don't want to re-tokenize and re-run the old tokens again right? Is there some easy way to do this?
The text was updated successfully, but these errors were encountered: