You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I disabled flash attention and try to change the batch size. Now I can load the model but I have this output: end_id_id_id_id_id_id_id_id_id_id_id_id_id_id_id_id_id_id_id_id_id_id_id_id_id_id_id_i
Is this a llama-cpp version issue?
The text was updated successfully, but these errors were encountered: