Skip to content

Conversation

@TedThemistokleous
Copy link
Collaborator

Description

Add the quantization, tuning and memory limits as inputs to the final hashed output name fora model. This ensure that we're not reusing a different quantized, or tuned model from a previous session.

Motivation and Context

@TedThemistokleous TedThemistokleous self-assigned this Oct 3, 2025
@TedThemistokleous TedThemistokleous added the Bugfix Fix to a bug or reported issue label Oct 3, 2025
turns out we weren't passing exhaustive tune flags for the recompile in along with some other iflags like mem_limit
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Bugfix Fix to a bug or reported issue

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants