Skip to content

Commit

Permalink
Merge branch 'keras-team:master' into falcon-causal
Browse files Browse the repository at this point in the history
  • Loading branch information
SamanehSaadat committed May 23, 2024
2 parents 3e195a2 + 0bf204d commit 9609004
Show file tree
Hide file tree
Showing 3 changed files with 4 additions and 4 deletions.
4 changes: 2 additions & 2 deletions keras_nlp/src/models/causal_lm.py
Original file line number Diff line number Diff line change
Expand Up @@ -352,9 +352,9 @@ def generate(
raise ValueError(
'A `preprocessor` must be attached to the model if `stop_token_ids="auto"`. '
"Currently `preprocessor=None`. To call `generate()` with preprocessing "
"detached, either pass `stop_tokens_ids=None` to always generate until "
"detached, either pass `stop_token_ids=None` to always generate until "
"`max_length` or pass a tuple of token ids that should terminate generation "
"as `stop_tokens_ids`."
"as `stop_token_ids`."
)
elif stop_token_ids == "auto":
stop_token_ids = [self.preprocessor.tokenizer.end_token_id]
Expand Down
2 changes: 1 addition & 1 deletion keras_nlp/src/models/phi3/phi3_presets.py
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@
"model. The model was trained using the Phi-3 datasets. This "
"dataset includes both synthetic data and filtered publicly "
"available website data, with an emphasis on high-quality and "
"reasoning-dense properties.",
"reasoning-dense properties."
),
"params": 3821079552,
"official_name": "Phi-3",
Expand Down
2 changes: 1 addition & 1 deletion tools/checkpoint_conversion/convert_llama_checkpoints.py
Original file line number Diff line number Diff line change
Expand Up @@ -257,7 +257,7 @@ def main(_):
test_model(keras_nlp_model, keras_nlp_tokenizer, hf_model, hf_tokenizer)
print("\n-> Tests passed!")

keras_nlp_model.save_to_preset(keras_nlp_model, preset)
keras_nlp_model.save_to_preset(preset)
print("\n-> Saved the model preset in float16")

# === Save the tokenizer ===
Expand Down

0 comments on commit 9609004

Please sign in to comment.