Skip to content

Commit

Permalink
llama : the WPM vocabs use the CLS token as BOS (#10930)
Browse files Browse the repository at this point in the history
* llama : the WPM vocabs use the CLS token as BOS

ggml-ci

* llama : add comment
  • Loading branch information
ggerganov authored Dec 24, 2024
1 parent 60cfa72 commit 30caac3
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 2 deletions.
2 changes: 1 addition & 1 deletion src/llama-vocab.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -1657,7 +1657,7 @@ bool llama_token_is_control_impl(const struct llama_vocab & vocab, llama_token t
}

llama_token llama_token_bos_impl(const struct llama_vocab & vocab) {
return vocab.special_bos_id;
return vocab.type != LLAMA_VOCAB_TYPE_WPM ? vocab.special_bos_id : vocab.special_cls_id;
}

llama_token llama_token_eos_impl(const struct llama_vocab & vocab) {
Expand Down
2 changes: 1 addition & 1 deletion src/llama-vocab.h
Original file line number Diff line number Diff line change
Expand Up @@ -45,7 +45,7 @@ struct llama_vocab {
id special_unk_id = 0;
id special_sep_id = LLAMA_TOKEN_NULL;
id special_pad_id = LLAMA_TOKEN_NULL;
id special_cls_id = LLAMA_TOKEN_NULL;
id special_cls_id = LLAMA_TOKEN_NULL; // TODO: revisit if this is really needed https://github.com/ggerganov/llama.cpp/pull/10930
id special_mask_id = LLAMA_TOKEN_NULL;

id linefeed_id = 13;
Expand Down

0 comments on commit 30caac3

Please sign in to comment.