Replies: 2 comments 1 reply
-
@CoCoNuTeK let's use discussions for such open ended questions. You can check the implementation of the T5 model in transformers and modify as needed. |
Beta Was this translation helpful? Give feedback.
-
The output from the model has shape [batch_size, pred_len, 4096] the logits tensor; so to get the predicted token, i can just commit to the highest logit and get dimension [batch_size, pred_len] but i am still stuck with token values + the tokenizer i used here
to create the tokens from float values; does it have backwards operation?? And i dont see that working as the tokens encode a range of values not just one value. |
Beta Was this translation helpful? Give feedback.
-
Hello there,
I would like to ask about the loss function given i wanted to create my own loss function lets say moving average MASE as loss function to the model, everything is there except the outputs of the model contains: loss, logits and other stuff but there are no predicted values directly,
so is there a way to use the tokenizer used for creating the input_ids, labels and attention_mask to somehow turn the logits into the predicted values, so the reverse operation?
Beta Was this translation helpful? Give feedback.
All reactions