Replies: 2 comments 3 replies
-
@CoCoNuTeK let's use discussions for such open ended questions. You can check the implementation of the T5 model in transformers and modify as needed. |
Beta Was this translation helpful? Give feedback.
1 reply
-
The output from the model has shape [batch_size, pred_len, 4096] the logits tensor; so to get the predicted token, i can just commit to the highest logit and get dimension [batch_size, pred_len] but i am still stuck with token values + the tokenizer i used here
to create the tokens from float values; does it have backwards operation?? And i dont see that working as the tokens encode a range of values not just one value. |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hello there,
I would like to ask about the loss function given i wanted to create my own loss function lets say moving average MASE as loss function to the model, everything is there except the outputs of the model contains: loss, logits and other stuff but there are no predicted values directly,
so is there a way to use the tokenizer used for creating the input_ids, labels and attention_mask to somehow turn the logits into the predicted values, so the reverse operation?
Beta Was this translation helpful? Give feedback.
All reactions