-
Notifications
You must be signed in to change notification settings - Fork 5
computing the inference FLOPs #4
Copy link
Copy link
Open
Description
Hello,
Thanks for your talented work! I was trying to run your model in the test set. However, I have some questions about how did you compute FLOPs.
To have the amazing acceleration rates as you mentioned in the paper, does it need to train an end-to-end sparsified model, or running your finetuning codes for several epoches is just enough? As is mentioned in the paper, "In the training procedure, Transkimmer does not prune the hidden state tensors as it does in the inference time." So in the inference time, by which means you prune the hidden state tensors to turn it to the inference mode? Can I use torchfile directly on the trained transkimmer model to compute the inference FLOPs?
Looking forward to your reply.
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels