Skip to content

computing the inference FLOPs #4

@twwwwx

Description

@twwwwx

Hello,
Thanks for your talented work! I was trying to run your model in the test set. However, I have some questions about how did you compute FLOPs.
To have the amazing acceleration rates as you mentioned in the paper, does it need to train an end-to-end sparsified model, or running your finetuning codes for several epoches is just enough? As is mentioned in the paper, "In the training procedure, Transkimmer does not prune the hidden state tensors as it does in the inference time." So in the inference time, by which means you prune the hidden state tensors to turn it to the inference mode? Can I use torchfile directly on the trained transkimmer model to compute the inference FLOPs?
Looking forward to your reply.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions