Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Problem about Inference Time #13

Open
Hatins opened this issue Sep 28, 2024 · 0 comments
Open

Problem about Inference Time #13

Hatins opened this issue Sep 28, 2024 · 0 comments

Comments

@Hatins
Copy link

Hatins commented Sep 28, 2024

Hi @NikolaZubic

I have some trouble testing the inference time. I set both sequence length and batch size equal to 1 and tested the inference time on a single RTX4090 GPU. I found the time for lstm is about 4 ms, while the s5 block takes 7 ms.

I tried to exchange the forward to forward_rnn to evaluate the inference time as you mentioned, but I got the same error as @yuyangpoi reported.

Therefore, I wonder how to test the inference time of S5 block is correctly?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant