Replies: 1 comment 1 reply
-
Experiencing a similar issue when running Pytorch vs. ONNX on Mac M2... curious what's going on here. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I am testing inference performance between
roberta-base
original model vs onnx on CPU (Apple MacBook M1/16G RAM). From the test, it shows that while the sequence length is increasing, ONNX model inference is getting slower comparing to the original PyTorch model. The example code I can find from googling were mostly working with short sequences which demos ONNX is faster than original PyTorch model but no tests I have seen comparing different sequence lengths all the way to the maximum.Is this expected or something I need to do to get ONNX running faster? Or if I did anything wrong, please also let me know. Thanks for any comments. I have attached my code and package versions in the below.
Beta Was this translation helpful? Give feedback.
All reactions