Integrating PyTorch models with TVM makes it worse from time perspective

744 Views Asked by At

I was recently told that integrating/compiling my PyTorch model with the help of Apache TVM should speed inference up, but am very confused about the usage of TVM as there seem to be multiple ways to use it.

This blog on their website says

Usage is simple:

import torch_tvm
torch_tvm.enable()

That’s it! PyTorch will then attempt to convert all operators it can to known Relay operators during its JIT compilation process.

If just this much does the job, why are other specific tutorials so much more comprehensive? After going through a fair few tutorials another thing i assumed (not sure if i am right) was that TVM cannot really use the default way torch code is run and needs the 'jit' compiler (non-default way of running torch code) to work over. This assumption was based off of the fact that the github repo tutorial shows the following snippet

import torch
import torch_tvm

torch_tvm.enable()

# The following function will be compiled with TVM
@torch.jit.script
def my_func(a, b, c):
    return a * b + c   

where the function my_func is wrapped with a decorator that looks to be compiling the function with jit. But using this exact same function and timing its usage with and without the wrapper + TVM shows that the normal usage of this function is far more time efficient. If its not speeding things up, what is exactly does jit+tvm help in? and if it is supposed to be speeding things up, why is that not the case here?

P.S. I apologise for the comprehensive description of my problem. I was not able to figure this out and get whats happening even after reading and tinkering around a fair amount with torch_tvm. Would appreciate any link to resources or any explanation that might help me out.

0

There are 0 best solutions below