The basic flow is
Load the Model (SSD,CoreML,Onnx,Keras,MXNet,Caffee2,TFLite,Tensorflow)
Compile use Relay with support external Lib or not.
Generate the execution graph.(I think)
TVM will run the graph.
https://docs.tvm.ai/langref/index.html <--- Study Relay is the Key
I need to know, where I can Quantize the weight and reduce the computation graph size.