Must accept:
- GpuMat, CuPy, or Torch GPU tensor in a batched format.
- Build and load model engines.
- Image Preprocessor.
- Output Postprocessor.
It is beneficial for complex inference scenarios when we want to infer on historical data allocated in a GPU, not on streaming data.