TensorRT is a C++ library for high performance inference on NVIDIA GPUs and deep learning accelerators.
NVIDIA TensorRT is a platform for high-performance deep learning inference. It includes a deep learning inference optimizer and runtime that delivers low latency and high-throughput for deep learning inference applications.