Build a Triton Inference Server with MNIST Example: Part (1/4)
Author(s): Hao WEI Originally published on Towards AI. a step-by-step coding practice In this series of blogs, we will walk through how to build a triton inference server, host it in AWS ECS, and accelerate the speed using tensorRT. We will use …
Build a Triton Inference Server with Preprocess and Postprocess: Part (2/4)
Author(s): Hao WEI Originally published on Towards AI. A step-by-step coding practice a step-by-step coding practice pub.towardsai.net In my last blog, we built a bare minimum triton server using an MNIST example. In the example, we did preprocessing and postprocessing in the …