모델 서빙에 가장 적합한 도구를 찾자
- triton + tensorflow
- triton + tensorflow-onnx
- triton + tensorflow-onnx-tensorRT
- bentoML + tensorflow
- bentoML + tensorflow-onnx
- akka + tensorflow
- akka + tensorflow-onnx
docker build -t benjamin/nsmc/triton-onnx:0.0.1 -f triton-serving/onnx/Dockerfile .
docker build -t benjamin/nsmc/triton-onnx-tensorrt:0.0.1 -f triton-serving/onnx-tensorrt/Dockerfile .
docker run --gpus=all -p 9000:8000 -p 9001:8001 -p 9002:8002 --rm benjamin/nsmc/triton-onnx:0.0.1 tritonserver --model-repository=/models
docker run --gpus=all -p 9000:8000 -p 9001:8001 -p 9002:8002 --rm benjamin/nsmc/triton-onnx-tensorrt:0.0.1 tritonserver --model-repository=/models