Antonina
Antonina

Reputation: 604

Triton Inference Server - tritonserver: not found

I try to run NVIDIA’s Triton Inference Server. I pulled the pre-built container nvcr.io/nvidia/pytorch:22.06-py3 and then run it with the command

run --gpus=1 --rm -p8000:8000 -p8001:8001 -p8002:8002 -v/F/models:/models nvcr.io/nvidia/pytorch:22.06-py3 tritonserver --model-repository=/models

and got the error

/opt/nvidia/nvidia_entrypoint.sh: line 49: exec: tritonserver: not found

I googled and have not found something to catch this. I tried to change tritonserver to trtserver as recommended but it did not help. Please give some advice how it can be solved.

Upvotes: 0

Views: 3882

Answers (1)

Sysanin
Sysanin

Reputation: 1793

Looks like you're trying to run a tritonserver using a pytorch image but according to the triton-server quick start guide, the image should be:

$ docker run --gpus=1 --rm -p8000:8000 -p8001:8001 -p8002:8002 -v/full/path/to/docs/examples/model_repository:/models nvcr.io/nvidia/tritonserver:<xx.yy>-py3 tritonserver --model-repository=/models

Where <xx.yy> is the version of Triton that you want to use

In your case it should be nvcr.io/nvidia/tritonserver:22.06-py3 and the full command:

run --gpus=1 --rm -p8000:8000 -p8001:8001 -p8002:8002 -v/F/models:/models nvcr.io/nvidia/tritonserver:22.06-py3 tritonserver --model-repository=/models

Upvotes: 1

Related Questions