You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
===============================
== TensorRT Inference Server ==
===============================
NVIDIA Release 19.09 (build 8086825)
Copyright (c) 2018-2019, NVIDIA CORPORATION. All rights reserved.
Various files include modifications (c) NVIDIA CORPORATION. All rights reserved.
NVIDIA modifications are covered by the license terms that apply to the underlying
project or file.
WARNING: Detected NVIDIA NVIDIA GeForce RTX 4090 D GPU, which is not yet supported in this version of the container
ERROR: No supported GPU(s) detected to run this container
NOTE: The SHMEM allocation limit is set to the default of 64MB. This may be
insufficient for the inference server. NVIDIA recommends the use of the following flags:
nvidia-docker run --shm-size=1g --ulimit memlock=-1 --ulimit stack=67108864 ...
I0102 04:04:05.631646 1 metrics.cc:160] found 1 GPUs supporting NVML metrics
I0102 04:04:05.636899 1 metrics.cc:169] GPU 0: NVIDIA GeForce RTX 4090 D
I0102 04:04:05.637062 1 server.cc:110] Initializing TensorRT Inference Server
E0102 04:04:05.637134 1 main.cc:943] error: creating server: INTERNAL - Unable to create GCS client. Check account credentials.
Expected behavior
triton-inference-server can run success in rtx4090D
The text was updated successfully, but these errors were encountered:
Description
nvidia-smi
kubectl describe node xxx
kubectl logs triton-inference-server-tritoninferenceserver-544f6c6d77-78pgq
Expected behavior
triton-inference-server can run success in rtx4090D
The text was updated successfully, but these errors were encountered: