-
Notifications
You must be signed in to change notification settings - Fork 1.5k
Issues: triton-inference-server/server
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Milestones
Assignee
Sort
Issues list
OpenAI-Compatible Frontend should support world_size larger than 1
#7914
opened Jan 3, 2025 by
cocodee
vllm_backend: What is the right way to use downloaded model +
model.json
together?
#7912
opened Jan 2, 2025 by
kyoungrok0517
Python backend with multiple instances cause unexpected and non-deterministic results
#7907
opened Dec 25, 2024 by
NadavShmayo
MIG deployment of triton cause "CacheManager Init Failed. Error: -17"
#7906
opened Dec 25, 2024 by
LSC527
How Triton inference server always compare the current frame infer result with the previous one
#7893
opened Dec 19, 2024 by
Komoro2023
Error when using ONNX with TensorRT (ORT-TRT) Optimization on Multi-GPU
#7885
opened Dec 16, 2024 by
efajardo-nv
Manual warmup per model instance / specify warmup config dynamically using c api
#7884
opened Dec 16, 2024 by
asaff1
Segfault/Coredump in grpc::ModelInferHandler::InferResponseComplete
#7877
opened Dec 12, 2024 by
andyblackheel
Core was generated by /opt/tritonserver/backends/python/triton_python_backend_stub
#7875
opened Dec 12, 2024 by
powerpistn
No content returned with OpenAI-Compatible Frontend Beta (ensemble & bls)
#7868
opened Dec 11, 2024 by
njaramish
[Feature]: ORCA format reporting for KV-Cache metrics in Inference Response Header
#7865
opened Dec 10, 2024 by
BenjaminBraunDev
There is not a good way to call trtllm backend to initialize lora weights from Python BLS
#7856
opened Dec 6, 2024 by
ShuaiShao93
Torchscript Model can't have bfloat16 inputs / outputs in 24.09
#7853
opened Dec 5, 2024 by
MatthieuToulemont
Previous Next
ProTip!
Mix and match filters to narrow down what you’re looking for.