Whisper - Missing parameters for triton deployment using tensorrt_llm backend #672
Open
2 of 4 tasks
Labels
bug
Something isn't working
System Info
Hello,
I'm trying to deploy Whisper
large-v3
using Triton and tensorrtllm backend using this readme: https://github.com/triton-inference-server/tensorrtllm_backend/blob/v0.16.0/docs/whisper.mdContext
v0.16.0
Issues
However, I have some issues when I'm trying to go to step 3 (Prepare Tritonserver configs) due to missing parameters to fill in config file using the following script:
My questions are:
tensorrt_llm
"model" to run the triton server forwhisper_bls
?Thank you 🙂
Who can help?
@juney-nvidia
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction
In the https://github.com/triton-inference-server/tensorrtllm_backend/blob/v0.16.0/docs/whisper.md, at step 3:
Expected behavior
Variable not found when running the script:
Or in tritonserver logs:
actual behavior
Missing parameters to fill the config.pbtxtx
additional notes
Tried to add the parameters as follow but still missing other parameters
The text was updated successfully, but these errors were encountered: