@@ -156,7 +156,7 @@ Triton exposes some flags to control the execution mode of the TorchScript model
156156
157157 The section of model config file specifying this parameter will look like:
158158
159- ``` yaml
159+ ``` proto
160160 parameters: {
161161 key: "DISABLE_OPTIMIZED_EXECUTION"
162162 value: { string_value: "true" }
@@ -175,7 +175,7 @@ Triton exposes some flags to control the execution mode of the TorchScript model
175175
176176 To enable inference mode, use the configuration example below:
177177
178- ``` yaml
178+ ``` proto
179179 parameters: {
180180 key: "INFERENCE_MODE"
181181 value: { string_value: "true" }
@@ -195,7 +195,7 @@ Triton exposes some flags to control the execution mode of the TorchScript model
195195
196196 To disable cuDNN, use the configuration example below:
197197
198- ``` yaml
198+ ``` proto
199199 parameters: {
200200 key: "DISABLE_CUDNN"
201201 value: { string_value: "true" }
@@ -210,7 +210,7 @@ Triton exposes some flags to control the execution mode of the TorchScript model
210210
211211 To enable weight sharing, use the configuration example below:
212212
213- ``` yaml
213+ ``` proto
214214 parameters: {
215215 key: "ENABLE_WEIGHT_SHARING"
216216 value: { string_value: "true" }
@@ -228,7 +228,7 @@ Triton exposes some flags to control the execution mode of the TorchScript model
228228
229229 To enable cleaning of the CUDA cache after every execution, use the configuration example below:
230230
231- ``` yaml
231+ ``` proto
232232 parameters: {
233233 key: "ENABLE_CACHE_CLEANING"
234234 value: { string_value: "true" }
@@ -251,7 +251,7 @@ Triton exposes some flags to control the execution mode of the TorchScript model
251251
252252 To set the inter-op thread count, use the configuration example below:
253253
254- ``` yaml
254+ ``` proto
255255 parameters: {
256256 key: "INTER_OP_THREAD_COUNT"
257257 value: { string_value: "1" }
@@ -277,7 +277,7 @@ Triton exposes some flags to control the execution mode of the TorchScript model
277277
278278 To set the intra-op thread count, use the configuration example below:
279279
280- ``` yaml
280+ ``` proto
281281 parameters: {
282282 key: "INTRA_OP_THREAD_COUNT"
283283 value: { string_value: "1" }
@@ -321,7 +321,7 @@ where the input tensors are placed as follows:
321321
322322To set the model instance group, use the configuration example below:
323323
324- ``` yaml
324+ ``` proto
325325instance_group {
326326 count: 2
327327 kind: KIND_GPU
@@ -356,7 +356,7 @@ The following PyTorch settings may be customized by setting parameters on the
356356
357357For example:
358358
359- ``` yaml
359+ ``` proto
360360parameters: {
361361 key: "NUM_THREADS"
362362 value: { string_value: "4" }
0 commit comments