You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Below is a description of all the fields in config.json.
158
204
159
205
| Field | Description |
@@ -174,6 +220,7 @@ Below is a description of all the fields in config.json.
174
220
| backend_config.hf_model_class | HuggingFace model class. |
175
221
| backend_config.enable_lora | Set to true to enable loading LoRA adapters during inference. |
176
222
| backend_config.lora_adapters| A dictionary of LoRA adapters in the format `{name: path}`, where each path is a local or Hugging Face-hosted LoRA adapter directory. |
223
+
| backend_config.quantization_config| A dictionary specifying the desired `BitsAndBytesConfig`. Can be obtained by saving a `BitsAndBytesConfig` to JSON via `BitsAndBytesConfig.to_json_file(filename). Defaults to None.|
177
224
178
225
### sllm-cli delete
179
226
Delete deployed models by name, or delete specific LoRA adapters associated with a base model.
0 commit comments