You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
| model | This should be a deployed model name, used to identify the backend instance. |
324
+
| ft_backend | fine-tuning engine, only support `peft` now. |
325
+
| dataset_config | Config about the fine-tuning dataset |
326
+
| dataset_config.dataset_source | dataset is from `hf_hub` (huggingface_hub) or `local` file |
327
+
| dataset_config.hf_dataset_name | dataset name on huggingface_hub |
328
+
| dataset_config.tokenization_field | the field to tokenize |
329
+
| dataset_config.split | Partitioning of the dataset (`train`, `validation` and `test`), You can also split the selected dataset, e.g. take only the top 10% of the training data: train[:10%] |
330
+
| dataset_config.data_files | data files will be loaded from local|
331
+
| dataset_config.extension_type | extension type of data files (`csv`, `json`, `parquet`, `arrow`) |
332
+
| lora_config | Config about lora |
333
+
| lora_config.r |`r` defines how many parameters will be trained. |
334
+
| lora_config.lora_alpha | A multiplier controlling the overall strength of connections within a neural network, typically set at 1 |
335
+
| lora_config.target_modules | a list of the target_modules available on the [Hugging Face Documentation](https://github.com/huggingface/peft/blob/39ef2546d5d9b8f5f8a7016ec10657887a867041/src/peft/utils/other.py#L220) |
336
+
| lora_config.lora_dropout | used to avoid overfitting |
337
+
| lora_config.bias | use `none` or `lora_only`|
338
+
| lora_config.task_type | Indicates the task the model is begin trained for|
339
+
| training_config | Config about training parameters |
340
+
| training_config.auto_find_batch_size | Find a correct batch size that fits the size of Data. |
341
+
| training_config.num_train_epochs | Total number of training rounds |
0 commit comments