You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Feb 18, 2026. It is now read-only.
description="Scale the base LLM's context length by this factor using RoPE scaling. Only works if use_unsloth is set to True.",
1030
-
ge=1,
1031
-
le=6,
1053
+
description="Scale the base LLM's context length by this factor using RoPE scaling.",
1032
1054
title="rope_scaling_factor",
1033
1055
),
1034
-
] =1
1056
+
] ="auto"
1035
1057
1036
1058
1037
1059
classMaxSequencesPerExample(str, Enum):
@@ -1294,7 +1316,7 @@ class PeftParams(ConfigBase):
1294
1316
),
1295
1317
] =1
1296
1318
target_modules: Annotated[
1297
-
Optional[Union[str, List[str]]],
1319
+
Optional[Union[List[str], str]],
1298
1320
Field(
1299
1321
description="List of module names or regex expression of the module names to replace with LoRA. For example, ['q', 'v'] or '.*decoder.*(SelfAttention|EncDecAttention).*(q|v)$'. This can also be a wildcard 'all-linear' which matches all linear/Conv1D layers except the output layer. If not specified, modules will be chosen according to the model architecture. If the architecture is not known, an error will be raised -- in this case, you should specify the target modules manually.",
1300
1322
title="Target Modules",
@@ -1546,11 +1568,11 @@ class Column(ConfigBase):
1546
1568
Optional[str], Field(description="Rename to value.", title="Value")
0 commit comments