Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

"4-bit finetuning” is ambiguous #5

Open
jzhang38 opened this issue May 14, 2023 · 1 comment
Open

"4-bit finetuning” is ambiguous #5

jzhang38 opened this issue May 14, 2023 · 1 comment

Comments

@jzhang38
Copy link

jzhang38 commented May 14, 2023

Dear authors, can I ask what 4-bit finetuning exactly refer to?
A. Use 16-bit mixed precision to fine-tune LoRA on a 4-bit base model.
B. Use 4-bit mixed precision to fine-tune LoRA on a 4-bit base model.
C. Use 4-bit mixed prevision to fine-tune a full model.

@jzhang38
Copy link
Author

and maybe another slightly dumb qn: how do you store the 4-bit parameters in the GPU memory? Because 4-bit is not directly addressable.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant