-
Notifications
You must be signed in to change notification settings - Fork 4.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
DeepSeek 671B Fine-Tuning , how to merge lora model #6222
Comments
We used the following code for the conversion:
Our machine has 2TB of memory, The swap space is 2TB, and we have attempted to merge the LoRA weights, but it reported missing weights. There are over 1000 layers of weights missing after merging the LoRA weights compared to before.. |
what's your fine tune env? thks |
When merging LoRA weights on a machine with 8 GPUs and 1.7T memory, I keep getting OOM (Out Of Memory) errors. What should I do? |
Please have a look at this example: https://discuss.huggingface.co/t/help-with-merging-lora-weights-back-into-base-model/40968. We are working on providing a straightforward example on the github soon. |
after DeepSeek 671B Fine-Tuning , how to load the Original model and lora model to test ?
help ~
The text was updated successfully, but these errors were encountered: