-
Notifications
You must be signed in to change notification settings - Fork 47
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fix QLoRA weights and bias initialisation #833
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hi @caroteu,
I left some comments below. In general, I feel using strict=False
is a bit tricky (and error-prone) to set it for loading model checkpoints. We should discuss the issues and find an elegant way to take care of this!
@anwai98 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hi @caroteu,
I merged the previous PR here. Can we briefly touch down on the comments below?
Does the inference for QLoRA-finetuned models properly!
I added docstrings to our new export function (and thanks to @caroteu for making an elegant solution out of this). This is GTG from my side. PS. @caroteu @constantinpape It would be nice to go over this and see if everything makes sense now! |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This looks good to me. Feel free to merge.
@anwai98 This includes the QLoRA initialisation changes and this also works for inference now. The changes that made the inference work are a bit hacky though, so we should think about a better/safer way to do this.