-
Notifications
You must be signed in to change notification settings - Fork 1.9k
🤧 LD-DPO support #3458
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
🤧 LD-DPO support #3458
Conversation
congrats! can you also kindly add the method in the dpo documentation with some description? |
done |
also you can add a small test too? |
Sorry, i dont know how to test if it is effective. T_T |
no problem, I can do it |
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks, LGTM!
What does this PR do?
This PR adds LD-DPO implementations to TRL. This paper has been accepted by CoRR 2024
This paper proposed a novel method aims to desensitize DPO to data length by decoupling explicit length preference, which is relatively insignificant, from the other implicit preferences, thereby enabling more effective learning of the intrinsic preferences
Before submitting
Pull Request section?
to it if that's the case.
Who can review?
Anyone in the community is free to review the PR once the tests have passed. Feel free to tag
members/contributors who may be interested in your PR.