top | item 36452277

DPO: Direct Preference Optimization

3 points| Garcia98 | 2 years ago |github.com

discuss

order

No comments yet.