DPO: Direct Preference Optimization

computer science
Author

Luca Simonetti

Published

February 25, 2025