SafeRLHF

The dataset used in the paper for training and testing the DPO and PPO models.

BibTex: