Dataset Groups Activity Stream Anthropic red-team dataset The Anthropic red-team dataset is a significant open-access dataset aimed at improving AI safety through training preference models and assessing their safety. BibTex: @dataset{Bahareh_Harandizadeh_and_Abel_Salinas_and_Fred_Morstatter_2024, abstract = {The Anthropic red-team dataset is a significant open-access dataset aimed at improving AI safety through training preference models and assessing their safety.}, author = {Bahareh Harandizadeh and Abel Salinas and Fred Morstatter}, doi = {10.57702/bup1brhp}, institution = {No Organization}, keyword = {'AI safety', 'dataset', 'preference models', 'red-team'}, month = {dec}, publisher = {TIB}, title = {Anthropic red-team dataset}, url = {https://service.tib.eu/ldmservice/dataset/anthropic-red-team-dataset}, year = {2024} }