{"version.version":"0.0.7","version.cm_url":"https://github.com/ceurws/ceur-spt","spt.html_url":"/Vol-3808/paper7.html","spt.description":null,"spt.id":"Vol-3808/paper7","spt.wikidataid":null,"spt.title":"Mitigating Toxicity in Dialogue Agents through Adversarial Reinforcement Learning","spt.pdfUrl":"https://ceur-ws.org/Vol-3808/paper7.pdf","spt.volume":{"number":3808,"acronym":"AEQUITAS 2024","wikidataid":"Q131010072","title":"Proceedings of the 2nd Workshop on Fairness and Bias in AI","description":"Proceedings of AEQUITAS 2024 workshop","url":"http://ceur-ws.org/Vol-3808/","date":"2024-10-29","dblp":null,"k10plus":null,"urn":"urn:nbn:de:0074-3808-1"},"spt.session":null,"cvb.id":"Vol-3808/paper7","cvb.title":"Mitigating Toxicity in Dialogue Agents through Adversarial Reinforcement Learning","cvb.type":null,"cvb.position":null,"cvb.pagesFrom":null,"cvb.pagesTo":null,"cvb.authors":"Guillermo Villate-Castillo,Borja Sanz,Javier Del Ser","cvb.vol_number":"3808","cvb.pdf_name":"paper7.pdf","cvb.pages":null,"cvb.fail":null,"dblp.dblp_publication_id":"https://dblp.org/rec/conf/aequitas/Villate-Castillo24","dblp.dblp_proceeding_id":"https://dblp.org/rec/conf/aequitas/2024","dblp.volume_number":3808,"dblp.title":"Mitigating Toxicity in Dialogue Agents through Adversarial Reinforcement Learning.","dblp.authors":[{"dblp_author_id":"https://dblp.org/pid/94/1127","label":"Javier Del Ser","wikidata_id":"Q60532992","orcid_id":"0000-0002-1260-9775","gnd_id":null},{"dblp_author_id":"https://dblp.org/pid/86/7795","label":"Borja Sanz 0001","wikidata_id":"Q124609984","orcid_id":"0000-0003-2039-7773","gnd_id":null},{"dblp_author_id":"https://dblp.org/pid/389/5293","label":"Guillermo Villate-Castillo","wikidata_id":null,"orcid_id":null,"gnd_id":null}],"dblp.pdf_id":"Vol-3808/paper7"}