Mostrar el registro sencillo del ítem

Ponencia

dc.contributor.editorVarela Vaca, Ángel Jesúses
dc.contributor.editorCeballos Guerrero, Rafaeles
dc.contributor.editorReina Quintero, Antonia Maríaes
dc.creatorGhafouri, Vahides
dc.creatorAgarwal, Vibhores
dc.creatorZhang, Yonges
dc.creatorSastry, Nishanthes
dc.creatorSuch, Josées
dc.creatorSuarez Tangil, Guillermoes
dc.date.accessioned2024-08-26T11:09:53Z
dc.date.available2024-08-26T11:09:53Z
dc.date.issued2024
dc.identifier.citationGhafouri, V., Agarwal, V., Zhang, Y., Sastry, N., Such, J. y Suarez Tangil, G. (2024). AI in the Gray: Exploring Moderation Policies in Dialogic Large Language Models vs. Human Answers in Controversial Topics [Póster]. En Jornadas Nacionales de Investigación en Ciberseguridad (JNIC) (9ª.2024. Sevilla) (486-487), Sevilla: Universidad de Sevilla. Escuela Técnica Superior de Ingeniería Informática.
dc.identifier.isbn978-84-09-62140-8es
dc.identifier.urihttps://hdl.handle.net/11441/162049
dc.description.abstractThe increasing sophistication of Large Language Models (LLMs), particularly ChatGPT, has revolutionized how users interact with information and make decisions. However, when addressing controversial topics without universally ac cepted answers, such as religion, gender identity, or freedom of speech, these models face the challenge of potential bias. Biased responses in these complex domains can amplify misinformation, fuel harmful ideologies, and undermine trust in AI systems. This paper investigates the biases embedded within LLMs like ChatGPT when responding to controversial questions. We use the Kialo social debate platform as a benchmark, comparing AI generated responses to human discussions. Our analysis reveals significant progress in reducing explicit biases in recent ChatGPT versions. However, residual implicit biases, including subtle right-wing leanings, call for further moderation. These findings hold substantial cybersecurity implications, emphasizing the need to mitigate the spread of misinformation or the promotion of extremist viewpoints through AI-powered systems.es
dc.formatapplication/pdfes
dc.format.extent2es
dc.language.isoenges
dc.publisherUniversidad de Sevilla. Escuela Técnica Superior de Ingeniería Informáticaes
dc.relation.ispartofJornadas Nacionales de Investigación en Ciberseguridad (JNIC) (9ª.2024. Sevilla) (2024), pp. 486-487.
dc.rightsAttribution-NonCommercial-NoDerivatives 4.0 Internacional*
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/4.0/*
dc.subjectChatGPTes
dc.subjectLLMses
dc.subjectModeration Policieses
dc.subjectKialoes
dc.subjectSocial Networkses
dc.titleAI in the Gray: Exploring Moderation Policies in Dialogic Large Language Models vs. Human Answers in Controversial Topics [Póster]es
dc.typeinfo:eu-repo/semantics/conferenceObjectes
dc.type.versioninfo:eu-repo/semantics/publishedVersiones
dc.rights.accessRightsinfo:eu-repo/semantics/openAccesses
dc.publication.initialPage486es
dc.publication.endPage487es
dc.eventtitleJornadas Nacionales de Investigación en Ciberseguridad (JNIC) (9ª.2024. Sevilla)es
dc.eventinstitutionSevillaes
dc.relation.publicationplaceSevillaes

FicherosTamañoFormatoVerDescripción
JNIC24_504.pdf1.152MbIcon   [PDF] Ver/Abrir  

Este registro aparece en las siguientes colecciones

Mostrar el registro sencillo del ítem

Attribution-NonCommercial-NoDerivatives 4.0 Internacional
Excepto si se señala otra cosa, la licencia del ítem se describe como: Attribution-NonCommercial-NoDerivatives 4.0 Internacional