AI-Assisted and Explainable Hate Speech Detection for Social Media Moderators – A Design Science Approach
AI-Assisted and Explainable Hate Speech Detection for Social Media Moderators – A Design Science Approach
Files
Date
2021-01-05
Authors
Bunde, Enrico
Contributor
Advisor
Department
Instructor
Depositor
Speaker
Researcher
Consultant
Interviewer
Annotator
Journal Title
Journal ISSN
Volume Title
Publisher
Volume
Number/Issue
Starting Page
1264
Ending Page
Alternative Title
Abstract
To date, the detection of hate speech is still primarily carried out by humans, yet there is great potential for combining human expertise with automated approaches. However, identified challenges include low levels of agreement between humans and machines due to the algorithms’ missing expertise of, e.g., cultural, and social structures. In this work, a design science approach is used to derive design knowledge and develop an artifact, through which humans are integrated in the process of detecting and evaluating hate speech. For this purpose, explainable artificial intelligence (XAI) is utilized: the artifact will provide explanative information, why the deep learning model predicted whether a text contains hate. Results show that the instantiated design knowledge in form of a dashboard is perceived as valuable and that XAI features increase the perception of the artifact’s usefulness, ease of use, trustworthiness as well as the intention to use it.
Description
Keywords
Explainable Artificial Intelligence (XAI),
deep learning,
design science,
explainable artificial intelligence,
hate speech detection
Citation
Extent
10 pages
Format
Geographic Location
Time Period
Related To
Proceedings of the 54th Hawaii International Conference on System Sciences
Table of Contents
Rights
Attribution-NonCommercial-NoDerivatives 4.0 International
Rights Holder
Local Contexts
Collections
Email libraryada-l@lists.hawaii.edu if you need this content in ADA-compliant format.