Resolving the Chatbot Disclosure Dilemma: Leveraging Selective Self-Presentation to Mitigate the Negative Effect of Chatbot Disclosure

Date
2021-01-05
Authors
Mozafari, Nika
Weiger, Welf H.
Hammerschmidt, Maik
Contributor
Advisor
Department
Instructor
Depositor
Speaker
Researcher
Consultant
Interviewer
Annotator
Journal Title
Journal ISSN
Volume Title
Publisher
Volume
Number/Issue
Starting Page
2916
Ending Page
Alternative Title
Abstract
Chatbots are increasingly able to pose as humans. However, this does not hold true if their identity is explicitly disclosed to users—a practice that will become a legal obligation for many service providers in the imminent future. Previous studies hint at a chatbot disclosure dilemma in that disclosing the non-human identity of chatbots comes at the cost of negative user responses. As these responses are commonly attributed to reduced trust in algorithms, this research examines how the detrimental impact of chatbot disclosure on trust can be buffered. Based on computer-mediated communication theory, the authors demonstrate that the chatbot disclosure dilemma can be resolved if disclosure is paired with selective presentation of the chatbot’s capabilities. Study results show that while merely disclosing (vs. not disclosing) chatbot identity does reduce trust, pairing chatbot disclosure with selectively presented information on the chatbot’s expertise or weaknesses is able to mitigate this negative effect.
Description
Keywords
Mediated Conversation, chatbot disclosure, computer-mediated communication, selective self-presentation, trust
Citation
Extent
8 pages
Format
Geographic Location
Time Period
Related To
Proceedings of the 54th Hawaii International Conference on System Sciences
Table of Contents
Rights
Attribution-NonCommercial-NoDerivatives 4.0 International
Rights Holder
Local Contexts
Email libraryada-l@lists.hawaii.edu if you need this content in ADA-compliant format.