Investigating the Effects of Classification Model Error Type on Trust-relevant Criteria in a Human-Machine Learning Interaction Task
Loading...
Files
Date
Contributor
Advisor
Editor
Performer
Department
Instructor
Depositor
Speaker
Researcher
Consultant
Interviewer
Interviewee
Narrator
Transcriber
Annotator
Journal Title
Journal ISSN
Volume Title
Publisher
Journal Name
Volume
Number/Issue
Starting Page
297
Ending Page
Alternative Title
Abstract
Machine learning models have been critiqued for their opaqueness, so recent work has created models to accurately convey model confidence. High performance is the most important aspect of trust in the model. However, when performance drops, accurate decision confidence leads to higher trust outcomes. The current research expands upon this work investigating how incorrect, low confidence decisions differentially impact the trust process. Incorrect decisions were either made on stimuli the model was trained to classify or stimuli outside those classification categories. In a between-subjects design, participants monitored low performing models of varying low confidence error type in an online image classification task. Results demonstrated when the model flagged incorrect stimuli it was not trained to classify with low confidence, process perceptions increased, while decision time and task performance decreased. Our results extend the current framework regarding how model confidence influences the trust process. Implications, limitations, and future research are discussed.
Description
Citation
Extent
10
Format
Type
Conference Paper
Geographic Location
Time Period
Related To
Proceedings of the 58th Hawaii International Conference on System Sciences
Related To (URI)
Table of Contents
Rights
Attribution-NonCommercial-NoDerivatives 4.0 International
Rights Holder
Catalog Record
Local Contexts
Email libraryada-l@lists.hawaii.edu if you need this content in ADA-compliant format.
