Beyond RAG: A LLM-Based FAQ Matching Framework for Real-Time Decision Support in Contact Centers
Loading...
Files
Date
Contributor
Advisor
Editor
Performer
Department
Instructor
Depositor
Speaker
Researcher
Consultant
Interviewer
Interviewee
Narrator
Transcriber
Annotator
Journal Title
Journal ISSN
Volume Title
Publisher
Journal Name
Volume
Number/Issue
Starting Page
1786
Ending Page
Alternative Title
Abstract
In customer contact centers, human agents often face long average handling times (AHT) due to the need to manually interpret queries and search large knowledge bases (KBs). While retrieval-augmented generation (RAG) systems using large language models (LLMs) are increasingly adopted to support these tasks, they face limitations in real-time conversations—particularly with poorly formulated queries and repeated retrieval of frequently asked questions (FAQs). To address these issues, we propose a decision support framework that extends beyond RAG by combining real-time question identification with a dual-threaded FAQ matching and generation system. If the query matches a FAQ, the answer is retrieved instantly; otherwise a well-formed query is generated and routed to a RAG model. Deployed within Minerva CQ’s human-agent assist platform, our solution delivers sub-2-second responses for matched queries, significantly reduces unnecessary RAG calls, and lowers operational costs. We also introduce an automated, LLM-agentic pipeline for mining FAQs from historical transcripts, enabling continuous improvement of the FAQ knowledge base in the absence of manually curated QA pairs.
Description
Citation
DOI
Extent
8 pages
Format
Type
Conference Paper
Geographic Location
Time Period
Related To
Proceedings of the 59th Hawaii International Conference on System Sciences
Related To (URI)
Table of Contents
Rights
Attribution-NonCommercial-NoDerivatives 4.0 International
Rights Holder
Catalog Record
Local Contexts
Email libraryada-l@lists.hawaii.edu if you need this content in ADA-compliant format.
