Industry
Enterprise
Business Operations
Internal Tool
Skills
Prototyping
Conversational AI Design
AI-Native Product Thinking
Team
Timeline
Jan - Mar 2025
TL;DR

PROBLEM
UNDERSTANDING USER NEEDS
My two primary users were Managed Services leadership and Ops Managers. My PM interviewed our key users to understand how leadership currently reviews metrics and received direct stakeholder feedback. We discovered that leadership reviews metrics via static reports, workbooks, and exports, but data entry is time-consuming and disorganized, so tracking performance metrics was not consistent.
Trust was the bottleneck
Metrics were calculated using undocumented business rules and embedded in hidden worksheets. As a result, leadership lacked confidence in performance monitoring data.
Leadership wanted on-demand answers based on existing data
A conversational chatbot could interpret data and establish trust better than a dashboard
"AI features must solve real problems, not be implemented for novelty. Unnecessary AI chatbots and features can harm rather than help users."
Norman Nielson Group
"Be skeptical of the marketing claims being made by AI tools designed for UX researchers. Many of these systems are not able to do everything they claim."
Norman Nielson Group
Played around with this, but ultimately decided against it to control model behavior and prevent off-domain or unverifiable responses
Could help user track data entries used in calculations, boosting trust and credibility
Prioritized for future phase
Thinking/reasoning experience to show how the model is calculating the data
Chat responses followed structured template:
Time-based comparison (yesterday / 7d / 14d / 30d)
Trend interpretation
Operational explanation (what likely changed)
During review, we discovered that Ops Managers needed to ask micro-level and mid-level questions whereas leadership only needed to ask macro-level questions
We also discovered users needed to double-check the data used against the calculation. To fulfill this user need, I extended the design so that any metric answer could display calculation logic and source attribution.
Improving success rate of conversational experience & decreasing org costs for tokens
TESTING
Reduced ad-hoc reporting requests to analysts
Faster leadership decision cycles (staffing, tooling, escalation response)
Higher trust signals
Proof Mode usage rate










