Mika Okamoto
B.S. in Computer Science, Georgia Institute of Technology
I am a final-year B.S. student in Computer Science at Georgia Tech, advised by Dr. Mark Riedl.
My research focuses on explainability for AI systems — understanding and interpreting the behavior of large language models and agentic workflows. I am particularly interested in human-centered approaches to explainability, where the goal is not just to produce explanations, but to make them useful and actionable for people interacting with AI. I also study LLM behavior more broadly, including how models reason, make decisions, and fail.
After graduation, I will be joining Decagon as a Member of Technical Staff, working on AI agents for enterprise customer service.
news
| Mar 2026 | Two papers accepted to the Human-Centered Explainable AI (HCXAI) 2026 workshop at CHI: Explainable Model Routing for Agentic Workflows and Counterfactual Explanations for Agentic Workflows as spotlight posters! |
|---|---|
| May 2025 | FLaME (Holistic Finance Language Model Evaluation) accepted to ACL Findings 2025! |
| Apr 2025 | Trust by Design: Skill Profiles for Transparent, Cost-Aware LLM Routing (BELLA) accepted as a poster at MLSys YPS 2025! |
| Mar 2025 | Excited to be joining Two Sigma Investments as a Software Engineering Intern this summer (2025)! |
selected publications
† denotes equal contribution
- Explainable Model Routing for Agentic WorkflowsIn Workshop on Human-Centered Explainable AI (HCXAI) at CHI, 2026
Agentic AI systems increasingly route subtasks across multiple specialized models, but these routing decisions are opaque to end users. We propose a framework for generating natural language explanations of model routing decisions in agentic workflows, enabling users to understand why tasks are delegated to specific models and supporting meaningful human oversight of automated multi-step pipelines.
- Counterfactual Explanations for Agentic WorkflowsIn Workshop on Human-Centered Explainable AI (HCXAI) at CHI, 2026
We introduce counterfactual explanation methods for multi-step agentic AI systems, providing users with actionable insights into how workflow outcomes could have differed under alternative conditions. Our approach generates counterfactuals that help users identify failure modes and build appropriate trust in automated agentic pipelines.
- FLaME: Holistic Finance Language Model EvaluationIn Findings of the Association for Computational Linguistics (ACL), 2025
We introduce FLaME, a comprehensive benchmark for evaluating large language models on a broad range of financial NLP tasks, including sentiment analysis, numerical reasoning, document understanding, and question answering. Our evaluation of frontier models reveals significant gaps between general LLM capabilities and the demands of real-world financial analysis.
- Trust by Design: Skill Profiles for Transparent, Cost-Aware LLM RoutingIn MLSys Young Professionals Symposium (YPS), 2025
We introduce BELLA, a routing system that constructs interpretable skill profiles for LLMs to match incoming queries to the most capable and cost-efficient model. By making routing decisions transparent and grounded in empirically measured model competencies, BELLA achieves competitive task performance while reducing inference costs and providing users with clear rationale for model selection.