Mika Okamoto

B.S. in Computer Science, Georgia Institute of Technology

headshot.jpeg

I am a final-year B.S. student in Computer Science at Georgia Tech, advised by Dr. Mark Riedl.

My research focuses on explainability for AI systems — understanding and interpreting the behavior of large language models and agentic workflows. I am particularly interested in human-centered approaches to explainability, where the goal is not just to produce explanations, but to make them useful and actionable for people interacting with AI. I also study LLM behavior more broadly, including how models reason, make decisions, and fail.

After graduation, I will be joining Decagon as a Member of Technical Staff, working on AI agents for enterprise customer service.

news

Mar 2026 Two papers accepted to the Human-Centered Explainable AI (HCXAI) 2026 workshop at CHI: Explainable Model Routing for Agentic Workflows and Counterfactual Explanations for Agentic Workflows as spotlight posters!
May 2025 FLaME (Holistic Finance Language Model Evaluation) accepted to ACL Findings 2025!
Apr 2025 Trust by Design: Skill Profiles for Transparent, Cost-Aware LLM Routing (BELLA) accepted as a poster at MLSys YPS 2025!
Mar 2025 Excited to be joining Two Sigma Investments as a Software Engineering Intern this summer (2025)!

selected publications

See all publications

† denotes equal contribution

  1. Explainable Model Routing for Agentic Workflows
    Mika Okamoto, Ansel Erol, and Mark Riedl
    In Workshop on Human-Centered Explainable AI (HCXAI) at CHI, 2026

    Agentic AI systems increasingly route subtasks across multiple specialized models, but these routing decisions are opaque to end users. We propose a framework for generating natural language explanations of model routing decisions in agentic workflows, enabling users to understand why tasks are delegated to specific models and supporting meaningful human oversight of automated multi-step pipelines.

  2. Counterfactual Explanations for Agentic Workflows
    Madhuri Singh, Grace Kim, Mika Okamoto, Aarushi Ammavajjala, Amal Alabdulkarim, Gennie Mansi, and Mark Riedl
    In Workshop on Human-Centered Explainable AI (HCXAI) at CHI, 2026

    We introduce counterfactual explanation methods for multi-step agentic AI systems, providing users with actionable insights into how workflow outcomes could have differed under alternative conditions. Our approach generates counterfactuals that help users identify failure modes and build appropriate trust in automated agentic pipelines.

  3. ACL
    FLaME: Holistic Finance Language Model Evaluation
    Glenn Matlin, Mika Okamoto, Huzaifa Pardawala, Yang Yang, and Sudheer Chava
    In Findings of the Association for Computational Linguistics (ACL), 2025

    We introduce FLaME, a comprehensive benchmark for evaluating large language models on a broad range of financial NLP tasks, including sentiment analysis, numerical reasoning, document understanding, and question answering. Our evaluation of frontier models reveals significant gaps between general LLM capabilities and the demands of real-world financial analysis.

  4. Trust by Design: Skill Profiles for Transparent, Cost-Aware LLM Routing
    Mika Okamoto, Ansel Erol, and Glenn Matlin
    In MLSys Young Professionals Symposium (YPS), 2025

    We introduce BELLA, a routing system that constructs interpretable skill profiles for LLMs to match incoming queries to the most capable and cost-efficient model. By making routing decisions transparent and grounded in empirically measured model competencies, BELLA achieves competitive task performance while reducing inference costs and providing users with clear rationale for model selection.

  5. DeepVerify: Evidence-Based Expert-Level Scientific Claim Verification
    Siheng Xiong, Oguzhan Gungordu, Blair Johnson, Mika Okamoto, James C. Kerce, and Faramarz Fekri
    2026

    DeepVerify is an agentic framework for expert-level scientific claim verification, combining search, tool use, and structured reasoning to ground decisions in retrieved evidence.