Lead the creation of a unified framework that lets MARL agents decide how much explanation to produce, generate counterfactual scenarios on the fly, and embed audit‑ready logs—all while keeping inference cost minimal. Your work will be the safety net that keeps agents trustworthy under adversarial conditions.
You will integrate lightweight uncertainty estimation (MC‑Dropout, ensembles) with LLM inference in a single, latency‑bounded pipeline, a capability that has never been demonstrated at scale in multi‑agent RL.
Uncertainty‑Driven Explanation Budgeting, LLM‑Generated Counterfactual Reward Shaping, and Continuous Auditing
From: Explainability Budget Optimization for Sample Efficiency
This role orchestrates the dynamic allocation of explanation resources, LLM‑guided counterfactual generation, and real‑time compliance logging—critical for safety, regulatory alignment, and sample‑efficiency.
An end‑to‑end system that estimates per‑decision uncertainty, decides token budget allocation, generates counterfactual explanations via LLMs, shapes rewards, and logs immutable audit trails.
PhD in Machine Learning, Computer Science, or related field.
Within 12 months, deliver a fully autonomous budget‑aware explanation engine that reduces human‑review workload by 70% and cuts sample complexity by 40% on a live MARL deployment.
Expand the framework to support multi‑modal agents, scale to thousands of concurrent users, and lead the company’s compliance‑first AI strategy.
If this sounds like the challenge you have been looking for, we want to hear from you. We value what you can build over where you have been.