Design and implement a self‑learning LLM system that crafts semantic adversarial prompts to stress‑test multi‑agent policies, driving robust learning and uncovering hidden failure modes.
You’ll create an LLM‑based red‑team that can autonomously generate instruction‑level attacks, a frontier concept in AI safety and policy robustness that pushes beyond gradient‑based adversaries.
LLM‑Driven Adversarial Curriculum (LLM‑AC) for semantic adversarial scenario generation
From: Adversarial Observation Perturbations and Policy Inference
Key to expose policy brittleness through high‑level semantic attacks and to generate diverse, maximally regretful scenarios.
Attacker‑target‑judge loop harnessing LLM‑TOC, retrieval‑augmented generation pipeline, curriculum scheduling, and integration with MARL training.
PhD or MS in NLP, AI Safety, or Robotics with experience in LLM fine‑tuning and adversarial generation.
Within a year, produce a curriculum that increases policy resilience by 40% against unseen semantic attacks in benchmark multi‑agent environments, directly enhancing mission reliability.
Expand the curriculum framework to other domains (e.g., autonomous driving, cyber‑defense) and lead a growing team of LLM engineers focused on adversarial robustness.
If this sounds like the challenge you have been looking for, we want to hear from you. We value what you can build over where you have been.