Formal verification is often dismissed as too rigid, complex, or unscalable for frontier AI systems, including LLMs, VLMs, and agentic systems. In practice, this perception has led many researchers and developers to rely on less rigorous alternatives such as benchmarking, red teaming, and adversarial attacks to evaluate safety and performance.
In this talk, I will present a new class of efficient formal verification methods for frontier AI systems that certify safety properties such as secure code generation and catastrophic conversational risks, delivering stronger generalization guarantees than standard evaluation approaches. These results position formal verification as a necessary foundation for building reliable and trustworthy systems at scale, and demand sustained, substantial investment.
Papers / readings:
- BEAVER: An Efficient Deterministic LLM Verifier
https://arxiv.org/abs/2512.05439