showSidebars ==
showTitleBreadcrumbs == 1
node.field_disable_title_breadcrumbs.value ==

Pre-Conference Talk by WANG Haoyu | AgentSpec: Customizable Runtime Enforcement for Safe and Reliable LLM Agents

Please click here if you are unable to view this page.

 


AgentSpec: Customizable Runtime Enforcement for Safe and Reliable LLM Agents

Speaker (s):


WANG Haoyu
PhD Candidate
School of Computing and Information Systems
Singapore Management University

Date:

Time:

Venue:

 

7 April 2026, Tuesday

1:00pm – 1:30pm

Meeting room 4.4, Level 4
School of Computing and
Information Systems 1, 
Singapore Management University, 
80 Stamford Road,
Singapore 178902

We look forward to seeing you at this research seminar.

Please register by 5 April 2026.

About the Talk

Agents built on LLMs are increasingly deployed across diverse domains, automating complex decision-making and task execution. However, their autonomy introduces safety risks, including security vulnerabilities, legal violations, and unintended harmful actions. Existing mitigation methods, such as model-based safeguards and early enforcement strategies, fall short in robustness, interpretability, and adaptability. To address these challenges, we propose AgentSpec, a lightweight domain-specific language for specifying and enforcing runtime constraints on LLM agents. With AgentSpec, users define structured rules that incorporate triggers, predicates, and enforcement mechanisms, ensuring agents operate within predefined safety boundaries. We implement AgentSpec across multiple domains, including code execution, embodied agents, and autonomous driving, demonstrating its adaptability and effectiveness. Our evaluation shows that AgentSpec successfully prevents unsafe executions in over 90% of code agent cases, eliminates all hazardous actions in embodied agent tasks, and enforces 100% compliance by autonomous vehicles (AVs). Despite its strong safety guarantees, AgentSpec remains computationally lightweight, with overheads in milliseconds. By combining interpretability, modularity, and efficiency, AgentSpec provides a practical and scalable solution for enforcing LLM agent safety across diverse applications. We also automate the generation of rules using LLMs and assess their effectiveness. Our evaluation shows that the rules generated by OpenAI o1 achieve a precision of 95.56% and recall of 70.96% for embodied agents, successfully identify 87.26% of the risky code, and prevent AVs from breaking laws in 5 out of 8 scenarios.

This is a Pre-Conference talk for IEEE/ACM International Conference on Software Engineering (ICSE 2026).

About the speaker

Haoyu WANG is currently a PhD candidate at the Singapore Management University co-advised by Prof. C.M. Poskitt and Prof. Jun Sun. His current research interest mainly focus on ensuring the safety and security for AI software. Besides, he also interested in fuzzing fundamental software systems including compilers, smart contracts, etc.