There is a growing trend towards autonomy in present and forthcoming computing applications, including web-services and autonomous vehicles. Many of these applications are based on the concept of autonomous agent.
The group focuses on developing methods aimed at verifying that autonomous multi-agent systems meet their specifications. Specifically, the group is concerned with developing efficient model checking techniques and tools to verify multi-agent systems specified by agent-based logics. The research draws from areas such as modal logic, multi-agent systems, and model checking.
More recently, systems based on neural networks have become of increasing importance. The group is now also investigating techniques to allow verification of systems based on neural networks. This research is partly funded by DARPA’s Assured Autonomy program.
We are always looking for passionate new PhD students, Postdocs, and Master students to join the team (more info) !
Verification of ReLU-based Neural Networks
21 October 2019, 1pm
Room 361, Huxley Building
Three new PhD students join the group!01 March 2019
Safe & Trusted AI CDT23 January 2019
VAS members have two papers accepted at AAMAS 201901 November 2018
VAS members have two papers accepted at AAAI 201908 October 2018
Elena Botoeva joins the group as a postdoctoral researcher