Zizhan Zheng, Computer Science, Tulane University

Learning to Secure Cooperative Multi-Agent Learning Systems: Advanced Attacks and Robust Defenses

Personnel

Goals

Cooperative multi-agent learning (MAL), where multiple intelligent agents learn to coordinate with each other and with humans, is emerging as a promising paradigm for solving some of the most challenging problems in various security and safety-critical domains, including transportation, power systems, robotics, and healthcare. The decentralized nature of MAL systems and agents' exploration behavior, however, introduce new vulnerabilities unseen in standalone machine learning systems and traditional distributed systems. This project aims to develop a data-driven approach to MAL security that can provide an adequate level of protection even in the presence of persistent, coordinated, and stealthy insiders or external adversaries. The main novelty of the project is to go beyond heuristics-based attack and defense schemes by incorporating opponent modeling and adaptation into security-related decision-making in a principled way. The project contributes to the emerging fields of science of security and trustworthy artificial intelligence via a cross-disciplinary approach that integrates cybersecurity, multi-agent systems, machine learning, and cognitive science.

Tasks

First, we develop learning-based targeted and untargeted attacks against federated and decentralized machine learning. These attacks first infer a "world model" from publicly available data and then apply reinforcement learning with properly defined state and action spaces and reward function to identify an adaptive attack policy that can fully exploit the system vulnerabilities.

Second, we investigate Stackelberg Markov games with asymmetric observations as a principled framework for achieving proactive defenses for cooperative learning systems against unknown/uncertain attacks. Our defenses integrate adversarial training and meta-learning by utilizing the automated attack framework developed in the first task as a simulator of (strong) adversaries.

Third, we study security in multi-agent reinforcement learning systems by addressing a set of new challenges, including complicated interactions among agents, non-stationarity, and partial observability. The goal is to understand how malicious attacks and deceptions can prevent benign agents from reaching a socially preferred outcome and how accounting for a higher order of beliefs can help an agent (benign or malicious) in both fully cooperative and mixed-motive settings.

Publications

Support

The project is funded by National Science Foundation (NSF) CAREER award CNS-2146548.

Disclaimer: Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation.