Abstract

In this article, we consider how to increase trust in development processes in which there is a risk for adversarial manipulation, and the adversary's objectives and resources are either ill-specified, imprecisely specified, or unknown. In such problems, we must hedge against the risk of misapprehension of attacker objectives and resources, which is further complicated in the absence of adversarial training data. We show how to model dynamic agent interaction, on the basis of partially observed or noise-corrupted data, using a partially observable Markov game (POMG) framework. We then propose a threefold heuristic solution procedure that: 1) uses the POMG to generate potential adversarial policies; 2) explicitly incorporates these adversarial policies in the construction of a robust defender policy by solving a robust dynamic program (DP); and 3) employs a probability matching heuristic in partially observable environments.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.