ServiceNow Research

Adversarial Attacks

Constraining Representations Yields Models That Know What They Don't Know
A well-known failure mode of neural networks is that they may confidently return erroneous predictions. Such unsafe behaviour is …
Maximal Jacobian-based Saliency Map Attack
The Jacobian-based Saliency Map Attack is a family of adversarial attack methods for fooling classification models, such as deep neural …