23/06/2021

Fast and Precise Certification of Transformers

Gregory Bonaert, Dimitar I. Dimitrov, Maximilian Baader, Martin Vechev

Keywords: Abstract Interpretation, Robustness Certification, Deep Learning, Adversarial attacks, Transformer Networks

Abstract: We present DeepT, a novel method for certifying Transformer networks based on abstract interpretation. The key idea behind DeepT is our new Multi-norm Zonotope abstract domain, an extension of the classical Zonotope designed to handle ℓ1 and ℓ2-norm bound perturbations. We introduce all Multi-norm Zonotope abstract transformers necessary to handle these complex networks, including the challenging softmax function and dot product. Our evaluation shows that DeepT can certify average robustness radii that are 28× larger than the state-of-the-art, while scaling favorably. Further, for the first time, we certify Transformers against synonym attacks on long sequences of words, where each word can be replaced by any synonym. DeepT achieves a high certification success rate on sequences of words where enumeration-based verification would take 2 to 3 orders of magnitude more time.

The video of this talk cannot be embedded. You can watch it here:
https://slideslive.com/38956281
(Link will open in new window)
 0
 0
 0
 0
This is an embedded video. Talk and the respective paper are published at PLDI 2021 virtual conference. If you are one of the authors of the paper and want to manage your upload, see the question "My papertalk has been externally embedded..." in the FAQ section.

Comments

Post Comment
no comments yet
code of conduct: tbd

Similar Papers