Is BERT Really Robust? A Strong Baseline for Natural Language Attack on Text Classification and Entailment

27 Jul 2019Di JinZhijing JinJoey Tianyi ZhouPeter Szolovits

Machine learning algorithms are often vulnerable to adversarial examples that have imperceptible alterations from the original counterparts but can fool the state-of-the-art models. It is helpful to evaluate or even improve the robustness of these models by exposing the maliciously crafted adversarial examples... (read more)

PDF Abstract

Evaluation results from the paper


  Submit results from this paper to get state-of-the-art GitHub badges and help community compare results to other papers.