Ruoqi Liu, Pin-Yu Chen, et al.
Patterns
Adversarial robustness studies the worst-case performance of a machine learning model to ensure safety and reliability. With the proliferation of deep-learning-based technology, the potential risks associated with model development and deployment can be amplified and become dreadful vulnerabilities. This paper provides a comprehensive overview of research topics and foundational principles of research methods for adversarial robustness of deep learning models, including attacks, defenses, verification, and novel applications.
Ruoqi Liu, Pin-Yu Chen, et al.
Patterns
Saiteja Utpala, Alex Gu, et al.
NAACL 2024
Megh Thakkar, Quentin Fournier, et al.
ACL 2024
Chih-kai Ting, Karl Munson, et al.
AAAI 2023