Adversarial Robustness and Privacy
Even advanced AI systems can be vulnerable to adversarial attacks. We’re making tools to protect AI and certify its robustness, including quantifying the vulnerability of neural networks and designing new attacks to make better defenses. And we’re helping AI systems adhere to privacy requirements.
Our work
What is red teaming for generative AI?
ExplainerKim MartineauAn open-source toolkit for debugging AI models of all data types
Technical noteKevin Eykholt and Taesung LeeDid an AI write that? If so, which one? Introducing the new field of AI forensics
ExplainerKim MartineauManipulating stock prices with an adversarial tweet
ResearchKim MartineauSecuring AI systems with adversarial robustness
Deep DivePin-Yu Chen8 minute readResearchers develop defenses against deep learning hack attacks
ReleaseAmbrish Rawat, Killian Levacher, and Mathieu Sinn7 minute read- See more of our work on Adversarial Robustness and Privacy
Publications
- Kristjan Greenewald
- Yuancheng Yu
- et al.
- 2024
- NeurIPS 2024
- Noam Koren
- Abigail Goldsteen
- et al.
- 2024
- ACML 2024
- 2024
- AIES 2024
- Ching-yun Ko
- Pin-Yu Chen
- et al.
- 2024
- COLM 2024
- Zhiyuan He
- Yijun Yang
- et al.
- 2024
- ICML 2024
- Zhi-yi Chin
- Chieh-ming Jiang
- et al.
- 2024
- ICML 2024