CMU Artificial Intelligence Seminar Series sponsored by


Back to Seminar Schedule

Tuesday, Nov 23, 2021

Time: 12:00 - 01:00 PM ET
Recording of this Online Seminar on Youtube

Eric Wallace -- What Can We Learn From Vulnerabilities of ML Models?

Relevant Paper(s):

Abstract: Today's neural network models achieve high accuracy on in-distribution data and are being widely deployed in production systems. This talk will discuss attacks on such models that not only expose worrisome security and privacy vulnerabilities, but also provide new perspectives into how and why the models work. Concretely, I will show how realistic adversaries can extract secret training data, steal model weights, and manipulate test predictions, all using black-box access to models at either training- or test-time. These attacks will reveal different insights, including how NLP models rely on dataset biases and spurious correlations, and how their training dynamics impact memorization of examples. Finally, I will discuss defenses against these vulnerabilities and suggest practical takeaways for developing secure ML systems.

Bio: Eric Wallace is a 3rd year PhD student at UC Berkeley advised by Dan Klein and Dawn Song. His research interests center around large language models and making them more secure, private, and robust. Eric's work received the best demo award at EMNLP 2019.