Organizing Committee

Deep learning is profoundly reshaping the research directions of entire scientific communities across mathematics, computer science, and statistics, as well as the physical, biological and medical sciences . Yet, despite their indisputable success, deep neural networks are known to be universally unstable. That is, small changes in the input that are almost undetectable produce significant changes in the output. This happens in applications such as image recognition and classification, speech and audio recognition, automatic diagnosis in medicine, image reconstruction and medical imaging as well as inverse problems in general. This phenomenon is now very well documented and yields non-human-like behaviour of neural networks in the cases where they replace humans, and unexpected and unreliable behaviour where they replace standard algorithms in the sciences.

The many examples produced over the last years demonstrate the intricacy of this complex problem and the questions of safety and security of deep learning become crucial. Moreover, the ubiquitous phenomenon of instability combined with the lack of interpretability of deep neural networks makes the reproducibility of scientific results based on deep learning at stake.

For these reasons, the development of mathematical foundations aimed at improving the safety and security of deep learning is of key importance. The goal of this workshop is to bring together experts from mathematics, computer science, and statistics in order to accelerate the exploration of breakthroughs and of emerging mathematical ideas in this area.

This workshop is fully funded by a Simons Foundation Targeted Grant to Institutes.

Image for "VIRTUAL ONLY: Safety and Security of Deep Learning"

Confirmed Speakers & Participants

  • Speaker
  • Poster Presenter
  • Attendee
  • Virtual Attendee
  • Ben Adcock
    Simon Fraser University
  • Ibrahim Olalekan Alabi
  • Elie Alhajjar
    US Military Academy
  • Genevera Allen
    Rice University
  • jenny baglivo
    boston college
  • Nicola Bastianello
    University of Padova
  • Getachew Befekadu
    Morgan State University
  • Aaron Berk
    University of British Columbia
  • Alex Bespalov
    University of Birmingham
  • Shivam Bhatt
    University of Toronto
  • Ghanshyam Bhatt
    Tennessee State University
  • Ralph Bording
    Alabama A&M University
  • Nicolas Boulle
    University of Oxford
  • Elisa Bravo
    Florida International University
  • Simone Brugiapaglia
    Concordia University
  • Emmanuel Candes
    Stanford University
  • Aaron Charous
  • Ke Chen
    University of Liverpool
  • Janhavi Chitale
    Florida International University
  • Matthew Colbrook
    University of Cambridge
  • Rachel Cummings
    Columbia University
  • Marta D'Elia
    Sandia National Laboratories, NM
  • Anil Damle
    Cornell University
  • Ronald DeVore
    Texas A&M University
  • Nick Dexter
    Simon Fraser University
  • George Dulikravich
    Florida International University
  • Thomas Fel
    Artificial and Natural Intelligence Toulouse Institute, Brown University
  • Marija Furdek
    Chalmers University of Technology
  • Gustavo Gasperazzo
    Federal University of Rio de Janeiro
  • Horacio Gomez-Acevedo
    University of Arkansas for Medical Sciences
  • Pedro González Rodelas
    University of Granada
  • Zach Grey
    National Institute of Standards and Technology
  • Suman Guha
    Presidency University
  • Anders Hansen
    University of Cambridge
  • Fengxiang He
    The University of Sydney
  • Kaveh Heidary
    Alabama A&M University
  • Fred Hickernell
    Illinois Institute of Technology
  • James Hyman
    Tulane University
  • Rajesh Jha
    Florida International University, Miami, Florida, USA
  • Frederic Jurie
  • Avleen Kaur
    University of Manitoba
  • Abdul Khaliq
    Middle Tennessee State University
  • Tamara Kolda
    Sandia National Labs
  • Boris Krämer
    University of California San Diego
  • Amit Kumar
    Indian Institute of Technology Kharagpur
  • Gitta Kutyniok
    LMU Munich
  • Henry Kvinge
    Pacific Northwest National Lab
  • Christopher Lehnig
    San Diego State University
  • Wenyuan Liao
    University of Calgary
  • En-Bing Lin
    Central Michigan University
  • Jie Long
    Middle Tennessee State University
  • Kathryn Lund
  • Aleksander Madry
    Massachusetts Institute of Technology
  • Jodi Mead
    Boise State University
  • Kevin Miller
    University of California, Los Angeles
  • Pablo Moriano
    Oak Ridge National Laboratory
  • Marshall Mueller
    Tufts University
  • Reshma Munbodh
    Alpert Medical School of Brown University
  • Basim Mustafa
    University of Granada
  • Evangelos Nastas
  • Linda Ness
    Rutgers University
  • Maksym Neyra-Nesterenko
    Simon Fraser University
  • Shobhit Nigam
    Pandit Deendayal Petroleum University Gandhinagar
  • Evi Ofekeze
    Boise State University
    Federal University of Rio de Janeiro, UFRJ
  • Jun Sur Park
    University of Iowa
  • Vivak Patel
    University of Wisconsin -- Madison
  • Sandhya Prabhakaran
    Moffitt Cancer Center
  • Jing Qin
    University of Kentucky
  • Jason Quinones
    Gallaudet University
  • Viktor Reshniak
    Oak Ridge National Laboratory
  • Jacob Rezac
    National Institute of Standards and Technology
  • Cynthia Rudin
    Duke University
  • Quratulan Sabir
    National College of business administration and economics
  • Tarik Sahin
    Bundeswehr University Munich
  • Giovanni Samaey
    KU Leuven
  • Ruchi Sandilya
    TIFR Centre For Applicable Mathematics
  • Thomas Serre
    Brown University
  • Qin Sheng
    Baylor University
  • Yeonjong Shin
    Brown University
  • Mansi Sood
    Carnegie Mellon University, Pittsburgh
  • Varsha Srivastava
    Quantum Integrators Group LLC
  • Ömer Sümer
    University of Tubingen
  • Li-yeng Sung
    Louisiana State University
  • THOMAS Torku
    Middle Tennessee State University
  • Alex Townsend
    Cornell University
  • Ivan Tyukin
    University of Leicester
  • Marilyn Vazquez Landrove
    Ohio State University
  • Abhinav Verma
    The University of Texas at Austin
  • Clayton Webster
    University of Texas
  • Colby Wight
    Pacific Northwest National Laboratory
  • Joab Winkler
    Sheffield University
  • Eliyas Woldegeorgis
    University of Leicester
  • Karamatou Yacoubou Djima
    Amherst College
  • Masanao Yajima
    Boston Universisty
  • Ming Yan
    Michigan State University
  • Yunan Yang
    New York University
  • Vladimir Yushutin
    University of Maryland
  • Vasilis Zafiris
    University of Houston-Downtown
  • Longbin Zhang
    KTH royal institute of technology
  • Bentuo Zheng
    University of Memphis
  • Philipp Zilk
    Bundeswehr Universität München

Workshop Schedule

Saturday, April 10, 2021
  • 10:00 - 10:15 am EDT
    Welcome - Virtual
    • Brendan Hassett, ICERM/Brown University
  • 10:15 - 10:55 am EDT
    An Information Theoretic Approach to Validate Deep Learning-Based Algorithms
    • Speaker
    • Gitta Kutyniok, LMU Munich
    • Session Chair
    • Simone Brugiapaglia, Concordia University
    In this talk, we provide a theoretical framework for interpreting neural network decisions by formalizing the problem in a rate-distortion framework. The solver of the associated optimization, which we coin Rate-Distortion Explanation (RDE), is then accessible to a mathematical analysis. We will discuss theoretical results as well as present numerical experiments showing that our algorithmic approach outperforms established methods, in particular, for sparse explanations of neural network decisions.
  • 11:55 am - 1:30 pm EDT
    Lunch/Free Time
  • 1:30 - 2:10 pm EDT
    Data Matters in Robust ML
    • Speaker
    • Aleksander Madry, Massachusetts Institute of Technology
    • Session Chair
    • Clayton Webster, University of Texas
  • 2:20 - 3:00 pm EDT
    Breaking into a Deep Learning box
    • Speaker
    • Ivan Tyukin, University of Leicester
    • Session Chair
    • Clayton Webster, University of Texas
    Recent decade brought explosive progress in the applications of Machine Learning and data-driven Artificial Intelligence (AI) to real-life problems across sectors. Autonomous cars and automated passport control are examples of the new reality. Deep Learning models, or more generally, models with multiple learnable processing stages constitute a large class of models to which a significant part of the recent successes has been apportioned. Notwithstanding these successes, there are emerging challenges too. In this talk we will discuss a set of vulnerabilities which may typically arise in large Deep Learning models. These vulnerabilities are extreme sensitivities of the models to data or structure perturbations. We will present a formal theoretical framework for assessing and analysing two classes of such vulnerabilities. The first class is linked with adversarial examples. Vulnerabilities of the second class are linked with purposeful malicious structure perturbations which may be, with high probability, undetectable through input-output validation. We name these perturbations “stealth attacks”. We will show how to construct stealth attacks on Deep Learning models that are hard to spot unless the validation set is made exponentially large. For both classes of attacks, the high dimensionality of the AI’s decision-making space appears to be a major contributor to the AI’s vulnerability. We conclude with recommendations of how robustness to malicious perturbations of data and structure can be mitigated by ensuring that the data dimensionality at relevant processing stages in Deep Learning models is kept sufficiently small.
  • 3:10 - 4:00 pm EDT
    Gathertown Afternoon Coffee Break
    Coffee Break - Virtual
Sunday, April 11, 2021
  • 9:10 - 9:50 am EDT
    Deep Learning and Neural Networks: The Mathematical View
    • Speaker
    • Ronald DeVore, Texas A&M University
    • Session Chair
    • Anders Hansen, University of Cambridge
    Deep Learning is much publicized and has had great empirical success on challenging problems in learning. Yet there is no quantifiable proof of performance and certified guarantees for these methods. This talk will give an overview of Deep Learning from the viewpoint of mathematics and numerical computation.
  • 10:00 - 10:30 am EDT
    Gathertown Morning Coffee Break
    Coffee Break - Virtual
  • 10:30 - 11:10 am EDT
    Can we design deep learning models that are inherently interpretable?
    • Speaker
    • Cynthia Rudin, Duke University
    • Session Chair
    • Anders Hansen, University of Cambridge
    Black box deep learning models are difficult to troubleshoot. In practice, it can be difficult to tell whether their reasoning process is correct, and ""explanations"" have repeatedly been shown to be ineffective. In this talk I will discuss two possible approaches to create deep learning methods that are inherently interpretable. The first is to use case-based reasoning, through a neural architecture called ProtoPNet, where an extra ""prototype"" layer in the network allows it to reason about an image based on how similar it looks to other images (the network says ""this looks like that""). Second, I will describe ""concept whitening,"" a method for disentangling the latent space of a neural network by decorrelating concepts in the latent space and aligning them along the axes. This Looks Like That: Deep Learning for Interpretable Image Recognition. NeurIPS spotlight, 2019. Concept Whitening for Interpretable Image Recognition. Nature Machine Intelligence, 2020.
  • 11:20 am - 12:00 pm EDT
    Differential privacy, deep learning, and synthetic data generation
    • Speaker
    • Rachel Cummings, Columbia University
    • Session Chair
    • Anders Hansen, University of Cambridge
    Differential privacy is a parameterized notion of database privacy that gives a mathematically rigorous worst-case bound on the maximum amount of information that can be learned about an individual's data from the output of a computation. Recent work has provided tools for differentially private stochastic gradient decent, which enables differentially private deep learning. These in turn enable differentially private synthetic data generation, to provide synthetic versions of sensitive datasets that share statistical properties with the original data while additionally providing formal privacy guarantees for the training dataset. This talk will first give an introduction to differential privacy, and then survey recent advances in differentially private deep learning and its application to synthetic data generation.
  • 12:10 - 1:30 pm EDT
    Lunch/Free Time
  • 1:30 - 2:10 pm EDT
    Reliability, Robustness and Minipatch Learning
    • Speaker
    • Genevera Allen, Rice University
    • Session Chair
    • Ben Adcock, Simon Fraser University
    Many have noted and lamented a reproducibility crisis in science with more recent discussion and interest on the reproducibility and reliability of data science and machine learning techniques.  In this talk, I will introduce the Four R's, a tiered framework for discussing and assessing the reproducibility, replicability, reliability, and robustness of a data science or machine learning pipeline.  Then, I will introduce a new minipatch learning framework that helps to improve the reliability and robustness of machine learning procedures.  Inspired by stability approaches from high-dimensional statistics, random forests, and dropout training in deep learning, minipatch learning is an ensemble approach where we train on very tiny randomly or adaptively chosen subsets of both observations and features or parameters.  Beyond the obvious computational and memory efficiency advantages, we show that minipatch learning also yields more reliable and robust solutions by providing implicit regularization.  
  • 2:20 - 3:00 pm EDT
    Reliable predictions? Counterfactual predictions? Equitable treatment? Some recent progress in predictive inference
    • Speaker
    • Emmanuel Candes, Stanford University
    • Session Chair
    • Ben Adcock, Simon Fraser University
    Recent progress in machine learning provides us with many potentially effective tools to learn from datasets of ever increasing sizes and make useful predictions. How do we know that these tools can be trusted in critical and high-sensitivity systems? If a learning algorithm predicts the GPA of a prospective college applicant, what guarantees do I have concerning the accuracy of this prediction? How do we know that it is not biased against certain groups of applicants? This talk introduces statistical ideas to ensure that the learned models satisfy some crucial properties, especially reliability and fairness (in the sense that the models need to apply to individuals in an equitable manner). To achieve these important objectives, we shall not “open up the black box” and try understanding its underpinnings. Rather we discuss broad methodologies that can be wrapped around any black box to produce results that can be trusted and are equitable. We also show how our ideas can inform causal inference predictive; for instance, we will answer counterfactual predictive problems: i.e. predict the outcome of a treatment would have been given that the patient was actually not treated.
  • 3:10 - 4:00 pm EDT
    Gathertown Afternoon Coffee Break
    Coffee Break - Virtual

All event times are listed in ICERM local time in Providence, RI (Eastern Daylight Time / UTC-4).

All event times are listed in .

Request Reimbursement

Acceptable Costs
  • 1 roundtrip between your home institute and ICERM
  • Flights in economy class to either Providence airport (PVD) or Boston airport (BOS)
  • Ground Transportation to and from airports and ICERM.
Unacceptable Costs
  • Seats in economy plus, business class, or first class
  • Change ticket fees of any kind
  • Multi-use bus passes
  • Meals or incidentals
Advance Approval Required
  • Personal car travel to ICERM from outside New England
  • Multiple-destination plane ticket; does not include layovers to reach ICERM
  • Arriving or departing from ICERM more than a day before or day after the program
  • Multiple trips to ICERM
  • Rental car to/from ICERM
  • Arriving or departing from airport other than PVD/BOS or home institution's local airport
  • 2 one-way plane tickets to create a roundtrip (often purchased from Expedia, Orbitz, etc.)
Reimbursement Requests

Request Reimbursement with Cube

Refer to the back of your ID badge for more information. Checklists are available at the front desk and in the Reimbursement section of Cube.

Reimbursement Tips
  • Scanned original receipts are required for all expenses
  • Airfare receipt must show full itinerary and payment
  • ICERM does not offer per diem or meal reimbursement
  • Allowable mileage is reimbursed at prevailing IRS Business Rate and trip documented via pdf of Google Maps result
  • Keep all documentation until you receive your reimbursement!
Reimbursement Timing

6 - 8 weeks after all documentation is sent to ICERM. All reimbursement requests are reviewed by numerous central offices at Brown who may request additional documentation.

Reimbursement Deadline

Submissions must be received within 30 days of ICERM departure to avoid applicable taxes. Submissions after thirty days will incur applicable taxes. No submissions are accepted more than six months after the program end.