AI Safety

Summary

This Projet Long (litteraly Long Project) was a 3 months project in a group of 5 students on the topic of AI Safety, supervised by Guillaume Dupont, Aurélie Hurault and Philippe Quéinnec. This work consisted first in an extended litterature review and overview of the subject of robustness, verification and explainability of AI models such as Deep Neural Networks (DNNs) and Decision Trees (DTs). It then focused, with an approach more practical, on adversarial attacks and defenses of DNNs, formal verification of DNNs and DTs, and explainablity of more broad Machine Learning models.

My Contribution

I contributed in this project to the setup and running of the adversarial attacks and training experiments, as well as a study on confidence learning, where the model outputs in addition a confidence bit in order to validate or invalidate its predicton, to measure the impact of an accurate confidence prediction on adversarial examples. I also carried out a test on Parametric Noise Injection, another robust training method, which did not perform better than adversarial training. My experiments code is primarily in the Colab Notebook.

Joceran Gouneau
Joceran Gouneau
CS Engineering Graduate - Listening for opportunities