One of the main open problems in deep learning is the lack of explainability of the models. It is unquestionable that these algorithms outperform traditional techniques, but there is a real risk of introducing unintentional biases during the training process. The goal of this project is to evaluate different visualization solutions that have been proposed in the literature for convolutional neural networks and try to make sense of what they learn.
Óscar Llorente González