Overview

Activation maximisation is a locally explainable method that focuses on input patterns which maximise a given hidden unit activation. Activation maximisation helps understand the layer-wise feature importance to an input instance.

xai-activation-maximisation-demoHow convolutional neural networks see the world

Let be the parameters of the model, and be the activation of a particular unit from layer . The activation map can be defined as the problem:

where is fixed.

Algorithm

The above process consists of four steps:

  1. An image with random pixel values is set to be the input to the activation computation.
  2. The gradients with respect to the noise image, , are computed through backpropagation.
  3. Each pixel of the noise image is changed iteratively to maximise the activation of the neuron, guided by the direction of the gradient:
  1. This process terminates at a specific pattern image , which can be seen as the preferred input for this neuron.

References

  • Erhan, D., Courville, A., & Bengio, Y. (2010). Understanding representations learned in deep architectures.
  • Qin, Z., Yu, F., Liu, C., & Chen, X. (2018). How convolutional neural network see the world-A survey of convolutional neural network visualization methods. arXiv preprint arXiv:1804.11191.
  • https://blog.keras.io/how-convolutional-neural-networks-see-the-world.html