Interpreting Deep Neural Networks with the Package innsight

Dublin Core

Title

Interpreting Deep Neural Networks with the Package innsight

Subject

neural networks, feature attribution, interpretable machine learning, explainable
artificial intelligence, XAI, IML, torch, keras, R.

Description

The R package innsight offers a general toolbox for revealing variable-wise interpretations of deep neural networks’ predictions with so-called feature attribution methods.
Aside from the unified and user-friendly framework, the package stands out in three ways:
It is generally the first R package implementing feature attribution methods for neural
networks. Secondly, it operates independently of the deep learning library, allowing the
interpretation of neural networks from any R package, including keras, torch, neuralnet,
and even custom models. Despite its flexibility, innsight benefits internally from the torch
package’s fast and efficient array calculations, which builds on LibTorch – PyTorch’s C++
backend – without a Python dependency. Finally, it offers a variety of visualization tools
for tabular, signal, image data, or a combination of these. Additionally, the plots can be
rendered interactively using the plotly package.

Creator

Niklas Koenen

Source

Leibniz Institute for Prevention
Research and Epidemiology – BIPS,
University of Bremen

Publisher

Leibniz Institute for Prevention
Research and Epidemiology – BIPS,
University of Bremen

Date

November 2024

Contributor

Fajar bagus W

Format

PDF

Language

English

Type

Text

Files

Citation

Niklas Koenen, “Interpreting Deep Neural Networks with the Package innsight,” Repository Horizon University Indonesia, accessed April 4, 2025, https://repository.horizon.ac.id/items/show/8352.