[HTML][HTML] Methods for interpreting and understanding deep neural networks

G Montavon, W Samek, KR Müller - Digital signal processing, 2018 - Elsevier
This paper provides an entry point to the problem of interpreting a deep neural network
model and explaining its predictions. It is based on a tutorial given at ICASSP 2017. As a …

Explainable sentiment analysis with applications in medicine

C Zucco, H Liang, G Di Fatta… - 2018 IEEE International …, 2018 - ieeexplore.ieee.org
Sentiment Analysis can help to extract knowledge related to opinions and emotions from
user generated text information. It can be applied in medical field for patients monitoring …

[PDF][PDF] Interpreting and explaining deep neural networks for classification of audio signals

S Becker, M Ackermann, S Lapuschkin… - arXiv preprint arXiv …, 2018 - researchgate.net
Interpretability of deep neural networks is a recently emerging area of machine learning
research targeting a better understanding of how models perform feature selection and …

Extracting automata from recurrent neural networks using queries and counterexamples

G Weiss, Y Goldberg, E Yahav - International Conference on …, 2018 - proceedings.mlr.press
We present a novel algorithm that uses exact learning and abstraction to extract a
deterministic finite automaton describing the state dynamics of a given trained RNN. We do …

Evaluating neural network explanation methods using hybrid documents and morphological agreement

N Poerner, B Roth, H Schütze - arXiv preprint arXiv:1801.06422, 2018 - arxiv.org
The behavior of deep neural networks (DNNs) is hard to understand. This makes it
necessary to explore post hoc explanation methods. We conduct the first comprehensive …

Interpreting recurrent and attention-based neural models: a case study on natural language inference

R Ghaeini, XZ Fern, P Tadepalli - arXiv preprint arXiv:1808.03894, 2018 - arxiv.org
Deep learning models have achieved remarkable success in natural language inference
(NLI) tasks. While these models are widely explored, they are hard to interpret and it is often …

Explaining therapy predictions with layer-wise relevance propagation in neural networks

Y Yang, V Tresp, M Wunderle… - 2018 IEEE International …, 2018 - ieeexplore.ieee.org
In typical data analysis projects in biology and healthcare, simpler predictive models, such
as regressions and decision trees, enjoy more popularity than more complex and expressive …

Learning finite state representations of recurrent policy networks

A Koul, S Greydanus, A Fern - arXiv preprint arXiv:1811.12530, 2018 - arxiv.org
Recurrent neural networks (RNNs) are an effective representation of control policies for a
wide range of reinforcement and imitation learning problems. RNN policies, however, are …

Before name-calling: Dynamics and triggers of ad hominem fallacies in web argumentation

I Habernal, H Wachsmuth, I Gurevych… - arXiv preprint arXiv …, 2018 - arxiv.org
Arguing without committing a fallacy is one of the main requirements of an ideal debate. But
even when debating rules are strictly enforced and fallacious arguments punished, arguers …

Extractive adversarial networks: High-recall explanations for identifying personal attacks in social media posts

S Carton, Q Mei, P Resnick - arXiv preprint arXiv:1809.01499, 2018 - arxiv.org
We introduce an adversarial method for producing high-recall explanations of neural text
classifier decisions. Building on an existing architecture for extractive explanations via hard …