\documentclass{beamer} \usetheme[progressbar=frametitle]{metropolis} % g \in G is explanation as a model % f is the model we're trying to explain % does, being model agnostic, means we do not care about specifics of f. % We use Locally Weighted Square Loss as L, where I suspect pi is the weight and we thus estimate the difference between the actual model % and our explanation, and multiply this with the proximity of the data point z, to x. % Spørg lige Lasse hvorfor min(L(f,g,pi_x(z)) + omega(g)) bliver intractable, når omega(g) er en konstant! \usepackage{bbm} \usepackage{setspace} \usepackage[T1]{fontenc} \usepackage[sfdefault,scaled=.85]{FiraSans} %\usepackage{newtxsf} \usepackage[ruled, linesnumbered]{algorithm2e} \SetKwInput{kwRequire}{Require} \SetKw{kwExpl}{explain} \title{Why Should I Trust You?} \subtitle{Explaining the Predictions of Any Classifier} \author{Casper Vestergaard Kristensen \and Alexander Munch-Hansen} \institute{Aarhus University} \date{\today} \begin{document} \begin{frame} \titlepage \end{frame} \begin{frame} \setbeamertemplate{section in toc}[sections numbered] \frametitle{Outline} \setstretch{0.5} \tableofcontents \end{frame} \section{Meta information} %\subsection{Authors} \begin{frame} \frametitle{Authors} \begin{itemize} \item Marco Tulio Ribeiro \item Sameer Singh \item Carlos Guestrin \end{itemize} \end{frame} %\subsection{Publishing} \begin{frame}[fragile]{Metropolis} \frametitle{Publishing} \begin{itemize} \item Conference Paper, Research \item KDD '16 Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining \begin{itemize} \item A premier interdisciplinary conference, brings together researchers and practitioners from data science, data mining, knowledge discovery, large-scale data analytics, and big data. \item Sigkdd has the highest h5 index of any conference involving databases or data in general \item Highly trusted source \end{itemize} \end{itemize} \end{frame} \section{Article} %\subsection{Problem} \begin{frame} \frametitle{Problem definition} \begin{itemize} \item People often use Machine Learning models for predictions \item Blindly trusting a prediction can lead to poor decision making \item We seek to understand the reasons behind predictions \begin{itemize} \item As well as the model doing the predictions \end{itemize} \end{itemize} \end{frame} \begin{frame} \frametitle{Problem definition} \begin{itemize} \item People often use Machine Learning models for predictions \item Blindly trusting a prediction can lead to poor decision making \item We seek to understand the reasons behind predictions \begin{itemize} \item As well as the model doing the predictions \end{itemize} \end{itemize} \center \includegraphics[scale=0.2]{graphics/doctor_pred.png} \end{frame} %\subsection{Previous Solutions} \begin{frame} \frametitle{Previous Solutions} \begin{itemize} \item Relying on accuracy based on validation set \end{itemize} \end{frame} \begin{frame} \frametitle{A look into two predictions} \includegraphics[scale=0.25]{graphics/christ_vs_atheism.png} \end{frame} \begin{frame} \frametitle{A look into two predictions} \includegraphics[scale=0.25]{graphics/christ_vs_atheism_annotated_1.png} \end{frame} \begin{frame} \frametitle{A look into two predictions} \includegraphics[scale=0.25]{graphics/christ_vs_atheism_annotated_2.png} \end{frame} \subsection{The LIME framework} \begin{frame} \frametitle{LIME} \begin{itemize} \item The algorithm created \item Explains the predictions of \emph{any} classifier or regressor in a faithful way, by approximating it locally with an \emph{interpretable} model. \end{itemize} \end{frame} \begin{frame} \frametitle{Properties of a good explanation} \begin{itemize} \item It should be \emph{intepretable}: \begin{itemize} \item They must provide qualitative understanding between the input variables and the response \item They must take into account the users limitations \item Use a representation understandable to humans \item Could be a binary vector indicating presence or absence of a word \item Could be a binary vector indicating presence of absence of super-pixels in an image \end{itemize} \item It should have \emph{fidelity}: \begin{itemize} \item Essentially means the model should be faithful. \item Local fidelity does not imply global fidelity \item The explanation should aim to correspond to how the model behaves in the vicinity of the instance being predicted \end{itemize} \item It should be \emph{model-agnostic}: \begin{itemize} \item The explanation should be blind to what model is underneath \end{itemize} \end{itemize} \end{frame} \subsection{Explaining Predictions} \begin{frame}[shrink=20] \frametitle{The Fidelity-Interpretability Trade-off} We want a simple explanation, still capable of displaying fidelity \begin{itemize} \item Let an explanation be defined as a model $g \in \{0,1\}^{d^{\prime}} \in G$, where $G$ is a class of \emph{potentially interpretable} models \begin{itemize} \item Linear models, decision trees \item $g$ is a vector showing presence or absence of \emph{interpretable components} \end{itemize} \item $\Omega(g)$ explains the \emph{complexity} of an explanation $g$ \begin{itemize} \item Could be height of a decision tree or number of non-zero weights of a linear model \end{itemize} \item The model we try to explain is $f : \mathbb{R}^d \rightarrow \mathbb{R}$ \begin{itemize} \item In classification, $f(x)$ is the probability or binary indicator that x belongs to a certain class \end{itemize} \item $\pi_x(z)$ is a proximity measure between instance $z$ and $x$ and defines the locality around $x$ \item $\mathcal{L}(f,g,\pi_x)$ defines how \emph{unfaithful} $g$ is in approximating $f$ in the locality around $\pi_x$. \item Ensuring both \emph{interpretability} and \emph{local fidelity}, we minimize $\mathcal{L}$ while having $\Omega(g)$ be low as well \end{itemize} % So a more complex g will achieve a more faithful interpretation (a lower L), but will increase the value of Omega(g) $$\xi(x) = \operatornamewithlimits{argmin}_{g \in G} \mathcal{L}(f,g,\pi_x) + \Omega(g)$$ \end{frame} \begin{frame} % Note: WTF is x' here? Should it not just be x? % Note: g acts in d' while f acts in d, so when we say that we have z' in dimension d', it's the model g, we can recover the z in the original representation i.e. explained by f in dimension d. \frametitle{Sampling for Local Exploration} Goal: Minimizing $\mathcal{L}(f,g,\pi_x)$ without making assumptions on $f$ \begin{itemize} \item For a sample $x$, we need to draw samples around $x$ \item Accomplished by drawing non-zero elements of $x$, resulting in perturbed samples $z^\prime$ \item Given $z^\prime \in \{0,1\}^{d^\prime}$, we compute un-pertubed $z \in R^d$, $f(z)$, so we have a label for $z^\prime$. \end{itemize} \center \includegraphics[scale=0.15]{graphics/sample_points.png} \end{frame} % \subsubsection{Examples} \begin{frame} \frametitle{More definitions} \begin{itemize} \item $G = $ Class of linear models: $g(z^\prime) = w_g \cdot z^\prime$ \item $L = $ The locally weighted square loss \item $\pi_x(z) = \exp(-D(x,z)^2 / \sigma^2)$ \begin{itemize} \item An exponential kernel function based on some distance function D (could be L2 distance for images) \end{itemize} \item Thus; $L(f, g, \pi_x) = \sum\limits_{z,z^\prime \in \mathcal{Z}} \pi_x(z) (f(z) - g(z^\prime))^2$ \end{itemize} \end{frame} \begin{frame} % \frametitle{Sparse Linear Explanations} % Talk through the algorithm, discussing the sampling and K-Lasso \frametitle{Explaining an individual prediction} \begin{algorithm}[H] \setstretch{0.9} \SetAlgoLined \kwRequire{Classifier $f$, Number of samples $N$} \kwRequire{Instance $x$, and its intepretable version $x^{\prime}$} \kwRequire{Similarity kernel $\pi_x$, Length of explanation $K$} \Indp $\mathcal{Z} \leftarrow \{\}$ \\ \For{$i \in \{1,2,3,\dots, N\}$}{ $z_i^{\prime} \leftarrow sample\_around(x^{\prime})$ \\ $\mathcal{Z} \leftarrow \mathcal{Z} \cup \langle z_i^{\prime}, f(z_i), \pi_{x}(z_i) \rangle$ \\ } $w \leftarrow \text{K-Lasso}(\mathcal{Z},K) \vartriangleright \text{with } z_i^{\prime} \text{ as features, } f(z) \text{ as target}$ \\ \Return $w$ \caption{Sparse Linear Explanations using LIME} \end{algorithm} % This algorithm approximates the minimization problem of computing a single individual explanation of a prediction. % K-Lasso is the procedure of learning the weights via least squares. Wtf are these weights??? \end{frame} \begin{frame} \frametitle{Text Classification} \end{frame} \begin{frame} \frametitle{Deep Networks for Images} \end{frame} \subsection{Explaining Models} \begin{frame} \frametitle{Explaining models} Idea: We give a global understanding of the model by explaining a set of individual instances \begin{itemize} \item Still model agnositc (since the indiviudal explanations are) \item Instances need to be selected in a clever way \item Some definitions \begin{itemize} \item Time/patience of humans is explained by a budget \emph{B} which denotes number of explanations a human will sit through. \item Given a set of instances \textbf{X}, we define the \emph{pick step} as the task of selecting \emph{B} instances for the user to inspect. \end{itemize} \end{itemize} \end{frame} \begin{frame} \frametitle{The pick step} The task of selecting \emph{B} instances for the user to inspect \begin{itemize} \item Not dependent on the existence of explanations \item So it should not assist users in selecting instances themselves \item Looking at raw data is not enough to understand predicitions and get insights \item Should take into account the explanations that accompany each prediction \item Should pick a diverse, representative set of explanations to show the user, so non-redundant explanations that represent how the model behaves globally. \end{itemize} \end{frame} % This is a matrix explaining instances and their features explained by a binary list s.t. an instance either has a feature or does not. The blue line explains the most inherent feature, which is important, as it is found in omst of the instances. The red lines indicate the two samples which are most important in explaining the model. Thus, explaining importance, is done by: I_j = sqrt(sum_i=1^n W_ij) \begin{frame} \frametitle{Picking instances} \center \includegraphics[scale=0.81]{graphics/picker_first.png} \end{frame} \begin{frame} \frametitle{Picking instances} \center \includegraphics[scale=0.32]{graphics/picker_second.png} \end{frame} \begin{frame} \frametitle{Picking instances} \center \includegraphics[scale=0.32]{graphics/picker_third.png} \end{frame} \begin{frame} \frametitle{Picking instances} \center \includegraphics[scale=0.32]{graphics/picker_final.png} \end{frame} \begin{frame} \frametitle{Definitions} \begin{itemize} \item $I_j = \sqrt{\sum_{i=1}^n W_{ij}}$ \item $c(V,W,I) = \sum\limits_{j=1}^{d^\prime} \mathbbm{1}_{[\exists i \in V : W_{ij} > 0]}\ I_j$ \item $Pick(W,I) = \operatornamewithlimits{argmax}\limits_{V,|V| \leq B} c(V,W,I)$ \end{itemize} \end{frame} \begin{frame} \frametitle{Explanation of algorithm 2} \begin{itemize} \item Given explanations for set of instances $X$, $(|X| = n)$. Construct $n \times d^\prime$ \emph{explanation matrix} $W$ \end{itemize} \end{frame} \begin{frame} \frametitle{Submodular Picks} \begin{algorithm}[H] \setstretch{0.9} \SetAlgoLined \kwRequire{Instances $X$, Budget $B$} \Indp \ForAll{$x_i \in X$}{ $W_i \leftarrow \mathbf{explain}(x_i, x_i^{\prime})$ \qquad \qquad $\vartriangleright$ Using Algorithm 1 } \For{$j \in \{1\dots d^{\prime}$} { $I_j \leftarrow \sqrt{\sum_{i=1}^n |W_{ij}|}$ \qquad \qquad \quad $\vartriangleright$ Compute feature importances } $V \leftarrow \{\}$ \\ \While(\qquad \qquad \qquad \quad \ \ $\vartriangleright$ Greedy optimisation of Eq 4){$|V| < B$} { $V \leftarrow V \cup \text{argmax}_i \ c(V \cup \{i\}, W, i)$ } \Return $V$ \caption{Submodular pick (SP) algorithm} \end{algorithm} \end{frame} \section{Experiments} %\subsection{Simulated User Experiments} %\subsubsection{Setup} \begin{frame} \frametitle{Experiments} Interested in three questions: \begin{itemize} \item Are the explanations faithful to the model? \item Can the explanations aid users in ascertaining trust in the predictions \item Are the explanations useful for evaluating the model as a whole \end{itemize} \end{frame} \begin{frame} \frametitle{Setup} \begin{itemize} \item Use two datasets, \emph{books} and \emph{DVDs}, both of $2000$ instances. \begin{itemize} \item Task is to classify reviews as \emph{positive} or \emph{negative} \end{itemize} \item Decision Trees (\textbf{DT}), Logistic Regression (\textbf{LR}), Nearest Neighbours (\textbf{NN}), and SVMs with RBF kernel (\textbf{SVM}), all used BoW as features, are trained. \begin{itemize} % Note, random forest will make no sense without any explanation system, such as LIME \item Also train random forest (\textbf{RF}) with $1000$ trees. \end{itemize} \item Each dataset used for training will consist of $1600$ instances and $400$ will be used for testing. \item Explanations of \textbf{LIME} is compared with \textbf{parzen} \begin{itemize} \item \textbf{parzen} approximates black box classifier globally and explains individual predictions by taking the gradient of the prediction probability function. \item Both are also compared to a greedy method where features are picked by removing most contributing ones until prediction change, as well as a random procedure. % K explains the amount of words in the BoW model and the complexity of the model. Higher K => More complex but more faithful, lower k => Less complex, potentially less faithful \item $K = 10$ for the experiments \end{itemize} \end{itemize} \end{frame} \begin{frame} \frametitle{Faithfulness} \centering % Books faithfulness \includegraphics[height=0.35\textheight]{graphics/books_dt_lr.png}{ } % Dvds faithfulness \includegraphics[height=0.35\textheight]{graphics/dvds_dt_lr.png} \end{frame} \begin{frame} \frametitle{Can users select the best classifier?} \begin{itemize} \item Faithfulness of explanations is measured on classifiers that are interpretable, \textbf{LR} and \textbf{DT}. Both are trained s.t. the max no. of features is $10$, so features found by these are the \emph{gold standard} of features, in regards to which are important. \item For each prediction on the test set, explanations are produced and the fraction of the gold features found, is computed. \end{itemize} \end{frame} \begin{frame} \frametitle{Can non-experts improve a classifier?} \end{frame} %\subsection{Human Subjects} \section{Conclusion} \begin{frame} \frametitle{Conclusion} \end{frame} \begin{frame} \frametitle{Future work} \end{frame} \section{Recap} \begin{frame} \frametitle{Recap} \end{frame} \end{document}