Investigating the interpretability of hidden layers in deep text mining

conference paper
In this short paper, we address the interpretability of hidden layer representations in deep text mining: deep neural networks applied to text mining tasks. Following earlierwork predating deep learning methods, we exploit the internal neural network activation (latent) space as a source for performing k-nearest neighbor search, looking for representative, explanatory training data examples with similar neural layer activations as test inputs. We deploy an additional semantic document similarity metric for establishing document similarity between the textual representations of these nearest neighbors and the test inputs. We argue that the statistical analysis of the output of this measure provides insight to engineers training the networks, and that nearest neighbor search in latent space combined with semantic document similarity measures offers a mechanism for presenting explanatory, intelligible examples to users. © 2017 Copyright.
et al.; Kadaster; MarkLogic; Ordina; Taxonic; Wolters Kluwer
TNO Identifier
787735
ISBN
9781450352963
Publisher
Association for Computing Machinery
Source title
13th International Conference on Semantic Systems, SEMANTiCS 2017. 12 September 2017 through 13 September 2017
Editor(s)
Hoekstra, R.
Boer, V. de
Pellegrini, T.
Hoekstra, R.
Faron-Zucker, C.
Pages
177-180
Files
To receive the publication files, please send an e-mail request to TNO Repository.