Interpretable structure induction via sparse attention

B Peters, V Niculae, AFT Martins - Proceedings of the 2018 …, 2018 - aclanthology.org
Proceedings of the 2018 EMNLP workshop blackboxnlp: analyzing and …, 2018aclanthology.org
Neural network methods are experiencing wide adoption in NLP, thanks to their empirical
performance on many tasks. Modern neural architectures go way beyond simple
feedforward and recurrent models: they are complex pipelines that perform soft,
differentiable computation instead of discrete logic. The price of such soft computing is the
introduction of dense dependencies, which make it hard to disentangle the patterns that
trigger a prediction. Our recent work on sparse and structured latent computation presents a …
Abstract
Neural network methods are experiencing wide adoption in NLP, thanks to their empirical performance on many tasks. Modern neural architectures go way beyond simple feedforward and recurrent models: they are complex pipelines that perform soft, differentiable computation instead of discrete logic. The price of such soft computing is the introduction of dense dependencies, which make it hard to disentangle the patterns that trigger a prediction. Our recent work on sparse and structured latent computation presents a promising avenue for enhancing interpretability of such neural pipelines. Through this extended abstract, we aim to discuss and explore the potential and impact of our methods.
aclanthology.org
Showing the best result for this search. See all results