51 - A Regularized Framework for Sparse and Structured Neural Attention, with Vlad Niculae

51 - A Regularized Framework for Sparse and Structured Neural Attention, with Vlad Niculae

NIPS 2017 paper by Vlad Niculae and Mathieu Blond…
17 Minuten
Podcast
Podcaster
**The podcast is currently on hiatus. For more ac…

Beschreibung

vor 7 Jahren
NIPS 2017 paper by Vlad Niculae and Mathieu Blondel. Vlad comes on
to tell us about his paper. Attentions are often computed in neural
networks using a softmax operator, which maps scalar outputs from a
model into a probability space over latent variables. There are
lots of cases where this is not optimal, however, such as when you
really want to encourage a sparse attention over your inputs, or
when you have additional structural biases that could inform the
model. Vlad and Mathieu have developed a theoretical framework for
analyzing the options in this space, and in this episode we talk
about that framework, some concrete instantiations of attention
mechanisms from the framework, and how well these work.

Kommentare (0)

Lade Inhalte...

Abonnenten

15
15