Write For Us

[SAIF 2019] Day 2: Rational Recurrences for Empirical Natural Language Processing – Noah Smith

E-Commerce Solutions SEO Solutions Marketing Solutions
91 Views
Published
Despite their often-discussed advantages, deep learning methods largely disregard theories of both learning and language. This makes their prediction behavior hard to understand and explain. In this talk, I will present a path toward more understandable (but still "deep") natural language processing models, without sacrificing accuracy. Rational recurrences comprise a family of recurrent neural networks that obey a particular set of rules about how to calculate hidden states, and hence correspond to parallelized weighted finite-state pattern matching. Many recently introduced models turn out to be members of this family, and the weighted finite-state view lets us derive some new ones. I'll introduce rational RNNs and present some of the ways we have used them in NLP. My collaborators on this work include Jesse Dodge, Hao Peng, Roy Schwartz, and Sam Thomson.

#SAIF
#SamsungAIForum
#SR
#SamsungResearch
Category
Phone
Sign in or sign up to post comments.
Be the first to comment