Fwd: [clunch] CLUNCH 9/30: Adhiguna Kuncoro (Oxford/DeepMind)

18 views
Skip to first unread message

João Sedoc

unread,
Aug 25, 2017, 12:15:21 PM8/25/17
to PennStatNLP
---------- Forwarded message ----------
From: "Anne Cocos" <aco...@seas.upenn.edu>
Date: Aug 25, 2017 10:44 AM
Subject: [clunch] CLUNCH 9/30: Adhiguna Kuncoro (Oxford/DeepMind)
To: <clu...@lists.seas.upenn.edu>
Cc:


images.jpg

Computational Linguistics and Lunch

Upcoming

CLunch




9/6: TBD


9/13: Anne Cocos (UPenn)


9/20: TBD






Adhiguna Kuncoro

Oxford University / DeepMind


Towards More Robust and Interpretable Models for Structured Prediction and Language Generation


Wednesday, August 30th

12:00 pm– 1:30 pm

Levine 307

3330 Walnut St


ABSTRACT


Many important tasks in natural language processing involve problems with structured output spaces, such as machine translation and syntactic parsing. In line with recent advances in representation learning, I will describe two ways to improve state of the art neural methods in language modeling and syntactic parsing. Despite the expressive power of recurrent neural networks, I explore the hypothesis that such models can still benefit from linguistically-motivated inductive biases, thereby facilitating better generalization given a limited amount of labeled data. I will also argue for the importance of cost functions in learning, where a refined notion of costs, based on the idea of ensemble distillation, improves the performance of a strong neural dependency parser.


In the opposite direction, neural representation learners can be thought of as “mini-linguists”, in the sense that they often need to come up with generalizations and theories about certain latent aspects of language. The learners’ findings can then be used as empirical means to confirm or refute certain linguistic hypothesis. To this end, I demonstrate that the findings of Recurrent Neural Network Grammars (RNNG), a state of the art model for parsing and language modeling, mostly align with certain linguistic theories of syntax, while also discovering some syntactic phenomena that are different from our intuition, but are interesting nevertheless.




BIOGRAPHY


Adhiguna Kuncoro is an incoming DPhil student in computer science at the University of Oxford and research scientist at DeepMind. His primary research interest lies at the intersection of natural language processing and machine learning, particularly on designing statistical models of natural language that are: i.) robust, ii.) interpretable, and iii.) able to learn more with less amount of annotated data. His co-authored work received an Outstanding Long Paper award at EACL 2017. He holds a Master’s degree in language technologies from CMU LTI, where he worked on low-resource language processing under the DARPA LORELEI project, and a Master’s degree in computer science from the University of Oxford. His DPhil study at Oxford will be jointly funded by an EPSRC studentship under the flexibility to support the very best students scheme and a Balliol Mark Sadler scholarship award. He completed his undergraduate degree in Informatics Engineering from Institut Teknologi Bandung, Indonesia.


Lunch will be served.


Want to share CLunch? New users can register for the mailing list here.

\



_______________________________________________
CLUNCH mailing list
CLU...@LISTS.SEAS.UPENN.EDU
https://LISTS.SEAS.UPENN.EDU/mailman/listinfo/clunch

Reply all
Reply to author
Forward
0 new messages