Toward a Better Understanding of Relational Knowledge in Language Models

2022/01/26 (Wed) 21:00 (JST)

潮旭 / Asahi Ushio (Cardiff University)

[Webサイト]

Asahi Ushio is a Ph.D student in the School of Computer Science and Informatics at Cardiff University, co-advised by Jose Camacho-Collados and Steven Schockaert. He studies natural language processing (NLP), focusing on language model probing and application of language model in tasks such as named-entity recognition (NER) and question generation (QG). In 2021, he did research internships at Amazon supervised by Danushka Bollegala, and Snapchat co-supervised by Francesco Barbieri and Leonardo Neves. Before joining Cardiff University, he had been a full-time research engineer at Cogent Labs from 2018 to 2020. Visit his personal website for more info https://asahiushio.com

概要

Language model pretraining plays a significant role in natural language processing nowadays, yet we are still far from fully understanding the mechanism of language models. In this talk, I will present our research about relational knowledge in language models, which is one of the important aspects to understand the ability of language models but less studied than others objective such as factual knowledge or syntactic analysis. In short, relational knowledge can be regarded as a capacity to predict the relation among concepts, and utilize it to make a better decision in downstream tasks. In our first work [1], we point out that the current state-of-the-art language models are not capable of solving a simple word analogy task, and our second work [2] presents a finetuning framework to distill the relational knowledge from language models that establishes the best performance in two tasks, which require relation understanding (even better than GPT-3). The series of our researches indicate that language models actually acquire relational knowledge to some extent at the pretraining phase, but we need a specific methodology such as the finetuning scheme presented in our work [2] to extract such knowledge.

※トークは日本語です。

[動画] [スライド1] [論文1] (ACL 2021) [スライド2] [論文2] (EMNLP 2021)

メーリングリストへの登録: 参加用URLなどNLPコロキウムに関するお知らせを受け取りたい方はメーリングリストへのご登録をお願いします。

メーリングリスト登録フォーム

[トップページへ戻る]