A Case Study on Learning a Unified Encoder of Relations

Lisheng Fu, Bonan Min, Thien Huu Nguyen, Ralph Grishman

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Typical relation extraction models are trained on a single corpus annotated with a pre-defined relation schema. An individual corpus is often small, and the models may often be biased or overfitted to the corpus. We hypothesize that we can learn a better representation by combining multiple relation datasets. We attempt to use a shared encoder to learn the unified feature representation and to augment it with regularization by adversarial training. The additional corpora feeding the encoder can help to learn a better feature representation layer even though the relation schemas are different. We use ACE05 and ERE datasets as our case study for experiments. The multi-task model obtains significant improvement on both datasets.

Original languageEnglish (US)
Title of host publication4th Workshop on Noisy User-Generated Text, W-NUT 2018 - Proceedings of the Workshop
PublisherAssociation for Computational Linguistics (ACL)
Pages202-207
Number of pages6
ISBN (Electronic)9781948087797
StatePublished - 2018
Event4th Workshop on Noisy User-Generated Text, W-NUT 2018 - Brussels, Belgium
Duration: Nov 1 2018 → …

Publication series

Name4th Workshop on Noisy User-Generated Text, W-NUT 2018 - Proceedings of the Workshop

Conference

Conference4th Workshop on Noisy User-Generated Text, W-NUT 2018
Country/TerritoryBelgium
CityBrussels
Period11/1/18 → …

ASJC Scopus subject areas

  • Information Systems
  • Computer Science Applications
  • Computational Theory and Mathematics

Fingerprint

Dive into the research topics of 'A Case Study on Learning a Unified Encoder of Relations'. Together they form a unique fingerprint.

Cite this