NUIG-DSI’s submission to the GEM Benchmark 2021

View/ Open
Date
2021-08-05Author
Pasricha, Nivranshu
Arcan, Mihael
Buitelaar, Paul
Metadata
Show full item recordUsage
This item's downloads: 25 (view details)
Recommended Citation
Pasricha, Nivranshu, Arcan, Mihael, & Buitelaar, Paul. (2021). NUIG-DSI’s submission to the GEM Benchmark 2021. Paper presented at the 1st Workshop on Natural Language Generation, Evaluation, and Metrics (GEM 2021), Online, 05-06 August. doi:10.18653/v1/2021.gem-1.13
Published Version
Abstract
This paper describes the submission by NUIG-DSI to the GEM benchmark 2021. We participate in the modeling shared task where we submit outputs on four datasets for data-to-text generation, namely, DART, WebNLG (en), E2E and CommonGen. We follow an approach similar to the one described in the GEM benchmark paper where we use the pre-trained T5-base model for our submission. We train this model on additional monolingual data where we experiment with different masking strategies specifically focused on masking entities, predicates and concepts as well as a random masking strategy for pre-training. In our results we find that random masking performs the best in terms of automatic evaluation metrics, though the results are not statistically significantly different compared to other masking strategies.