Skip to main navigation Skip to search Skip to main content

Unilmv2: Pseudo-masked language models for unified language model pre-Training

  • Hangbo Bao
  • , Li Dong
  • , Furu Wei*
  • , Wenhui Wang
  • , Nan Yang
  • , Xiaodong Liu
  • , Yu Wang
  • , Songhao Piao
  • , Jianfeng Gao
  • , Ming Zhou
  • , Hsiao Wuen Hon
  • *Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

We propose to pre-Train a unified language model for both autoencoding and partially autoregressive language modeling tasks using a novel training procedure, referred to as a pseudo-masked language model (PMLM). Given an input text with masked tokens, we rely on conventional masks to learn inter-relations between corrupted tokens and context via autoencoding, and pseudo masks to learn intra-relations between masked spans via partially autoregressive modeling. With welldesigned position embeddings and self-Attention masks, the context encodings are reused to avoid redundant computation. Moreover, conventional masks used for autoencoding provide global masking information, so that all the position embeddings are accessible in partially autoregressive language modeling. In addition, the two tasks pre-Train a unified language model as a bidirectional encoder and a sequence-To-sequence decoder, respectively. Our experiments show that the unified language models pre-Trained using PMLM achieve new state-of-The-Art results on a wide range of language understanding and generation tasks across several widely used benchmarks. The code and pre-Trained models are available at https://github.com/ microsoft/unilm.

Original languageEnglish
Title of host publication37th International Conference on Machine Learning, ICML 2020
EditorsHal Daume, Aarti Singh
PublisherInternational Machine Learning Society (IMLS)
Pages619-629
Number of pages11
ISBN (Electronic)9781713821120
StatePublished - 2020
Event37th International Conference on Machine Learning, ICML 2020 - Virtual, Online
Duration: 13 Jul 202018 Jul 2020

Publication series

Name37th International Conference on Machine Learning, ICML 2020
VolumePartF168147-1

Conference

Conference37th International Conference on Machine Learning, ICML 2020
CityVirtual, Online
Period13/07/2018/07/20

Fingerprint

Dive into the research topics of 'Unilmv2: Pseudo-masked language models for unified language model pre-Training'. Together they form a unique fingerprint.

Cite this