Ask a Question

Prefer a chat interface with context about you and your work?

ProphetNet: Predicting Future N-gram for Sequence-to-SequencePre-training

ProphetNet: Predicting Future N-gram for Sequence-to-SequencePre-training

This paper presents a new sequence-to-sequence pre-training model called ProphetNet, which introduces a novel self-supervised objective named future n-gram prediction and the proposed n-stream self-attention mechanism. Instead of optimizing one-step-ahead prediction in the traditional sequence-to-sequence model, the ProphetNet is optimized by n-step ahead prediction that predicts the next n tokens …