Penalty Decoding: Well Suppress the Self-Reinforcement Effect in Open-Ended Text Generation

Type: Article

Publication Date: 2023-01-01

Citations: 0

DOI: https://doi.org/10.18653/v1/2023.emnlp-main.78

Abstract

The decoding algorithm is critical for open-ended text generation, transforming latent representations into coherent and meaningful outputs. This paper investigates the self-reinforcement effect in text generation and the effectiveness of a repetition penalty to mitigate it. However, determining the optimal repetition penalty value is challenging. To tackle this, we propose a forgetting mechanism that disregards distant tokens, reducing the burden of penalty selection. In addition, we introduce a length penalty to address overly short sentences caused by excessive penalties. Our penalty decoding approach incorporating three strategies helps resolve issues with sampling methods deviating from factual information. Experimental results demonstrate the efficacy of our approach in generating high-quality sentences resembling human output.

Locations

  • arXiv (Cornell University) - View - PDF
  • Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing - View - PDF

Similar Works

Action Title Year Authors
+ Penalty Decoding: Well Suppress the Self-Reinforcement Effect in Open-Ended Text Generation 2023 Wen-Hong Zhu
Hongkun Hao
Rui Wang
+ Implicit Unlikelihood Training: Improving Neural Text Generation with Reinforcement Learning 2021 Evgeny Lagutin
Daniil Gavrilov
Pavel Kalaidin
+ Implicit Unlikelihood Training: Improving Neural Text Generation with Reinforcement Learning 2021 Evgeny Lagutin
Daniil Gavrilov
Pavel Kalaidin
+ Implicit Unlikelihood Training: Improving Neural Text Generation with Reinforcement Learning 2021 Evgeny Lagutin
Daniil Gavrilov
Pavel Kalaidin
+ PDF Chat The Hyperfitting Phenomenon: Sharpening and Stabilizing LLMs for Open-Ended Text Generation 2024 Fredrik Carlsson
Fangyu Liu
D M Ward
Murathan Kurfali
Joakim Nivre
+ Language Model Evaluation in Open-ended Text Generation 2021 An Nguyen
+ Reward-Augmented Decoding: Efficient Controlled Text Generation With a Unidirectional Reward Model 2023 Haikang Deng
Colin Raffel
+ Learning to Break the Loop: Analyzing and Mitigating Repetitions for Neural Text Generation 2022 Jin Xu
Xiaojiang Liu
Jianhao Yan
Deng Cai
Huayang Li
Jian Li
+ PDF Chat Improving Open-Ended Text Generation via Adaptive Decoding 2024 Wen-Hong Zhu
Hongkun Hao
Zhiwei He
Yiming Ai
Rui Wang
+ Contrastive Decoding: Open-ended Text Generation as Optimization 2022 Xiang Lisa Li
Ari Holtzman
Daniel Fried
Percy Liang
Jason Eisner
Tatsunori Hashimoto
Luke Zettlemoyer
Michael Lewis
+ PDF Chat Controllable Text Generation for Large Language Models: A Survey 2024 Xun Liang
Hanyu Wang
Yezhaohui Wang
Shichao Song
Jiawei Yang
Simin Niu
Jie Hu
Dan Liu
Shunyu Yao
Feiyu Xiong
+ Informed Sampling for Diversity in Concept-to-Text NLG 2021 Giulio Zhou
Γεράσιμος Λάμπουρας
+ PDF Chat Informed Sampling for Diversity in Concept-to-Text NLG 2020 Giulio Zhou
Γεράσιμος Λάμπουρας
+ Informed Sampling for Diversity in Concept-to-Text NLG 2020 Giulio Zhou
Γεράσιμος Λάμπουρας
+ PDF Chat Decoding Decoded: Understanding Hyperparameter Effects in Open-Ended Text Generation 2024 Esteban Garces Arias
Meimingwei Li
Christian Heumann
Matthias Aßenmacher
+ PDF Chat Critic-Guided Decoding for Controlled Text Generation 2023 Minbeom Kim
Hwanhee Lee
Kang Min Yoo
Joonsuk Park
Hwaran Lee
Kyomin Jung
+ The Stable Entropy Hypothesis and Entropy-Aware Decoding: An Analysis and Algorithm for Robust Natural Language Generation 2023 Kushal Arora
Timothy J. O’Donnell
Doina Precup
Jason Weston
Jackie Chi Kit Cheung
+ PDF Chat Look-back Decoding for Open-Ended Text Generation 2023 Nan Xu
Chunting Zhou
Aslı Çelikyılmaz
Xuezhe Ma
+ Look-back Decoding for Open-Ended Text Generation 2023 Nan Xu
Chunting Zhou
Aslı Çelikyılmaz
Xuezhe Ma
+ Reward Gaming in Conditional Text Generation 2022 Richard Yuanzhe Pang
Vishakh Padmakumar
Thibault Sellam
Ankur P. Parikh
He He

Works That Cite This (0)

Action Title Year Authors

Works Cited by This (20)

Action Title Year Authors
+ The Curious Case of Neural Text Degeneration 2019 Ari Holtzman
Jan Buys
Li Du
Maxwell Forbes
Yejin Choi
+ PDF Chat Hierarchical Neural Story Generation 2018 Angela Fan
Mike Lewis
Yann Dauphin
+ CTRL: A Conditional Transformer Language Model for Controllable Generation 2019 Nitish Shirish Keskar
Bryan McCann
Lav R. Varshney
Caiming Xiong
Richard Socher
+ HuggingFace's Transformers: State-of-the-art Natural Language Processing 2019 Thomas Wolf
Lysandre Debut
Victor Sanh
Julien Chaumond
Clément Delangue
Anthony Moi
Pierric Cistac
Tim Rault
Rémi Louf
Morgan Funtowicz
+ Don’t Say That! Making Inconsistent Dialogue Unlikely with Unlikelihood Training 2020 Margaret Li
Stephen Roller
Ilia Kulikov
Sean Welleck
Y-Lan Boureau
Kyunghyun Cho
Jason Weston
+ Mirostat: A Neural Text Decoding Algorithm that Directly Controls Perplexity 2020 Sourya Basu
Govardana Sachitanandam Ramachandran
Nitish Shirish Keskar
Lav R. Varshney
+ A Contrastive Framework for Neural Text Generation 2022 Yixuan Su
Lü Tian
Yan Wang
Dani Yogatama
Lingpeng Kong
Nigel Collier
+ LaMDA: Language Models for Dialog Applications 2022 Romal Thoppilan
Daniel De Freitas
Jamie Hall
Noam Shazeer
Apoorv Kulshreshtha
Heng-Tze Cheng
Alicia Jin
Taylor Bos
Leslie L. Baker
Yu Du
+ OPT: Open Pre-trained Transformer Language Models 2022 Susan Zhang
Stephen Roller
Naman Goyal
Mikel Artetxe
Moya Chen
Shuohui Chen
Christopher Dewan
Mona Diab
Xian Li
Xi Victoria Lin
+ Learning to Break the Loop: Analyzing and Mitigating Repetitions for Neural Text Generation 2022 Jin Xu
Xiaojiang Liu
Jianhao Yan
Deng Cai
Huayang Li
Jian Li