Typical Decoding for Natural Language Generation (Get more human-like outputs from language models!)

First published at 14:20 UTC on March 28th, 2022.
subscribers

#deeplearning #nlp #sampling

Modern language models like T5 or GPT-3 achieve remarkably low perplexities on both training and validation data, yet when sampling from their output distributions, the generated text often seems dull and uninteresting.…

MORE
CategoryScience & Technology
SensitivityNormal - Content that is suitable for ages 16 and over
DISCUSS THIS VIDEO