Today’s transformer-based large language models (LLMs) have proven a game-changer in natural language processing, achieving state-of-the-art performance on reading comprehension, question answering and common sense reasoning benchmarks. Given a prompt, LLMs can also generate coherent and sensible completions — but they struggle with infilling, where they are tasked with generating…