Transformer-based encoder-decoder models that generate outputs in a
left-to-right fashion have become standard for sequence-to-sequence tasks. In
this paper, we propose a framework for decoding that produces sequences from
the "outside-in": at each step, the model chooses to generate a