Abstract
                                                                        Fact-to-text generation can allow for the generation of high-quality, informative texts such  as Wikipedia articles. Cross-lingual fact-to-text  generation (XF2T) involves using facts available  in a language, typically English, and generating texts in a different language based on these  facts. This is particularly relevant for low and  medium-resource languages, which have relatively structured informative content. This work  explores the problem of XF2T for generating  long text from given facts with a specific focus on generating factually grounded content.  Unfortunately, previous work either focuses on  cross-lingual facts to short text or monolingual  graph to text generation. In this paper, we propose a novel solution to the multi-sentence XF2T  task, which addresses these challenges by training multilingual Transformer-based models with  coverage prompts and rebalanced beam search,  and further improving the quality by defining  task-specific reward functions and training on  them using reinforcement learning.  Keywords: XF2T, text generation, cross-lingual, NLG  evaluation, low resource NLG