Cross-lingual extreme summarization of scholarly documents

The number of scientiﬁc publications nowadays is rapidly increasing, causing information overload for researchers and making it hard for scholars to keep up to date with current trends and lines of work. Recent work has tried to address this problem by developing methods for automated summarization in the scholarly domain, but concentrated so far only on monolingual settings, primarily English. In this paper, we consequently explore how state-of-the-art neural abstract summarization models based on a multilingual encoder–decoder architecture can be used to enable cross-lingual extreme summaries of scholarly texts. To this end, we compile a new abstractive cross-lingual summarization dataset for the scholarly domain in four different languages, which enables us to train and evaluate models that process English papers and generate summaries in German, Italian, Chinese and Japanese. We present our new X-SCITLDR dataset for multilingual summarization and thoroughly benchmarkdifferentmodelsbasedonastate-of-the-artmultilingualpre-trainedmodel,includingatwo-stagepipelineapproach that independently summarizes and translates, as well as a direct cross-lingual model. We additionally explore the beneﬁts of intermediate-stage training using English monolingual summarization and machine translation as intermediate tasks and analyze performance in zero-and few-shot scenarios. Finally, we investigate how to make our approach more efﬁcient on the basis of knowledge distillation methods, which make it possible to shrink the size of our models, so as to reduce the computational complexity of the summarization inference.


Introduction
For years, the number of scholarly documents has been steadily increasing [9], thus making it difficult for researchers to keep up to date with current publications, trends and lines of work.Because of this problem, approaches based on Natural Language Processing (NLP) have been developed to niques using supervision from labeled data.For the task of summarizing research papers, most available datasets are in English only, e.g., CSPubSum/CSPubSumExt [17] and ScisummNet [95], with community-driven shared tasks also having concentrated on English as de facto the only language of interest [12,40].But while English is the main language in most of the research communities, especially those in the science and technology domain, this limits the accessibility of summarization technologies for researchers who do not use English as the main language (e.g., many scholars in a variety of areas of humanities and social and political sciences).We accordingly focus on the problem of cross-lingual summarization of scientific articles-i.e., produce summaries of research papers in languages different than the one of the original paper-and benchmark the ability of state-ofthe-art multilingual transformers to produce summaries for English research papers in different languages.Specifically, we propose the new task of cross-lingual extreme summarization of scientific papers (CL-TLDR), since TLDR-like summaries have shown much promise in real-world applications such as search engines for academic publications like Semantic Scholar. 2n order to evaluate the difficulty of CL-TLDR and provide a benchmark to foster further research on this task, we create a new multilingual dataset of TLDRs in a variety of different languages (i.e., German, Italian, Chinese and Japanese).Our dataset consists of two main portions: (a) a translated version of the original dataset from Cachola et al. [10] in German, Italian and Chinese to enable comparability across languages on the basis of post-edited automatic translations; (b) a dataset of human-generated TLDRs in Japanese from a community-based summarization platform to test performance on a second, comparable human-generated dataset.Our work complements seminal efforts from Fatima and Strube [26], who compile an English-German cross-lingual dataset from the Spektrum der Wissenschaft/Scientific American and Wikipedia.We focus on extreme summarization, build a dataset of expert-derived multilingual TLDRs (as opposed to leads from Wikipedia) and provide additional languages.

Contributions.
Our work provides the following contributions on the research topic of cross-lingual summarization for the scholarly domain.
• We propose the new task of cross-lingual extreme summarization of scientific articles (CL-TLDR).• We create the first multilingual dataset for extreme summarization of scholarly papers from computer science in four different languages.
• We use our dataset to benchmark the difficulty of crosslingual extreme summarization with different models built on top of state-of-the-art pre-trained language models [49,57].• We additionally investigate whether cross-lingual summarization models using large pre-trained language models can be improved with intermediate finetuning techniques, which have shown to be effective to improve performance of pre-trained multilingual language models on many downstream NLP tasks [29, 32, 70, 71, inter alia].
We build upon our original paper [86] and extend it in a number of ways: • We benchmark the choice of the multilingual encoderdecoder by comparing performance of our original models using mBART [57] with those using mT5 [92].• We study the role of the stacking order in the summarization and translation pipeline approach, so as to establish whether we can achieve better cross-lingual summaries by first translating and then summarizing, or vice versa.• We further analyze the code-switching capabilities of our model by quantifying how much our multilingual models are able to retain English technical terminology in the translated summaries.• We investigate the application of a knowledge distillation method [82] on our direct cross-lingual summarization models to explore the possibility of shrinking the model sizes while keeping the original summarization output quality.
While the first three new contributions are meant to extend the experimental part so as to provide a more complete and in-depth analysis of our original experiments, the last one focuses on improving its scope of application.This is because the large size of the cross-lingual models we use in our experiments can hinder building scalable real-world applications around them.To address this point, we follow the recent trend in 'green' and scalable NLP [65] and explore how to reduce the computational inference costs of our summarization models using knowledge distillation.This is especially essential for our overarching future vision of coupling summarization with semantification techniques within the broader vision of the VADIS project, which aims at improving accessibility of social science publications by connecting survey data and text from research papers [44].The remainder of this paper is organized as follows.Section 2 provides an overview of relevant previous work in monolingual and multilingual summarization, as well as the broader field of scholarly document mining.We summarize in Sect. 3 seminal work on monolingual extreme summa-rization for English from Cachola et al. [10], on which our multilingual extension builds upon.We next introduce our new dataset for cross-lingual TLDR generation in Sect. 4. We present our cross-lingual models and benchmarking experiments in Sects.5 and 6, respectively.We wrap up our work with concluding remarks and directions for future work in Sect.7.

Datasets and resources
General-domain summarization datasets.News article platforms play a major role when collecting data for summarization [35,78], since article headlines provide ground-truth summaries.Narayan et al. [66] propose a news domain summarization dataset with highly compressed summaries to provide a more challenging summarization task (i.e., extreme summarization).Sotudeh et al. [84] propose TLDR9+, another extreme summarization dataset that was collected automatically from a social network service.
Cross-lingual summarization datasets.While there are growing numbers of cross-lingual datasets for natural language understanding tasks [18,53,74], few datasets for cross-lingual summarization are available.Zhu et al. [99] propose to use machine translation to extend English news summarization to Chinese.To ensure dataset quality, they adopt round-trip translation by translating the original summary into the target language and back-translating the result to the original language for comparison, keeping the ones that meet a predefined similarity threshold.Ouyang et al. [68] create cross-lingual summarization datasets by using machine translation for low-resource languages such as Somali, and show that they can generate better summaries in other languages by using noisy English input documents with English reference summaries.Our work differs from these prior attempts in that our automatically translated summaries are corrected by human annotators, as opposed to providing silver standards in the form of automatic translations without any human correction.Recently, Ladhak et al. [46] presented a large-scale multilingual dataset for the evaluation of cross-lingual abstractive summarization systems that are built out of parallel data from WikiHow.Even though it is a large high-quality resource of parallel data for cross-lingual summarization, this corpus is built from how-to guides: our dataset focuses instead on scholarly documents.Perez-Beltrachini and Lapata [69] automatically constructed datasets for cross-lingual summarization in four European languages by exploiting the structure of Wikipedia.Besides cross-lingual corpora, there are also large-scale multilingual summarization datasets for the news domain [80,87].The work we present here differs in that we focus on extreme summarization for the scholarly domain and we look specifically at the problem of cross-lingual summarization in which source and target language differ.

Datasets for summarization in the scholarly domain.
There are only a few existing summarization datasets for the scholarly domain and most of them are in English.SCITLDR [10], the basis for our work on multilingual summarization, presents a dataset for research papers (see Sect. 3 for more details).Collins et al. [17] use author-provided summaries to construct an extractive summarization dataset from computer science papers, with over 10,000 documents.Cohan et al. [14] regard abstract sections in papers as summaries and create large-scale datasets from two open-access repositories (arXiv and PubMed).Yasunaga et al. [95] efficiently create a dataset for the computational linguistics domain by manually exploiting the structure of papers.Meng et al. [62] present a dataset which contains four summaries from different aspects for each paper, which makes it possible to provide summaries depending on requests by users.Lu et al. [59] release a large-scale dataset for multi-document summarization for scientific papers, for which models need to summarize multiple documents.
The work closest to ours has been recently presented by Fatima and Strube [26], who introduce an English-German cross-lingual summarization dataset collected from German scientific magazines and Wikipedia.This resource is complementary to ours in many different aspects.While both datasets are in the scientific domain, their data include either articles from the popular science magazine Scientific American/Spektrum der Wissenschaft or articles from the Wikipedia Science Portal.In contrast, our dataset includes scientific publications written by researchers for a scientific audience.Second, our dataset focuses on extreme, TLDRlike summarization, which we argue is more effective in helping researchers browse through many potentially relevant publications in search engines for scholarly documents.Finally, our summaries are expert-generated, as opposed to relying on the 'wisdom of the crowds' from Wikipedia, and are available in three additional languages.

Models
Scholarly document mining.In recent years, there has been much interest from the NLP community in developing text mining techniques that bring order and provide novel ways to better access scientific publications [76].Previous work has addressed a wide range of tasks, including citation linking [2,3] and recommendation [34,38], summarization [1,77] (inter alia, see below) and argumentation mining [4,5,31].But while there have been full-fledged projects on mining scientific publications [72], scholarly document processing has arguably gained much traction lately [7,16], due to the ever growing need to efficiently access large amounts of published information, e.g., in the COVID-19 pandemic [24,89].Most recent contributions range from scholarly specific search platforms [47] all the way through novel reading interfaces [27] and full-fledged infrastructures [11,44] leveraging advancements in data-driven AI, NLP and semantification techniques (e.g., document understanding and information extraction).

Automated text summarization.
Summarization is a longstanding task in NLP [33,67].While early efforts focused mostly on extractive summarization [55], e.g., using an unsupervised graph-based approach [63], abstractive summarization has gained ever more traction in recent years starting with work using sequence-to-sequence models [75].Just like in virtually all areas of NLP research, most successful current approaches to summarization rely on neural techniques using supervision from labeled data.This includes neural models to summarize documents in general domains such as news articles [56,81], including cross-and multilingual models and datasets [80,87], as well as specialized ones e.g., the biomedical domain [64].Work on cross-lingual summarization has historically received little attention until recent years [90], arguably to due to the availability of new resources (Sect.2.1) as well as neural multilingual summarizers.
Summarization of scientific documents.In recent years, there has been much work on the problem of summarizing scientific publications and community-driven evaluation campaigns such as the CL-SciSumm shared tasks [12,40].Previous work on summarization has focused on specific features of scientific documents such as using citation contexts [13,97] or document structure [15,19].Complementary to these efforts is a recent line of work on automatically generating visual summaries or graphical abstracts [93,94].In our work, we build upon recent contributions on using multilingual pre-trained language models for cross-lingual summarization [46] and extreme summarization for English [10] and bring these two lines of research together to propose the new task of cross-lingual extreme summarization of scientific documents.
Knowledge distillation for summarization models.While massively large pretrained language models achieve strong results on various summarization tasks, the enormous sizes hinder their deployment in real-world applications.Knowledge distillation [36] offers a chance to reduce the model size by transferring knowledge of the original teacher model to a smaller student without large performance drops.Because of its practicality, there has been a lot of work exploring how to utilize this framework for various NLP tasks [41,79] as well as for summarization.Shleifer and Rush [82] perform comparative experiments of three different knowl-edge distillation methods for summarization models to better understand how they affect training and inference time as well as final summary quality.Zhang et al. [98], on the basis of their observation of how attention layers behave in summarization models, propose to modify the attention temperature parameter in the teacher model to generate pseudo-labels that are easier to learn for the student model.Li et al. [52] present a controlled study to understand the interaction between model quantization and distillation and report significant speed improvements.In our work, we utilize a simple yet effective knowledge distillation method called 'shrink and fine-tune' investigated by Shleifer and Rush [82] to understand its effects on our new cross-lingual extreme summarization task.

SCITLDR: English monolingual extreme summarization of scientific documents
Our work builds heavily on seminal work on extreme summarization of scientific publications from Cachola et al. [10], who first introduced an English monolingual dataset for this task and used it to benchmark a variety of state-of-the-art summarization models.SCITLDR is a dataset composed of pairs of research papers and corresponding summaries: in contrast to other existing datasets, this dataset is unique because of its focus on extreme summarization, i.e., very short, TLDR-like summaries and consequently high compression ratios-cf.the compression ratio of 238.1% of SCITLDR versus 36.5% of CLPubSum [17].An example of a TLDR summary is presented in Table 1, where we see how information from different summary-relevant sections of the paper (typically, in the abstract, introduction and conclusions) is often merged The paper provides a full characterization of permutation invariant and equivariant linear layers for graph data.Automatic Translation L'articolo fornisce una caratterizzazione completa degli strati lineari invarianti di permutazione ed equivarianti per i dati del grafico .Postedited Version L'articolo fornisce una caratterizzazione completa dei layer lineari invarianti o equivarianti per la permutazione per i dati del grafo .We present a novel iterative algorithm based on generalized low rank models for computing and interpreting word embedding models.
to provide a very short summary that is meant to help readers quickly understand the key message and contribution of the paper.
The original SCITLDR dataset consists of 5411 TLDRs for 3229 scientific papers in the computer science domain: it is divided into a training set of 1992 papers, each with a single gold-standard TLDR, and dev and test sets of 619 and 618 papers each, with 1452 and 1967 TLDRs, respectively (thus being multi-target in that a document can have multiple gold-standard TLDRs).The summaries consist of TLDRs written by authors and collected from the peer review plat-form OpenReview, 3 as well as human-generated summaries from peer-review comments found on the same platform.
In the following, we extend the original work of Cachola et al. [10] in two different ways, namely in terms of: (a) a new multilingual dataset for TLDR-like extreme summarization in languages other than English and (b) benchmarking of multilingual transformer-based pre-trained generative language models.We achieve this by creating a new crosslingual dataset that consists of an automatically translated, post-edited version of the SCITLDR dataset to support four additional languages, namely German, Italian, Chinese and BART is fine-tuned to convert a given text (e.g., paper abstract) into a shorter summary in a regressive manner by generating one token at a time Japanese.We then use these reference summaries to fine-tune pre-trained language models and produce multilingual summarization systems that are able to support languages other than English as the target language.

X-SCITLDR: a new dataset for cross-lingual extreme summarization of scientific papers
We first describe the creation of our X-SCITLDR dataset and briefly present some statistics to provide a quantitative overview.Our dataset is composed of two main sources: • An automatically translated, manually post-edited version of the original SCITLDR dataset [10] for German, Italian and Chinese (X-SCITLDR-PostEdit). • A manually generated dataset of expert-authored TLDRs harvested from a community-based summarization platform for Japanese (X-SCITLDR-Human).
Besides allowing us to evaluate our models across languages with different sizes of pre-training data (e.g., mBART has been exposed to half as much Italian than Japanese or German, cf.Table 1 from [57]), using two different sources allows us to perform a 'cross-domain'-like evaluation between datasets from different sources, namely conference reviews (X-SCITLDR-PostEdit) versus expert community efforts (X-SCITLDR-Human), so as to evaluate the generalization capabilities of our models across different domains.Moreover, having a dataset comprising post-edited translated summaries and human-generated ones makes it possible to investigate performance across different summarization styles-since post-edited summaries are not guaranteed to be the same as the ones humans would have generated from scratch.

X-SCITLDR-PostEdit.
Given the overall quality of automatic translators [20], we opt for a hybrid machine-human translation process of post-editing [30] in which human annotators correct machine-generated translations as postprocessing to achieve higher quality than when only using an automatic system.Although current machine translation systems arguably provide nowadays high-quality translations, a manual correction process is still necessary for our data, especially given their domain specificity.In Tables 2 and 3, we present examples of how translations are corrected by human annotators and the reasons for the correction.These can be grouped into two cases: (a) Wrong translation due to selected wrong sense (Table 2).In this case, the machine translation system has problems selecting the domain-specific sense and translation of the source term.(b) Translation of technical terms (Table 3).To avoid having the same technical term being translated in different ways, we reduce the sparsity of the translated summaries and simplify the translation task by preserving technical terms in English.
Both cases indicate the problems of the translation system with domain-specific terminology.For the underlying translation system, we use DeepL. 4After the automatic translation process, we asked graduate students in computer science courses who are native speakers in the target language to fix incorrect translations.) translates an English abstract to the target language and then summarizes it with a monolingual summarization model in the same target language X-SCITLDR-Human.We complement the translated portion of the original TLDR dataset with a new dataset in Japanese crawled from the Web.For this, we collect TLDRs of scientific papers from a community-based summarization platform, arXivTimes. 5This Japanese online platform is actively updated by users who voluntarily add links to papers and a corresponding user-provided short summary.The posted papers cover a wide range of machine learning-related topics (e.g., computer vision, natural language processing and reinforcement learning).This second dataset portion allows us to test with a dataset for extreme summarization of research papers in an additional language and, crucially, with data entirely written by humans, which might result in a writing style different from the one in X-SCITLDR-PostEdit.That is, we can use these data not only to test the capabilities of multilingual summarization in yet another language but, more importantly, test how much our models are potentially overfitting by too closely optimizing to learn the style of the X-SCITLDR-PostEdit summaries or vice versa.In Table 4, we present various statistics of our X-SCITLDR dataset for both documents and summaries from the original English (EN) SCITLDR data and our new dataset in four target languages. 6SCITLDR and X-SCITLDR-PostEdit (DE/IT/ZH) have a comparably high compression ratio (namely, the average number of words per document to the average number of words per summary) across all four languages, thus indeed requiring extreme cross-lingual compression capabilities.While summaries in German, Italian and Chinese keep the compression ratio close to the original Fig. 3 mBART learns to take an English abstract and generate a summary in the target language (here, German).We can control the target language by providing a language token (< de > in the figure) Base models: mBART and mT5.In our experiments, we use BART [49] and its multilingual variant mBART [57] as underlying summarization models.They are both transformer-based [88] pre-trained generative language models, which are trained with an objective to reconstruct noised text in an unsupervised sequence-to-sequence fashion.While BART only uses an English corpus for pre-training, mBART learns from a corpus containing multiple languages.These pre-trained BART/mBART models can be further trained (i.e., fine-tuned) in order to be applied to downstream tasks of interest like, for instance, summarization, translation or dialogue generation-cf.Fig. 1.
We use BART/mBART as our underlying models, since these have been shown in previous work to perform well on the task of extreme summarization [49].We follow Ladhak et al. [46] and use BART/mBART as components of two different architectures, namely: (a) two-step approach to cross-lingual summarization, i.e., summarization via BART and translation using machine translation (MT) (Sect.5.1); (b) a direct cross-lingual summarization system obtained by fine-tuning mBART with input articles from English and summaries from the target language (Sect.5.2).
In addition to mBART, we quantify performance using different pre-trained language models, in order to benchmark how stable our results are across different transformer-based encoder-decoder models.For this we additionally evaluate performance using mT5 [92], which, akin to mBART, is a large pre-training language model designed to generate texts in multiple languages.When compared with mBART, mT5 has two major differences, namely a) the noising function used during pre-training-i.e., while mBART learns to recover masked spans and shuffled sentences in texts, mT5 only uses span masking as a noising function-and b) the overall model size-i.e., mBART and mT5-contains 610 and 1229 million of parameters, respectively, which impacts computational costs (e.g., required computational time, memory consumption).

Two-stage cross-lingual summarization
A first solution to the CL-TLDR task is to combine a monolingual summarization model with a machine translation system.This approach is composed of two stages, namely translation and (monolingual) summarization.In this work, we investigate two variants of this setting, namely 'summarize and translate' and 'translate and summarize', whose difference is the different stacking order of the two modules, i.e., whether we first translate and then (monolingually) summarize or vice versa.Summarize and Translate.One variant of two-stage crosslingual summarization is to have the model first take an English text as input and then generate a summary in English (we call this approach EnSum-MT): the English summary is then automatically translated into the target language using machine translation (Fig. 2a). 7This model does not rely on any cross-lingual signal: it merely consists of two independent modules for translation and summarization and does not require any cross-lingual dataset to train the summarization model.
While this system is conceptually simple, such a pipeline approach is known to cause an error propagation problem [99], since errors of the first stage (i.e., summarization) get amplified in the second stage (i.e., translation) leading to overall performance degradation.
Translate and Summarize.An alternative two-stage summarization pipeline consists in training monolingual summarization models for each target language by translating English input documents to match the language of reference summaries (we call this method TGTSum-MT).During testing, input documents are then similarly first translated and then summarized using the corresponding monolingual summarization model.
While having a model observe more text in the target language is known to help its performance [39], the overall cost of translating texts is higher for TGTSum-MT than in EnSum-MT as it requires (1) to translate the entire input documents, as opposed to (shorter) generated summaries, and (2) to perform automatic translation not only of the test documents (for evaluation) but also of the train and development sets (for training monolingual summarization models and tune their hyperparameters).

Direct cross-lingual summarization
A third approach to CL-TLDR is to directly perform crosslingual summarization using a pre-trained multilingual language model (we call this method CLSum).For this, we investigate the use of pre-trained multilingual denoising autoencoders like mBART [57] and mT5 [92] and use the cross-lingual training data provided by our new X-SCITLDR dataset to fine-tune them and generate summaries in the target languages given abstracts in English, as depicted in Fig. 3.We follow Liu et al. [57] and control the target language by providing a language token to the decoder.
Intermediate task and cross-lingual fine-tuning.Our training dataset is relatively small compared to datasets for general domain summarization [35,66].To mitigate this data scarcity problem, we investigate the effectiveness of intermediate fine-tuning, which has been reported to improve a wide range of downstream NLP tasks (see, among others, [29,32,70,71]).Gururangan et al. [32], for instance, show that training pre-trained language models on texts in a domain/task similar to the target domain/task can boost the performance on the downstream task by injecting additional related knowledge into the models.Based on this observation, in our experiments, we investigate two strategies for intermediate fine-tuning: intermediate task and cross-lingual fine-tuning.
• Intermediate task fine-tuning (CLSum+EnSum).We explore the benefits of using additional summarization data other than the summaries in the target language and augment the training dataset with English data, i.e., the original SCITLDR data.That is, before fine-tuning on summaries in the target language (e.g., German), we train the model on English TLDR summarization as auxiliary monolingual summarization task to provide additional summarization capabilities (Fig. 4a).

• Cross-lingual intermediate fine-tuning (CLSum+MT).
Direct cross-lingual summarization requires the model to learn both translation and summarization skills, arguably a difficult task given our small dataset. 8To alleviate this problem, we investigate training our model on machine translation before fine-tuning it on the summarization task.For this, we automatically translate English abstracts into the target language and use these synthetic data as training data for fine-tuning on the task of automatically translating abstracts (Fig. 4b).

Knowledge distillation.
Akin to other recent pre-trained language models, the demanding computational requirements of mBART hinder its deployment in real-world applications.To tackle this issue, there are works that aim to reduce the size of large summarization models [52,83,98].In our work, we evaluate one of the knowledge distillation approaches proposed by Shleifer and Rush [83], dubbed 'shrink and fine-tune,' which takes a trained mBART as a teacher model and uses some of its parameters for the initialization of a smaller version of the model (called student) and finally fine-tunes the student model again on the target dataset.Clearly, training teacher and student takes more time than fine-tuning the teacher alone but this provides us with a smaller model which can be more easily deployed.

Experiments
Input documents.We follow Cachola et al. [10] and rely in all our experiments on an input consisting of abstracts only, since they showed that it yields similar results when compared to using the abstract, introduction and conclusion sections together.Even more importantly, using only abstracts enables the applicability of our models also to those cases where only the abstracts are freely available and we do not have open access to the complete manuscripts.The average length of an abstract is 185.9 words for X-SCITLDR-PostEdit (EN/DE/IT/ZH) with an average compression ratio of 12.64% and 190.9 words for X-SCITLDR-Human (Japanese) and a compression ratio of 39.76%.
Evaluation metrics.We compute performance using a standard metric to automatically evaluate summarization systems, namely ROUGE [54].In the case of the post-edited portion of the X-SCITLDR dataset (X-SCITLDR-PostEdit, Sect.4), the gold standard can contain multiple reference summaries for a given paper and abstract.Consequently, for Italian, German and Chinese we calculate ROUGE scores in two ways (avg and max) to account for these multiple references [10].For avg, we compute ROUGE F1 scores with respect to the different references and take the average score, whereas for max we select the highest scoring one.The Japanese dataset does not contain multiple reference TLDRs: hence, we compute standard ROUGE F1 only.We test for statistical significance using sample level Wilcoxon signed-rank test with α = 0.05 [23].
Hyperparameter tuning.To find the best hyperparameters for each model, we use the development data and run a grid search using ROUGE-1 avg as a reference metric.We run experiments with learning rate ∈ {1 • 10 −5 , 3 • 10 −5 } and random seed ∈ {1122, 22} during fine-tuning, number of beams for beam search ∈ {2, 3} and repetition penalty rate ∈ {0.8, 1.0} during decoding.For all settings, we set the batch size equal to 1 and perform 8 steps of gradient accumulation.We use the AdamW optimizer [58] with weight decay of 0.01 for 5 epochs without warm-up.
Training strategy.To prevent the model from losing the knowledge acquired in pre-training during fine-tuning (i.e., catastrophic forgetting [45]), we freeze the parameters of the embedding and decoder layers during intermediate task and cross-lingual fine-tuning [60], while we update the entire model during the final fine-tuning for downstream CL-TLDR.Since mBART requires large memory space when we update the entire model, we utilize the DeepSpeed library 9 to meet our infrastructure requirements.Our models are built using PyTorchLightning [25] and HuggingFace Transformers [91].
Research questions.We organize the presentation and discussion of our results in the remainder of this section using the following research questions:

Architecture
• RQ1: Which pre-trained multilingual language model, mBART or mT5, is best suitable for performing direct cross-lingual summarization on our dataset?

Analysis
• RQ7: How challenging is code-switching summary generation for direct cross-lingual summarization models?• RQ8: How much data do we need to perform crosslingual TLDR summarization?What is the performance in a zero-shot or few-shot setting?• RQ9: What are the major kinds of errors that can be found by manual inspection of the summaries generated by direct cross-lingual models?
RQ1: mBART vs mT5.All our models, be it either a pipeline architecture or a direct cross-lingual summarizer, rely for summarization on a transformer-based multilingual pre-trained language model (cf.Sect.5).Since different large-scale multilingual generative pre-training language models available to perform cross-lingual text generation, we first investigate which model to use: consequently, in this first set of experiments, we compare two popular models, namely mBART [51] and mT5 [92], which support all languages contained in our X-SCITLDR dataset and have been shown to achieve state-of-the-performance on automated text summarization.
We compare these two models along two different dimensions, namely overall summarization performance using ROUGE scores, as well as efficiency, here measured as number of summaries produced per second.Comparison of ROUGE scores and summary generation speed for both models on the direct cross-lingual summarization setup (Sect.5.2) is shown in Table 5.We observe that the two models perform on par across different languages, e.g., mBART shows better performance on German, whereas mT5 slightly outperforms it on Italian and Chinese (possibly because of a larger exposure during pre-training to these languages). 10mT5, however, has twice more parameters, which leads to more expensive computational costs, as highlighted by the model being able to generate half the summaries per second when compared with mBART.Given the comparable performance, yet this major difference in text generation speed, we opt for mBART in the remainder of our experiments.RQ2: summarize and translate versus translate and summarize.We next conduct experiments to compare two ways to implement a pipeline architecture for cross-lingual summarization using a summarization and translation module in different orders, namely a 'summarize and translate' 10 ROUGE-2 scores are (much) lower than other ROUGE metrics since it computes matches of consecutive bigrams, which is harder and thus less frequent than matching unigrams (evaluated by ROUGE-1) or nonconsecutive longest common sequences (evaluated by ROUGE-L), and which is in line with previous works [56,75] Table 7 Results on the X-SCITLDR-Human portion of our crosslingual TLDR dataset (Rouge-1,-2 and -L): English to Japanese TLDR-like summarization using human-generated summaries from ArXivTimes.Best results per metric are bolded
Results on X-SCITLDR-PostEdit (German, Italian and Chinese) and X-SCITLDR-Human (Japanese) portions of our dataset are shown in Tables 6 and 7, respectively.We notice that performance varies greatly across the two different portions of the dataset.On the automatically translated, post-edited portion of the data (Table 6) we observe no major difference in ROUGE scores between the two different architectures.On the contrary, on the manually generated, expert-authored portion (Table 7) TGTSum-MT outperforms EnSum-MT by a large margin.This is because EnSum-MT relies on English monolingual summarization using Englishonly SCITLDR as training data for fine-tuning.On the contrary, TGTSum-MT uses target language-specific models, which have been fine-tuned on the respective languages using multilingual summaries from X-SCITLDR.Using language-specific fine-tuning can, in turn, better account for different styles across different subsets of our data (cf.different summary length and compression ratio of JA vs. DE/IT/ZH in Table 4), thus allowing the generation of summaries that are best aligned with the test data.

RQ3: two-stage versus direct cross-lingual summarization.
We next compare our two main architectures, namely the pipeline models (Sect.5.1, EnSum-MT/TGTSum) with a multilingual encoder-decoder architecture based on mBART (Sect.5.2, CLSum/+EnSum/+MT), again on the basis of the results on the two main portions of our dataset, i.e., postedited translations and human-generated summaries, from Tables 6 and 7, respectively.
Similarly to the case of RQ2, we again see major performance differences across the two dataset portions.While MT-based summarization (EnSum-MT/TGTSum) is superior or comparable when evaluated on translated/postedited TLDRs in German, Italian and Chinese (Table 6), the direct cross-lingual summarization model (CLSum) improves results on native Japanese summaries by a large margin when compared to the 'summarize and translate' EnSum-MT pipeline (Table 7).The differences in perfor- mance figures between X-SCITLDR-PostEdit and X-SCITLDR-Human are due to the different nature of the multilingual data, and how they were created.Post-edited data like those in German, Italian and Chinese are indeed automatically translated and tend to better align to the also automatically translated English summaries, as provided as output of the EnSum-MT system.That is, since both summaries-the post-edited reference ones and those automatically generated and translated-go through the same process of automatic machine translation, they naturally tend to have a higher lexical overlap, i.e., a higher overlap in terms of shared word sequences.This, in turn, receives a higher reward from ROUGE, since this metric relies on n-gram overlap between system and reference summaries.While EnSum-MT seems to better align with post-edited reference TLDRs, for human-generated Japanese summaries we observe an opposite behavior.Japanese summaries indeed have a different style than those in English (and their postedited multilingual versions from X-SCITLDR-PostEdit) and accordingly have a lower degree of lexical overlap with translated English summaries from EnSum-MT.As a result of this, models that have been trained on target languagespecific data like 'translate and summarize' (TGTSum-MT) and direct cross-lingual summarization (CLSum) are better adapted to style variations across different portions and languages of our cross-lingual dataset and thus achieve better results.

RQ4: post-edited versus human-generated cross-lingual summaries.
To better understand the behavior of the system in light of the different performance on post-edited versus human-generated data, we manually inspected the output of the two systems.Table 8 shows an example of automatically generated summaries for a given input abstract: it highlights that summaries generated using our crosslingual models (CLSum) tend to be shorter and consequently 'abstracter' than those created by translating English summaries (EnSum-MT).This, in turn, can hurt the performance of the cross-lingual models more in that, while we follow standard practices and use ROUGE F1, this metric has been found unable to address the problems with ROUGE recall, which rewards longer summaries, in the ranges of typical summary lengths produced by neural systems [85].and thus allow for language-and style-specific adaptation.Within the post-edited portion of our dataset, EnSum-MT performs significantly better than the cross-lingual models in German and Chinese; however, there is generally no significant difference with cross-lingual models in Italian, where CLSum+MT is even able to achieve statistically significant improvements on average Rouge-1 and Rouge-L.To better understand such different behavior across languages, we computed for each language the word-level Jaccard coefficients between the automatically translated summaries and their post-edited versions in X-SCITLDR-PostEdit.
The Jaccard coefficient takes two sets, i.e., in our case words from automatically translated English TLDR summaries and their human-corrected version, and computes the ratio of overlapping elements to all elements in both sets, thus making it possible to quantify the rate of human correction as the ratio of words that are fixed during the process of post-editing.As Table 10 shows, the Italian post-edited translations contain more edits than the other two languages.This, in turn, seems to disadvantage the two-stage EnSum-MT pipeline, whose output aligns more with the 'vanilla' automatic translations.We notice also differences in absolute numbers between German, Italian and Chinese, which could be due to the distribution of training data used to train the multilingual transformer [57], with mBART being trained on more Chinese than Italian data.However, German performs worst among the three languages, despite mBART being trained on more German data than Chinese or Italian.Manual inspection reveals that German summaries tend to be penalized more because of differences in word compounding between refer-ence and generated summaries: while there exist proposals to address this problem in terms of language-specific preprocessing [28], we opt here for a standard evaluation setting equal for all languages.Moreover, German summaries tend to contain less English terms than, for instance, Italian summaries (6.78 vs. 4.88 English terms per summary on average in the test data), which seems to put the cross-lingual model at an advantage (cf.English terminology in EnSum-MT vs. CL-Sum in Table 8).The performance gap between EnSum-MT and CLSum is the largest on the Chinese dataset, which shows that it is more challenging for mBART to learn to summarize from English into a more distant language [48].

RQ5: The potential benefits of intermediate fine-tuning.
In the next set of experiments, we investigate whether intermediate-stage training, which aims at learning the summarization and translation tasks from additional data, can improve the performance of a vanilla cross-lingual model.Specifically, we compare target-language fine-tuning of mBART (CLSum) with additional intermediate task finetuning on English monolingual summarization (CLSum+EnSum) and cross-lingual intermediate fine-tuning using machine translation on synthetic data (CLSum+MT).The rationale behind these experiments is that in the direct cross-lingual setting the model needs to acquire both summarization and translation capabilities, which requires a large amount of cross-lingual training data, and thus might be hindered by the limited size of our dataset.
Including additional training on summarization based on English data (CLSum+EnSum) has virtually no effects on the translated portion of SCITLDR (Table 6) for German and even degrades performance on Italian and Chinese.This is likely because English TLDR summaries are well aligned with their post-edited translations and virtually bring no additional signal while requiring the decoder to additionally translate into a new language (i.e., English and the target language).On the contrary, CLSum+EnSum improves on all ROUGE metrics for Japanese (Table 7).This is because, as previously mentioned, the Japanese data have a different style from those from SCITLDR consequently, English TLDRs provide an additional training signal that helps to improve results for the summarization task.
Including MT-based pre-training, i.e., fine-tuning mBART on texts that have been automatically translated from English into the target language, and then on cross-lingual summarization (CLSum+MT) improves over simple direct crosslingual summarization (CLSum) on all languages-a finding in line with results from Ladhak et al. [46] for WikiHow summarization.This highlights the importance of fine-tuning the encoder-decoder for translation before actual fine-tuning for the specific cross-lingual task, thus injecting general translation capabilities into the model.

RQ6:
The effects of knowledge distillation.We now investigate the effect of knowledge distillation for cross-lingual summarization models on our dataset.In our experiments we use the 'shrink and fine-tune' (SFT) distillation introduced in Shleifer and Rush [83], as it has been shown to perform well on summarization while being conceptually simple.Using this method, we initialize a smaller model by taking the parameters from a fine-tuned full-sized model before fine-tuning it again on our dataset.Specifically, this method extracts the smaller student model by taking the maximally spaced layers of a fine-tuned teacher (with ties being broken arbitrarily), copying the selected layers from teacher to student and re-fine-tuning the student model (see Fig. 5 for a schematic overview).For example, when creating a distilled BART model with three encoder and three decoder layers (referred to as a 3-3 student model) from a full 12-12 BART teacher, we copy layers 0, 6 and 11 of both encoder and decoder to the student before re-fine-tuning it.Consequently, to understand the effect of layer selection, we conduct experiments with several combinations of layers of the teacher model to be used to initialize the student model and analyze their performance on the cross-lingual summarization task.
The results of layer selection for SFT knowledge distillation are presented in Table 11, where we denote with N-M the number of layers copied from the teacher encoder and decoder to the student.To highlight the reduction in computational costs provided by the student, i.e., distilled, models, we complement ROUGE scores with the number of parameters and the number of summaries that a model can generate in a second in our infrastructure.To obtain the latter, we take the best-performing CLSum model for each target language and generate summaries on a single NVIDIA RTX A6000 GPU with batch size 32, number of beams equal to 3 and 1.0 as the repetition penalty rate.The number of summaries per second is only comparable within a language as the generating speed is highly dependent on the output sequence length which varies for each language and data source.Reducing the number of layers in half (i.e.,  does not reduce the parameters by half since mBART contains, in addition to the encoder and decoder, an embedding layer and a final prediction head.As shown in the comparison of 12-3 and 3-12, removing layers in the decoder results in greater parameter reduction and faster inference since one decoder layer contains more parameters than an encoder layer, due to the additional encoder-decoder attention parameters which are not present in the encoder layers.By removing the same number of layers from the encoder and decoder, as in the 12-3 versus 3-12 and 12-6 versus 6-12 setups, we observe that removing layers from the encoder has a stronger negative impact on the ROUGE score and provides lower inference time speedup, in line with previous findings by Shleifer and Rush [83].From this we can draw a useful practical tip for future work: when distilling summarization models, it is more cost-and performance-efficient to reduce the layers in the decoder.
The impact of distillation on the ROUGE score is highly dependent on the target language.For German and Japanese, the performance drop is minor even when one-fourth of the layers is removed.In contrast, in Italian and Chinese ROUGE scores can drop up to 6.35 points which reflects a large degradation of the summary quality.
Table 11 Comparison between the original models and the distilled students: scores in parentheses report the difference with the original model.For each language, we order the models by the number of parameters.The second column shows how many layers are transferred from the encoder and decoder: e.g., 12-3 indicates that twelve layers from the encoder and three layers from the decoder were copied into the student, while 12 3).To better understand how much our direct cross-lingual models are able to generate these codeswitched summaries, 11 we collect words 'copied' from the original English summaries by extracting overlapping words between original English summaries and post-edited versions, and compute the coverage of such words within summaries generated by direct cross-lingual summarization models (CLSum) on X-SCITLDR-PostEdit.Concretely, we compute coverage as follows: where N , E i , T i , H i are the number of summaries in the dataset, the set of words (i.e., unigrams) from the original English reference summary, its post-edited version in the target language and a generated summary, respectively.We present the coverage of code-switching words in Table 12.
Overall, the remarkably low scores indicate the difficulty of including those words in summaries, especially in Chinese.
With respect to Italian and German, Chinese is the language most typologically distant from English, which is the most common language that takes part in code-switching in our dataset.Generally, manual inspection of the output reveals how our CL summarization models have only limited ability to generate (often, English) domain-specific terminology, a point to which we come back later in the qualitative evaluation.
RQ8: Zero-and few-shot experiments.To better understand the contribution of intermediate fine-tuning and to analyze performance in the absence of multilingual summarization training data (i.e., in zero-shot settings), we present experiments in which we compare: a) using mBART with no fine-tuning (CLSum); b) fine-tuning mBART on English SCITLDR data only and evaluating performance on X-SCITLDR in our four languages; c) fine-tuning mBART on synthetic translations of abstracts only and testing on X-SCITLDR.These experiments are meant to quantify the zero-shot cross-lingual transfer capabilities of the crosslingual models (i.e., can we train on English summarization data only without the need of a multilingual dataset?) as well as to explore how much we can get away with constructing cross-lingual summarization data at all (i.e., what is the performance of a system that is trained to simply translate abstracts?).We present our results in Table 13.The performance figures indicate that the zero-shot cross-lingual transfer performance of CLSum+EnSum is extremely low for all our four languages, with reference performance on English TLDRs from Cachola et al. [10] being 31.1/10.7/24.4R1/R2/RL (cf.Table 10, BART 'abstract-only'), and barely improves over no fine-tuning at all (CLSum).This suggests that robust cross-lingual transfer in our summarization task is more difficult than in other language understanding tasks (see for example the much higher average performance on the XTREME tasks [37]).The overall very good performance of CLSum+MT seems to suggest that robust cross-lingual summarization performance can still be achieved without multilingual summarization data through the 'shortcut' of fine-tuning a multilingual pre-trained model to translate English abstracts, since these can indeed be seen as summaries (albeit of a longer length than our TLDR-like summaries) and thus provide a strong baseline.
We finally present results for our models in a few-shot scenario to investigate performance using cross-lingual data with a limited number of example summaries (shots) in the target language.Figure 6 shows few-shot results averaged across all four target languages (detailed per-language results are given in Table 14) for different sizes of the training data from the X-SCITLDR dataset.The results highlight that, while CL-TLDR is a difficult task with the models having little cross-lingual transfer capabilities (as shown in the zero-shot experiments), performance can be substantially improved when combining a small amount of cross-lingual data, i.e., as little as 1% of examples for each target language, and intermediate training.As the amount of cross-lingual training data increases, the benefits of intermediate finetuning become smaller and results for all models tend to converge.This indicates the benefits of intermediate finetuning in the scenario of limited training data such as, e.g., low-resource languages other than those in our resource.Our few-shot results indicate that we can potentially generate TLDRs for a multitude of languages by creating a small amount of labeled data in those languages, and at the same time leverage via intermediate fine-tuning labeled resources for English summarization and machine translation (for which there exist plenty of resources).

RQ9: Common errors of direct cross-lingual summarization models.
To shed light on the main errors found in the summaries generated by direct cross-lingual summarization models, we conclude our empirical analysis by presenting a list of limitations revealed through manual quality evaluation.Such inspection of the output of the CLSum model shows that, while using pre-trained language models ensures that the output is generally fluent, the generated summaries still have limitations along two main dimensions: • Domain specificity and technical terminology: summaries fail to include important 'keywords' that define fine-grained aspects of the employed models, developed methodology or experimental evaluation they perform.For instance, in Table 15a) both abstract and reference summary contain the expression 'catastrophic forgetting'  in English and German, respectively, which plays an important role to explain the corresponding paper.Such notion, however, is missing in the generated summary.• Overly generic summaries: summaries are correct, but are too generic in that they do not cover enough specific details of the scientific paper.For instance, in the automatically generated Italian summary from Table 15b), there is no mention of 'confidence thresholding' or 'defense against adversarial examples.' Notice how both of these errors are critical for summarization in the scholarly domain, since information seeking in this domain is heavily focused on domain-specific information and technical lingo that is specific to certain research communities.

Conclusion
In this paper, we presented X-SCITLDR, the first dataset for cross-lingual summarization of scientific papers.Our new dataset makes it possible to train and evaluate NLP models that can generate summaries in German, Italian, Chinese and Japanese from input papers in English.We used our dataset to investigate the performance of different architectures based on multilingual transformers, including two-stage 'summarize and translate' (or vice-versa) approaches and a direct cross-lingual approach with two different underlying models.We additionally explored the potential benefits of intermediate task and cross-lingual fine-tuning and analyzed the performance in zero-and few-shot scenarios as well as the model's behavior on 'code-switched' texts.Furthermore, we conducted extensive experiments with a knowledge distillation approach aimed at reducing model size in a performance-preserving way.For future work, we plan to extend X-SCITLDR to include papers from research communities other than computer science or other STEM-fields, specifically those that use languages other than English for professional communication (e.g., humanities in Germanspeaking countries).From a methodological perspective, we plan to investigate how to apply additional techniques designed for cross-lingual text generation such as training with multiple decoders [99], automatically complementing our multilingual TLDRs with visual summaries [93], as well as devising new methods to include background knowledge such as, in our case, technical terminology and domain adaptation capabilities [96], into multilingual pre-trained models.
Our work crucially builds upon recent advances in multilingual pre-trained models [57] and cross-lingual summa-  rization [46] and investigates how these methodologies can be applied for multilingual scholarly document processing.
In future, we propose to explore the direction of keywordoriented summarization systems [22,50] to enforce our models to include domain-specific terminology and have better overall focus.
The application of NLP techniques for mining scientific papers has been primarily focused on the English language: with this work we want to put forward the vision of enabling scholarly document processing for a wider range of languages, ideally including both resource-rich and resourcepoor languages in the longer term.Our vision of 'Scholarly Document Processing for all languages' is in line with current trends in NLP (cf., e.g., [73] and [6], inter alia): while our initial effort concentrated here on fairly resource-rich languages, in future work we plan to focus specifically on Fig. 7 A screenshot of a web page of papers from COLING 2022 with one sentence summaries in German at https://sotaro.io/info/2022_coling/2022.coling.deresource-poor languages where multilingual NLP can and is indeed expected to make a difference in enabling wider (and consequently more diverse and fairer [43]) accessibility to scholarly resources.

Limitations
In this section, we follow recent proposals from major conferences in the Natural Language Processing community, 12 and present the limitations of our work.We hope to help researchers who plan to conduct further studies on cross-lingual summarization systems by hinting at possible extensions of this work.
Our work's first and foremost limitation is the X-SciTLDR dataset itself, as it only contains paper in the computer science domain.We are currently evaluating models trained with X-SciTLDR on social science papers in German for our ongoing VADIS project [44].This, however, still covers a tiny fraction of all scholarly documents that could be processed by crosslingual summarization systems.
Another limitation of this work is in the evaluation of model-generated summaries.Following prior research, we used ROUGE-1/2/L to compute performance scores on the summarization task.However, while ROUGE is the most widely used metric to evaluate summarization, several works from the literature show its problems and limitations, which call for more rigorous means of evaluation for summarization systems.One of the most reliable ways of evaluating is to hire human annotators to assess generated summaries.However, our experiments involve summaries in the scholarly domain for multiple languages, making manual evaluation highly expensive.Since the main objective of this work is constructing a multilingual gold standard and benchmarking existing models, we did not perform human evaluations on generated summaries.We plan to further investigate summary quality evaluation in detail in our future work.
Another limitation is that our analysis for research question 9 ("Common errors of direct cross-lingual summarization models") is qualitative rather than quantitative.However, we think such a level of analysis can shed light on future lines of research by identifying the weaknesses of the current summarization systems through manual inspections.We accordingly include our observations to facilitate further studies on cross-lingual extreme summarization.
Finally, this paper focuses on abstractive summarization using a multilingual pre-trained generative model as the underlying architecture.Much in line with other work in NLP, state-of-the-art models for summarization rely on transfer learning and the pre-trained model paradigm.We leave the exploration of cross-lingual extractive summarization, e.g., using multilingual embeddings, for future work.

Fig. 2
Fig. 2 Two pipelines for cross-lingual summarization using automatic machine translation and monolingual summarization in stacking orders: (a) 'summarize and translate' (Sect.5.1) first summarizes an English abstract and then translates the generated summary into the target lan-

Fig. 4
Fig. 4 Two approaches to perform the intermediate fine-tuning to mitigate the training data scarcity issue.(a) CLSum+Ensum inserts an additional training stage between pre-training and cross-lingual summarization fine-tuning in which the pre-trained language model learns

Fig.
Fig. Overview of the 'shrink and fine-tune' procedure: a) we fine-tune the teacher model in its original size; b) we shrink its size by selecting subsets of layers from the encoder and/or decoder and copying them into the student model; c) we fine-tune the student

Fig. 6
Fig.6 Few-shot results without (CLSum) and with intermediate task (+EnSum) and cross-lingual fine-tuning (+MT) for different sizes of the training data in the target language (i.e., different number of shots): 1%, 5%, 30% and 100% of the X-SCITLDR training set of each target language found in summaries generated by a multilingual pre-trained language model like BART: (a) domain specificity and technical terminology; (b) overly generic summaries a) missing keywords Abstract: Incremental class learning involves sequentially learning classes in bursts of examples from the same class.This violates the assumptions that underlie methods for training standard deep neural networks, and will cause them to suffer from catastrophic forgetting.Arguably, the best method for incremental class learning is iCaRL, but it requires storing training examples for each class, making it challenging to scale.Here, we propose FearNet for incremental class learning.[...] TLDR: FearNet ist ein speichereffizientes neuronales Netz, das von der Gedächtnisbildung im Säugetiergehirn inspiriert ist und inkrementelles Klassenlernen ohne katastrophales Vergessen ermöglicht.CLSum: Wir schlagen FearNet vor, ein neuronales Lernsystem, das neue Erinnerungen aus einem Netund das basolaterale amygdala verwendet.b)too generic summary Abstract: Current machine learning algorithms can be easily fooled by adversarial examples.One possible solution path is to make models that use confidence thresholding to avoid making mistakes.Such models refuse to make a prediction when they are not confident of their answer.We propose to evaluate such models in terms of tradeoff curves with the goal of high success rate on clean examples and low failure rate on adversarial examples.[...] TLDR: Presentiamo una metrica e un attacco ottimale per valutare i modelli che si difendono dagli adversarial example usando confidence thresholding.CLSum: Proponiamo una famiglia di attacchi Max-Confidence per i modelli di apprendimento automatico.

Table 2
Example of a post-editing correction (wrong sense): 'Papier' means a generic piece of paper but not a research paper in German ('Artikel').Similarly, English 'graph' needs to be translated as 'grafo' as opposed to 'grafico' (English: 'diagram')

Table 4
Statistics of our dataset (X-SCITLDR) Fig. 1 Using monolingual BART for English text summarization:

Table 5
Comparison between mBART and mT5 on our cross-lingual TLDR dataset: we report ROUGE-1,-2 and -L scores as well as how many summaries each model can generate in a second (# Sum/S).Best results per language and metric are in bold, and significant difference within the sub-table of each language is marked with † Which architecture-i.e., two-stage or direct crosslingual summarization (Sects.5.1 and 5.2)-is overall best suited for the CL-TLDR task?
• RQ4: How do the results compare for different kinds of cross-lingual data, i.e., different portions of our dataset (X-SCITLDR-PostEdit vs. -Human, Sect.4)? • RQ5: Does intermediate-stage fine-tuning help improve direct CL-TLDR summarization? • RQ6: Can we reduce model size while keeping its summarization ability by knowledge distillation?

Table 6
[10]lts on the X-SCITLDR-PostEdit portion of our crosslingual TLDR dataset (ROUGE-1,-2 and -L): English to German, Italian or Chinese TLDR-like summarization using post-edited, automatically translated summaries of the English data from Cachola et al.[10].Best results per language and metric are bolded.Statistically significant improvements of the cross-lingual models (CLSum/+EnSum/+MT) with respect to the 'summarize and translate' pipeline (EnSum-MT) are marked with †

Table 8
Example of gold-standard summaries and automatically gen-

Table 9
Average summary length (number of tokens) of pipeline-based versus cross-lingual models

Table 10
Word-level Jaccard coefficients between automatically translated summaries and their post-edited versions -12 represents the original model

Table 12
CLSum model coverage of 'copied' words inclusion in generated summaries on the test split of the post-edited portion of our dataset

Table 14
Detailed few-shot performance on downstream CL-TLDR for each language and cross-lingual model with different percentages of X-SCITLDR training data for each target language

Table 15
Two main categories of errors