This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Adaptive Cross-lingual Text Classification through In-Context One-Shot Demonstrations

Emilio Villa-Cueva1,2, A. Pastor López-Monroy1,
Fernando Sánchez-Vega1,4, Thamar Solorio2,3

1Mathematics Research Center (CIMAT), Gto, Mexico,
2Mohamed bin Zayed University of Artificial Intelligence, UAE,
3University of Houston, USA,
4Consejo Nacional de Humanidades, Ciencias y Tecnologías, México
Abstract

Zero-Shot Cross-lingual Transfer (ZS-XLT) utilizes a model trained in a source language to make predictions in another language, often with a performance loss. To alleviate this, additional improvements can be achieved through subsequent adaptation using examples in the target language. In this paper, we exploit In-Context Tuning (ICT) for One-Shot Cross-lingual transfer in the classification task by introducing In-Context Cross-lingual Transfer (IC-XLT). The novel concept involves training a model to learn from context examples and subsequently adapting it during inference to a target language by prepending a One-Shot context demonstration in that language. Our results show that IC-XLT successfully leverages target-language examples to improve the cross-lingual capabilities of the evaluated mT5 model, outperforming prompt-based models in the Zero and Few-shot scenarios adapted through fine-tuning. Moreover, we show that when source-language data is limited, the fine-tuning framework employed for IC-XLT performs comparably to prompt-based fine-tuning with significantly more training data in the source language.

Adaptive Cross-lingual Text Classification through In-Context One-Shot Demonstrations


Emilio Villa-Cueva1,2, A. Pastor López-Monroy1, Fernando Sánchez-Vega1,4, Thamar Solorio2,3 1Mathematics Research Center (CIMAT), Gto, Mexico, 2Mohamed bin Zayed University of Artificial Intelligence, UAE, 3University of Houston, USA, 4Consejo Nacional de Humanidades, Ciencias y Tecnologías, México


1 Introduction

The recent progress in the development of multilingual Language Models (LMs) has allowed for effective cross-lingual transfer (XLT) with minimal need for architectural modifications Pires et al. (2019); Xue et al. (2020). By simply training a multilingual model in a language with abundant resources its acquired knowledge can be extended to target languages, in either Zero-Shot or Few-Shot scenarios. Cross-lingual transfer is a significant topic as it addresses the prevalent challenge of data scarcity in languages other than widely resourced ones, such as English Joshi et al. (2020). The ability to leverage the extensive linguistic resources available in high-resource languages for languages with limited training data enables the deployment of truly inclusive NLP systems.

Zero-Shot Cross-lingual Transfer (ZS-XLT) involves transferring a model trained in a source language to a target language without any demonstration of target-language examples Chen et al. (2021); Pires et al. (2019). This approach is highly modular, as it requires no adaptations specific to the target language. On the other hand, Few-Shot Cross-lingual Transfer (FS-XLT) enhances target-language accuracy by further fine-tuning a model using labeled target data Lauscher et al. (2020); Zhao et al. (2021); Schmidt et al. (2022). However, this method faces limitations when the available target-language data is limited, especially for higher-level tasks, leading to negligible enhancements or even detrimental effects on performance Lauscher et al. (2020); Zhao et al. (2021). Furthermore, this approach incurs higher computational costs due to the fine-tuning step and diminishes the modularity that characterizes the Zero-Shot method.

Our perspective is that adapting to a target language should prioritize resource efficiency and modularity, where we can seamlessly deploy a single model trained in English (or another source language) across different languages without any fine-tuning. In this work, we aim to improve this aspect for text classification –a high-level task– by leveraging the language-specific abilities of a multilingual model by prepending a One-Shot text-label target-language demonstration to the input text to predict the correct label. Specifically, we propose In-Context Cross-lingual transfer (IC-XLT), a simple yet effective method for One-Shot Cross-Lingual Transfer in Text Classification.

This novel approach employs In-Context Tuning (ICT) Chen et al. (2022) to train an encoder-decoder model in the source language tasking it to predict input texts with information derived from context demonstrations. ICT is a meta-learning strategy that optimizes a model’s ability to learn from in-context examples, originally designed for facilitating swift adaptation to new tasks by prepending target-task in-context demonstrations to the input during the adaptation process. To the best of our knowledge, this is the first study of ICT application in the context of cross-lingual transfer.

The proposed method is composed of a fine-tuning and an adaptation stage. First, we fine-tune on the source language through ICT, where the model is trained for the classification task and also to learn from context demonstrations. Then, we adapt to the target language at inference time by prepending a One-Shot111One-Shot per label demonstration in that language to the input. This method is modular and cost-effective at the adaptation stage as it does not require any gradient update.

We evaluate IC-XLT on two multilingual text classification datasets, spanning 5 and 13 target languages, with English as the source language. We consider two distinct settings. First, we assume access to the entire source-language training dataset. For the second setting, we deliberately constrain the amount of source training data available. This limitation aims to gauge the robustness of the proposed approach in scenarios where the availability of source data is restricted. We hypothesize that leveraging context information may prove particularly beneficial in tasks where source data is limited.

The contributions of this work are the following:

  1. 1.

    IC-XLT as an effective strategy for One-Shot Cross-lingual transfer: By measuring the reduction in the transfer gap of IC-XLT against standard approaches, and the performance improvement after introducing target-language examples, we present empirical evidence that training a model in a source language with In-Context Tuning allows it to leverage a One-Shot demostration through In-Context Learning to adapt to a target language. This results in a One-Shot XLT approach that adapts to a target language at inference without requiring gradient updates.

  2. 2.

    ICT improves mT5 fine-tuning when resources are limited. We observe that for the evaluated tasks, ICT training yields better performance compared to traditional fine-tuning when (source language) training data consists of only few-shots per label. In particular IC-XLT models trained on this scenario (1) benefit from this behavior at the adaptation and (2) leverage target-language in-context examples, achieving comparable performance to Prompt Tuning transfer methods with significantly less source-language data.

2 Related work

2.1 Zero and Few-Shot Cross-lingual Transfer

Multilingual transformers, such as mBERT Devlin et al. (2018), XLMR Conneau et al. (2019), and mT5 Xue et al. (2020), have showcased notable ability in Zero-Shot Cross-lingual Transfer (ZS-XLT) Pires et al. (2019). In this paradigm, these models are trained using abundant data in a source language and subsequently undergo evaluation in a target language without exposure to any training data in that specific language. However, this methodology is susceptible to significant performance variance Keung et al. (2020), and the transfer performance gap is contingent upon the linguistic proximity between the source and target languages Pires et al. (2019).

Furthermore, recent studies indicate that incorporating a small number of annotated examples in the target language can mitigate the performance gap between the source and target languages Lauscher et al. (2020); Zhao et al. (2021); Schmidt et al. (2022). This methodology, termed Few-Shot Cross-Lingual Transfer (FS-XLT), involves first fine-tuning a model on an extensive source dataset (as in ZS-XLT), and then subjecting it to a second fine-tuning on the reduced target-language data, facilitating its adaptation to this target language. This approach yields a noticeable improvement in performance at a relatively low labeling cost across various NLP tasks Lauscher et al. (2020).

However, empirical evidence Lauscher et al. (2020); Zhao et al. (2021); Schmidt et al. (2022) indicates that FS-XLT yields the most significant benefits for low-level tasks, such as Named Entity Recognition (NER) or Part-of-Speech (POS) tagging. When applied to high-level tasks, like Natural Language Inference (NLI) or text classification, the performance only shows improvement with a substantial number of examples from the target language. When adapted with small datasets in the target language (<100 samples), FS-XLT tends to offer minimal performance gains or may even lead to a decline in model efficacy.

Additionally, according to Schmidt et al. (2022), sequential FS-XLT can also exhibit unreliability in the Few-Shot scenario due to considerable variance in performance at different checkpoints during training. To address this issue, they propose jointly training the model using both source and target data in the adaptation stage of the process, which improves stability in the Few-Shot setting. This fine-tuned FS-XLT approach, however, has two notable drawbacks. Firstly, it lacks modularity, as the models are trained specifically for the selected target language during the adaptation stage. Secondly, there is a substantial increase in computational cost compared to Zero-Shot Cross-lingual Transfer due to the adaptation fine-tuning, whose cost scales with the size of the base model.

Moreover, existing methods predominantly address the XLT task under the assumption of abundant data in the source language. Although this is a fair assumption for many cases, as in general it is much more likely to find labeled datasets in high resource languages, there are scenarios where the source domain itself is limited. Instances of this include domain-specific tasks with a scarcity of annotated samples or tasks related to rapidly emerging trends and language patterns originated from social media, where due to its emerging nature, there is no labeled data available. In such cases, it might be more feasible to secure annotations for high-resource languages, which can then be transferred to other languages.

Given these considerations, we believe it is pertinent to investigate how the XLT performance scales as the quantity of available source data is systematically reduced. The intuition behind this is that the introduction of target-language shots may alleviate the performance decrease associated with a reducing source training data.

2.2 In-Context Learning and Language Models

LMs have demonstrated an aptitude for learning from a small number of demonstrations through a method known as In-Context Learning (ICL) Brown et al. (2020), where the model is tasked with predicting an input prepended with labeled examples. Particularly, Winata et al. (2021) observed that it is possible to achieve satisfactory performance in a cross-lingual setting when evaluating an mT5 model with a target-language input prefixed with labeled English demonstrations. This Zero-Shot approach, although efficient, can be sub-optimal as it does not take full advantage of resources in the source language due to the lack of fine-tuning.

Recent findings indicate that transformers Vaswani et al. (2017) can perform model selection on functions encountered during pre-training through in-context demonstrations. Yet, they still find it challenging in generalizing effectively to out-of-distribution classes, as highlighted by Yadlowsky et al. (2023). Given that most pre-trained LMs have not been explicitly trained for ICL, they might exhibit sub-optimal behavior when presented with Few-Shot demonstrations. In response to this challenge, Chen et al. (2022) introduced In-Context Tuning (ICT), a meta-learning approach designed to train a model to effectively learn from in-context demonstrations222Also, ICT consistently improves performance of ICL and is less sensitive to the shot selection when compared to raw, pre-trained LMs. Chen et al. (2022). ICT meta-trains a language model across a range of tasks, enhancing its ability to swiftly adapt to new tasks through ICL.

Still, In-Context Tuning has not yet been implemented for language transfer, as opposed to task transfer. We hypothesize that fine-tuning a multilingual model concurrently for learning from input context and the downstream task can leverage multilingual knowledge acquired during pretraining. This, we anticipate, will result in enhanced classification performance in a target language when provided with examples in that language. Therefore, in this study we showcase the efficacy of this idea for One-Shot Cross-lingual Transfer, particularly, for adapting to a target language through a One-Shot demonstration in-context. This adaptation method proves effective in improving the text classification performance by better leveraging target-language examples compared to the the fine-tuned FS-XLT. Moreover, we delve into the advantages of employing this approach in scenarios where source task data is not abundant.

3 Our proposed approach: In-Context Cross-Lingual Transfer

Our method aims to simultaneously train a pretrained multilingual encoder-decoder model for (1) a downstream text classification task, and (2) learning from context demonstrations. Then, we expect it to be able to generate predictions in a target language by including context demonstrations in this language. Therefore, we leverage ICT fine-tuning to transfer between languages at inference. As described above, our proposed procedure, called In-Context Cross-lingual Transfer (IC-XLT), is comprised of two stages:

In-Context Tuning

During the meta-training stage, we fine-tune the base multilingual model for a specific task using data from the source language. Let the set of pairs Dsrc={(x1src,y1src),,(x|D|src,y|D|src)}D^{src}=\{(x^{src}_{1},y^{src}_{1}),\ldots,(x^{src}_{|D|},y^{src}_{|D|})\} represent the source-language training dataset. The objective is to train the model to predict the label yisrcy_{i}^{src} for a given text xisrcx_{i}^{src} with the following input\Rightarrowoutput format:

Xsrc,xisrcyisrcX^{src},x_{i}^{src}\Rightarrow y_{i}^{src}

Here, Xsrc=((xj1,yj1),,(xjM,yjM))X^{src}=\left((x_{j_{1}},y_{j_{1}}),\ldots,(x_{j_{M}},y_{j_{M}})\right) is a random sequence of MM text-label pairs randomly sampled from DsrcD^{src} without replacement, which excludes the pair (xisrc,yisrc)(x_{i}^{src},y_{i}^{src}). In simpler terms, MM is the number of source-language demonstrations prepended to each input during fine-tuning, not the number of KtgtK_{tgt}-shot examples prepended during inference.

In-Context Learning

At inference, we adapt to a target language by prepending the samples from the target language training dataset D~tgt={(x~1tgt,y~1tgt),,(x~NKtgttgt,y~NKtgttgt)}\widetilde{D}^{tgt}=\{(\widetilde{x}^{tgt}_{1},\widetilde{y}^{tgt}_{1}),\ldots,(\widetilde{x}^{tgt}_{NK_{tgt}},\widetilde{y}^{tgt}_{NK_{tgt}})\} to each entry xitgtx_{i}^{tgt} of the test set to predict yitgty_{i}^{tgt}. Consequently, the input format mirrors the structure observed in the ICT stage:

X~tgt,xitgtyitgt\widetilde{X}^{tgt},x_{i}^{tgt}\Rightarrow y_{i}^{tgt}

Where NN is the number of classes and the sequence X~tgt\widetilde{X}^{tgt} is a concatenation of D~tgt\widetilde{D}^{tgt} entries comprising the KtgtK_{tgt}-shot samples per class, prepended to each xitgtx_{i}^{tgt} entry at the inference stage.

The intuitive idea for this approach is that, after the meta-training stage, we expect the model to understand both the classification task and the contextual relationships relevant to it. During the adaptation stage, the model leverages its multilingual pretraining to interpret context examples in the target language. Note that the adaptation to the target language in this context is gradient free, as it occurs during inference, and thus no model weights are updated.

4 Experimental Methodology

In this section, we outline the methodology employed to evaluate the proposed approach. We assess IC-XLT effectiveness in adapting to a target language for the classification task and compare its performance in cross-lingual transfer under (1) full training data on the source language and (2) various source-language data budgets. We conduct these limited data experiments to assess how much IC-XLT improves over a traditional fine-tuning method by leveraging the One-Shot demonstration.

4.1 Data and Evaluation Metrics

We conduct evaluations on two mutlilingual text classification datasets. The first dataset is Aspect Category Detection (ACD) on Restaurant Reviews Pontiki et al. (2016), a multi-label dataset comprising 12 classes representing different aspects mentioned in reviews. The second dataset is Domain Classification on assistant utterances from the MASSIVE dataset FitzGerald et al. (2022), a single-label classification dataset with 18 possible domain classes. The main difference between these datasets is that MASSIVE assigns only one label per entry, whereas ACD allows for multiple labels per entry, presenting a more challenging task. The datasets were chosen for their larger number of labels and their availability in multiple languages with shared labels (See Appendix A.3 for further details).

We select F1F_{1} micro as our evaluation metric, following Pontiki et al. (2016). For both datasets, our model is trained in English as the source language, and its performance is evaluated across 5 target languages: Dutch, Turkish, Russian, French, and Spanish for ACD, and 13 target languages for MASSIVE: French, Spanish, Turkish, Russian, Thai, Japanese, Indonesian, Icelandic, Amharic, Arabic, Azeri, Swahili and Urdu

To evaluate the performance of our proposed In-Context Cross-Lingual Transfer (IC-XLT) approach in a resource-constrained scenario with limited source-language data, we construct synthetically reduced datasets by sampling subsets of the training datasets following various KK-shot configurations, specifically Ksrc{8,16,32,64}K_{src}\in\{8,16,32,64\}. The objective of these evaluations is to assess IC-XLT’s ability to leverage target-language demonstrations for enhancing performance in situations where the source-language task has limited resources.

Regarding the shot selection, our KK-shot approach selects KK examples per class. Considering the class imbalance and the multi-label nature of the ACD dataset, the total number of examples will be in the range [K,K×N]\left[K,K\times N\right]. For a detailed explanation on this refer to Appendix A.1

4.2 Experimental Setting

As our multilingual base model, we utilize mT5-large Xue et al. (2020), an encoder-decoder model pre-trained on a diverse corpus encompassing over 100 languages. We employ LoRA Hu et al. (2021) for fine-tuning the model on the source-language data with full training data and varying numbers of shots KsrcK_{src}. During the inference stage, label predictions are generated through text generation, which facilitates multi-label inference. We adopt a greedy decoding strategy as implemented in Wolf et al. (2020).

In this work, we set Ktgt=1K_{tgt}=1, using only a One-Shot demonstration for the proposed IC-XLT approach333We provide IC-XLT implementation in the following repository: https://github.com/villacu/ic_xlt. We train the ICT models in the source language with different number of context examples, specifically M=10M=10 and M=20M=20. All models are trained on an NVIDIA Titan RTX GPU, the hyperparameter selection is discussed in Appendix A.2.1.

For the experiments with reduced source-language data, we conduct evaluations using two seeds for each of the following: the fine-tuning process, KsrcK_{src} shot selection, and KtgtK_{tgt} shot selection. Since Zero-Shot approaches do not require selecting target shots, we run a total of 4 and 8 runs for Zero-Shot and One-Shot respectively, using seeds within {1,2}\{1,2\}. For the models trained with full source-language data, we trained 5 models with seeds for the fine-tuning process within {1,,5}\{1,...,5\} and selected the best 3 in the English validation set. At adaptation, we employ two seeds for KtgtK_{tgt} shot selection in {1,2}\{1,2\}.

4.3 Baselines

We benchmark our proposed approach against the following baseline methods:

(1S) One-shot Prediction

Leveraging mT5’s pretraining objective, we task the model with predicting the missing span corresponding to the correct label given an input text prepended with a One-Shot demonstration. This is a traditional In-Context Learning approach where we expect the model to deduce label meanings from the examples without undergoing source-language fine-tuning, similar to the idea introduced in Winata et al. (2021), serving as the lower bound when Ksrc=0K_{src}=0.

(ZS-XLT) Zero Shot XLT

The standard Zero-Shot (Ktgt=0K_{tgt}=0) Cross-lingual Transfer approach, where the model is initially trained on a source language, and subsequent inference is conducted on the target language without any additional tuning. In this case, we train the mT5 model through prompt-based fine-tuning (PFT), with the input-output form:

xiyix_{i}\Rightarrow y_{i}

Hence, training is performed at the source and inference at target languages.

(1S-XLT and 8S-XLT) 1-Shot and 8-Shot XLT

Using the same training scheme (PFT), we continue fine-tuning on the checkpoints from ZS-XLT, training with Ktgt=1K_{tgt}=1 and Ktgt=8K_{tgt}=8 shots per label in the target language. This approach is the standard gradient-based approach for adapting to a target language in Few-Shot Cross-Lingual Transfer Lauscher et al. (2020). For this baseline, the target-language shots are not prepended as in IC-XLT and 1S, but used to further fine-tune the model following the same input-output as ZS-XLT.

(1S-XLTmacro{}^{\nabla}_{macro}) macro-averaging

In this baseline, we build upon the methodology established in 1S-XLT with Ktgt=1K_{tgt}=1 by adopting a strategy that incorporates fine-tuning on both source and target-language data for adaptation to the target language. This method follows the approach described by Schmidt et al. (2022), where the loss for each batch is calculated using a weighted combination of source and target-language losses:

L=βLsrc+(1β)LtgtL=\beta L_{src}+(1-\beta)L_{tgt}

Here, LsrcL_{src} and LtgtL_{tgt} represent the source and target-language losses, respectively. We select a value of β=0.5\beta=0.5 following the original implementation.

(IC-XLTSRC) IC-XLT with source-language context

We use the same models trained for IC-XLT, however, in this method In-Context examples are drawn from the training set of the source language (English). In essence, this can be considered a Zero-Shot baseline since no target language (or unseen source language) is involved for adaptation. Through this baseline we aim to evaluate the relevance of the target language One-Shot samples at the adaptation stage, assessing whether they are necessary for successful transfer to that target language.

5 Results and analysis

IC-XLT performance at Cross-lingual transfer
Method (KtgtK_{tgt}) ENG(SRC) Target avg AMH AZE ISL SWA URD ARA IND THA TUR FRA JAP RUS SPA
1S (1) 38.54 30.48 24.26 30.38 31.19 32.78 29.48 31.15 32.26 30.83 33.24 33.34 29.69 25.80 31.90
ZS-XLT (0) 90.06 76.12 62.53 71.63 73.77 65.75 67.72 70.98 84.23 79.88 77.7 84.87 83.39 83.98 83.09
1S-XLT (1) 90.08 76.21 62.78 71.63 73.9 65.85 68.34 71.11 84.17 79.95 77.68 84.86 83.45 83.89 83.15
8S-XLT (8) 89.87 76.78 65.38 72.13 74.73 66.78 69.24 71.45 84.08 80.46 78.02 85.16 83.25 84.17 83.29
1S-XLTmacro{}_{macro}^{\nabla} (1) 89.93 75.8 62.53 71.05 73.48 65.42 67.9 70.49 83.8 79.33 77.17 84.71 82.87 83.78 82.86
IC-XLTM=10SRC{}_{SRC}^{M=10} (0) 89.41 74.39 63.39 70.82 69.22 56.06 67.47 71.62 81.04 79.02 76.65 82.81 82.99 84.41 81.53
IC-XLTM=20SRC{}_{SRC}^{M=20} (0) 89.46 74.02 61.27 71.49 68.83 53.76 68.26 70.77 81.6 78.91 76.36 83.13 82.83 83.76 81.23
IC-XLTM=10 (1) 89.41 81.24 70.68 81.73 81.07 78.23 76.34 77.28 85.6 80.98 83.63 85.53 84.68 86.18 84.18
IC-XLTM=20 (1) 89.46 80.26 68.09 81.37 80.32 75.54 76.12 76.43 85.46 81.02 82.25 84.92 83.77 84.39 83.65
Table 1: Average F1F_{1} micro in the MASSIVE Domain Detection task, trained with full data in English. The number in parenthesis is the amount of samples per label in the target language used for the adaptation process (KtgtK_{tgt}). Standard deviations over different seeds per each language are shown in Appendix A.5.
Method (KtgtK_{tgt}) ENG(SRC) Target avg FRA NLD RUS SPA TUR
1S (1) 37.09 28.42 31.78 20.33 34.38 34.86 20.76
ZS-XLT (0) 79.14 70.15 67.96 69.24 73.6 70.01 69.95
1S-XLT (1) 79.16 70.39 67.98 69.56 73.44 70.23 70.72
8S-XLT (8) 78.78 70.26 67.87 68.81 72.32 69.84 72.47
1S-XLTmacro{}_{macro}^{\nabla} (1) 79.16 70.41 68.03 69.59 73.3 70.25 70.88
IC-XLTM=10SRC{}_{SRC}^{M=10} (0) 81.48 73.83 74.06 72.54 77.59 73.7 71.27
IC-XLTM=20SRC{}_{SRC}^{M=20} (0) 81.76 73.11 73.49 71.89 77.52 73.14 69.51
IC-XLTM=10 (1) 81.48 75.25 74.07 73.34 78.07 75.20 75.59
IC-XLTM=20 (1) 81.76 75.05 73.5 73.00 78.01 74.46 76.26
Table 2: Average F1F_{1} micro in the Aspect Category Detection dataset, trained with full data in English, the source language. Standard deviations over different seeds per each language are shown in Appendix A.5.

For the first experiment, we compare our proposed approach, IC-XLT, to the baselines detailed in Section 4.3 using the full training set in the source language. For Aspect Category Detection, we observed a general trend where mT5 models trained with In-Context Tuning, which employs the input-output setting X~,xiyi\widetilde{X},x_{i}\Rightarrow y_{i}, consistently outperformed models subjected to prompt-based fine-tuning with xiyix_{i}\Rightarrow y_{i} under the same training regimes, despite both models being trained for an equivalent number of steps and exact same data instances. We hypothesize that this superior performance may be attributed to the fact that the ICT-trained models see MM randomly ordered input-output examples at each instance, even though they are tasked with predicting only xix_{i}.

We present the F1F_{1} micro scores across five different languages on ACD and 13 languages on MASSIVE in Tables 1 and 2 respectively. The numbers in these tables represent average metrics calculated across various seeds, as detailed in Section 4.2. The standard deviation for each method and language is provided in the complete tables located in Appendix A.5. We observe that our proposed approach, In-Context Cross-Lingual Transfer (IC-XLT), effectively outperforms the baselines by a substantial margin in the evaluated datasets, greatly improving mT5 cross-lingual transfer performance. A crucial observation is that for both of the evaluated datasets there is a noticeable increase in performance from IC-XLTSRC to IC-XLT. This means that the proposed approach adapts successfully, during inference time, to the target language by taking advantage of the One-Shot in context demonstration.

On the other hand, the 1S-XLT approach did not improve over ZS-XLT by a considerable margin, which is consistent with previous results for high-level tasks Lauscher et al. (2020); Zhao et al. (2021). Even when increasing target-language samples to Ktgt=8K_{tgt}=8 (baseline 8S-XLT) on this gradient-based baseline, we observe a performance decline in the ACD dataset for most languages except Turkish. The MASSIVE dataset shows improvements across nearly all target languages with the increased number of target shots on 8S-XLT, yet, these gains are modest compared to those achieved with IC-XLT using only Ktgt=1K_{tgt}=1, highlighting its superior efficiency in utilizing minimal target-language data for improved cross-lingual transfer.

The gradient-based adaptation that mixes target and source languages (1S-XLTmacro{}^{\nabla}_{macro}) marginally outperforms 1S-XLT in the ACD dataset. However, it does not improve, and even drops, performance on the MASSIVE dataset. This approach also requires more computational resources, due to the larger data volume when including source-language examples.

The observed variability in the effectiveness of gradient-based methods across datasets and languages may be related to differences between the datasets; ACD represents a more complex, non-parallel task with classes that are easier to confuse, unlike the parallel and simpler nature of MASSIVE. Nonetheless, IC-XLT consistently outperforms the baselines across languages and datasets, demonstrating its robustness and cost-effectiveness in terms of computational resources. We find that M=10M=10 (the number of in-context demonstrations during ICT training) performs slightly better than M=20M=20 in the experiments that employ the full training set.

Performance with limited source-language data

We conduct experiments to quantify the ability of IC-XLT to perform at scenarios with limited source-language resources. For this we evaluate ZS-XLT, 1S-XLT, IC-XLTSRC, and IC-XLT models trained with Ksrc{8,16,32,64}K_{src}\in\{8,16,32,64\}. We noticed that models trained with the ICT framework generally perform better compared to PFT for low values of KsrcK_{src}. In Figures 1(b) and 1(a), we illustrate the average performance on the target languages at different source-language resource availability regimes. In both Figures, we can observe that IC-XLT makes better use of resources than prompt-based fine-tuning specially at smaller values for KsrcK_{src}. Furthermore, the performance difference with the source language (English) is visibly smaller for IC-XLT, more discussion on this can be found below.

The F1F_{1}-micro averages for the target languages are shown in Tables 3 and 4 for ACD and MASSIVE, respectively. For all the language-specific performance metrics at different KsrcK_{src} budgets refer to Appendix A.5. Results on these tables show that IC-XLT trained on limited data (Ksrc=64K_{src}=64) achieves competitive or superior performance compared to baseline methods (ZS-XLT and 1S-XLT) trained with full source datasets.

Given that the target language adaptation occurs at inference time, the improvement over the Zero-Shot approach comes at no extra computational cost and at a minimal data cost. This allows to achieve good performance with limited computational and data resources.

For the experiments with limited source-language data, M=20M=20 achieves a better performance in MASSIVE. We believe that since ACD contains only 12 labels, in this scenario a context length of 20 will inevitably prepend more repeated context examples than the MASSIVE dataset444Which contains 18. when training with limited data. This reduced variability may hurt the model’s performance compared to M=10M=10.

Refer to caption
((a)) MASSIVE performance with different source data availability. IC-XLT trained with M=10M=10.
Refer to caption
((b)) ACD performance with different source data availability. IC-XLT trained with M=10M=10.
Figure 1: Comparison of IC-XLT and 1S-XLT performance at different source-language data budgets. Pink lines employ target-language examples for adaptation while blue lines do not. We can observe that, in general, the IC-XLT models yield better performance compared to ZS-XLT and 1S-XLT. This is especially notable at lower resource scenarios.
Measuring the transfer gap with the source language.

By measuring the performance gap between the source language and the target language, we aim to quantify the contribution of the ICT training framework and target-language demonstrations for mitigating this gap. As we provide the model with target-language examples, we anticipate a smaller decrease in performance from the source language when evaluating on a new language, compared to Zero-Shot approaches. We can measure this by computing the average transfer gap Δ¯%\bar{\Delta}\%, which is the average percentage decrease in performance relative to the evaluations on the test set in the source language (English). This is defined as:

Δ¯%=100×𝐄[PTL/PSL1]\bar{\Delta}\%=100\times\mathbf{E}\left[P_{TL}/P_{SL}-1\right]

Where PTLP_{TL} and PSLP_{SL} represent the evaluation performance of the exact same method on the target and source-language test sets, respectively. The performance gap values are shown in Figure 2. We can observe that for most source-language data budgets, we obtain a reduced average transfer gap Δ¯%\bar{\Delta}\% through IC-XLT compared to ZS-XLT, 1S-XLT and IC-XLTSRC.

KsrcK_{src} 1S
0 28.42
ZS-XLT 1S-XLT IC-XLTM=10 IC-XLTM=20
8 25.40 27.12 33.34 16.64
16 30.84 32.63 48.66 47.04
32 43.56 43.36 58.91 61.00
64 55.15 53.85 65.64 65.28
Full 70.15 70.39 75.25 75.05
Table 3: Average F1F_{1} micro across 5 target languages for Aspect Category Detection.
KsrcK_{src} 1S
0 30.48
ZS-XLT 1S-XLT IC-XLTM=10 IC-XLTM=20
8 44.05 45.86 62.46 66.29
16 51.13 51.97 70.30 73.14
32 64.74 65.28 73.27 76.72
64 63.45 63.68 77.55 79.09
Full 76.12 76.21 81.24 80.26
Table 4: Average F1F_{1} micro across 13 target languages for MASSIVE (Domain Classification).
Refer to caption
Refer to caption
Figure 2: The average transfer gap Δ¯%\bar{\Delta}\% of IC-XLT, IC-XLTSRC, 1S-XLT and ZS-XLT at different source-language data budgets. (IC-XLT M=10M=10). We can observe that, for most cases, IC-XLT yields the smallest drop in performance after transfering to a target language compared to the baselines.
Improvement due to target-language demonstrations.
<5 B 5-100 B > 100B
Method (KtgtK_{tgt}) Target avg AMH AZE ISL SWA URD ARA IND TUR THA FRA JAP RUS SPA
δ%\delta\% 8S-XLT (8) 0.87 4.56 0.7 1.3 1.57 2.24 0.66 -0.18 0.41 0.73 0.34 -0.17 0.23 0.24
IC-XLT (1) 9.81 11.5 15.41 17.12 39.55 13.15 7.9 5.63 9.11 2.48 3.28 2.04 2.1 3.25
Table 5: δ%\delta\% in MASSIVE. The first row refers to the number of tokens of the target languages in mT5 pretraining corpora. We use IC-XLT with M=10M=10.

Aiming to assess the proposed method’s capacity to utilize demonstrations in the target language to enhance performance in that language, we compute the average percentage improvement in performance δ%\delta\% after introducing target-language shots. The formula for computing this value is described in Appendix A.4.

For the MASSIVE dataset, since 1S-XLT offers minimal improvements with Ktgt=1K_{tgt}=1, we conduct a comparison between IC-XLT with Ktgt=1K_{tgt}=1 and 8S-XLT with Ktgt=8K_{tgt}=8.

Values for δ%\delta\% are detailed in Tables 6 and 5 for the Aspect Category Detection (ACD) and MASSIVE datasets, respectively. We observe that IC-XLT consistently achieves the highest (δ%\delta\%) compared to the fine-tuned approaches. This result underscores IC-XLT’s effectiveness for Few-Shot Cross-lingual text classification, a high-level task where methods that involve fine-tuning for language adaptation underperform when the amount of target-language data is limited.

Method (KtgtK_{tgt}) Target avg FRA NLD RUS SPA TUR
δ%\delta\% 1S-XLT (1) 0.34 0.03 0.46 -0.22 0.31 1.10
IC-XLT (1) 1.92 0.01 1.1 0.62 2.04 6.06
Table 6: δ%\delta\% in the Aspect Category Detection dataset. In here we use IC-XLT with M=10M=10.
Correlation of methods, pretraining data and linguistic distance.
mT5 pretraining Ling. Dist. vs ENG
IC-XLTM=10SRC{}_{SRC}^{M=10} -0.786∗∗ 0.232
8S-XLT -0.874∗∗ 0.331
Table 7: Improvement per language correlation (δ%\delta\%) with Linguistic Distance and language representation in mT5 pretraining corpora (MASSIVE dataset).

By analyzing results is the MASSIVE dataset across the 13 target languages, we observe that by introducing target-language demonstrations some languages, such as French, Japanese, or Russian, show modest enhancements of around 2%, while others, like Azeri, Icelandic, and Swahili, benefit from increases exceeding 15%. To understand the underlying factors contributing to these differences, we explore the relationship between the improvement observed (δ%\delta\%) and two variables: (1) the number of tokens representing each language in mT5 pretraining corpora Xue et al. (2020), and (2) the linguistic distance between each target language and English, the source language, according to the URIEL database Malaviya et al. (2017). The analysis is limited to the MASSIVE dataset because the ACD dataset includes only 5 target languages, mostly European and all well-represented in mT5 pretraining data, with no languages having low representation. This limits the scope for obtaining reliable correlation measures.

To quantitatively assess these relationships, we measure the Spearman Correlation to identify how the token counts of pretraining data and the linguistic proximity to English correlate with the effectiveness of target-language demonstrations in improving cross-lingual transfer performance. From the correlations presented in Table 7 we can observe that both IC-XLT and 8S-XLT exhibit a statistically significant negative Spearman correlation between the improvement (δ%\delta\%) and the representation of target languages in the mT5 pretraining corpora. This pattern indicates that languages with less representation in the pretraining data will experience more substantial improvements through target-language adaptation. Correlation with linguistic distance, on the other hand, is weaker and not statistically significant.

6 Conclusion

In this paper, we investigated the application of In-Context Tuning for One-Shot Cross-lingual transfer, introducing In-Context Cross-lingual Transfer (IC-XLT). Our evaluations conducted on a multilingual encoder-decoder model (mT5) demonstrate the efficacy of the proposed method in effectively adapting at inference time to target languages using only a One-Shot demonstration in-context, all without incurring additional computational expenses (gradient free). Furthermore, in comparison to ZS-XLT and 1S-XLT, IC-XLT demonstrated superior performance and smaller transfer gap for the task of text classification, a high-level task where FS-XLT tends to underperform.

In scenarios with limited source-language training data, we provide empirical evidence that IC-XLT learns better the source language at the meta-training stage and demonstrates a smaller transfer gap at the adaptation stage with the One-Shot demonstration, compared to ZS-XLT and 1S-XLT. This makes IC-XLT a valuable tool for cross-lingual transfer in resource-limited scenarios. Our findings also show a significant correlation between the performance improvements in target languages and their token count in the mT5 pretraining corpus, indicating that languages with lesser representation tend to benefit more from target-language adaptation through IC-XLT.

To our knowledge, this study represents the first exploration of In-Context Tuning for Cross-Lingual Transfer. For future work, we aim to explore the potential and limitations of this approach by evaluating its applicability to other architectures, such as decoder-only or encoder-only models, and examining the impact of training with a greater number of examples in-context.

7 Limitations

In this study, we implement our approach using an mT5-large encoder-decoder model. However, an evaluation of its applicability to encoder-only or decoder-only models remains unexplored and it is left for future work. Furthermore, due to storage and compute constraints and the need to conduct experiments across diverse seeds and training data budgets, we opted to fine-tune the models using LoRA Hu et al. (2021). While some variability compared to the fully trained model is expected with this architectural choice, empirical evidence from Hu et al. (2021) suggests that its impact is minimal. In this work, we do not compare with methods that translate the source-language training set into target languages. Such approaches require a separate machine translation system and thus are more expensive, falling beyond the scope of our research. Our focus remains on utilizing a single model in an end-to-end manner. Finally, it is important to outline that due to the maximum input length of mT5 (1024 tokens), scaling IC-XLT is to a larger number of target-language shots (e.g Ktgt{4,8,16}K_{tgt}\in\{4,8,16\}) may prove difficult using the current approach. This challenge is particularly pronounced in scenarios with a substantial number of labels, where input text may need to be truncated. Consequently, there is a need to devise a strategy to either reduce input length or integrate information from different example batches in order to address this limitation.

Acknowledgements

We thank the anonymous reviewers and the meta reviewer at ARR for their valuable feedback. Also, we thank CONAHCYT for the computer resources provided through the INAOE Supercomputing Laboratory’s Deep Learning Platform for Language Technologies and CIMAT Bajio Super-computing Laboratory (#300832). Sanchez-Vega acknowledges CONAHCYT for its support through the program “Investigadoras e Investigadores por México” (Project ID.11989, No.1311). Villa-Cueva (CVU 1019520) thanks CONAHCYT for the support through the master’s degree scholarship at CIMAT.

References

  • Brown et al. (2020) Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. 2020. Language models are few-shot learners. CoRR, abs/2005.14165.
  • Chen et al. (2021) Guanhua Chen, Shuming Ma, Yun Chen, Li Dong, Dongdong Zhang, Jia Pan, Wenping Wang, and Furu Wei. 2021. Zero-shot cross-lingual transfer of neural machine translation with multilingual pretrained encoders. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 15–26, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
  • Chen et al. (2022) Yanda Chen, Ruiqi Zhong, Sheng Zha, George Karypis, and He He. 2022. Meta-learning via language model in-context tuning. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 719–730, Dublin, Ireland. Association for Computational Linguistics.
  • Conneau et al. (2019) Alexis Conneau, Kartikay Khandelwal, Naman Goyal, Vishrav Chaudhary, Guillaume Wenzek, Francisco Guzmán, Edouard Grave, Myle Ott, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Unsupervised cross-lingual representation learning at scale. CoRR, abs/1911.02116.
  • Devlin et al. (2018) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. BERT: pre-training of deep bidirectional transformers for language understanding. CoRR, abs/1810.04805.
  • FitzGerald et al. (2022) Jack FitzGerald, Christopher Hench, Charith Peris, Scott Mackie, Kay Rottmann, Ana Sanchez, Aaron Nash, Liam Urbach, Vishesh Kakarala, Richa Singh, Swetha Ranganath, Laurie Crist, Misha Britan, Wouter Leeuwis, Gokhan Tur, and Prem Natarajan. 2022. Massive: A 1m-example multilingual natural language understanding dataset with 51 typologically-diverse languages.
  • Hu et al. (2021) Edward J. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, and Weizhu Chen. 2021. Lora: Low-rank adaptation of large language models. CoRR, abs/2106.09685.
  • Joshi et al. (2020) Pratik Joshi, Sebastin Santy, Amar Budhiraja, Kalika Bali, and Monojit Choudhury. 2020. The state and fate of linguistic diversity and inclusion in the NLP world. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 6282–6293, Online. Association for Computational Linguistics.
  • Keung et al. (2020) Phillip Keung, Yichao Lu, Julian Salazar, and Vikas Bhardwaj. 2020. Don’t use English dev: On the zero-shot cross-lingual evaluation of contextual embeddings. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 549–554, Online. Association for Computational Linguistics.
  • Lauscher et al. (2020) Anne Lauscher, Vinit Ravishankar, Ivan Vulić, and Goran Glavaš. 2020. From zero to hero: On the limitations of zero-shot language transfer with multilingual Transformers. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 4483–4499, Online. Association for Computational Linguistics.
  • Loshchilov and Hutter (2017) Ilya Loshchilov and Frank Hutter. 2017. Fixing weight decay regularization in adam. CoRR, abs/1711.05101.
  • Malaviya et al. (2017) Chaitanya Malaviya, Graham Neubig, and Patrick Littell. 2017. Learning language representations for typology prediction. In Conference on Empirical Methods in Natural Language Processing (EMNLP), Copenhagen, Denmark.
  • Pires et al. (2019) Telmo Pires, Eva Schlinger, and Dan Garrette. 2019. How multilingual is multilingual BERT? In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 4996–5001, Florence, Italy. Association for Computational Linguistics.
  • Pontiki et al. (2016) Maria Pontiki, Dimitris Galanis, Haris Papageorgiou, Ion Androutsopoulos, Suresh Manandhar, Mohammad AL-Smadi, Mahmoud Al-Ayyoub, Yanyan Zhao, Bing Qin, Orphée De Clercq, Véronique Hoste, Marianna Apidianaki, Xavier Tannier, Natalia Loukachevitch, Evgeniy Kotelnikov, Nuria Bel, Salud María Jiménez-Zafra, and Gülşen Eryiğit. 2016. SemEval-2016 task 5: Aspect based sentiment analysis. In Proceedings of the 10th International Workshop on Semantic Evaluation (SemEval-2016), pages 19–30, San Diego, California. Association for Computational Linguistics.
  • Schmidt et al. (2022) Fabian David Schmidt, Ivan Vulić, and Goran Glavaš. 2022. Don’t stop fine-tuning: On training regimes for few-shot cross-lingual transfer with multilingual language models. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 10725–10742, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
  • Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Ł ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc.
  • Winata et al. (2021) Genta Indra Winata, Andrea Madotto, Zhaojiang Lin, Rosanne Liu, Jason Yosinski, and Pascale Fung. 2021. Language models are few-shot multilingual learners. In Proceedings of the 1st Workshop on Multilingual Representation Learning, pages 1–15, Punta Cana, Dominican Republic. Association for Computational Linguistics.
  • Wolf et al. (2020) Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, Rémi Louf, Morgan Funtowicz, Joe Davison, Sam Shleifer, Patrick von Platen, Clara Ma, Yacine Jernite, Julien Plu, Canwen Xu, Teven Le Scao, Sylvain Gugger, Mariama Drame, Quentin Lhoest, and Alexander M. Rush. 2020. Huggingface’s transformers: State-of-the-art natural language processing.
  • Xue et al. (2020) Linting Xue, Noah Constant, Adam Roberts, Mihir Kale, Rami Al-Rfou, Aditya Siddhant, Aditya Barua, and Colin Raffel. 2020. mt5: A massively multilingual pre-trained text-to-text transformer. CoRR, abs/2010.11934.
  • Yadlowsky et al. (2023) Steve Yadlowsky, Lyric Doshi, and Nilesh Tripuraneni. 2023. Pretraining data mixtures enable narrow model selection capabilities in transformer models.
  • Zhao et al. (2021) Mengjie Zhao, Yi Zhu, Ehsan Shareghi, Ivan Vulić, Roi Reichart, Anna Korhonen, and Hinrich Schütze. 2021. A closer look at few-shot crosslingual transfer: The choice of shots matters. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 5751–5767, Online. Association for Computational Linguistics.

Appendix A Appendix

A.1 Shot selection

Similar to Zhao et al. (2021), with "KK-shot" we refer to selecting KK examples for each of the NN classes. The examples are randomly sampled from the training splits of the datasets. Note that the number of shots per label may not precisely be KK due to underrepresented classes in the training set. This holds true for the ACD dataset, where certain classes may have insufficient samples to meet the per-class KK value. In such cases, the total number of shots per i-th class is determined as min(K,Ni)\min{(K,N_{i})}, where NiN_{i} represents the total number of samples for the i-th class in the dataset.

Furthermore, since the ACD task involves a multi-label dataset, multi-label examples may add to more than one of the NN buckets simultaneously. Hence, the total number of examples in a KK-shot dataset is in the range [K,K×N]\left[K,K\times N\right].

A.2 Hyperparameter selection

Here, we outline the hyperparameters utilized for fine-tuning models across the two stages of our pipeline. Initially, we detail the hyperparameters specific to fine-tuning models in the source language for both PFT and ICT methods. Subsequently, we address those employed for the gradient-based XLT methods. The LoRA Hu et al. (2021) parameters are r=16,α=32,r=16,\alpha=32, with dropout of 10%. For all cases, we employ an AdamW optimizer Loshchilov and Hutter (2017) with a linear scheduler and a batch size of 8.

A.2.1 Fine-tuning on source data

For the fine-tuning process on both datasets, we explored learning rates within the range of lr{3,4,5,6,7,8,10}×104lr\in\{3,4,5,6,7,8,10\}\times 10^{-4}, selecting 4×1044\times 10^{-4} for as it performed adequately on both datasets in the source language.

Regarding the number of epochs for training on the full datasets: for the MASSIVE dataset, we fine-tuned models for 10 epochs under both prompt-based Fine-Tuning (PFT) and In-Context Tuning (ICT) training schemes. For the Aspect Category Detection (ACD) dataset, which is considerably smaller, we extended the training duration to 15 epochs for ICT and 25 epochs for PFT. The decision to train the PFT models for more epochs in the ACD dataset was taken because it underperformed when only 15 epochs were used.

For models trained with a reduced quantity of source-language data, we standardized the training process by setting the learning rate to 5×1045\times 10^{-4} for all models across both datasets. Given the similar volume of data in these constrained scenarios, we extended the training period to 35 epochs for both datasets to expose the models to sufficient training data.

A.2.2 Fine-tuned XLT baselines hyperparameter selection.

For the fine-tuned XLT baselines, we use the model checkpoint fine tuned in the source language, used for ZS-XLT. Training focuses only on the target language, incorporating only the specified number of target-language shots KtgtK_{tgt}, and for 1S-XLTmacro{}_{macro}^{\nabla}, an equal amount of source-language shots is added.

We evaluated the following learning rates lr{0.5,1,5}×105lr\in\{0.5,1,5\}\times 10^{-5}. We observed that the limited target-language examples often led to overfitting and reduced performance (notably when Ktgt=1K_{tgt}=1) with a large number of epochs, especially in some languages such as Russian. For the reported results, the selected learning rates and training durations are as follows:

  • For both 1S-XLT and 1S-XLTmacro{}_{macro}^{\nabla}, a learning rate of 5×1055\times 10^{-5} is used, training for 1 epoch for models trained with the full source-language dataset and 5 epochs for those trained with limited source-language data.

  • For the 8S-XLT baseline, given the increased number of examples, we opt for a learning rate of 1×1051\times 10^{-5} across 10 epochs.

All adaptations used a batch size of 8 and a constant scheduler.

A.3 Dataset description

In this Section we provide descriptive information of the employed datasets. MASSIVE features parallel language splits, each comprising 11.5k samples in the training partition and 2.97k in the test partition.

However, for the Aspect Category Detection dataset, which is non-parallel, the sample counts vary across languages. Detailed information on these counts is presented in Table 8.

Train Test
English 2000 676
Spanish 2070 881
French 1664 668
Turkish 1232 144
Russian 3655 1209
Dutch 1722 575
Table 8: Length of the training and test partitions in the Aspect Category Detection Dataset.

A.4 Computing the improvement due to target-language demonstrations

By computing δ%\delta\%, we aim to measure the average improvement on performance of the evaluated methods after introducing target-language demonstrations. For this we compute the ratio between the model with target-language demonstrations and the Zero-Shot approach. This value is computed using the following formula:

δ%=100×[PTLfewshot/PTLzerozhot1]\delta\%=100\times\left[P^{few-shot}_{TL}/P^{zero-zhot}_{TL}-1\right]

Where PTLfewshotP^{few-shot}_{TL} and PTLzeroshotP^{zero-shot}_{TL} represent the average evaluation performance of the model under the same training scheme (ICT or PFT), with and without the target-language (TLTL) examples, respectively. Hence, for prompt-based fine-tuning PTLfewshotP^{few-shot}_{TL} is 1S/8S-XLT and PTLzeroshotP^{zero-shot}_{TL} is ZS-XLT, while for In-Context Tuning PTLfewshotP^{few-shot}_{TL} is IC-XLT and PTLzeroshotP^{zero-shot}_{TL} is IC-XLTSRC.

A.5 Performance metrics per language of the evaluated method on different source language budgets.

This section presents the comprehensive results of our evaluations across different target languages and different source data availability settings of ZS-XLT, 1S-XLT, and IC-XLT, with English serving as the source language. Detailed performance metrics for cross-lingual transfer on the Aspect Category Detection (ACD) dataset are depicted in Table 9 and results for the MASSIVE dataset are provided in Tables 10 and 11.

Additionally, we illustrate the language-wise performance at different KsrcK_{src} values on Figures 3 and 4. Furthermore, for the MASSIVE dataset we illustrate the different behavior of groups of languages with different representation in the mT5 pretraining corpus. Figures 5(a) and 5(b) illustrate the performance metrics for languages with high representation (over 100 billion tokens) and low representation (under 100 billion tokens) in the mT5 pretraining data, respectively. We also include the Average Transfer Gaps Δ¯%\bar{\Delta}\% per language at the different source-language budgets in Figures 6 and 7.

Refer to caption
Figure 3: F1-micro for each of the 5 evaluated languages in the Aspect Category Detection dataset at different source-language data budget.
Refer to caption
Figure 4: F1-micro for each of the 13 evaluated languages in the MASSIVE dataset at different source-language data budget.
Refer to caption
((a)) Mean F1-micro in the the MASSIVE dataset across the target languages with large representation in mT5 pretraining corpora (>100B>100B). In this scenario, the gap between One-Shot (pink) and Zero-Shot (blue) lines is similar as the one observed for ACD in the Figure 1(b).
Refer to caption
((b)) Mean F1-micro in the the MASSIVE dataset across the target languages with smaller representation in mT5 pretraining corpora (<100B<100B). We can observe a larger improvement when introducing target-language demonstrations than the one observed for languages well-represented in pretraining (left).
Figure 5: Comparison of IC-XLT and 1S-XLT performance on the MASSIVE datasets at different source-language data budgets on languages with high representation (left), and with low representation (right) in the base model pretraining.
Refer to caption
Figure 6: Average Transfer Gap Δ¯%\bar{\Delta}\% per language on the Aspect Category Detection dataset.
Refer to caption
Figure 7: Average Transfer Gap Δ¯%\bar{\Delta}\% per language on the MASSIVE dataset.
KsrcK_{src} ENG FRA NLD RUS SPA TUR
ZS-XLT (Ktgt=0K_{tgt}=0)
8-shot 29.49±2.0629.49\pm_{2.06} 26.1±1.4626.1\pm_{1.46} 21.19±3.7421.19\pm_{3.74} 29.32±2.8429.32\pm_{2.84} 31.99±2.4631.99\pm_{2.46} 18.4±4.3418.4\pm_{4.34}
16-shot 33.73±4.1233.73\pm_{4.12} 32.24±3.6332.24\pm_{3.63} 28.73±6.7628.73\pm_{6.76} 33.02±3.2233.02\pm_{3.22} 34.96±4.4834.96\pm_{4.48} 25.27±7.6825.27\pm_{7.68}
32-shot 49.05±5.3349.05\pm_{5.33} 45.41±4.6845.41\pm_{4.68} 40.44±4.6840.44\pm_{4.68} 45.31±5.945.31\pm_{5.9} 43.37±4.7343.37\pm_{4.73} 43.28±3.9343.28\pm_{3.93}
64-shot 60.45±3.360.45\pm_{3.3} 55.07±2.4655.07\pm_{2.46} 53.49±1.8853.49\pm_{1.88} 56.35±1.2356.35\pm_{1.23} 53.9±3.453.9\pm_{3.4} 56.96±3.0256.96\pm_{3.02}
Full 79.14±0.9179.14\pm_{0.91} 67.96±0.7767.96\pm_{0.77} 69.24±1.2469.24\pm_{1.24} 73.6±1.6173.6\pm_{1.61} 70.01±0.4470.01\pm_{0.44} 69.95±1.1369.95\pm_{1.13}
1S-XLT (Ktgt=1K_{tgt}=1)
8-shot 28.97±1.6628.97\pm_{1.66} 26.65±1.0926.65\pm_{1.09} 25.87±1.5625.87\pm_{1.56} 29.21±0.5629.21\pm_{0.56} 33.23±1.3333.23\pm_{1.33} 20.66±1.4220.66\pm_{1.42}
16-shot 37.62±4.9837.62\pm_{4.98} 33.37±1.9833.37\pm_{1.98} 32.44±4.3232.44\pm_{4.32} 33.86±2.4833.86\pm_{2.48} 33.22±4.2333.22\pm_{4.23} 30.28±2.6630.28\pm_{2.66}
32-shot 51.94±4.0651.94\pm_{4.06} 45.07±4.6945.07\pm_{4.69} 41.88±2.2641.88\pm_{2.26} 45.18±5.8845.18\pm_{5.88} 37.69±4.0437.69\pm_{4.04} 46.97±3.146.97\pm_{3.1}
64-shot 62.32±2.5362.32\pm_{2.53} 54.5±3.3754.5\pm_{3.37} 53.62±2.3353.62\pm_{2.33} 55.34±1.4855.34\pm_{1.48} 48.09±3.1748.09\pm_{3.17} 57.68±2.6557.68\pm_{2.65}
Full 79.16±0.7879.16\pm_{0.78} 67.98±0.7467.98\pm_{0.74} 69.56±0.6369.56\pm_{0.63} 73.44±1.5573.44\pm_{1.55} 70.23±0.4770.23\pm_{0.47} 70.72±1.4370.72\pm_{1.43}
IC-XLTM=20SRC{}_{SRC}^{M=20} (Ktgt=0K_{tgt}=0)
8-shot 23.66±6.1623.66\pm_{6.16} 17.3±1.7617.3\pm_{1.76} 19.9±2.7819.9\pm_{2.78} 19.05±4.8919.05\pm_{4.89} 22.01±4.3122.01\pm_{4.31} 16.09±5.2416.09\pm_{5.24}
16-shot 41.19±9.2241.19\pm_{9.22} 44.39±5.3944.39\pm_{5.39} 38.9±7.1138.9\pm_{7.11} 43.25±6.6643.25\pm_{6.66} 40.82±8.5540.82\pm_{8.55} 36.22±13.036.22\pm_{13.0}
32-shot 63.25±2.3663.25\pm_{2.36} 60.92±1.7160.92\pm_{1.71} 55.11±2.2855.11\pm_{2.28} 59.7±2.0259.7\pm_{2.02} 59.69±1.3259.69\pm_{1.32} 56.42±2.9156.42\pm_{2.91}
64-shot 70.01±1.7770.01\pm_{1.77} 65.67±1.365.67\pm_{1.3} 60.08±2.2760.08\pm_{2.27} 64.09±2.1964.09\pm_{2.19} 63.55±0.3463.55\pm_{0.34} 62.23±4.6962.23\pm_{4.69}
Full 81.76±0.8181.76\pm_{0.81} 73.49±0.3773.49\pm_{0.37} 71.89±0.3871.89\pm_{0.38} 77.52±0.3877.52\pm_{0.38} 73.14±0.8273.14\pm_{0.82} 69.51±1.0469.51\pm_{1.04}
IC-XLTM=20 (Ktgt=1K_{tgt}=1)
8-shot 23.66±6.1623.66\pm_{6.16} 17.12±13.0517.12\pm_{13.05} 15.17±6.9615.17\pm_{6.96} 16.58±11.3316.58\pm_{11.33} 13.49±8.8713.49\pm_{8.87} 20.83±13.220.83\pm_{13.2}
16-shot 41.19±9.2241.19\pm_{9.22} 48.24±4.7548.24\pm_{4.75} 44.89±6.1344.89\pm_{6.13} 49.64±4.9549.64\pm_{4.95} 47.26±8.5947.26\pm_{8.59} 45.17±8.645.17\pm_{8.6}
32-shot 63.25±2.3663.25\pm_{2.36} 61.37±2.4961.37\pm_{2.49} 58.01±1.4458.01\pm_{1.44} 62.05±1.5362.05\pm_{1.53} 61.74±1.361.74\pm_{1.3} 61.85±5.9761.85\pm_{5.97}
64-shot 70.01±1.7770.01\pm_{1.77} 65.86±1.2865.86\pm_{1.28} 63.1±0.7963.1\pm_{0.79} 65.12±0.9665.12\pm_{0.96} 65.34±1.1865.34\pm_{1.18} 66.98±2.6666.98\pm_{2.66}
Full 81.76±0.8181.76\pm_{0.81} 73.5±0.6173.5\pm_{0.61} 73.0±0.5973.0\pm_{0.59} 78.01±0.4178.01\pm_{0.41} 74.46±0.4874.46\pm_{0.48} 76.26±1.0276.26\pm_{1.02}
IC-XLTM=10SRC{}_{SRC}^{M=10} (Ktgt=0K_{tgt}=0)
8-shot 34.71±7.3334.71\pm_{7.33} 34.05±9.3334.05\pm_{9.33} 32.13±6.4632.13\pm_{6.46} 38.42±5.9238.42\pm_{5.92} 37.47±7.737.47\pm_{7.7} 25.75±6.1525.75\pm_{6.15}
16-shot 52.08±11.8552.08\pm_{11.85} 48.49±10.4648.49\pm_{10.46} 45.12±11.0845.12\pm_{11.08} 50.81±8.6150.81\pm_{8.61} 46.89±11.1546.89\pm_{11.15} 44.34±16.0144.34\pm_{16.01}
32-shot 60.45±8.8460.45\pm_{8.84} 57.6±5.9557.6\pm_{5.95} 53.42±6.5353.42\pm_{6.53} 58.15±5.0258.15\pm_{5.02} 57.32±5.4857.32\pm_{5.48} 57.22±10.2257.22\pm_{10.22}
64-shot 69.84±1.3269.84\pm_{1.32} 66.96±1.4366.96\pm_{1.43} 60.41±0.7960.41\pm_{0.79} 65.14±1.0865.14\pm_{1.08} 64.17±1.564.17\pm_{1.5} 62.31±2.3362.31\pm_{2.33}
Full 81.48±0.3781.48\pm_{0.37} 74.06±1.0374.06\pm_{1.03} 72.54±0.8272.54\pm_{0.82} 77.59±0.8577.59\pm_{0.85} 73.7±0.7273.7\pm_{0.72} 71.27±1.1771.27\pm_{1.17}
IC-XLTM=10 (Ktgt=1K_{tgt}=1)
8-shot 34.71±7.3334.71\pm_{7.33} 32.24±4.1632.24\pm_{4.16} 30.62±6.0430.62\pm_{6.04} 40.4±7.8540.4\pm_{7.85} 31.04±12.5331.04\pm_{12.53} 32.41±8.2332.41\pm_{8.23}
16-shot 52.08±11.8552.08\pm_{11.85} 49.71±11.0449.71\pm_{11.04} 45.86±9.5545.86\pm_{9.55} 54.83±5.1854.83\pm_{5.18} 48.69±9.9248.69\pm_{9.92} 44.2±9.6844.2\pm_{9.68}
32-shot 60.45±8.8460.45\pm_{8.84} 59.38±5.5559.38\pm_{5.55} 55.23±5.655.23\pm_{5.6} 60.21±2.9860.21\pm_{2.98} 59.06±5.0659.06\pm_{5.06} 60.66±9.3360.66\pm_{9.33}
64-shot 69.84±1.3269.84\pm_{1.32} 67.2±1.4967.2\pm_{1.49} 62.32±1.162.32\pm_{1.1} 65.32±0.6265.32\pm_{0.62} 66.33±0.9866.33\pm_{0.98} 67.04±2.9267.04\pm_{2.92}
Full 81.48±0.3781.48\pm_{0.37} 74.07±0.5574.07\pm_{0.55} 73.34±0.8273.34\pm_{0.82} 78.07±0.7678.07\pm_{0.76} 75.2±1.3375.2\pm_{1.33} 75.59±2.8475.59\pm_{2.84}
Table 9: Average per language across the different runs for evaluations under different resource budgets for the Aspect Category Detection dataset. In here, ±\pm refers to the standard deviation of the performance on the conducted runs.
KsrcK_{src} ENG AMH ARA AZE FRA IND ISL JAP
ZS-XLT (Ktgt=0K_{tgt}=0)
8-shot 62.93±1.562.93\pm_{1.5} 32.53±0.4232.53\pm_{0.42} 36.81±0.9336.81\pm_{0.93} 40.3±0.8240.3\pm_{0.82} 52.11±0.7752.11\pm_{0.77} 47.24±0.6447.24\pm_{0.64} 44.9±0.9944.9\pm_{0.99} 50.96±0.6750.96\pm_{0.67}
16-shot 70.52±7.2470.52\pm_{7.24} 39.27±7.7939.27\pm_{7.79} 43.38±6.5543.38\pm_{6.55} 47.55±6.4847.55\pm_{6.48} 59.71±7.7559.71\pm_{7.75} 56.42±8.6256.42\pm_{8.62} 51.42±6.851.42\pm_{6.8} 57.27±6.157.27\pm_{6.1}
32-shot 81.72±1.3981.72\pm_{1.39} 52.4±1.7352.4\pm_{1.73} 58.98±2.2558.98\pm_{2.25} 59.58±1.6359.58\pm_{1.63} 73.72±1.8873.72\pm_{1.88} 71.08±1.4671.08\pm_{1.46} 63.67±3.8363.67\pm_{3.83} 71.65±1.3371.65\pm_{1.33}
64-shot 81.71±2.8181.71\pm_{2.81} 49.68±6.3249.68\pm_{6.32} 56.19±4.8556.19\pm_{4.85} 58.49±6.4158.49\pm_{6.41} 72.78±5.172.78\pm_{5.1} 71.34±2.671.34\pm_{2.6} 63.23±4.8863.23\pm_{4.88} 70.85±3.3470.85\pm_{3.34}
Full 90.06±0.4590.06\pm_{0.45} 62.53±0.7962.53\pm_{0.79} 70.98±1.9570.98\pm_{1.95} 71.63±1.8771.63\pm_{1.87} 84.87±0.2584.87\pm_{0.25} 84.23±0.0984.23\pm_{0.09} 73.77±0.6273.77\pm_{0.62} 83.39±0.6683.39\pm_{0.66}
1S-XLT (Ktgt=1K_{tgt}=1)
8-shot 63.3±1.463.3\pm_{1.4} 34.26±0.6234.26\pm_{0.62} 37.67±0.937.67\pm_{0.9} 41.96±0.341.96\pm_{0.3} 53.41±1.0253.41\pm_{1.02} 49.6±0.5749.6\pm_{0.57} 46.48±1.3446.48\pm_{1.34} 53.53±0.3253.53\pm_{0.32}
16-shot 70.52±5.7470.52\pm_{5.74} 39.33±5.8939.33\pm_{5.89} 44.65±5.7444.65\pm_{5.74} 48.4±4.7548.4\pm_{4.75} 59.49±5.5859.49\pm_{5.58} 57.42±5.9357.42\pm_{5.93} 51.4±4.0151.4\pm_{4.01} 60.13±4.8760.13\pm_{4.87}
32-shot 81.7±1.0681.7\pm_{1.06} 52.87±1.0752.87\pm_{1.07} 58.56±1.6558.56\pm_{1.65} 60.66±1.9260.66\pm_{1.92} 73.61±1.3173.61\pm_{1.31} 71.58±2.271.58\pm_{2.2} 64.26±2.9664.26\pm_{2.96} 72.56±1.2472.56\pm_{1.24}
64-shot 81.17±2.6181.17\pm_{2.61} 50.49±4.9150.49\pm_{4.91} 56.37±3.9256.37\pm_{3.92} 58.99±5.6758.99\pm_{5.67} 72.6±4.5372.6\pm_{4.53} 71.11±1.8571.11\pm_{1.85} 63.67±4.4963.67\pm_{4.49} 71.46±2.9571.46\pm_{2.95}
Full 90.08±0.490.08\pm_{0.4} 62.78±0.8262.78\pm_{0.82} 71.11±1.7771.11\pm_{1.77} 71.63±1.6771.63\pm_{1.67} 84.86±0.1884.86\pm_{0.18} 84.17±0.1684.17\pm_{0.16} 73.9±0.5373.9\pm_{0.53} 83.45±0.6383.45\pm_{0.63}
IC-XLTM=20SRC{}_{SRC}^{M=20} (Ktgt=0K_{tgt}=0)
8-shot 73.24±2.7173.24\pm_{2.71} 51.95±2.8451.95\pm_{2.84} 56.6±3.2956.6\pm_{3.29} 58.92±3.1858.92\pm_{3.18} 65.75±3.0465.75\pm_{3.04} 64.79±3.7464.79\pm_{3.74} 59.41±3.1259.41\pm_{3.12} 66.87±3.6166.87\pm_{3.61}
16-shot 82.0±1.3782.0\pm_{1.37} 57.2±2.9957.2\pm_{2.99} 62.97±2.7562.97\pm_{2.75} 65.03±3.1565.03\pm_{3.15} 74.38±2.174.38\pm_{2.1} 71.95±3.0371.95\pm_{3.03} 65.89±3.7265.89\pm_{3.72} 73.25±2.373.25\pm_{2.3}
32-shot 85.03±0.5285.03\pm_{0.52} 62.5±1.562.5\pm_{1.5} 68.45±1.8768.45\pm_{1.87} 69.26±1.6169.26\pm_{1.61} 78.52±1.1478.52\pm_{1.14} 77.93±1.8977.93\pm_{1.89} 70.62±1.5370.62\pm_{1.53} 77.94±1.1177.94\pm_{1.11}
64-shot 87.18±0.6687.18\pm_{0.66} 63.94±1.4463.94\pm_{1.44} 70.06±1.470.06\pm_{1.4} 71.18±1.8971.18\pm_{1.89} 81.37±1.1481.37\pm_{1.14} 80.31±1.4480.31\pm_{1.44} 71.16±1.4571.16\pm_{1.45} 80.6±1.1980.6\pm_{1.19}
Full 89.46±0.4389.46\pm_{0.43} 61.27±2.6661.27\pm_{2.66} 70.77±2.1870.77\pm_{2.18} 71.49±2.1271.49\pm_{2.12} 83.13±1.8883.13\pm_{1.88} 81.6±2.1581.6\pm_{2.15} 68.83±3.7268.83\pm_{3.72} 82.83±2.1982.83\pm_{2.19}
IC-XLTM=20 (Ktgt=1K_{tgt}=1)
8-shot 73.24±2.7173.24\pm_{2.71} 58.17±3.3958.17\pm_{3.39} 61.45±3.4161.45\pm_{3.41} 66.0±3.1466.0\pm_{3.14} 67.26±3.7267.26\pm_{3.72} 70.46±3.3570.46\pm_{3.35} 67.16±2.7267.16\pm_{2.72} 69.77±3.1669.77\pm_{3.16}
16-shot 82.0±1.3782.0\pm_{1.37} 63.5±3.6763.5\pm_{3.67} 67.55±1.9367.55\pm_{1.93} 72.78±1.8672.78\pm_{1.86} 75.98±1.8375.98\pm_{1.83} 77.77±1.7977.77\pm_{1.79} 74.27±2.174.27\pm_{2.1} 76.95±1.2976.95\pm_{1.29}
32-shot 85.03±0.5285.03\pm_{0.52} 67.12±2.3967.12\pm_{2.39} 72.76±1.3372.76\pm_{1.33} 75.42±1.7975.42\pm_{1.79} 80.06±1.0680.06\pm_{1.06} 81.6±1.2181.6\pm_{1.21} 77.81±1.0677.81\pm_{1.06} 80.13±0.8680.13\pm_{0.86}
64-shot 87.18±0.6687.18\pm_{0.66} 70.14±1.670.14\pm_{1.6} 74.69±0.9874.69\pm_{0.98} 78.02±1.6578.02\pm_{1.65} 83.29±0.7983.29\pm_{0.79} 83.9±0.8483.9\pm_{0.84} 79.74±1.0879.74\pm_{1.08} 82.09±0.782.09\pm_{0.7}
Full 89.46±0.4389.46\pm_{0.43} 68.09±4.0768.09\pm_{4.07} 76.43±1.8776.43\pm_{1.87} 81.37±1.9981.37\pm_{1.99} 84.92±1.484.92\pm_{1.4} 85.46±1.6285.46\pm_{1.62} 80.32±1.180.32\pm_{1.1} 83.77±1.7283.77\pm_{1.72}
IC-XLTM=10SRC{}_{SRC}^{M=10} (Ktgt=0K_{tgt}=0)
8-shot 73.36±0.9273.36\pm_{0.92} 48.95±1.9848.95\pm_{1.98} 52.93±1.5752.93\pm_{1.57} 55.65±1.6555.65\pm_{1.65} 65.77±1.1665.77\pm_{1.16} 60.89±2.1260.89\pm_{2.12} 59.64±0.9759.64\pm_{0.97} 62.74±1.2962.74\pm_{1.29}
16-shot 80.54±0.9980.54\pm_{0.99} 55.99±3.3255.99\pm_{3.32} 60.75±3.4460.75\pm_{3.44} 62.65±2.9162.65\pm_{2.91} 73.17±2.1873.17\pm_{2.18} 70.08±2.6670.08\pm_{2.66} 65.99±2.7865.99\pm_{2.78} 70.86±2.670.86\pm_{2.6}
32-shot 84.22±0.6284.22\pm_{0.62} 59.47±1.8459.47\pm_{1.84} 64.83±1.464.83\pm_{1.4} 66.48±1.866.48\pm_{1.8} 77.22±0.9977.22\pm_{0.99} 74.92±1.2974.92\pm_{1.29} 68.99±1.5468.99\pm_{1.54} 74.61±1.0874.61\pm_{1.08}
64-shot 86.75±0.2986.75\pm_{0.29} 62.92±1.2862.92\pm_{1.28} 68.55±1.4368.55\pm_{1.43} 69.91±1.769.91\pm_{1.7} 80.21±0.7880.21\pm_{0.78} 78.86±1.5478.86\pm_{1.54} 72.37±1.1672.37\pm_{1.16} 77.97±1.6977.97\pm_{1.69}
Full 89.41±0.489.41\pm_{0.4} 63.39±3.5463.39\pm_{3.54} 71.62±2.1371.62\pm_{2.13} 70.82±3.9570.82\pm_{3.95} 82.81±1.882.81\pm_{1.8} 81.04±2.3881.04\pm_{2.38} 69.22±3.469.22\pm_{3.4} 82.99±1.5382.99\pm_{1.53}
IC-XLTM=10 (Ktgt=1K_{tgt}=1)
8-shot 73.36±0.9273.36\pm_{0.92} 51.7±1.8851.7\pm_{1.88} 57.18±2.6757.18\pm_{2.67} 61.04±2.0161.04\pm_{2.01} 67.12±1.6267.12\pm_{1.62} 67.44±2.9867.44\pm_{2.98} 66.48±1.4566.48\pm_{1.45} 64.99±1.7264.99\pm_{1.72}
16-shot 80.54±0.9980.54\pm_{0.99} 60.89±3.5660.89\pm_{3.56} 65.16±2.8965.16\pm_{2.89} 68.59±2.6768.59\pm_{2.67} 74.81±1.8174.81\pm_{1.81} 75.34±1.7475.34\pm_{1.74} 73.26±1.3973.26\pm_{1.39} 73.03±2.1673.03\pm_{2.16}
32-shot 84.22±0.6284.22\pm_{0.62} 61.26±1.6761.26\pm_{1.67} 66.91±1.466.91\pm_{1.4} 70.99±0.9270.99\pm_{0.92} 80.0±0.7380.0\pm_{0.73} 79.44±1.0779.44\pm_{1.07} 75.56±1.1775.56\pm_{1.17} 76.27±0.4476.27\pm_{0.44}
64-shot 86.75±0.2986.75\pm_{0.29} 66.93±1.6266.93\pm_{1.62} 72.2±1.2272.2\pm_{1.22} 74.91±0.9474.91\pm_{0.94} 82.99±0.7882.99\pm_{0.78} 83.11±1.1883.11\pm_{1.18} 79.7±0.6779.7\pm_{0.67} 80.51±0.7880.51\pm_{0.78}
Full 89.41±0.489.41\pm_{0.4} 70.68±2.9470.68\pm_{2.94} 77.28±0.4577.28\pm_{0.45} 81.73±1.1381.73\pm_{1.13} 85.53±1.1185.53\pm_{1.11} 85.6±0.9285.6\pm_{0.92} 81.07±0.8981.07\pm_{0.89} 84.68±0.4484.68\pm_{0.44}
Table 10: Average performance for English, Amharic, Arabic, Azeri, French, Indonesian, Icelandic, and Japanese across the different runs for evaluations under different resource budgets in the MASSIVE Domain Classification Task. In here, ±\pm refers to the standard deviation of the performance on the conducted runs.
KsrcK_{src} RUS SPA SWA THA TUR URD
ZS-XLT (Ktgt=0K_{tgt}=0)
8-shot 48.8±1.0548.8\pm_{1.05} 51.05±0.851.05\pm_{0.8} 36.46±0.5836.46\pm_{0.58} 46.05±0.1546.05\pm_{0.15} 48.24±0.9948.24\pm_{0.99} 37.22±0.6537.22\pm_{0.65}
16-shot 57.1±7.6357.1\pm_{7.63} 58.39±7.0458.39\pm_{7.04} 41.78±6.3841.78\pm_{6.38} 53.49±7.9353.49\pm_{7.93} 53.6±5.4753.6\pm_{5.47} 45.26±7.8145.26\pm_{7.81}
32-shot 72.15±2.072.15\pm_{2.0} 72.12±1.6472.12\pm_{1.64} 52.09±2.3752.09\pm_{2.37} 69.0±2.7469.0\pm_{2.74} 66.26±1.7966.26\pm_{1.79} 58.88±2.0358.88\pm_{2.03}
64-shot 72.11±5.1772.11\pm_{5.17} 71.83±4.4371.83\pm_{4.43} 50.36±5.5450.36\pm_{5.54} 67.6±5.0167.6\pm_{5.01} 63.97±5.4663.97\pm_{5.46} 56.4±4.6456.4\pm_{4.64}
Full 83.98±0.4983.98\pm_{0.49} 83.09±0.7983.09\pm_{0.79} 65.75±2.1965.75\pm_{2.19} 79.88±2.3779.88\pm_{2.37} 77.7±1.1977.7\pm_{1.19} 67.72±0.9667.72\pm_{0.96}
1S-XLT (Ktgt=1K_{tgt}=1)
8-shot 50.65±1.0650.65\pm_{1.06} 52.26±1.0652.26\pm_{1.06} 37.31±0.6137.31\pm_{0.61} 49.75±0.3849.75\pm_{0.38} 49.86±0.8949.86\pm_{0.89} 39.4±0.3239.4\pm_{0.32}
16-shot 58.79±6.6158.79\pm_{6.61} 58.37±5.2858.37\pm_{5.28} 42.03±4.6442.03\pm_{4.64} 55.46±6.0255.46\pm_{6.02} 54.14±3.854.14\pm_{3.8} 46.04±5.946.04\pm_{5.9}
32-shot 72.77±2.1372.77\pm_{2.13} 72.65±1.0372.65\pm_{1.03} 52.88±2.0752.88\pm_{2.07} 69.34±2.8969.34\pm_{2.89} 66.52±1.9466.52\pm_{1.94} 60.4±1.3660.4\pm_{1.36}
64-shot 71.95±4.9371.95\pm_{4.93} 71.64±3.4871.64\pm_{3.48} 51.26±4.1951.26\pm_{4.19} 66.78±4.8466.78\pm_{4.84} 63.44±5.2563.44\pm_{5.25} 58.09±3.6158.09\pm_{3.61}
Full 83.89±0.4283.89\pm_{0.42} 83.15±0.5883.15\pm_{0.58} 65.85±1.8465.85\pm_{1.84} 79.95±1.9379.95\pm_{1.93} 77.68±1.0977.68\pm_{1.09} 68.34±0.7468.34\pm_{0.74}
IC-XLTM=20SRC{}_{SRC}^{M=20} (Ktgt=0K_{tgt}=0)
8-shot 64.98±3.4664.98\pm_{3.46} 63.79±3.1263.79\pm_{3.12} 51.51±4.1751.51\pm_{4.17} 63.69±3.5663.69\pm_{3.56} 62.52±3.3162.52\pm_{3.31} 57.68±2.6457.68\pm_{2.64}
16-shot 74.26±2.5974.26\pm_{2.59} 72.2±2.1772.2\pm_{2.17} 55.0±4.255.0\pm_{4.2} 69.61±2.2969.61\pm_{2.29} 68.52±2.3568.52\pm_{2.35} 63.29±2.763.29\pm_{2.7}
32-shot 79.78±0.8279.78\pm_{0.82} 75.58±1.7775.58\pm_{1.77} 59.21±3.6659.21\pm_{3.66} 75.42±1.8675.42\pm_{1.86} 73.58±1.6573.58\pm_{1.65} 68.19±1.268.19\pm_{1.2}
64-shot 81.58±1.0581.58\pm_{1.05} 78.44±1.6678.44\pm_{1.66} 60.89±2.8960.89\pm_{2.89} 77.9±1.377.9\pm_{1.3} 76.14±1.7576.14\pm_{1.75} 70.03±0.9570.03\pm_{0.95}
Full 83.76±1.6883.76\pm_{1.68} 81.23±2.5581.23\pm_{2.55} 53.76±5.1853.76\pm_{5.18} 78.91±1.6278.91\pm_{1.62} 76.36±2.9576.36\pm_{2.95} 68.26±2.4568.26\pm_{2.45}
IC-XLTM=20 (Ktgt=1K_{tgt}=1)
8-shot 69.41±3.0169.41\pm_{3.01} 67.04±3.4667.04\pm_{3.46} 65.27±3.165.27\pm_{3.1} 66.53±2.6566.53\pm_{2.65} 70.03±3.3170.03\pm_{3.31} 63.22±2.8263.22\pm_{2.82}
16-shot 77.11±1.5177.11\pm_{1.51} 75.4±1.575.4\pm_{1.5} 71.36±2.2871.36\pm_{2.28} 72.55±0.8172.55\pm_{0.81} 76.18±1.4376.18\pm_{1.43} 69.46±1.9869.46\pm_{1.98}
32-shot 81.43±1.1681.43\pm_{1.16} 78.68±1.278.68\pm_{1.2} 74.61±1.2674.61\pm_{1.26} 76.1±2.1976.1\pm_{2.19} 78.46±1.2878.46\pm_{1.28} 73.13±1.1973.13\pm_{1.19}
64-shot 83.24±0.883.24\pm_{0.8} 81.06±1.3381.06\pm_{1.33} 76.78±1.4776.78\pm_{1.47} 79.36±0.9779.36\pm_{0.97} 80.75±1.5280.75\pm_{1.52} 75.13±0.9775.13\pm_{0.97}
Full 84.39±2.284.39\pm_{2.2} 83.65±1.8483.65\pm_{1.84} 75.54±5.1475.54\pm_{5.14} 81.02±1.6781.02\pm_{1.67} 82.25±2.5182.25\pm_{2.51} 76.12±1.3876.12\pm_{1.38}
IC-XLTM=10SRC{}_{SRC}^{M=10} (Ktgt=0K_{tgt}=0)
8-shot 62.66±1.7162.66\pm_{1.71} 61.93±1.4761.93\pm_{1.47} 48.24±3.1848.24\pm_{3.18} 60.16±1.5960.16\pm_{1.59} 59.48±1.4459.48\pm_{1.44} 54.15±1.6654.15\pm_{1.66}
16-shot 70.82±3.0370.82\pm_{3.03} 69.05±2.369.05\pm_{2.3} 54.17±4.4154.17\pm_{4.41} 68.4±2.3368.4\pm_{2.33} 65.58±2.7665.58\pm_{2.76} 61.64±2.661.64\pm_{2.6}
32-shot 76.48±0.9376.48\pm_{0.93} 74.15±1.174.15\pm_{1.1} 54.89±3.254.89\pm_{3.2} 72.85±1.3472.85\pm_{1.34} 70.25±1.8970.25\pm_{1.89} 65.48±1.6265.48\pm_{1.62}
64-shot 81.02±0.8981.02\pm_{0.89} 78.2±1.2278.2\pm_{1.22} 58.61±3.6358.61\pm_{3.63} 76.28±0.9376.28\pm_{0.93} 73.64±0.5773.64\pm_{0.57} 67.74±1.2367.74\pm_{1.23}
Full 84.41±0.9384.41\pm_{0.93} 81.53±2.1781.53\pm_{2.17} 56.06±3.3356.06\pm_{3.33} 79.02±1.8679.02\pm_{1.86} 76.65±2.3276.65\pm_{2.32} 67.47±3.3567.47\pm_{3.35}
IC-XLTM=10 (Ktgt=1K_{tgt}=1)
8-shot 65.57±2.665.57\pm_{2.6} 65.0±1.3765.0\pm_{1.37} 61.48±1.7161.48\pm_{1.71} 61.81±2.6261.81\pm_{2.62} 63.79±2.4263.79\pm_{2.42} 58.32±2.4858.32\pm_{2.48}
16-shot 73.95±2.873.95\pm_{2.8} 71.74±2.771.74\pm_{2.7} 69.91±2.169.91\pm_{2.1} 70.48±2.2870.48\pm_{2.28} 70.72±2.470.72\pm_{2.4} 66.05±2.0866.05\pm_{2.08}
32-shot 76.83±0.9476.83\pm_{0.94} 76.54±0.6676.54\pm_{0.66} 70.82±1.8670.82\pm_{1.86} 74.33±1.0374.33\pm_{1.03} 74.68±0.9774.68\pm_{0.97} 68.89±1.2768.89\pm_{1.27}
64-shot 82.49±0.8982.49\pm_{0.89} 80.75±1.280.75\pm_{1.2} 75.7±2.275.7\pm_{2.2} 78.26±0.5678.26\pm_{0.56} 78.02±0.978.02\pm_{0.9} 72.57±1.372.57\pm_{1.3}
Full 86.18±0.3986.18\pm_{0.39} 84.18±1.5484.18\pm_{1.54} 78.23±1.8778.23\pm_{1.87} 80.98±0.4880.98\pm_{0.48} 83.63±1.183.63\pm_{1.1} 76.34±0.6176.34\pm_{0.61}
Table 11: Average performance for Russian, Spanish, Swahili, Thai, Turkish, and Urdu across the different runs for evaluations under different resource budgets in the MASSIVE Domain Classification Task. In here, ±\pm refers to the standard deviation of the performance on the conducted runs.

A.6 Ethics Statement

The proposed method helps to improve downstream cross-lingual performance on languages underrepresented in the multilingual model pretraining. However, we believe that data collection for low-resource languages should continue to be a priority for the NLP research community. It remains very important to better integrate linguistic diversity in multilingual models to avoid technological biases and avoid hindering technological development in societies whose language has a limited number of speakers, or limited funding for developing linguistic resources.

On the other hand, we acknowledge the challenge posed by language variants, including regional dialects, which often suffer from underrepresentation. This may result in multilingual models biased towards variants with a larger digital footprint, excluding linguistic features of communities less present in digital spaces. We wish to acknowledge this aspect and warn that it could increase the exclusion of these communities from integration into the digital content ecosystem and information technology tools.

A.7 Licences of systems and datasets

In this work, the tools utilized include an mT5 model and the transformers library Wolf et al. (2020), both of which use the Apache 2.0 license. The MASSIVE dataset, on the other hand, operates under a CC by 4.0 license. As for the Aspect Category Detection dataset, it employs a MS-NC-No ReD license, which limits its usage strictly to an academic scope. Since the aim of this work is to evaluate the performance of a proposed cross-lingual system, we adhere to all the licenses of the utilized material.

The research presented in this paper is intended for academic purposes, and therefore, we adhere to the licenses governing all utilized materials.