Chu-Cheng Lin , Xinyi Wang11footnotemark: 1 , Jonathan H. Clark, Han Lu, &Yun Zhu, Chenxi Whitehouse, Hongkun Yu Google
Abstract
Adapting pretrained large language models(LLMs) to various downstream tasks in tens or hundreds of human languages is computationally expensive.Parameter-efficient fine-tuning(PEFT) significantly reduces the adaptation cost, by tuning only a small amount of parameters. However, common PEFT methods LoRA(Hu etal., 2022) suffer from suboptimal performance on diverse dataset mixtures, due to aggressive parameter tying and negative interference among different datasets.In this work, we propose Featurized Low-rank Mixtures(FLix), a novel PEFT method designed for effective multitask multilingual adaptation.FLix associates each unique dataset feature, such as the dataset’s language or task, with its own low-rank weight update parameters.By composing feature-specific parameters for each dataset, FLix can accommodate diverse dataset mixtures and generalize better to unseen datasets. Our experiments show that FLix leads to significant improvements over a variety of tasks for both supervised learning and zero-shot settings with gains of up to in exact match points in zero-shot semantic parsing.
1 Introduction
Large language models(LLMs) have shown impressive performance on various real world applications in many different human languages (Brown etal., 2020; Soltan etal., 2022; Google etal., 2023).While there have been notable successes aligning an LLM to become a generalist which can follow human instructions to perform different tasks (Ouyang etal., 2022), there are also significant interests in adapting an LLM into specialists, each of which works on its specific task that is known a priori.
Intuitively, LLM adaptation can be done by continued training (or fine-tuning) on target languages and datasets. However, fine-tuning all model parameters on every dataset can be computationally and financially prohibitive.Parameter-efficient fine-tuning methods(PEFT) such as LoRA (Hu etal., 2022) and prompt tuning (Lester etal., 2021) reduce the computational costs of adapting LLMs to a downstream task. They parametrize LLM fine-tuning with a small set of trainable parameters, keeping the majority of LLM parameters frozen.
While PEFT has been widely used to adapt LLMs to a single dataset, very few prior works studied the best practices for adapting the model jointly on many different use cases.Vu etal. (2022) proposed adding a multilingual pretraining stage to prompt tuning using multilingual unlabeled data to improve zero-shot summarization. Wang etal. (2023b) proposed a multitask prompt tuning method that learns a single soft prompt which could be used to be adapt to other target tasks. However, these methods only considered either multilingual or multitask datasets, while we generally would like to adapt the model such that it generalizes along multiple axes(i.e. tasks and languages). Moreover, they require multiple tuning stages, which limits their applicability in practice.
In this paper, we propose Featurized Low-rank Mixtures (FLix), an extension of LoRA for modeling diverse dataset mixtures. Compared to LoRA which applies the same low-rank adaptation for all inputs, FLix parametrizes such updates to decompose linearly as a sum of feature-specific low-rank adaptations, each associated with an active dataset feature, such as language or task ID.Under FLix, different adaptations can be learned for different features. The compositional nature of FLix also provides an inductive bias for learning generalizable adaptations. Moreover,FLix is generally computationally efficient: it only needs to activate a tiny fraction of trainable parameters for each input, making both tuning and deployment efficient.FLix is related to prior works on sub-network composition(Lin etal., 2021; Ilharco etal., 2022), which show that models fine-tuned on different tasks could be composed together as a single model.
In this article, we contribute:
- •
a modeling formulation that disentangles learning of tasks and languages in a way that improves generalization quality while remaining highly efficient;
- •
experimental evidence that the model improves generalization quality across four very different tasks: named entity recognition, semantic parsing, in-language question answering, and cross-lingual question answering; and
- •
evidence for the hypothesis that imbuing models with meta-data such as task and language improves quality, even when compared with powerful modern adaptation methods such as LoRA.
The rest of this paper is structured as follows: In §2, we first formulate the multitask multilingual learning setting, and discuss the challenges associated with the current PEFT methods.In §3, we describe the design of FLix, and how FLix adapts to zero-shot scenarios.We also propose to train FLix with feature dropout, which encourages positive transfer and generalization. We evaluate FLix on multitask or multilingual learning setting(§5.2), and later on joint multilingual multitask tuning and zero-shot generalization(§5.1). The experiment results and ablations show that FLix brings significant improvements over standard PEFT methods for all settings, and it is especially effective when used with very diverse training mixture and at zero-shot generalizations (§5, §6).
2 Multitask multilingual learning
We consider a multitask learning (Caruana, 1997) setting where datasets are indexed using task-language tuples.We assume that there are tasks and languages . And our goal is to serve fine-tuned LLMs for each of these task-language combinations. It is worth noting that two special cases of this setup are multilingual modeling () and multitask modeling ().
Parameter-efficient tuning(PEFT) is a popular method to adapt a pretrained LLM to downstream tasks without incurring large computational cost. In this work, we want to use PEFT to support languages and tasks.
2.1 Low-rank Adaptation
We focus on Low-rank Adaptation(LoRA)(Hu etal., 2022), an effective PEFT method that incurs minimal additional inference cost.Let be the pretrained weight matrix, LoRA keeps the much larger weight matrix unchanged, and instead only optimizes low-rank factorized weight adaptation matrix . The final weight matrix is
(1) |
where , . Empirically LoRA often compares competitively with full fine-tuning, even when . LoRA can thus significantly reduce the size of trainable parameters.
2.2 Challenges
PEFT methods such as LoRA have been shown to be very effective for tuning LLMs, achieving comparable results to full fine-tuning while incurring only a fraction of the computational cost.However, there are several challenges with the multitask multilingual learning problem that the current PEFT methods might not be able to address.
Interference among different datasets
Multitask multilingual training with PEFT requires one to fine-tune an LLM on a diverse data mixture, from up to different datasets spanning tasks and languages. This approach has significantly lower overhead than modeling each dataset individually, and allows positive transfer. However, training a single set of PEFT parameters over all datasets could also lead to negative interference among tasks and languages that are dissimilar.
Generalizing to unseen task and language
Publicly available wide-coverage multitask and multilingual datasets are often incomplete. For example, many task-language combinations are missing in XTREME-UP. Moreover, some underrepresented languages may still be missing from such datasets.Standard PEFT methods could have difficulty generalizing to unseen task-language combinations and unseen languages: they simply optimize a single set of parameters on all tasks without explicit modeling assumptions that capture the relationships and similarities between different datasets. And it is not clear how to transform such PEFT parameters to a new task or language at inference time.
3 Featurized Low-Rank Mixtures
We propose Featurized Low-Rank Mixtures(FLix), an effective multitask PEFT method that supports diverse training data mixture, and excels at zero-shot generalization to new task-language combinations.Under FLix, NLP tasks and languages are featurized as discrete features. And each feature is associated with a low-rank weight update matrix. Figure1 shows the training and inference processes of FLix.
3.1 Model Architecture
Given a diverse data mixture from tasks where each tasks are in languages, first we define a set of features where each feature could represent either a task, a language, or any other data attribute. We assign a low-rank factorized weight matrix for each feature .
Let be an input to the model, let represent the features of , where indicates that has feature , and otherwise.
(2) |
where , , and is the maximum rank of the the -th feature’s adaptation matrix.Note that compared to LoRA in equation1 that applies the same (and therefore same ) for all inputs, FLix uses different adaptation matrices based on the features of the input data .
Feature dropout.
One potential problem of training FLix is that the model might become overly reliant on the feature annotation of the training dataset, limiting positive transfer and making the model brittle to inputs different from training distribution.We randomly turn off a subset of active features at training time with a predetermined feature dropout probability.Experiments in §6 show that feature dropout brings consistent gains to FLix.
Exploiting feature sparsity for low training and serving costs.
Note that equation2 implies that on input , is not a function of if . In other words, weights of unused features of input are not needed in either training or serving time. While the number of trainable parameters under FLix grows linearly with the number of features , the feature count of each input remains a constant; and this is a relatively small value in our multitask multilingual learning settings. Therefore, the compute costs of FLix could still remain constant when scaling to increasingly more tasks and languages.
3.2 Zero-shot Composition
We find FLix to be particularly effective at zero-shot generalization, likely because of the explicit feature composition assumption. While previous work proposed using language-specific modules at pretraining time to enhance crosslingual transfer (Vu etal., 2022; Pfeiffer etal., 2023), our work shows that sparse modularized PEFT architecture is also effective for directly adapting dense pretrained models to unseen datasets.
In this paper, we consider how FLix adapts to unseen datasets in two different zero-shot scenarios:
- Unseen combinations.
We want to do inference for a dataset that has a combination of active features that did not appear in the training mixture. For example, say the training data mixture contains QA data in French and semantic parsing data in Bengali; and we want to test the model on QA in Bengali. FLix naturally supports such features; and no change is required while applying equation2.
- Unseen languages.
The test data could have a subset of features that are not covered by any of the dataset in the training mixture. Specifically, we focus on the setting where the test data is from a new language unseen during training. In this case, we only use the task feature of the data to calculate the model weights in equation2.
In §5, we show that FLix significantly outperforms the baselines for both types of zero-shot settings.
4 Experiments
For all experiments, we use instruction fine-tuned PaLM-2 small Google etal. (2023) as the base model. We evaluate our method on a variety of tasks, languages, and data mixtures to verify that it generalizes to different use cases.
Datasets.
We use the XTREME-UP dataset (Ruder etal., 2023) and format the input data with the language and task information(Wang etal., 2023a).We use different language and task subsets from XTREME-UP. Dataset details are described in §5.111A list of all experiments’ training and evaluation tasks can also be found in AppendixB.
Metrics.
We report F1 scores for both in- and cross-lingual QA tasks, and NER tasks. For semantic parsing, we report exact-match accuracy. All numbers are normalized between . We report average normalized metric scores for (monolingual) multitask experiments.
Hyperparameters.
We use a batch size of during training for all experiments. For FLix, we also set the feature dropout probability to be .
Ranks.
Compute-matched LoRA baselines have rank- adaptation matrices222Strictly speaking their ranks ; we use this terminology loosely to reduce clutter. in all experiments. In this paper, every dataset has both task and language features; and we allocate feature-specific parameter counts such that the trainable parameters of active features under FLix always match compute-matched baselines. More specifically, we let task and language features have either rank- or rank- adaptation matrices. We allocate smaller matrices (rank-) for task features in multitask experiments (§5.2), and larger matrices in both multilingual (§5.2) and joint multitask-multilingual (§5.1) experiments. We adjust the ranks of language features’ adaptation matrices accordingly, to ensure that every dataset’s adaptation matrices have a total rank .
Model selection.
We evaluate on validation splits every 200 iterations, and train for a maximum of 2000 steps. For every multilingual and multitask task, we choose the checkpoint that has the best averaged metric numbers across languages or tasks, and subsequently evaluate on the test splits.
Baselines.
We compare to the vanilla LoRA method under several different settings to ensure the fairness of comparison:
- •
Compute-matched sets the rank of the vanilla LoRA model to be equivalent to the maximum sum of the rank of feature-specific adaptations under FLix. This ensures LoRA and FLix uses comparable computation during training and inference.
- •
Param-matched sets the rank of the vanilla LoRA model such that the total number of trainable parameters is the same as its FLix counterpart.333The number of trainable parameters in FLix grows linearly with the number of features in a dataset. Therefore the param-matched counterpart’s rank can be a relatively large number from around 20 to 90, depending on the task.
5 Results
5.1 Study A: Joint Multitask Multilingual Learning
Method NER Semantic Parsing QA InLang QA CrossLang Zero-shot Unseen Comb: QA CrossLang Zero-shot Unseen Comb: Semantic parsing Zero-shot Unseen Lang: QA CrossLang Avg. Compute-matched 76.3 35.4 86.6 83.1 65.1 Param-matched 47.1 87.3 68.9 FLix 84.0 89.4 85.2 82.8 42.6 77.2 72.4
In this section, we evaluate the performance of FLix and baselines on a diverse data mixture, where the tuning data contain different languages and tasks.
5.1.1 Multitask multilingual tuning
We conduct multitask multilingual tuning on both vanilla LoRA and our proposed FLix model. This setting shows one of the primary strengths of FLix: The ability to simultaneously generalize along multiple axes (tasks and languages) without suffering from negative task interference. We evaluate on four tasks from XTREME-UP covering a wide variety of use cases and languages: crosslingual QA, in-language QA, semantic parsing, NER. In addition, we also add training data from machine translation to the data mixture since it has the best language coverage which allows cross-lingual transfer.We use all languages in in-language QA, semantic parsing, NER, and a subset of languages in cross-lingual QA as training data.444We only use the languages included in the original XOR-TyDi QA(Asai etal., 2021): Arabic, Bengali, Japanese, Finish, Korean, Russian, Telugu. The rest of the cross-lingual QA languages — all Indic languages — are evaluated for zero-shot generalization in §5.1.2. In addition, we also include the corresponding machine translation datasets of languages from these 4 datasets.
FLix significantly outperforms baselines under multitask multilingual tuning.
The overall results are listed in Table1. We can see that FLix out-performs the the best LoRA baseline for all tasks other than semantic parsing. While it loses to param-matched LoRA on semantic parsing, FLixhas significantly less computational cost in comparison and it significantly out-performs the vanilla LoRA with the same computational cost.
5.1.2 Zero-shot Generalization
To evaluate both zero-shot generalizations (§3.2) we prepare two different training datasets:
- •
Holding out languages in cross-lingual QA. We reuse the training dataset from the joint multitask multilingual learning setup (§5.1) to evaluate both unseen feature combinations and unseen languages in the cross-lingual QA task. For unseen combinations, we evaluate the performance of the set of languages where the languages were present in other multilingual tasks in the training dataset; and for unseen languages, we evaluate on the set of languages that do not appear in other multilingual tasks in the training dataset.
- •
Holding out languages in semantic parsing. In this scenario, we use portions of crosslingual QA, in-language QA, semantic parsing, NER, and machine translation datasets from XTREME-UP as our training dataset. We include full crosslingual QA, in-language QA, NER datasets; but we do not include underrepresented languages in the semantic parsing portion of the training data. We also only include machine translation portions of languages that are already available in the other subsets, as in §5.1. And we evaluate the unseen combination performance of all held-out languages under semantic parsing.555These held-out languages are unseen combinations, rather than unseen languages, since they are available in the machine translation datasets of XTREME-UP.
Our method is very effective at both types of zero-shot generalization.
The comparison of FLix and baseline methods can be found in the rightmost 3 columns in Table1. We can see that FLix brings significant improvements to both unseen combinations and unseen languages on both cross-lingual QA and semantic parsing. This is likely because FLix allows one to select the subset of parameters most relevant to the features of the new test data, allowing more effective zero-shot generalization.
5.2 Study B: Multitask or multilingual learning
In this section, we examine the performance of our method and the baselines on data mixtures with a single type of feature. That is, we train and evaluate on datasets of a single task from several different languages(multilingual learning), or datasets of a single language with different tasks(multitask learning). We also add additional baselines for this setting where we train a separate LoRA model for each individual dataset (denoted as Single-Lang and Single-Task in Table2). This approach alleviates the capacity constraint of vanilla LoRA on diverse datasets, but adds more engineering overhead (as we briefly noted in §1) and cannot leverage positive transfer between datasets.
Specifically, we use subsets of the XTREME-UP dataset to construct these datasets:
- •
Multilingual learning. We experiment on four tasks: NER, semantic parsing, in-language QA, and cross-lingual QA. For each task, we train and evaluate on all languages available in the XTREME-UP dataset. Unlike §5.1, we train a separate multilingual model for each task in this setting.
- •
Multitask learning. We evaluate on two under-represented languages: Swahili and Bengali. We use the subset of tasks available for each language in XTREME-UP. For Swahili, we use semantic parsing, QA (in-language), and NER. For Bengali we use semantic parsing and QA (both in-language and cross-lingual).666The mismatch between task choices between these two languages is due to the sparsity of available datasets in XTREME-UP. Unlike §5.1, we train a separate multitask model for each language in this setting.
Method Multilingual Learning Multitask Learning Avg. NER Semantic Parsing QA InLang QA CrossLang Swahili Bengali Compute-matched 69.6 Param-matched 47.2 71.3 FLix 84.3 89.4 77.6 70.2 69.0 72.6
No baseline method consistently out-performs other baselines.
We report the performance of our method and the baselines in Table2. For each experiment, we list the average result over all languages or tasks in the datasets. First, we find that among the different baseline methods, there is no method that consistently out-performs others. Specifically, param-matched LoRA tends to have advantage for multitask learning and tasks that are very different from pretraining(semantic parsing and NER), while compute-matched LoRA appears to be superior on multilingual QA tasks. We suspect that this is because param-matched LoRA benefits from its higher capacity for the model to learn to generate into targets that are very different from the pretraining data, and it is also helpful for supporting generation into diverse target tasks. However, param-matched LoRA has significantly higher numbers of trainable parameters. This could lead to much higher computational overhead compared to compute-matched LoRA. Moreover, we observe that compute-matched LoRA is actually more competitive on tasks such as crosslingual QA, likely because it reduces over-fitting by tuning a much smaller number of parameters.
FLix achieves much better performance than vanilla LoRA with the same computation budgets.
FLix consistently outperforms compute-matched LoRA baseline on all settings, achieving significant gains without adding additional serving cost. Furthermore, our proposed method also outperforms param-matched LoRA on five out of the six data mixtures we evaluated.
We hypothesize that param-matched LoRA is better than FLix at semantic parsing but worse at other tasks because semantic parsing requires the LLM to generate into structured data that are very different from the pretraining data distribution, which might require particularly large model capacity to learn. In fact, Table2 shows that param-matched LoRA with a higher rank is much better than compute-matched LoRA for semantic parsing, while being worse or comparable on question answering tasks. While param-matched LoRA could be particularly helpful for the semantic parsing task, it requires more computational resources to scale to large number of datasets. These results indicate that our method is an effective and computationally-efficient strategy for tuning LLMs on diverse data mixtures.
6 Analysis and Ablations
6.1 Effect of feature dropout
Feature dropout strength | Validation | Test |
() | () | |
() | () | |
() | () |
In this section, we evaluate the effectiveness of feature dropout for FLix. We compare the performance of FLix with and without feature dropout for in-language QA task using multilingual training, and the results are in Table3. We can see that removing feature dropout leads to significant performance drop for both validation and test set. We hypothesize that this is because feature dropout is an effective regularization that encourages FLix to utilize the shared parameters for positive transfer between tasks and languages.
Rank | QA CrossLang | QA InLang | Semantic Parsing | NER |
77.6 | 89.4 | 45.0 | 84.3 | |
76.5() | 89.3() | 45.2() | 83.1() |
6.2 Effect of rank for FLix
In multilingual experiments (Table2) the language features only have rank- feature-specific weight update matrices.Such low rank configurations help reduce the overall parameter count.In this section, we examine how well FLix performs under an even smaller budget. Table4 shows the results of FLix on multitask multilingual tuning with rank set to 4 and 2. We can see that reducing the capacity of the feature-specific weight update matrices actually only lead to a small drop in performance on most tasks, indicating that FLix could be effective under even restrictive computational requirements.
6.3 FLix performs increasingly better on a diverse training dataset
In §5 we examined the performance of FLix and baselines under each data mixtures. Here we want to compare how different methods perform when using increasingly more diverse data mixtures. Specifically, we compare the change in task performance when using the multitask multilingual mixture as opposed to using only the multilingual data for each task. Figure2 shows the results of FLix and the baselines. We can see that vanilla LoRA suffers from negative transfer Wang etal. (2018). In particular, the compute-matched version with a smaller rank has significant decrease in performance on a diverse multitask multilingual training dataset. On the other hand, FLix is able to generally maintain similar or slightly higher performance when using the more diverse data mixture. This result indicates that FLix is a superior PEFT method to scale to large number of tasks and languages.
7 Related Work
Multilingual/multitask PEFT
While most prior works on parameter-efficient tuning focus on tuning the model for a single task, some recent works propose post-hoc composition pretrained PEFT modules to generalize to new task or language(Pfeiffer etal., 2020; Huang etal., 2023; Wang etal., 2021; Chronopoulou etal., 2023). These methods generally require separate training runs for each dataset, while our work propose a parameter-efficient tuning method that adapts the LLM using diverse data mixture in a single training run. Vu etal. (2022) proposed to add a multilingual pretraining stage to prompt tuning, which shows some improvements for zero-shot generalization when adapting to cross-lingual summarization task. In comparison, our proposed FLix focuses on tuning using multilingual data in many downstream tasks without additional training on unlabeled pretrainining data. Wang etal. (2023b) proposed a multitask prompt tuning method to learn a single prompt which could be used to adapt to many other target tasks. Similarly, this method requires multiple training stage while our method allows training end-to-end. Both of these methods are built upon prompt tuning using either multilingual training in a single task or multitask tuning in English, while our method supports diverse datasets in different tasks languages or any other arbitrary features.
Mixture-of-experts Models
Mixture-of-experts models (MoEs)(Shazeer etal., 2017; Lepikhin etal., 2020; Jawahar etal., 2023) are effective at increasing the model capacity by adding multiple expert parameters which could be activated differently to support different inputs. This architecture has been used to improve both pretrained models(Lepikhin etal., 2020; Jawahar etal., 2023) and parameter-efficient tuning methods(Zadouri etal., 2023; Zhu etal., 2023). Since MoEs often adds more computational cost, many works try to reduce the cost and improve the effectiveness of the model by either task-level routing (e.g., Task MoEs proposed by Kudugunta etal. (2021)) or encouraging sparsity of the experts(Shazeer etal., 2017; Lepikhin etal., 2020). FLix resembles Task MoEs in that FLix leverages task information as well. But FLix has additional composition capabilities thanks to its featurization.
Modularized Pretrained Models
Previous work proposed to add language-specific parameters to multilingual pretrained models(Pfeiffer etal., 2022; 2023) or machine translation models(Zhang etal., 2020). These works showed that language-specific modules often bring significant improvements to multilingual tasks. And they are especially helpful for zero-shot cross-lingual transfer. While prior works added language-specific modules during multilingual pretraining,in this paper we focus on the problem of adapting a pretrained model to a diverse mixture with many tasks and languages.
8 Future Work
There are several promising future directions for our work. While FLix achieves good zero-shot performance, it could potentially benefit from methods that automatically learn to select and compose parameters pretrained on different features for unseen data. It is also interesting to examine other applications of FLix: we encoded task and language informations as features in this work. But potentially other properties, such as modality, could also be featurized under FLix.
9 Conclusion
In this paper, we propose Featurized Low-Rank Mixtures(FLix), an effective parameter-efficient tuning method to fine-tune pretrained LLMs to data mixtures containing datasets in diverse tasks and languages. Our experiments show that FLix leads to significantly better performance on multitask multilingual fine-tuning compared to standard LoRA with little computational overhead. We also find that FLix achieves much better zero-shot generalization to new languages and task language combination unseen at training time.
References
- Asai etal. (2021)Akari Asai, Jungo Kasai, Jonathan Clark, Kenton Lee, Eunsol Choi, and HannanehHajishirzi.XOR QA: Cross-lingual open-retrieval question answering.In Kristina Toutanova, Anna Rumshisky, Luke Zettlemoyer, DilekHakkani-Tur, IzBeltagy, Steven Bethard, Ryan Cotterell, Tanmoy Chakraborty,and Yichao Zhou (eds.), NAACL, Online, June 2021. Association forComputational Linguistics.
- Brown etal. (2020)Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, JaredD Kaplan, PrafullaDhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell,Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, RewonChild, Aditya Ramesh, Daniel Ziegler, Jeffrey Wu, Clemens Winter, ChrisHesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess,Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever,and Dario Amodei.Language models are few-shot learners.In H.Larochelle, M.Ranzato, R.Hadsell, M.F. Balcan, and H.Lin(eds.), Advances in Neural Information Processing Systems, volume33,pp. 1877–1901. Curran Associates, Inc., 2020.URLhttps://proceedings.neurips.cc/paper_files/paper/2020/file/1457c0d6bfcb4967418bfb8ac142f64a-Paper.pdf.
- Caruana (1997)Rich Caruana.Multitask learning.Machine Learning, 28:41–75, 1997.URL https://api.semanticscholar.org/CorpusID:45998148.
- Chronopoulou etal. (2023)Alexandra Chronopoulou, Jonas Pfeiffer, Joshua Maynez, Xinyi Wang, SebastianRuder, and Priyanka Agrawal.Language and task arithmetic with parameter-efficient layers forzero-shot summarization, 2023.
- Google etal. (2023)Google, Rohan Anil, AndrewM. Dai, Orhan Firat, Melvin Johnson, DmitryLepikhin, Alexandre Passos, Siamak Shakeri, Emanuel Taropa, Paige Bailey,Zhifeng Chen, Eric Chu, JonathanH. Clark, LaurentEl Shafey, Yanping Huang,Kathy Meier-Hellstern, Gaurav Mishra, Erica Moreira, Mark Omernick, KevinRobinson, Sebastian Ruder, YiTay, Kefan Xiao, Yuanzhong Xu, Yujing Zhang,GustavoHernandez Abrego, Junwhan Ahn, Jacob Austin, Paul Barham, Jan Botha,James Bradbury, Siddhartha Brahma, Kevin Brooks, Michele Catasta, Yong Cheng,Colin Cherry, ChristopherA. Choquette-Choo, Aakanksha Chowdhery, ClémentCrepy, Shachi Dave, Mostafa Dehghani, Sunipa Dev, Jacob Devlin, Mark Díaz,Nan Du, Ethan Dyer, Vlad Feinberg, Fangxiaoyu Feng, Vlad Fienber, MarkusFreitag, Xavier Garcia, Sebastian Gehrmann, Lucas Gonzalez, Guy Gur-Ari,Steven Hand, Hadi Hashemi, LeHou, Joshua Howland, Andrea Hu, Jeffrey Hui,Jeremy Hurwitz, Michael Isard, Abe Ittycheriah, Matthew Jagielski, WenhaoJia, Kathleen Kenealy, Maxim Krikun, Sneha Kudugunta, Chang Lan, KatherineLee, Benjamin Lee, Eric Li, Music Li, Wei Li, YaGuang Li, Jian Li, HyeontaekLim, Hanzhao Lin, Zhongtao Liu, Frederick Liu, Marcello Maggioni, AromaMahendru, Joshua Maynez, Vedant Misra, Maysam Moussalem, Zachary Nado, JohnNham, Eric Ni, Andrew Nystrom, Alicia Parrish, Marie Pellat, Martin Polacek,Alex Polozov, Reiner Pope, Siyuan Qiao, Emily Reif, Bryan Richter, ParkerRiley, AlexCastro Ros, Aurko Roy, Brennan Saeta, Rajkumar Samuel, ReneeShelby, Ambrose Slone, Daniel Smilkov, DavidR. So, Daniel Sohn, SimonTokumine, Dasha Valter, Vijay Vasudevan, Kiran Vodrahalli, Xuezhi Wang,Pidong Wang, Zirui Wang, Tao Wang, John Wieting, Yuhuai Wu, Kelvin Xu, YunhanXu, Linting Xue, Pengcheng Yin, Jiahui Yu, Qiao Zhang, Steven Zheng,CeZheng, Weikang Zhou, Denny Zhou, Slav Petrov, and Yonghui Wu.Palm 2 technical report, 2023.
- Hu etal. (2022)EdwardJ Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, SheanWang, LuWang, and Weizhu Chen.LoRA: Low-rank adaptation of large language models.In International Conference on Learning Representations, 2022.URL https://openreview.net/forum?id=nZeVKeeFYf9.
- Huang etal. (2023)Chengsong Huang, Qian Liu, BillYuchen Lin, Tianyu Pang, Chao Du, and Min Lin.Lorahub: Efficient cross-task generalization via dynamic loracomposition.arXiv preprint arXiv:2307.13269, 2023.
- Ilharco etal. (2022)Gabriel Ilharco, MarcoTulio Ribeiro, Mitchell Wortsman, Suchin Gururangan,Ludwig Schmidt, Hannaneh Hajishirzi, and Ali Farhadi.Editing models with task arithmetic.arXiv preprint arXiv:2212.04089, 2022.
- Jawahar etal. (2023)Ganesh Jawahar, Subhabrata Mukherjee, Xiaodong Liu, YoungJin Kim, MuhammadAbdul-Mageed, Laks Lakshmanan, V.S., AhmedHassan Awadallah, SebastienBubeck, and Jianfeng Gao.AutoMoE: Heterogeneous mixture-of-experts with adaptivecomputation for efficient neural machine translation.In Anna Rogers, Jordan Boyd-Graber, and Naoaki Okazaki (eds.),Findings of the Association for Computational Linguistics: ACL 2023,pp. 9116–9132, Toronto, Canada, July 2023. Association for ComputationalLinguistics.doi: 10.18653/v1/2023.findings-acl.580.URL https://aclanthology.org/2023.findings-acl.580.
- Kudugunta etal. (2021)Sneha Kudugunta, Yanping Huang, Ankur Bapna, Maxim Krikun, Dmitry Lepikhin,Minh-Thang Luong, and Orhan Firat.Beyond distillation: Task-level mixture-of-experts for efficientinference.In Marie-Francine Moens, Xuanjing Huang, Lucia Specia, and ScottWen-tau Yih (eds.), Findings of the Association for ComputationalLinguistics: EMNLP 2021, pp. 3577–3599, Punta Cana, Dominican Republic,November 2021. Association for Computational Linguistics.doi: 10.18653/v1/2021.findings-emnlp.304.URL https://aclanthology.org/2021.findings-emnlp.304.
- Lepikhin etal. (2020)Dmitry Lepikhin, HyoukJoong Lee, Yuanzhong Xu, Dehao Chen, Orhan Firat, YanpingHuang, Maxim Krikun, Noam Shazeer, and Zhifeng Chen.Gshard: Scaling giant models with conditional computation andautomatic sharding, 2020.
- Lester etal. (2021)Brian Lester, Rami Al-Rfou, and Noah Constant.The power of scale for parameter-efficient prompt tuning.In Marie-Francine Moens, Xuanjing Huang, Lucia Specia, and ScottWen-tau Yih (eds.), Proceedings of the 2021 Conference on EmpiricalMethods in Natural Language Processing, Online and Punta Cana, DominicanRepublic, November 2021. Association for Computational Linguistics.URL https://aclanthology.org/2021.emnlp-main.243.
- Lin etal. (2021)Zehui Lin, Liwei Wu, Mingxuan Wang, and Lei Li.Learning language specific sub-network for multilingual machinetranslation.In Chengqing Zong, Fei Xia, Wenjie Li, and Roberto Navigli (eds.),ACL, Online, August 2021. Association for Computational Linguistics.
- Ouyang etal. (2022)Long Ouyang, Jeff Wu, XuJiang, Diogo Almeida, CarrollL. Wainwright, PamelaMishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, JohnSchulman, Jacob Hilton, Fraser Kelton, LukeE. Miller, Maddie Simens, AmandaAskell, Peter Welinder, PaulFrancis Christiano, Jan Leike, and RyanJ. Lowe.Training language models to follow instructions with human feedback.ArXiv, abs/2203.02155, 2022.URL https://api.semanticscholar.org/CorpusID:246426909.
- Pfeiffer etal. (2020)Jonas Pfeiffer, Ivan Vulić, Iryna Gurevych, and Sebastian Ruder.MAD-X: An Adapter-Based Framework for Multi-TaskCross-Lingual Transfer.In Bonnie Webber, Trevor Cohn, Yulan He, and Yang Liu (eds.),Proceedings of the 2020 Conference on Empirical Methods in NaturalLanguage Processing (EMNLP), pp. 7654–7673, Online, November 2020.Association for Computational Linguistics.doi: 10.18653/v1/2020.emnlp-main.617.URL https://aclanthology.org/2020.emnlp-main.617.
- Pfeiffer etal. (2022)Jonas Pfeiffer, Naman Goyal, XiLin, Xian Li, James Cross, Sebastian Riedel,and Mikel Artetxe.Lifting the curse of multilinguality by pre-training modulartransformers.In Marine Carpuat, Marie-Catherine deMarneffe, and IvanVladimirMezaRuiz (eds.), Proceedings of the 2022 Conference of the NorthAmerican Chapter of the Association for Computational Linguistics: HumanLanguage Technologies, pp. 3479–3495, Seattle, United States, July 2022.Association for Computational Linguistics.doi: 10.18653/v1/2022.naacl-main.255.URL https://aclanthology.org/2022.naacl-main.255.
- Pfeiffer etal. (2023)Jonas Pfeiffer, Francesco Piccinno, Massimo Nicosia, Xinyi Wang, Machel Reid,and Sebastian Ruder.mmt5: Modular multilingual pre-training solves source languagehallucinations, 2023.
- Ruder etal. (2023)Sebastian Ruder, JonathanH. Clark, Alexander Gutkin, Mihir Kale, Min Ma,Massimo Nicosia, Shruti Rijhwani, Parker Riley, Jean-MichelA. Sarr, XinyiWang, John Wieting, Nitish Gupta, Anna Katanova, Christo Kirov, DanaL.Dickinson, Brian Roark, Bidisha Samanta, Connie Tao, DavidI. Adelani, VeraAxelrod, Isaac Caswell, Colin Cherry, Dan Garrette, Reeve Ingle, MelvinJohnson, Dmitry Panteleev, and Partha Talukdar.Xtreme-up: A user-centric scarce-data benchmark for under-representedlanguages, 2023.
- Shazeer etal. (2017)Noam Shazeer, Azalia Mirhoseini, Krzysztof Maziarz, Andy Davis, Quoc Le,Geoffrey Hinton, and Jeff Dean.Outrageously large neural networks: The sparsely-gatedmixture-of-experts layer, 2017.
- Soltan etal. (2022)Saleh Soltan, Shankar Ananthakrishnan, Jack G.M. FitzGerald, Rahul Gupta, WaelHamza, Haidar Khan, CharithS. Peris, Stephen Rawls, Andrew Rosenbaum, AnnaRumshisky, Chandan Prakash, Mukund Sridhar, Fabian Triefenbach, Apurv Verma,Gokhan Tur, and Premkumar Natarajan.Alexatm 20b: Few-shot learning using a large-scale multilingualseq2seq model.ArXiv, abs/2208.01448, 2022.URL https://api.semanticscholar.org/CorpusID:251253416.
- Vu etal. (2022)TuVu, Aditya Barua, Brian Lester, Daniel Cer, Mohit Iyyer, and Noah Constant.Overcoming catastrophic forgetting in zero-shot cross-lingualgeneration.In EMNLP, 2022.
- Wang etal. (2021)Xinyi Wang, Yulia Tsvetkov, Sebastian Ruder, and Graham Neubig.Efficient test time adapter ensembling for low-resource languagevarieties.In Marie-Francine Moens, Xuanjing Huang, Lucia Specia, and ScottWen-tau Yih (eds.), EMNLP, Punta Cana, Dominican Republic, November2021.
- Wang etal. (2023a)Xinyi Wang, John Wieting, and JonathanH. Clark.Fiat: Fusing learning paradigms with instruction-accelerated tuning,2023a.
- Wang etal. (2023b)Zhen Wang, Rameswar Panda, Leonid Karlinsky, Rogerio Feris, Huan Sun, and YoonKim.Multitask prompt tuning enables parameter-efficient transferlearning.In ICLR, 2023b.
- Wang etal. (2018)Zirui Wang, Zihang Dai, Barnabás Póczos, and JaimeG. Carbonell.Characterizing and avoiding negative transfer.2019 IEEE/CVF Conference on Computer Vision and PatternRecognition (CVPR), pp. 11285–11294, 2018.URL https://api.semanticscholar.org/CorpusID:53748459.
- Zadouri etal. (2023)Ted Zadouri, Ahmet Üstün, Arash Ahmadian, Beyza Ermiş, Acyr Locatelli, andSara Hooker.Pushing mixture of experts to the limit: Extremely parameterefficient moe for instruction tuning, 2023.
- Zhang etal. (2020)Biao Zhang, Ankur Bapna, Rico Sennrich, and Orhan Firat.Share or not? learning to schedule language-specific capacity formultilingual translation.In ICLR, 2020.
- Zhu etal. (2023)Yun Zhu, Nevan Wichers, Chu-Cheng Lin, Xinyi Wang, Tianlong Chen, Lei Shu, HanLu, Canoee Liu, Liangchen Luo, Jindong Chen, etal.Sira: Sparse mixture of low rank adaptation.arXiv preprint arXiv:2311.09179, 2023.
Appendix A Constant parameter sharing hurts FLix
Configuration | QA InLang | QA CrossLang | Semantic parsing | NER |
FLix | ||||
+ shared parameters | ||||
Compute-matched LoRA |
Our proposed FLix routes each input to their feature-specific adaptations, based on the dataset features.While constant parameter sharing happens when some features are always active for all datasets in a mixture (e.g., in experiments in §5.2), this is not generally true.In this section, we look into the effects of enforcing constant parameter-sharing in FLix. Specifically, we design a dummy feature that is always active to all datasets in the joint multitask multilingual training mixture used in §5.1.1, which implies constant parameter sharing.We then compare FLix models both without and with this dummy feature, and also with a vanilla LoRA baseline, all under comparable compute budgets.777Specifically, we rearrange the rank sizes of all features such that the sum is . The dummy feature has rank, and the language and task features have rank. The compute-matched LoRA baseline does not make use of dataset features, and has rank.
In Table5, we can see that adding a shared parameter actually leads to worse performance for FLix with similar computational cost. However, FLix with constantly shared features still outperforms computed-matched LoRA, which does not leverage task or language features at all.
Appendix B Training and evaluation datasets
B.1 Multilingual learning
The (mono-task) multilingual learning experiments described in §5.2 train and evaluate on the same languages. Their language and locale codes are:
- Semantic parsing
am, be, bn, fi, ha, hu, ja, pt_br, ru, sw, ta, tr, yo, zu, de_localized, en, de, es, fr, hi, th
- In-language QA
ar, bn, fi, id, ko, ru, sw, te, en
- Cross-lingual QA
ar, bn, fi, ko, ru, te, as, bho, brx, gbm, gom, gu, hi, hne, kn, mai, ml, mni, mr, mwr, or, pa, ps, sa, ta, ur
- NER
am, bm, bbj, ee, ha, ig, rw, lg, luo, mos, ny, pcm, sn, sw, tn, tw, wo, xh, yo, zu
B.2 Multitask learning
The (mono-lingual) multitask learning experiments described in §5.2 train and evaluate on different tasks, for the languages we evaluate on respectively. They are:
- Swahili (sw)
Semantic parsing, NER, in-language QA
- Bengali (bn)
Semantic parsing, in-language QA, cross-lingual QA
B.3 Joint multitask multilingual learning
B.3.1 Training
The joint multitask multilingual learning experiments in §5.1 use the union of the following datasets:
- Semantic parsing
am, be, bn, fi, ha, hu, ja, pt_br, ru, sw, ta, tr, yo, zu, de_localized, en, de, es, fr, hi, th
- In-language QA
ar, bn, fi, id, ko, ru, sw, te, en
- Cross-lingual QA
ar, bn, fi, ko, ru, te
- NER
am, bm, bbj, ee, ha, ig, rw, lg, luo, mos, ny, pcm, sn, sw, tn, tw, wo, xh, yo, zu
- Machine translation
id, es, hi, yo, ja, lg, ny, ru, be, ar, de, bn, fr, tr, ig, th, fi, zu, te, ko, sw, xh, hu, ha, sn, ta, am
B.3.2 Multilingual evaluation
We evaluate models trained on the dataset described in §B.3.1 on the multilingual subsets of the training dataset. The results are reported in the first columns of Table1.
B.4 Zero-shot generalization
B.4.1 Training
For zero-shot evaluations on cross-lingual QA datasets in §5.1.2, we reuse the joint multitask multilingual training dataset described in §B.3.1.For evaluation of zero-shot unseen combinations on semantic parsing datasets in the same section, we use the union of the following datasets:
- Semantic parsing
fi, hu, ja, pt_br, ru, tr, de_localized, en, de, es, fr, hi
- In-language QA
ar, bn, fi, id, ko, ru, sw, te, en
- Cross-lingual QA
ar, bn, fi, ko, ru, te, as, bho, brx, gbm, gom, gu, hi, hne, kn, mai, ml, mni, mr, mwr, or, pa, ps, sa, ta, ur
- NER
am, bm, bbj, ee, ha, ig, rw, lg, luo, mos, ny, pcm, sn, sw, tn, tw, wo, xh, yo, zu
- Machine translation
de, mni, mr, ig, id, mwr, ko, fi, ta, bn, ar, es, ja, zu, be, gbm, tr, as, bho, yo, ml, sw, hi, am, bbj, lg, pcm, ny, tw, hu, luo, rw, brx, pt_br, gu, or, gom, te, sn, wo, fr, ps, ha, hne, xh, en, sa, tn, de_localized, ur, bm, ee, th, ru, pa, kn, mai, mos
B.4.2 Zero-shot unseen combinations in cross-lingual QA
In the ‘Zero-shot Unseen Comb: QA CrossLang’ column of Table1 we report the average cross-lingual QA F1 scores of languages that are missing from the cross-lingual QA portion of the joint multitask multilingual training set (§B.3.1), but present in other datasets of the same training set. The language and locale codes of these languages are: hi, ta.
B.4.3 Zero-shot unseen languages in cross-lingual QA
In the ‘Zero-shot Unseen Lang’ column of Table1 we report the average cross-lingual QA F1 scores of languages that are missing from the joint multitask multilingual training set. The language and locale codes of these languages are: bho, brx, gbm, gom, hne, mai, mni, mr, mwr, sa, as, gu, kn, ml, or, pa, ps, ur.
B.4.4 Zero-shot unseen combinations in semantic parsing
In the ‘Zero-shot Unseen Comb: Semantic parsing’ column of Table1 we report the average exact match accuracy scores of languages that are designated as underrepresented in the XTREME-UP dataset, and held out from the training set in this experiment. These languages however appear in the machine translation portion of the training dataset. The language and locale codes of these languages are: am, be, bn, ha, sw, ta, yo, zu, th.
Appendix C Full results of Table2
In addition to the test results reported in Table2, we include validation results as well.
C.1 Multilingual experiments
C.1.1 Cross-lingual QA
Language / Locale ID | Validation | Test |
ar | ||
bn | ||
fi | ||
ko | ||
ru | ||
te | ||
as | ||
bho | ||
brx | ||
gbm | ||
gom | ||
gu | ||
hi | ||
hne | ||
kn | ||
mai | ||
ml | ||
mni | ||
mr | ||
mwr | ||
or | ||
pa | ||
ps | ||
sa | ||
ta | ||
ur | ||
Average |
Language / Locale ID | Validation | Test |
ar | ||
bn | ||
fi | ||
ko | ||
ru | ||
te | ||
as | ||
bho | ||
brx | ||
gbm | ||
gom | ||
gu | ||
hi | ||
hne | ||
kn | ||
mai | ||
ml | ||
mni | ||
mr | ||
mwr | ||
or | ||
pa | ||
ps | ||
sa | ||
ta | ||
ur | ||
Average |
Language / Locale ID | Validation | Test |
ar | ||
bn | ||
fi | ||
ko | ||
ru | ||
te | ||
as | ||
bho | ||
brx | ||
gbm | ||
gom | ||
gu | ||
hi | ||
hne | ||
kn | ||
mai | ||
ml | ||
mni | ||
mr | ||
mwr | ||
or | ||
pa | ||
ps | ||
sa | ||
ta | ||
ur | ||
Average |
Language / Locale ID | Validation | Test |
ar | ||
bn | ||
fi | ||
ko | ||
ru | ||
te | ||
as | ||
bho | ||
brx | ||
gbm | ||
gom | ||
gu | ||
hi | ||
hne | ||
kn | ||
mai | ||
ml | ||
mni | ||
mr | ||
mwr | ||
or | ||
pa | ||
ps | ||
sa | ||
ta | ||
ur | ||
Average |
C.1.2 In-language QA
Language / Locale ID | Validation | Test |
ar | ||
bn | ||
fi | ||
id | ||
ko | ||
ru | ||
sw | ||
te | ||
en | ||
Average |
Language / Locale ID | Validation | Test |
ar | ||
bn | ||
fi | ||
id | ||
ko | ||
ru | ||
sw | ||
te | ||
en | ||
Average |
Language / Locale ID | Validation | Test |
ar | ||
bn | ||
fi | ||
id | ||
ko | ||
ru | ||
sw | ||
te | ||
en | ||
Average |
Language / Locale ID | Validation | Test |
ar | ||
bn | ||
fi | ||
id | ||
ko | ||
ru | ||
sw | ||
te | ||
en | ||
Average |
C.1.3 Semantic parsing
Language / Locale ID | Validation | Test |
am | ||
be | ||
bn | ||
fi | ||
ha | ||
hu | ||
ja | ||
pt-br | ||
ru | ||
sw | ||
ta | ||
tr | ||
yo | ||
zu | ||
de-localized | ||
en | ||
de | ||
es | ||
fr | ||
hi | ||
th | ||
Average |
Language / Locale ID | Validation | Test |
am | ||
be | ||
bn | ||
fi | ||
ha | ||
hu | ||
ja | ||
pt-br | ||
ru | ||
sw | ||
ta | ||
tr | ||
yo | ||
zu | ||
de-localized | ||
en | ||
de | ||
es | ||
fr | ||
hi | ||
th | ||
Average |
Language / Locale ID | Validation | Test |
am | ||
be | ||
bn | ||
fi | ||
ha | ||
hu | ||
ja | ||
pt-br | ||
ru | ||
sw | ||
ta | ||
tr | ||
yo | ||
zu | ||
de-localized | ||
en | ||
de | ||
es | ||
fr | ||
hi | ||
th | ||
Average |
Language / Locale ID | Validation | Test |
am | ||
be | ||
bn | ||
fi | ||
ha | ||
hu | ||
ja | ||
pt-br | ||
ru | ||
sw | ||
ta | ||
tr | ||
yo | ||
zu | ||
de-localized | ||
en | ||
de | ||
es | ||
fr | ||
hi | ||
th | ||
Average |
C.1.4 Named-entity recognition (NER)
Language / Locale ID | Validation | Test |
am | ||
bm | ||
bbj | ||
ee | ||
ha | ||
ig | ||
rw | ||
lg | ||
luo | ||
mos | ||
ny | ||
pcm | ||
sn | ||
sw | ||
tn | ||
tw | ||
wo | ||
xh | ||
yo | ||
zu | ||
Average |
Language / Locale ID | Validation | Test |
am | ||
bm | ||
bbj | ||
ee | ||
ha | ||
ig | ||
rw | ||
lg | ||
luo | ||
mos | ||
ny | ||
pcm | ||
sn | ||
sw | ||
tn | ||
tw | ||
wo | ||
xh | ||
yo | ||
zu | ||
Average |
Language / Locale ID | Validation | Test |
am | ||
bm | ||
bbj | ||
ee | ||
ha | ||
ig | ||
rw | ||
lg | ||
luo | ||
mos | ||
ny | ||
pcm | ||
sn | ||
sw | ||
tn | ||
tw | ||
wo | ||
xh | ||
yo | ||
zu | ||
Average |
Language / Locale ID | Validation | Test |
am | ||
bm | ||
bbj | ||
ee | ||
ha | ||
ig | ||
rw | ||
lg | ||
luo | ||
mos | ||
ny | ||
pcm | ||
sn | ||
sw | ||
tn | ||
tw | ||
wo | ||
xh | ||
yo | ||
zu | ||
Average |
C.2 Multitask experiments
Task | Validation | Test |
SemParse | ||
QA-InLang | ||
NER | ||
Average |
Task | Validation | Test |
SemParse | ||
QA-InLang | ||
NER | ||
Average |
Task | Validation | Test |
SemParse | ||
QA-InLang | ||
NER | ||
Average |
Task | Validation | Test |
SemParse | ||
QA-InLang | ||
NER | ||
Average |
Task | Validation | Test |
SemParse | ||
QA-InLang | ||
QA-CrossLang | ||
Average |
Task | Validation | Test |
SemParse | ||
QA-InLang | ||
QA-CrossLang | ||
Average |
Task | Validation | Test |
SemParse | ||
QA-InLang | ||
QA-CrossLang | ||
Average |
Task | Validation | Test |
SemParse | ||
QA-InLang | ||
QA-CrossLang | ||
Average |
Appendix D Full results of Table1
As in AppendixC, we include validation results along with test results.
D.1 Supervised learning results
D.1.1 Cross-lingual QA
Language / Locale ID | Validation | Test |
ar | ||
bn | ||
fi | ||
ko | ||
ru | ||
te | ||
Average |
Language / Locale ID | Validation | Test |
ar | ||
bn | ||
fi | ||
ko | ||
ru | ||
te | ||
Average |
Language / Locale ID | Validation | Test |
ar | ||
bn | ||
fi | ||
ko | ||
ru | ||
te | ||
Average |
D.1.2 In-language QA
Language / Locale ID | Validation | Test |
ar | ||
bn | ||
fi | ||
id | ||
ko | ||
ru | ||
sw | ||
te | ||
en | ||
Average |
Language / Locale ID | Validation | Test |
ar | ||
bn | ||
fi | ||
id | ||
ko | ||
ru | ||
sw | ||
te | ||
en | ||
Average |
Language / Locale ID | Validation | Test |
ar | ||
bn | ||
fi | ||
id | ||
ko | ||
ru | ||
sw | ||
te | ||
en | ||
Average |
D.1.3 Semantic parsing
Language / Locale ID | Validation | Test |
am | ||
be | ||
bn | ||
fi | ||
ha | ||
hu | ||
ja | ||
pt-br | ||
ru | ||
sw | ||
ta | ||
tr | ||
yo | ||
zu | ||
de-localized | ||
en | ||
de | ||
es | ||
fr | ||
hi | ||
th | ||
Average |
Language / Locale ID | Validation | Test |
am | ||
be | ||
bn | ||
fi | ||
ha | ||
hu | ||
ja | ||
pt-br | ||
ru | ||
sw | ||
ta | ||
tr | ||
yo | ||
zu | ||
de-localized | ||
en | ||
de | ||
es | ||
fr | ||
hi | ||
th | ||
Average |
Language / Locale ID | Validation | Test |
am | ||
be | ||
bn | ||
fi | ||
ha | ||
hu | ||
ja | ||
pt-br | ||
ru | ||
sw | ||
ta | ||
tr | ||
yo | ||
zu | ||
de-localized | ||
en | ||
de | ||
es | ||
fr | ||
hi | ||
th | ||
Average |
D.1.4 NER
Language / Locale ID | Validation | Test |
am | ||
bm | ||
bbj | ||
ee | ||
ha | ||
ig | ||
rw | ||
lg | ||
luo | ||
mos | ||
ny | ||
pcm | ||
sn | ||
sw | ||
tn | ||
tw | ||
wo | ||
xh | ||
yo | ||
zu | ||
Average |
Language / Locale ID | Validation | Test |
am | ||
bm | ||
bbj | ||
ee | ||
ha | ||
ig | ||
rw | ||
lg | ||
luo | ||
mos | ||
ny | ||
pcm | ||
sn | ||
sw | ||
tn | ||
tw | ||
wo | ||
xh | ||
yo | ||
zu | ||
Average |
Language / Locale ID | Validation | Test |
am | ||
bm | ||
bbj | ||
ee | ||
ha | ||
ig | ||
rw | ||
lg | ||
luo | ||
mos | ||
ny | ||
pcm | ||
sn | ||
sw | ||
tn | ||
tw | ||
wo | ||
xh | ||
yo | ||
zu | ||
Average |
D.2 Zero-shot results
D.2.1 Zero-shot unseen combinations: cross-lingual QA
Language / Locale ID | Validation | Test |
hi | ||
ta | ||
Average |
Language / Locale ID | Validation | Test |
hi | ||
ta | ||
Average |
Language / Locale ID | Validation | Test |
hi | ||
ta | ||
Average |
D.2.2 Zero-shot unseen combinations: semantic parsing
Language / Locale ID | Validation | Test |
am | ||
be | ||
bn | ||
ha | ||
sw | ||
ta | ||
yo | ||
zu | ||
th | ||
Average |
Language / Locale ID | Validation | Test |
am | ||
be | ||
bn | ||
ha | ||
sw | ||
ta | ||
yo | ||
zu | ||
th | ||
average |
Language / Locale ID | Validation | Test |
am | ||
be | ||
bn | ||
ha | ||
sw | ||
ta | ||
yo | ||
zu | ||
th | ||
Average |
D.2.3 Zero-shot unseen languages
Language / Locale ID | Validation | Test |
as | ||
bho | ||
brx | ||
gbm | ||
gom | ||
gu | ||
hne | ||
kn | ||
mai | ||
ml | ||
mni | ||
mr | ||
mwr | ||
or | ||
pa | ||
ps | ||
sa | ||
ur | ||
Average |
Language / Locale ID | Validation | Test |
as | ||
bho | ||
brx | ||
gbm | ||
gom | ||
gu | ||
hne | ||
kn | ||
mai | ||
ml | ||
mni | ||
mr | ||
mwr | ||
or | ||
pa | ||
ps | ||
sa | ||
ur | ||
Average |
Language / Locale ID | Validation | Test |
as | ||
bho | ||
brx | ||
gbm | ||
gom | ||
gu | ||
hne | ||
kn | ||
mai | ||
ml | ||
mni | ||
mr | ||
mwr | ||
or | ||
pa | ||
ps | ||
sa | ||
ur | ||
Average |