How Many Languages Make Good Multilingual Instruction Tuning? A Case Study on BLOOM

Shaoxiong Ji, Pinzhen Chen

Research output: Chapter in Book/Report/Conference proceedingConference contributionScientificpeer-review

Abstract

Instruction tuning a large language model with multiple languages can prepare it for multilingual downstream tasks. Nonetheless, it is yet to be determined whether having a handful of languages is sufficient, or whether the benefits increase with the inclusion of more. By finetuning large multilingual models on 1 to 52 languages, we present a case study on BLOOM to understand three pertinent factors affecting performance: the number of languages, language exposure, and similarity between training and test languages. Overall we found that 1) expanding language coverage in multilingual instruction tuning proves to be beneficial; 2) accuracy often significantly boots if the test language appears in the instruction mixture; 3) languages' genetic features correlate with cross-lingual transfer more than merely the number of language but different languages benefit to various degrees.

Original languageEnglish
Title of host publicationProceedings of the 31st International Conference on Computational Linguistics
EditorsOwen Rambow, Leo Wanner, Marianna Apidianaki, Hend Al-Khalifa, Barbara Di Eugenio, Steven Schockaert
Number of pages7
Place of PublicationStroudsburg
PublisherAssociation for Computational Linguistics (ACL)
Publication date2025
Pages2575-2581
ISBN (Electronic)979-8-89176-196-4
Publication statusPublished - 2025
MoE publication typeA4 Article in conference proceedings
EventInternational Conference on Computational Linguistics - Abu Dhabi, United Arab Emirates
Duration: 19 Jan 202524 Jan 2025
Conference number: 31
https://coling2025.org

Publication series

NameInternational Conference on Computational Linguistics
PublisherAssociation for Computational Linguistics
ISSN (Print)2951-2093

Fields of Science

  • 6121 Languages
  • 113 Computer and information sciences

Cite this