Back to Search Start Over

Jais and Jais-chat: Arabic-Centric Foundation and Instruction-Tuned Open Generative Large Language Models

Authors :
Sengupta, Neha
Sahu, Sunil Kumar
Jia, Bokang
Katipomu, Satheesh
Li, Haonan
Koto, Fajri
Marshall, William
Gosal, Gurpreet
Liu, Cynthia
Chen, Zhiming
Afzal, Osama Mohammed
Kamboj, Samta
Pandit, Onkar
Pal, Rahul
Pradhan, Lalit
Mujahid, Zain Muhammad
Baali, Massa
Han, Xudong
Bsharat, Sondos Mahmoud
Aji, Alham Fikri
Shen, Zhiqiang
Liu, Zhengzhong
Vassilieva, Natalia
Hestness, Joel
Hock, Andy
Feldman, Andrew
Lee, Jonathan
Jackson, Andrew
Ren, Hector Xuguang
Nakov, Preslav
Baldwin, Timothy
Xing, Eric
Publication Year :
2023

Abstract

We introduce Jais and Jais-chat, new state-of-the-art Arabic-centric foundation and instruction-tuned open generative large language models (LLMs). The models are based on the GPT-3 decoder-only architecture and are pretrained on a mixture of Arabic and English texts, including source code in various programming languages. With 13 billion parameters, they demonstrate better knowledge and reasoning capabilities in Arabic than any existing open Arabic and multilingual models by a sizable margin, based on extensive evaluation. Moreover, the models are competitive in English compared to English-centric open models of similar size, despite being trained on much less English data. We provide a detailed description of the training, the tuning, the safety alignment, and the evaluation of the models. We release two open versions of the model -- the foundation Jais model, and an instruction-tuned Jais-chat variant -- with the aim of promoting research on Arabic LLMs. Available at https://huggingface.co/inception-mbzuai/jais-13b-chat<br />Comment: Arabic-centric, foundation model, large-language model, LLM, generative model, instruction-tuned, Jais, Jais-chat

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2308.16149
Document Type :
Working Paper