Jais is the worldβs leading Arabic large language model, developed by Inception (a G42 company), Cerebras Systems, and MBZUAI in the United Arab Emirates. The latest version, Jais 2, is a 70B parameter model trained on the largest Arabic-first dataset ever assembled.
Why Jais matters
Most LLMs are trained primarily on English data. Arabic speakers (400M+ people) get second-class AI that struggles with Arabic grammar, dialects, and cultural context. Jais was built from the ground up for Arabic, with English as a strong secondary language.
Jais 2 specs:
| Spec | Value |
|---|---|
| Parameters | 70B |
| Training data | 126B Arabic + 251B English + 50B code tokens |
| Languages | Arabic (MSA + dialects) + English |
| License | Open weights |
| Inference speed | 2,000 tokens/second on Cerebras hardware |
| Architecture | GPT-3 decoder-only |
Sovereign AI
Jais represents a growing trend: countries building their own AI models instead of depending on US/Chinese providers. The UAE invested in Jais to ensure Arabic-speaking populations have AI that understands their language and culture natively.
This is the same motivation behind:
- GLM-5.1 (China, trained on Huawei chips)
- Mistral (France/EU, European AI sovereignty)
- Aleph Alpha (Germany, enterprise-focused)
For developers building Arabic-language applications, Jais is the obvious choice over general-purpose models that treat Arabic as an afterthought.
How to use Jais
Jais models are available on HuggingFace:
# Via Ollama (if available)
ollama pull jais:13b
# Via HuggingFace Transformers
pip install transformers
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("inceptionai/jais-13b")
tokenizer = AutoTokenizer.from_pretrained("inceptionai/jais-13b")
Jais vs multilingual models for Arabic
| Model | Arabic quality | English quality | Open weights |
|---|---|---|---|
| Jais 2 70B | β Best | Good | β |
| GPT-5 | Good | Best | β |
| Qwen 3.5 | Decent | Excellent | β |
| Llama 4 | Decent | Excellent | β |
| Claude | Good | Best | β |
For Arabic-first applications, Jais outperforms every general-purpose model. For English-first with some Arabic, a multilingual model like Qwen or Llama is fine.
The sovereign AI movement
Jais is part of a global trend where countries invest in their own AI models instead of depending on US or Chinese providers:
| Country/Region | Model | Lab | Why |
|---|---|---|---|
| UAE | Jais 2 | G42/MBZUAI | Arabic language sovereignty |
| China | GLM-5.1 | Zhipu AI | Trained on Huawei chips, no NVIDIA |
| France/EU | Mistral | Mistral AI | European AI independence |
| China | Qwen | Alibaba | Largest open model ecosystem |
| China | DeepSeek | DeepSeek | Open reasoning models |
| China | Yi | 01.AI | Apache 2.0 bilingual models |
| Germany | Aleph Alpha | Aleph Alpha | Enterprise, anti-hallucination |
For developers, this means more choice, more competition, and lower prices. The days of depending on OpenAI or Google alone are over.
Jais for developers
If youβre building applications for Arabic-speaking users:
- Customer support chatbots β Jais understands Arabic dialects (Egyptian, Gulf, Levantine) that GPT struggles with
- Content generation β marketing copy, social media, documentation in natural Arabic
- Code documentation β generate Arabic docs for Arabic-speaking dev teams
- Translation β Arabic-English translation with cultural context
Falcon: the other UAE model
Worth mentioning alongside Jais is Falcon, also from the UAE (Technology Innovation Institute). Falcon 2 is a general-purpose open model, while Jais is Arabic-specialized. If you need Arabic, use Jais. If you need a general-purpose model from the UAE ecosystem, consider Falcon.
Getting started
The fastest way to try Jais:
- Visit HuggingFace for model downloads
- Use the 13B model for testing (fits on 16GB RAM)
- For production Arabic applications, use the Jais 2 70B via API
The Arabic AI space is growing fast. Jais is the current leader, but expect more competition as other Middle Eastern and North African countries invest in their own models.
FAQ
Does Jais only work for Arabic?
No. Jais is bilingual β it handles both Arabic and English well. It was trained on 126B Arabic tokens and 251B English tokens, so English is a strong secondary language. However, its primary advantage over general-purpose models is Arabic quality, including understanding of dialects (Egyptian, Gulf, Levantine) that models like GPT and Claude struggle with.
Can I use Jais for free?
Yes. Jais models are available on HuggingFace with open weights. The 13B model fits on 16GB RAM and is suitable for testing and smaller applications. For production Arabic applications, the Jais 2 70B model is recommended but requires more substantial hardware or API access.
How does Jais compare to GPT-5 for Arabic tasks?
For Arabic-first applications (customer support, content generation, translation with cultural context), Jais significantly outperforms GPT-5 because it was trained specifically on Arabic data including dialects and cultural nuances. For English-first applications with occasional Arabic, GPT-5 or Qwen 3.5 may be sufficient since they handle basic Arabic adequately.
Related: What is Yi? Β· GLM-5.1 Complete Guide Β· What is Mistral AI? Β· Best Open Source Coding Models Β· Yi vs Qwen vs DeepSeek Β· Self-Hosted AI for Enterprise