AI21 Jamba 1.5 Large
AI21 Jamba 1.5 Large
Version: 1
AI21 LabsLast updated October 2024
A 398B parameters (94B active) multilingual model, offering a 256K long context window, function calling, structured output, and grounded generation.
RAG
Multilingual
Large context
Jamba 1.5 Large is a state-of-the-art, hybrid SSM-Transformer instruction following foundation model. It's a Mixture-of-Expert model with 94B total parameters and 398B active parameters. The Jamba family of models are the most powerful & efficient long-context models on the market, offering a 256K context window, the longest available.. For long context input, they deliver up to 2.5X faster inference than leading models of comparable sizes. Jamba supports function calling/tool use, structured output (JSON), and grounded generation with citation mode and documents API. Jamba officially supports English, French, Spanish, Portuguese, German, Arabic and Hebrew, but can also work in many other languages. Model Developer Name: Jamba 1.5 Large

Model Architecture

Jamba 1.5 Large is a state-of-the-art, hybrid SSM-Transformer instruction following foundation model

Model Variations

94B total parameters and 398B active parameters

Model Input

Models input text only.

Model Output

Models generate text only.

Model Dates

Jamba 1.5 Large was trained in Q3 2024 with data covering through early March 2024.

Model Information Table

NameParamsContent Length
Jamba 1.5 Mini52B (12B active)256K
Jamba 1.5 Large398B (94B active)256K

License

Please refer to this link for AI21's Terms of Use For more information about how Azure processes data, for privacy and security in relation to Models-as-a-Service (MaaS), please see this Microsoft Docs link .

Ethical Considerations and Limitations

AI21 Labs is on a mission to make AI-first experiences, with machines working alongside humans as thought partners, thereby promoting human welfare and prosperity. To deliver its promise, this technology must be deployed and used in a responsible and sustainable way, taking into consideration potential risks, including malicious use by bad actors, accidental misuse and broader societal harms. We take these risks extremely seriously and put measures in place to mitigate them. AI safety is an important challenge with a large surface area, which we believe can be addressed most effectively by working together. We invite anyone interested in conducting research or otherwise promoting AI safety to contact us at safety@ai21.com and explore opportunities for collaboration.

Content Filtering

Prompts and completions in GitHub Models are passed through a default configuration of Azure AI Content Safety classification models to detect and prevent the output of harmful content. Learn more about Azure AI Content Safety . Configuration options for content filtering vary when you deploy a model for production in Azure AI; learn more .

Training Data

Jamba is trained on an in-house dataset that contains text data from the web, books, and code. The knowledge cutoff date is March 5, 2024.

Evaluation Results

CategoryMetricScore
GeneralArena Hard65.4
MMLU (CoT)81.2
MMLU Pro (CoT)53.5
IFEval81.5
BBH65.5
WildBench48.4
ReasoningARC-C93
GPQA36.9
Math, Code & Tool useGSM8K87
HumanEval71.3
BFCL85.5

Evaluation of pretrained LLMs on automatic safety benchmarks*

TruthfulQA
Jamba 1.5 Mini54.1
Jamba 1.5 Large58.3

Evaluation of fine-tuned LLMs on different safety datasets*

RealToxicity*
Jamba 1.5 Mini8.1
Jamba 1.5 Large6.7
* Lower score is better
Model Specifications
Context Length262144
LicenseCustom
Last UpdatedOctober 2024
Input TypeText
Output TypeText
PublisherAI21 Labs
Languages7 Languages