Jamba

Overview

Introducing the Jamba Family of Open Models.
Built on the novel Mamba-Transformer architecture, these highly efficient and powerful models push the boundaries of AI, delivering unmatched speed and quality, with the longest context window (256K tokens) among open models.

Jamba Large: Our most powerful and advanced model, designed for handling complex tasks at enterprise scale with superior performance.

Jamba Mini: The most efficient model in its size class, optimized for speed and quality with a smaller footprint.


Key Benefits

  • Optimize performance for business-critical tasks
    Jamba 1.6 leads the open model enterprise market in both speed and quality, ensuring your applications run faster without sacrificing accuracy. Its hybrid Mamba-Transformer architecture delivers enterprise-grade performance, even on complex workloads like RAG and real-time decisioning.
  • Ensure better alignment with enterprise privacy and deployment needs
    Jamba can be deployed in your VPC or on-premise, offering full data ownership with no vendor data visibility. This makes it ideal for industries with strict data regulations like finance, healthcare, or legal.
  • Long context performance
    With a 256K context window and hybrid SSM-Transformer architecture, Jamba excels at tasks like grounded question answering and multi-document summarization. Unlike other models that degrade with longer inputs, Jamba maintains high accuracy and recall across massive datasets.

Self Deployment Options

Jamba models are ideal for enterprises that need to maintain full control over data and performance. Deployment options include:

  • Cloud-hosted VPC – Isolate and scale your deployment in a virtual private cloud environment.
  • On-premise environments – Run Jamba entirely within your own infrastructure to meet strict compliance or latency requirements.
  • Custom hybrid solutions – Combine cloud and on-premise capabilities to fit your architecture, performance, and governance needs.

Supported Languages

Jamba models officially support 9 languages:

  • English, Spanish, French, Portuguese, Italian, Dutch, German, Arabic, Hebrew.

Model Details

ModelModel SizeMax TokensVersionSnapshotAPI Endpoint
Jamba Large398B parameters (94B active)256K1.62025-03jamba-large
Jamba Mini52B parameters (12B active)256K1.62025-03jamba-mini

Engineers and data scientists at AI21 labs created the model to help developers and businesses leverage AI to build real-world products with tangible value. Jamba Mini and Jamba Large support zero-shot instruction-following and multi-language support. The Jamba models also provides developers with industry-leading APIs that perform a wide range of productivity tasks designed for commercial use.

  • Organization developing model: AI21 Labs
  • Model date: August 22nd, 2024
  • Model type: Joint Attention and Mamba (Jamba)
  • Knowledge cutoff date March 5th, 2024
  • Input Modality: Text
  • Output Modality: Text
  • License: Jamba open model license
  • Contact: [email protected]

API Versioning

We advise using dated versions of the Jamba API to avoid disruptions from model updates and breaking changes.

Here are the details of the available versions:

  • jamba-largecurrently points to jamba-large-1.6-2025-03
  • jamba-minicurrently points to jamba-mini-1.6-2025-03
  • jamba-large-1.6 points to jamba-large-1.6-2025-03
  • jamba-mini-1.6 points to jamba-mini-1.6-2025-03

Model Deprecation

ModelSnapshotAPI EndpointDeprecation Date
Jamba Large 1.52024-08jamba-large-1.5-2024-082025-05-06
Jamba Mini 1.52024-08jamba-mini-1.5-2024-082025-05-06

Model Compliance and Certifications


Ethical Considerations

AI21 Labs is on a mission to supercharge human productivity with machines working alongside humans as thought partners, thereby promoting human welfare and prosperity. To deliver its promise, this technology must be deployed and used in a responsible and sustainable way, taking into consideration potential risks, including malicious use by bad actors, accidental misuse and broader societal harms. We take these risks extremely seriously and put measures in place to mitigate them.

AI21 provides open access to Jamba that can be used to power a large variety of useful applications. We believe it is important to ensure that this technology is used in a responsible way, while allowing developers the freedom they need to experiment rapidly and deploy solutions at scale. Overall, we view the safe implementation of this technology as a partnership and collaboration between AI21 and our customers and encourage engagement and dialogue to raise the bar on responsible usage.

In order to use Jamba, you are required to comply with our Terms of Service and with the following usage guidelines.

Please check these usage guidelines periodically, as they may be updated from time to time. For any questions, clarifications or concerns, please contact [email protected].


Limitations

There are a number of limitations inherent to neural networks technology that apply to Jamba. These limitations require explanation and carry important caveats for the application and usage of Jamba.

  • Accuracy: Jamba, like other large pretrained language models, lacks important context about the world because it is trained on textual data and is not grounded in other modalities of experience such as video, real-world physical interaction, and human feedback. Like all language models, Jamba is far more accurate when responding to inputs similar to its training datasets. Novel inputs have a tendency to generate higher variance in its output.

  • Coherence and consistency: Responses from Jamba are sometimes inconsistent, contradictory, or contain seemingly random sentences and paragraphs.

  • Western/English bias: Jamba is trained primarily on English language text from the internet, and is best suited to classifying, searching, summarizing, and generating English text. Furthermore, Jamba has a tendency to hold and amplify the biases contained in its training dataset. As a result, groups of people who were not involved in the creation of the training data can be underrepresented, and stereotypes and prejudices can be perpetuated. Racial, religious, gender, socioeconomic, and other categorizations of human groups can be considered among these factors.

  • Explainability: It is difficult to explain or predict how Jamba will respond without additional training and fine tuning. This is a common issue with neural networks of this scope and scale.

  • Recency: Jamba was trained on a dataset created in March 2024, and therefore has no knowledge of events that have occurred after that date. We update our models regularly to keep them as current as possible, but there are notable gaps and inaccuracies in responses as a result of this lack of recency.

Ready to take your projects to the next level?
Access Jamba via the AI21 Studio API or deploy privately.
See our Quick Start Guide for integration steps, SDKs, and sample code.