Jamba

Overview

Introducing the Jamba Family of Open Models.
Built on the novel Mamba-Transformer architecture, these highly efficient and powerful models push the boundaries of AI, delivering unmatched speed and quality, with the longest context window (256K tokens) among open models.

Jamba Large: Our most powerful and advanced model, designed for handling complex tasks at enterprise scale with superior performance.

Jamba Mini: The most efficient model in its size class, optimized for speed and quality with a smaller footprint.


Why Use Our Models?

Privately deployable without compromising on quality: Retain total control over your data, with zero data visibility for the model vendor by deploying Jamba privately in your VPC or on-premise. Ideal for organizations or use cases handling regulated data (e.g. in finance or healthcare) or confidential proprietary data. With comparable quality to market-leading models, you get superior quality with total data privacy.

Open models: With the weights available to be downloaded into your environment, Jamba can be customized to your domain and use case. This is a model that's fully yours.

Long context performance: With a 256K context window, Jamba excels on the kinds of use cases enterprises rely on in their workflows, from analyzing lengthy documents to enhancing RAG workflows at the retrieval stage. And due to its efficiency-optimized hybrid architecture, it can do all this for less money than competitors.


Supported Languages

Jamba models officially support 9 languages:

  • English, Spanish, French, Portuguese, Italian, Dutch, German, Arabic, Hebrew.

Model Details

ModelModel SizeMax TokensVersionSnapshotAPI Endpoint
Jamba Large398B parameters (94B active)256K1.62025-03jamba-large
Jamba Mini52B parameters (12B active)256K1.62025-03jamba-mini

Engineers and data scientists at AI21 labs created the model to help developers and businesses leverage AI to build real-world products with tangible value. Jamba Mini and Jamba Large support zero-shot instruction-following and multi-language support. The Jamba models also provides developers with industry-leading APIs that perform a wide range of productivity tasks designed for commercial use.

  • Organization developing model: AI21 Labs
  • Model date: August 22nd, 2024
  • Model type: Joint Attention and Mamba (Jamba)
  • Knowledge cutoff date March 5th, 2024
  • Input Modality: Text
  • Output Modality: Text
  • License: Jamba open model license
  • Contact: [email protected]

API Versioning

We advise using dated versions of the Jamba API to avoid disruptions from model updates and breaking changes. Also, please note that some endpoints will be deprecated soon.

Here are the details of the available versions:

  • jamba-largecurrently points to jamba-large-1.6-2025-03
  • jamba-minicurrently points to jamba-mini-1.6-2025-03
  • jamba-large-1.6 points to jamba-large-1.6-2025-03
  • jamba-mini-1.6 points to jamba-mini-1.6-2025-03
  • (Deprecated) jamba-large-1.5 points to jamba-large-1.5-2024-08
  • (Deprecated) jamba-mini-1.5 points to jamba-mini-1.5-2024-08

Model Deprecation

ModelSnapshotAPI EndpointDeprecation Date
Jamba Large 1.52024-08jamba-large-1.5-2024-082025-05-06
Jamba Mini 1.52024-08jamba-mini-1.5-2024-082025-05-06

Model Compliance and Certifications


Ethical Considerations

AI21 Labs is on a mission to supercharge human productivity with machines working alongside humans as thought partners, thereby promoting human welfare and prosperity. To deliver its promise, this technology must be deployed and used in a responsible and sustainable way, taking into consideration potential risks, including malicious use by bad actors, accidental misuse and broader societal harms. We take these risks extremely seriously and put measures in place to mitigate them.

AI21 provides open access to Jamba that can be used to power a large variety of useful applications. We believe it is important to ensure that this technology is used in a responsible way, while allowing developers the freedom they need to experiment rapidly and deploy solutions at scale. Overall, we view the safe implementation of this technology as a partnership and collaboration between AI21 and our customers and encourage engagement and dialogue to raise the bar on responsible usage.

In order to use Jamba, you are required to comply with our Terms of Service and with the following usage guidelines.

Please check these usage guidelines periodically, as they may be updated from time to time. For any questions, clarifications or concerns, please contact [email protected].


Limitations

There are a number of limitations inherent to neural networks technology that apply to Jamba. These limitations require explanation and carry important caveats for the application and usage of Jamba.

  • Accuracy: Jamba, like other large pretrained language models, lacks important context about the world because it is trained on textual data and is not grounded in other modalities of experience such as video, real-world physical interaction, and human feedback. Like all language models, Jamba is far more accurate when responding to inputs similar to its training datasets. Novel inputs have a tendency to generate higher variance in its output.

  • Coherence and consistency: Responses from Jamba are sometimes inconsistent, contradictory, or contain seemingly random sentences and paragraphs.

  • Western/English bias: Jamba is trained primarily on English language text from the internet, and is best suited to classifying, searching, summarizing, and generating English text. Furthermore, Jamba has a tendency to hold and amplify the biases contained in its training dataset. As a result, groups of people who were not involved in the creation of the training data can be underrepresented, and stereotypes and prejudices can be perpetuated. Racial, religious, gender, socioeconomic, and other categorizations of human groups can be considered among these factors.

  • Explainability: It is difficult to explain or predict how Jamba will respond without additional training and fine tuning. This is a common issue with neural networks of this scope and scale.

  • Recency: Jamba was trained on a dataset created in March 2024, and therefore has no knowledge of events that have occurred after that date. We update our models regularly to keep them as current as possible, but there are notable gaps and inaccuracies in responses as a result of this lack of recency.

Given these known challenges presented by neural network technology, we've developed usage guidelines to address them.