Jamba Mamba-based model represents a groundbreaking advancement in language model technology, combining Mamba Structured State Space (SSM) with traditional Transformer architecture to deliver a 256K context window and superior performance. Licensed under Apache 2.0, Jamba offers remarkable throughput and efficiency, fitting up to 140K context on a single GPU, and is available on Hugging Face and NVIDIA API catalog.