open source

Mamba-2.8B

Provided by: Framework: PyTorch

Mamba-2.8B is a powerful open-source natural language processing model developed by Albert Gu and collaborators. Built using PyTorch and licensed under Apache 2.0, this model supports a wide range of NLP applications, including text generation, summarization, and understanding, making it suitable for research and production use.

Model Performance Statistics

13

Views

December 4, 2023

Released

Jul 20, 2025

Last Checked

2.0

Version

Capabilities
  • Long-context Processing
  • Efficient Inference
Performance Benchmarks
PG-1918.2 perplexity
Throughput5x faster than Transformers
Technical Specifications
Parameter Count
N/A
Training & Dataset

Dataset Used

The Pile, Books3

Related AI Models

Discover similar AI models that might interest you

Modelopen source

GPT-Neo

GPT-Neo

GPT-Neo

EleutherAI

GPT-Neo is an open-source large language model developed by EleutherAI, designed as an alternative to OpenAI’s GPT-3. It uses the Transformer architecture to generate coherent, human-like text based on a given prompt. GPT-Neo is trained on the Pile dataset, which is a diverse and large-scale text corpus, making it capable of many NLP tasks such as text generation, summarization, translation, and question answering. GPT-Neo models come in different sizes, the most popular being the 1.3B and 2.7B parameter versions.

Natural Language Processingnlp
42
Modelopen source

Fairseq

Fairseq

Fairseq

Meta AI

Fairseq is Meta AI’s open-source PyTorch-based toolkit for training sequence-to-sequence models, widely used in machine translation, text summarization, and other NLP applications.

Natural Language Processingnlptranslation
38
Modelopen source

Llama 2

Llama 2

Llama 2

Meta AI

Llama 2 is Meta AI’s open-source large language model optimized for a wide range of natural language processing tasks, including chatbots, text generation, and comprehension.

Natural Language Processingnlp
34
Mamba-2.8B NLP Model by Albert Gu | Free API Hub