Nomic Embed
Nomic AI
• Framework: PyTorchNomic Embed is an open-source text embedding model built with PyTorch and Apache 2.0 license. With support for up to 8192-token context length, it achieves state-of-the-art performance on tasks like semantic search and retrieval using benchmarks such as MTEB and LoCo. It fully open-source the model weights, training data, and code, making it ideal for production and research usage.
Nomic Embed AI Model

Model Performance Statistics
Views
Released
Last Checked
Version
- Semantic Search
- Clustering
- Parameter Count
- N/A
Dataset Used
BEIR, MassiveText
Related AI Models
Discover similar AI models that might interest you
E5-Mistral

E5-Mistral
Microsoft
E5-Mistral is an open-source embeddings model developed by Microsoft, released under the MIT license. Built with PyTorch, it generates high-quality vector representations useful for semantic search, information retrieval, and clustering tasks. E5-Mistral enables efficient and accurate AI applications requiring text similarity and understanding.
Emu2-Chat

Emu2-Chat
Beijing Academy of AI
Emu2-Chat is a conversational AI model designed for engaging and context-aware chat interactions. It is optimized for natural language understanding and generating human-like responses across various domains. Ideal for chatbots, virtual assistants, and customer support automation.
GPT-Neo

GPT-Neo
EleutherAI
GPT-Neo is an open-source large language model developed by EleutherAI, designed as an alternative to OpenAI’s GPT-3. It uses the Transformer architecture to generate coherent, human-like text based on a given prompt. GPT-Neo is trained on the Pile dataset, which is a diverse and large-scale text corpus, making it capable of many NLP tasks such as text generation, summarization, translation, and question answering. GPT-Neo models come in different sizes, the most popular being the 1.3B and 2.7B parameter versions.