• Home
  • Categories
  • Pricing
  • Submit
    Built with
    Ever Works
    Ever Works

    Connect with us

    Stay Updated

    Get the latest updates and exclusive content delivered to your inbox.

    Product

    • Categories
    • Pricing
    • Help

    Clients

    • Sign In
    • Register
    • Forgot password?

    Company

    • About Us
    • Admin
    • Sitemap

    Resources

    • Blog
    • Submit
    • API Documentation
    All product names, logos, and brands are the property of their respective owners. All company, product, and service names used in this repository, related repositories, and associated websites are for identification purposes only. The use of these names, logos, and brands does not imply endorsement, affiliation, or sponsorship. This directory may include content generated by artificial intelligence.
    Copyright © 2025 Awesome Vector Databases. All rights reserved.·Terms of Service·Privacy Policy·Cookies
    Decorative pattern
    Decorative pattern
    1. Home
    2. Machine Learning Models
    3. mGTE

    mGTE

    Generalized long-context text representation and reranking models from Alibaba supporting 75 languages and context length up to 8192. Built on transformer++ encoder with RoPE and GLU for enhanced multilingual retrieval.

    Overview

    Alibaba's Tongyi Lab has introduced the GTE-Multilingual (mGTE) series which offers high performance, long-context handling, multilingual support, and elastic embedding, significantly improving retrieval and ranking efficiency.

    Key Features

    The mGTE series includes new generalized text encoder, embedding and reranking models that support 75 languages and the context length of up to 8192 tokens.

    Architecture

    Transformer++ Encoder Backbone

    The models are built upon the transformer++ encoder backbone (BERT + RoPE + GLU) as well as the vocabulary of XLM-R.

    Key Improvements

    • Rotary Position Encoding (RoPE): Replaces BERT's absolute position embeddings to better support long-context training
    • Gated Linear Units (GLU): Replaces the original feed-forward network (FFN) in BERT

    Model Specifications

    • Base model size: 305M parameters
    • Embedding dimension: 768
    • Max input tokens: 8192
    • Language support: 75 languages

    Performance

    The text encoder outperforms the same-sized previous state-of-the-art XLM-R, while the embedding and reranker match the performance of large-sized state-of-the-art BGE-M3 models and achieve better results on long-context retrieval benchmarks.

    Model Variants

    • gte-multilingual-base: For embeddings
    • gte-multilingual-reranker-base: For reranking tasks
    • gte-multilingual-mlm-base: Masked language model variant

    Availability

    The models are available on:

    • Hugging Face under the Alibaba-NLP organization
    • Alibaba Cloud as commercial API services (text-embedding-v1/v2/v3)

    Use Cases

    • Multilingual RAG applications
    • Long-context retrieval
    • Cross-lingual search
    • Enterprise multilingual search
    • Question answering across languages

    Integration

    Supports integration with Milvus, LangChain, and other popular vector database and LLM frameworks.

    Pricing

    Available as open-source models on Hugging Face or through Alibaba Cloud APIs with commercial pricing.

    Surveys

    Loading more......

    Information

    Websitewww.alibabacloud.com
    PublishedMar 26, 2026

    Categories

    1 Item
    Machine Learning Models

    Tags

    3 Items
    #multilingual#long-context#alibaba

    Similar Products

    6 result(s)

    Cohere Rerank v3.5

    State-of-the-art foundational model for ranking with 4096 context length and multilingual support for 100+ languages. Offers exceptional performance on BEIR benchmarks and specialized domains including finance, e-commerce, and enterprise search.

    Featured

    BGE-M3

    A versatile embedding model from BAAI that simultaneously supports dense retrieval, sparse retrieval, and multi-vector retrieval, with multilingual support for 100+ languages and multi-granularity processing from short sentences to 8192-token documents.

    Featured

    Qwen3 Embedding

    Multilingual embedding model supporting over 100 languages and ranking #1 on MTEB multilingual leaderboard. Offers flexible model sizes from 0.6B to 8B parameters with user-defined instructions.

    Featured

    Nomic Embed Text

    First fully reproducible open-source text embedding model with 8,192 context length. v2 introduces Mixture-of-Experts architecture for multilingual embeddings. Outperforms OpenAI models on benchmarks. This is an OSS model under Apache 2.0 license.

    Featured

    Jina Reranker v2

    Transformer-based cross-encoder model fine-tuned for text reranking with Flash Attention 2 architecture. Features multilingual support for 100+ languages, function-calling capabilities, code search, and 6x speedup over v1 with only 278M parameters.

    Cohere Embed Multilingual v3

    High-performance multilingual embedding model from Cohere supporting 100+ languages with 1024 dimensions, optimized for semantic search, RAG, and cross-lingual retrieval tasks.