

General Text Embeddings from Alibaba DAMO Academy trained on large-scale relevance pairs. Available in three sizes (large, base, small) with GTE-v1.5 supporting 8192 context length.
Loading more......
The GTE (General Text Embeddings) models are trained by Alibaba DAMO Academy and are mainly based on the BERT framework. They are trained on a large-scale corpus of relevance text pairs, covering a wide range of domains and scenarios.
GTE offers three different sizes to balance performance and efficiency:
GTE models were compared with other popular text embedding models on the MTEB benchmark:
Upgraded GTE embeddings with:
Introduced by Alibaba's Tongyi Lab featuring:
GTE models enable various downstream tasks:
GTE models provide strong performance while maintaining efficiency, making them suitable for production deployments where both quality and resource constraints matter.