• Home
  • Categories
  • Pricing
  • Submit
    Built with
    Ever Works
    Ever Works

    Connect with us

    Stay Updated

    Get the latest updates and exclusive content delivered to your inbox.

    Product

    • Categories
    • Pricing
    • Help

    Clients

    • Sign In
    • Register
    • Forgot password?

    Company

    • About Us
    • Admin
    • Sitemap

    Resources

    • Blog
    • Submit
    • API Documentation
    All product names, logos, and brands are the property of their respective owners. All company, product, and service names used in this repository, related repositories, and associated websites are for identification purposes only. The use of these names, logos, and brands does not imply endorsement, affiliation, or sponsorship. This directory may include content generated by artificial intelligence.
    Copyright © 2025 Awesome Vector Databases. All rights reserved.·Terms of Service·Privacy Policy·Cookies
    Decorative pattern
    Decorative pattern
    1. Home
    2. Search & Retrieval
    3. RankZephyr

    RankZephyr

    Open-source reranking model based on fine-tuned decoder-only LLMs (LLaMA family), designed for listwise document reranking in RAG pipelines. RankZephyr leverages supervised fine-tuning on ranking datasets to improve query-document relevance scoring beyond what zero-shot LLM prompts can achieve.

    Overview

    RankZephyr is an open-source reranking model that fine-tunes decoder-only large language models (specifically LLaMA-based architectures) for document reranking tasks. It addresses the lack of ranking awareness in pre-trained LLMs by supervised fine-tuning on task-specific ranking datasets.

    Reranking Approach

    • Listwise Method: Takes a query and a list of documents as input and instructs the LLM to output reranked document identifiers
    • Sliding Window Strategy: Due to LLM input length limits, employs a sliding window to rerank subsets of candidate documents at a time, ranking from back to front
    • Supervised Fine-Tuning: Fine-tuned on ranking datasets such as MS MARCO passage ranking to enable accurate query-document relevance measurement

    Architecture

    • Backbone: Decoder-only model structure (LLaMA family)
    • Training: Supervised fine-tuning on passage ranking datasets
    • Comparison: Alternative to encoder-decoder rerankers like RankT5 that formulate ranking as a classification/generation task

    Use Cases

    • RAG pipeline document reranking
    • Search result relevance improvement
    • Reducing hallucinations in retrieval-augmented generation by prioritizing most relevant context

    Performance

    • Outperforms zero-shot LLM-based rerankers without API costs
    • More computationally demanding than cross-encoder models but offers competitive effectiveness
    • Part of the family of supervised LLM rerankers showing promise for retrieval tasks
    Surveys

    Loading more......

    Information

    Websitegithub.com
    PublishedApr 4, 2026

    Categories

    1 Item
    Search & Retrieval

    Tags

    3 Items
    #open-source#LLM-reranking#listwise-ranking

    Similar Products

    6 result(s)

    RankT5

    Open-source reranking model that uses an encoder-decoder (T5) architecture, fine-tuned to generate classification tokens indicating whether query-document pairs are relevant or irrelevant. Formulates document ranking as a generation task.

    OpenSearch

    AWS-backed open-source search and analytics suite with vector search capabilities via faiss and nmslib engines, neural search plugins, and tight AWS ecosystem integration.

    BGE-VL

    State-of-the-art multimodal embedding model from BAAI supporting text-to-image, image-to-text, and compositional visual search. Trained on the MegaPairs dataset with over 26 million retrieval triplets.

    Featured

    Qwen3 Embedding

    Multilingual embedding model supporting over 100 languages and ranking #1 on MTEB multilingual leaderboard. Offers flexible model sizes from 0.6B to 8B parameters with user-defined instructions.

    Featured

    Apache Cassandra Vector Search

    Distributed NoSQL database with vector search capabilities via Storage-Attached Indexes (SAI) in Cassandra 5.0+. Uses Lucene HNSW for approximate nearest neighbor search. This is an OSS database under Apache 2.0 license.

    Featured

    Elasticsearch Vector Search

    Search and analytics engine with k-nearest neighbor (kNN) search for semantic similarity. Features approximate and exact kNN, HNSW indexing, and advanced quantization. This is commercial with OSS version available.

    Featured