• Home
  • Categories
  • Pricing
  • Submit
    Built with
    Ever Works
    Ever Works

    Connect with us

    Stay Updated

    Get the latest updates and exclusive content delivered to your inbox.

    Product

    • Categories
    • Pricing
    • Help

    Clients

    • Sign In
    • Register
    • Forgot password?

    Company

    • About Us
    • Admin
    • Sitemap

    Resources

    • Blog
    • Submit
    • API Documentation
    All product names, logos, and brands are the property of their respective owners. All company, product, and service names used in this repository, related repositories, and associated websites are for identification purposes only. The use of these names, logos, and brands does not imply endorsement, affiliation, or sponsorship. This directory may include content generated by artificial intelligence.
    Copyright © 2025 Awesome Vector Databases. All rights reserved.·Terms of Service·Privacy Policy·Cookies
    Decorative pattern
    Decorative pattern
    1. Home
    2. Research Papers & Surveys
    3. Residual Quantization with Implicit Neural Codebooks

    Residual Quantization with Implicit Neural Codebooks

    ICML 2024 paper presenting a novel residual quantization approach using implicit neural codebooks for vector compression in high-dimensional similarity search, replacing traditional fixed codebooks with learned representations.

    Overview

    Proposes residual quantization with implicit neural codebooks as an alternative to traditional product quantization.

    Key Contributions

    • Neural codebooks replace quantization lookup tables
    • Residual encoding with learned implicit representations
    • Improved compression accuracy for vector search
    • Published in ICML 2024

    Publication

    • Venue: ICML 2024
    • Authors: Huijben et al.
    Surveys

    Loading more......

    Information

    Websiteproceedings.mlr.press
    PublishedApr 4, 2026

    Categories

    1 Item
    Research Papers & Surveys

    Tags

    3 Items
    #quantization#neural-networks#compression

    Similar Products

    6 result(s)

    CommVQ

    A commutative vector quantization method for KV cache compression that reduces FP16 cache size by 87.5% with 2-bit quantization and enables 1-bit quantization, allowing LLaMA-3.1 8B to run with 128K context on a single RTX 4090 GPU.

    Featured

    Leech Lattice Vector Quantization

    Advanced vector quantization technique that explores the Leech lattice's optimal sphere packing properties at 24 dimensions. Delivers state-of-the-art LLM quantization performance, outperforming recent methods like Quip#, QTIP, and PVQ for extreme vector compression.

    Binary Quantization for Vector Search

    Compression technique that converts full-precision vectors to binary representations, achieving 32x storage reduction while maintaining 90-95% recall for efficient large-scale vector search.

    Statistical Binary Quantization

    Compression method developed by Timescale researchers that improves on standard Binary Quantization, reducing vector memory footprint by 32x while maintaining high accuracy for filtered searches.

    BBQ Binary Quantization

    Elasticsearch and Lucene's implementation of RaBitQ algorithm for 1-bit vector quantization, renamed as BBQ. Provides 32x compression with asymptotically optimal error bounds, enabling efficient vector search at massive scale with minimal accuracy loss.

    Locally-Adaptive Vector Quantization

    Advanced quantization technique that applies per-vector normalization and scalar quantization, adapting the quantization bounds individually for each vector. Achieves four-fold reduction in vector size while maintaining search accuracy with 26-37% overall memory footprint reduction.