



Maximum number of tokens an embedding model or LLM can process in a single input. Critical parameter for vector databases affecting chunk sizes, with modern models supporting 512 to 32,000+ tokens for long-document understanding.
Loading more......
Context window refers to the maximum number of tokens a model can process in a single input. For embedding models, it determines how much text can be encoded at once. For LLMs in RAG, it affects how much retrieved context can be used.
Context window directly impacts chunking decisions:
Larger context windows enable:
Advantages:
Disadvantages:
Context window affects: