ai-engineer-rag-implementation-chunking


id: ai-engineer-rag-implementation-chunking aliases: [ ] tags: - roadmap - ai-engineer - ai-engineer-rag-implementation - ready - –

# ai-engineer-rag-implementation-chunking

## Contents

__Roadmap info from [ roadmap website ] (https://roadmap.sh/ai-engineer/chunking@mX987wiZF7p3V_gExrPeX) __

  ## Chunking

  The
  chunking
  step in Retrieval-Augmented Generation (RAG) involves breaking down large documents or data sources into smaller, manageable chunks. This is done to ensure that the retriever can efficiently search through large volumes of data while staying within the token or input limits of the model. Each chunk, typically a paragraph or section, is converted into an embedding, and these embeddings are stored in a vector database. When a query is made, the retriever searches for the most relevant chunks rather than the entire document, enabling faster and more accurate retrieval.

Learn more from the following resources: