basic template
Vector-only RAG pipeline for developers building their first production-grade document Q&A system.
What you get
project/
├── src/
│ ├── pipeline.py # Main RAG pipeline — ingest and query
│ └── config.py # Pipeline configuration
├── eval/
│ ├── golden_set.json # Evaluation dataset (add your Q&A pairs)
│ └── config.yaml # Evaluation thresholds and CI gate settings
├── pyproject.toml # Python dependencies and RAG-Forge config
└── README.mdDefault configuration
The basic template uses recursive chunking with dense vector retrieval backed by Qdrant. There is no BM25 or reranker — retrieval is purely embedding-based.
# pyproject.toml [tool.rag-forge] section
[tool.rag-forge]
template = "basic"
chunk_strategy = "recursive"
chunk_size = 512
overlap_ratio = 0.1
vector_db = "qdrant"
embedding_model = "BAAI/bge-m3"Evaluation is gated on faithfulness >= 0.85 and tracks context_relevance, faithfulness, and answer_relevance.
Recommended next steps
- Add your documents to
./docsand runrag-forge index --source ./docsto build the vector index. - Add representative question-answer pairs to
eval/golden_set.json, then runrag-forge audit --golden-set eval/golden_set.jsonto measure baseline quality. - Swap the embedding model in
pyproject.tomlto match the language or domain of your documents (e.g., a multilingual model for non-English content).
When to upgrade
Move to the hybrid template when vector-only retrieval misses keyword-heavy queries — for example, exact product codes, technical identifiers, or any domain where BM25 recall complements dense retrieval.