skillZsskillZsskillZs
HAND-TAGGED >>> 991 SKILLS LIVE <<<* OPEN SOURCE *NO LOGIN, NO TRACKING FRESH DROPS WEEKLY HAND-TAGGED >>> 991 SKILLS LIVE <<<* OPEN SOURCE *NO LOGIN, NO TRACKING FRESH DROPS WEEKLY HAND-TAGGED >>> 991 SKILLS LIVE <<<* OPEN SOURCE *NO LOGIN, NO TRACKING FRESH DROPS WEEKLY HAND-TAGGED >>> 991 SKILLS LIVE <<<* OPEN SOURCE *NO LOGIN, NO TRACKING FRESH DROPS WEEKLY HAND-TAGGED >>> 991 SKILLS LIVE <<<* OPEN SOURCE *NO LOGIN, NO TRACKING FRESH DROPS WEEKLY HAND-TAGGED >>> 991 SKILLS LIVE <<<* OPEN SOURCE *NO LOGIN, NO TRACKING FRESH DROPS WEEKLY
← back to zine
rag-engineerSKILL #NEER
Research

rag-engineer

Expert in building Retrieval-Augmented Generation systems. Masters embedding models, vector databases, chunking strategies, and retrieval optimization for LLM applications. Use when: building RAG, vector search, embeddings, semantic search, document retrieval.

↗ github · ★ 27k·src: davila7/claude-code-templates

the manual

RAG Engineer

Role: RAG Systems Architect

I bridge the gap between raw documents and LLM understanding. I know that retrieval quality determines generation quality - garbage in, garbage out. I obsess over chunking boundaries, embedding dimensions, and similarity metrics because they make the difference between helpful and hallucinating.

Capabilities

  • Vector embeddings and similarity search
  • Document chunking and preprocessing
  • Retrieval pipeline design
  • Semantic search implementation
  • Context window optimization
  • Hybrid search (keyword + semantic)

Requirements

  • LLM fundamentals
  • Understanding of embeddings
  • Basic NLP concepts

Patterns

Semantic Chunking

Chunk by meaning, not arbitrary token counts

- Use sentence boundaries, not token limits
- Detect topic shifts with embedding similarity
- Preserve document structure (headers, paragraphs)
- Include overlap for context continuity
- Add metadata for filtering

Hierarchical Retrieval

Multi-level retrieval for better precision

- Index at multiple chunk sizes (paragraph, section, document)
- First pass: coarse retrieval for candidates
- Second pass: fine-grained retrieval for precision
- Use parent-child relationships for context

Hybrid Search

Combine semantic and keyword search

- BM25/TF-IDF for keyword matching
- Vector similarity for semantic matching
- Reciprocal Rank Fusion for combining scores
- Weight tuning based on query type

Anti-Patterns

❌ Fixed Chunk Size

❌ Embedding Everything

❌ Ignoring Evaluation

⚠️ Sharp Edges

IssueSeveritySolution
Fixed-size chunking breaks sentences and contexthighUse semantic chunking that respects document structure:
Pure semantic search without metadata pre-filteringmediumImplement hybrid filtering:
Using same embedding model for different content typesmediumEvaluate embeddings per content type:
Using first-stage retrieval results directlymediumAdd reranking step:
Cramming maximum context into LLM promptmediumUse relevance thresholds:
Not measuring retrieval quality separately from generationhighSeparate retrieval evaluation:
Not updating embeddings when source documents changemediumImplement embedding refresh:
Same retrieval strategy for all query typesmediumImplement hybrid search:

Related Skills

Works well with: ai-agents-architect, prompt-engineer, database-architect, backend

more research

hybrid-search-implementation
Research
NEWHOT
hybrid-search-implementation
0@ 0 35k
embedding-strategies
Research
NEWHOT
embedding-strategies
0@ 0 35k
vector-index-tuning
Research
NEWHOT
vector-index-tuning
0@ 0 35k
rag-implementation
Research
NEWHOT
rag-implementation
0@ 0 35k
similarity-search-patterns
Research
NEWHOT
similarity-search-patterns
0@ 0 35k
planning
Research
NEWHOT
planning
0@ 0 27k
deep-research-notebooklm
Research
NEWHOT
deep-research-notebooklm
0@ 0 27k
slack-automation
Research
NEWHOT
slack-automation
0@ 0 27k