Data & AI
Vector Search Designer - Claude MCP Skill
Design vector similarity search systems for semantic retrieval at scale
SEO Guide: Enhance your AI agent with the Vector Search Designer tool. This Model Context Protocol (MCP) server allows Claude Desktop and other LLMs to design vector similarity search systems for semantic retrieval at scale... Download and configure this skill to unlock new capabilities for your AI workflow.
Documentation
SKILL.md# Vector Search Designer
The Vector Search Designer skill helps you architect and implement vector similarity search systems that power semantic search, recommendation engines, and AI applications. It guides you through selecting the right vector database, designing index structures, optimizing query performance, and scaling to millions or billions of vectors.
Vector search has become foundational to modern AI systems, from RAG pipelines to product recommendations. This skill covers the full stack: understanding approximate nearest neighbor (ANN) algorithms, choosing between database options, tuning recall vs latency tradeoffs, and implementing production-ready search infrastructure.
Whether you are building on Pinecone, Weaviate, Qdrant, pgvector, or implementing your own solution, this skill ensures your vector search system meets your performance and accuracy requirements.
## Core Workflows
### Workflow 1: Select Vector Database
1. **Gather** requirements:
- Scale: How many vectors?
- Query patterns: Single vs batch, filters needed?
- Latency requirements: Real-time vs batch?
- Update frequency: Static vs dynamic?
- Infrastructure: Managed vs self-hosted?
2. **Compare** options:
| Database | Scale | Managed | Features | Best For |
|----------|-------|---------|----------|----------|
| Pinecone | Billion+ | Yes | Hybrid, namespaces | Production, zero-ops |
| Weaviate | 100M+ | Both | GraphQL, modules | Multi-modal, complex queries |
| Qdrant | 100M+ | Both | Rust perf, filtering | High performance, self-hosted |
| Milvus | Billion+ | Both | GPU support, clustering | Large scale, ML teams |
| pgvector | 10M | No | PostgreSQL native | Existing Postgres, small scale |
| Chroma | 1M | No | Simple API | Prototyping, embedded |
3. **Evaluate** with your data
4. **Document** decision rationale
### Workflow 2: Design Index Architecture
1. **Choose** ANN algorithm:
- HNSW: Best recall/speed tradeoff, memory intensive
- IVF: Good for very large datasets, requires training
- PQ: Compression for scale, some accuracy loss
- Flat: Exact search, small datasets only
2. **Configure** index parameters:
```python
# HNSW example configuration
hnsw_config = {
"M": 16, # Connections per node (higher = better recall, more memory)
"efConstruction": 200, # Build-time search depth
"efSearch": 100, # Query-time search depth
}
# IVF example configuration
ivf_config = {
"nlist": 1024, # Number of clusters
"nprobe": 32, # Clusters to search at query time
}
```
3. **Plan** sharding strategy for scale
4. **Design** metadata schema for filtering
### Workflow 3: Optimize Search Performance
1. **Benchmark** baseline performance:
- Queries per second (QPS)
- Latency (p50, p95, p99)
- Recall@k accuracy
2. **Identify** bottlenecks:
- Index loading time
- Search latency
- Filter computation
- Network overhead
3. **Apply** optimizations:
- Tune index parameters (ef, nprobe)
- Implement query caching
- Optimize filter expressions
- Consider quantization for memory
4. **Validate** recall vs latency tradeoff
5. **Document** optimal configuration
## Quick Reference
| Action | Command/Trigger |
|--------|-----------------|
| Choose database | "Which vector database for [use case]" |
| Design index | "Design vector index for [scale]" |
| Optimize search | "Speed up vector search" |
| Add filtering | "Add metadata filters to vector search" |
| Scale vectors | "Scale to [N] million vectors" |
| Benchmark search | "Benchmark vector search performance" |
## Best Practices
- **Right-Size Your Database**: Don't over-engineer for scale you don't need
- <1M vectors: pgvector or Chroma often sufficient
- 1-100M vectors: Qdrant, Weaviate, or Pinecone
- 100M+ vectors: Milvus or Pinecone with careful design
- **Understand Recall vs Speed Tradeoff**: ANN is approximate by design
- Higher recall = slower queries
- Tune based on your accuracy requirements
- Measure actual recall, don't assume
- **Use Hybrid Search When Needed**: Combine vector and keyword search
- Vector: semantic similarity
- Keyword (BM25): exact terms, names, codes
- Typically improves results for mixed queries
- **Design Metadata for Filtering**: Plan your filter strategy upfront
- Indexed fields for frequent filters
- Avoid filtering on high-cardinality fields
- Pre-filter vs post-filter tradeoffs
- **Batch Operations When Possible**: Reduce network overhead
- Batch upserts for ingestion
- Batch queries when latency allows
- Use async operations
- **Monitor and Alert**: Production search needs observability
- Query latency percentiles
- Index size and memory usage
- Recall degradation over time
## Advanced Techniques
### Multi-Vector Search
Handle documents with multiple representations:
```python
class MultiVectorIndex:
def __init__(self):
self.title_index = VectorIndex(dim=768)
self.content_index = VectorIndex(dim=768)
self.summary_index = VectorIndex(dim=768)
def search(self, query_embedding, weights=None):
weights = weights or {"title": 0.3, "content": 0.5, "summary": 0.2}
results = {}
for field, weight in weights.items():
index = getattr(self, f"{field}_index")
field_results = index.search(query_embedding, k=20)
for doc_id, score in field_results:
results[doc_id] = results.get(doc_id, 0) + score * weight
return sorted(results.items(), key=lambda x: x[1], reverse=True)[:10]
```
### Filtered Vector Search Strategies
Optimize search with filters:
```python
def filtered_search(query_embedding, filters, k=10):
# Strategy 1: Pre-filter (for selective filters)
if estimate_selectivity(filters) < 0.1:
candidate_ids = apply_filters(filters)
return vector_search_subset(query_embedding, candidate_ids, k)
# Strategy 2: Post-filter (for non-selective filters)
elif estimate_selectivity(filters) > 0.5:
results = vector_search(query_embedding, k * 3)
filtered = [r for r in results if matches_filters(r, filters)]
return filtered[:k]
# Strategy 3: Hybrid (general case)
else:
return vector_search_with_filters(query_embedding, filters, k)
```
### Quantization for Scale
Reduce memory with acceptable accuracy loss:
```python
# Product Quantization configuration
pq_config = {
"nbits": 8, # Bits per sub-quantizer
"m": 16, # Number of sub-quantizers
# 768-dim * 4 bytes = 3KB/vector -> 16 * 1 byte = 16 bytes/vector
}
# Binary quantization (extreme compression)
binary_config = {
"threshold": 0, # Values > 0 -> 1, else -> 0
# 768-dim * 4 bytes = 3KB/vector -> 768 bits = 96 bytes/vector
}
```
### Incremental Index Updates
Handle dynamic data efficiently:
```python
class DynamicVectorIndex:
def __init__(self, rebuild_threshold=10000):
self.main_index = build_optimized_index()
self.delta_index = [] # Recent additions
self.rebuild_threshold = rebuild_threshold
def add(self, vector, metadata):
self.delta_index.append((vector, metadata))
if len(self.delta_index) >= self.rebuild_threshold:
self.rebuild()
def search(self, query, k):
main_results = self.main_index.search(query, k)
delta_results = brute_force_search(self.delta_index, query, k)
return merge_results(main_results, delta_results, k)
def rebuild(self):
all_data = self.main_index.get_all() + self.delta_index
self.main_index = build_optimized_index(all_data)
self.delta_index = []
```
## Common Pitfalls to Avoid
- Using exact (flat) search at scale instead of ANN
- Not measuring actual recall, assuming ANN is "good enough"
- Over-indexing metadata fields, slowing down updates
- Ignoring the memory requirements of HNSW indexes
- Not planning for index rebuilds and maintenance windows
- Assuming vector databases handle all use cases (sometimes Elasticsearch is better)
- Forgetting to normalize vectors for cosine similarity
- Mixing embeddings from different models in the same indexSignals
Information
- Repository
- eddiebe147/claude-settings
- Author
- eddiebe147
- Last Sync
- 1/18/2026
- Repo Updated
- 1/16/2026
- Created
- 1/13/2026
Reviews (0)
No reviews yet. Be the first to review this skill!
Related Skills
upgrade-nodejs
Upgrading Bun's Self-Reported Node.js Version
cursorrules
CrewAI Development Rules
cn-check
Install and run the Continue CLI (`cn`) to execute AI agent checks on local code changes. Use when asked to "run checks", "lint with AI", "review my changes with cn", or set up Continue CI locally.
CLAUDE
CLAUDE.md
Related Guides
Bear Notes Claude Skill: Your AI-Powered Note-Taking Assistant
Learn how to use the bear-notes Claude skill. Complete guide with installation instructions and examples.
Mastering tmux with Claude: A Complete Guide to the tmux Claude Skill
Learn how to use the tmux Claude skill. Complete guide with installation instructions and examples.
OpenAI Whisper API Claude Skill: Complete Guide to AI-Powered Audio Transcription
Learn how to use the openai-whisper-api Claude skill. Complete guide with installation instructions and examples.