mirror of
https://github.com/NousResearch/hermes-agent.git
synced 2026-04-26 01:01:40 +00:00
The skills directory was getting disorganized — mlops alone had 40 skills in a flat list, and 12 categories were singletons with just one skill each. Code change: - prompt_builder.py: Support sub-categories in skill scanner. skills/mlops/training/axolotl/SKILL.md now shows as category 'mlops/training' instead of just 'mlops'. Backwards-compatible with existing flat structure. Split mlops (40 skills) into 7 sub-categories: - mlops/training (12): accelerate, axolotl, flash-attention, grpo-rl-training, peft, pytorch-fsdp, pytorch-lightning, simpo, slime, torchtitan, trl-fine-tuning, unsloth - mlops/inference (8): gguf, guidance, instructor, llama-cpp, obliteratus, outlines, tensorrt-llm, vllm - mlops/models (6): audiocraft, clip, llava, segment-anything, stable-diffusion, whisper - mlops/vector-databases (4): chroma, faiss, pinecone, qdrant - mlops/evaluation (5): huggingface-tokenizers, lm-evaluation-harness, nemo-curator, saelens, weights-and-biases - mlops/cloud (2): lambda-labs, modal - mlops/research (1): dspy Merged singleton categories: - gifs → media (gif-search joins youtube-content) - music-creation → media (heartmula, songsee) - diagramming → creative (excalidraw joins ascii-art) - ocr-and-documents → productivity - domain → research (domain-intel) - feeds → research (blogwatcher) - market-data → research (polymarket) Fixed misplaced skills: - mlops/code-review → software-development (not ML-specific) - mlops/ml-paper-writing → research (academic writing) Added DESCRIPTION.md files for all new/updated categories.
224 lines
5 KiB
Markdown
224 lines
5 KiB
Markdown
---
|
||
name: faiss
|
||
description: Facebook's library for efficient similarity search and clustering of dense vectors. Supports billions of vectors, GPU acceleration, and various index types (Flat, IVF, HNSW). Use for fast k-NN search, large-scale vector retrieval, or when you need pure similarity search without metadata. Best for high-performance applications.
|
||
version: 1.0.0
|
||
author: Orchestra Research
|
||
license: MIT
|
||
dependencies: [faiss-cpu, faiss-gpu, numpy]
|
||
metadata:
|
||
hermes:
|
||
tags: [RAG, FAISS, Similarity Search, Vector Search, Facebook AI, GPU Acceleration, Billion-Scale, K-NN, HNSW, High Performance, Large Scale]
|
||
|
||
---
|
||
|
||
# FAISS - Efficient Similarity Search
|
||
|
||
Facebook AI's library for billion-scale vector similarity search.
|
||
|
||
## When to use FAISS
|
||
|
||
**Use FAISS when:**
|
||
- Need fast similarity search on large vector datasets (millions/billions)
|
||
- GPU acceleration required
|
||
- Pure vector similarity (no metadata filtering needed)
|
||
- High throughput, low latency critical
|
||
- Offline/batch processing of embeddings
|
||
|
||
**Metrics**:
|
||
- **31,700+ GitHub stars**
|
||
- Meta/Facebook AI Research
|
||
- **Handles billions of vectors**
|
||
- **C++** with Python bindings
|
||
|
||
**Use alternatives instead**:
|
||
- **Chroma/Pinecone**: Need metadata filtering
|
||
- **Weaviate**: Need full database features
|
||
- **Annoy**: Simpler, fewer features
|
||
|
||
## Quick start
|
||
|
||
### Installation
|
||
|
||
```bash
|
||
# CPU only
|
||
pip install faiss-cpu
|
||
|
||
# GPU support
|
||
pip install faiss-gpu
|
||
```
|
||
|
||
### Basic usage
|
||
|
||
```python
|
||
import faiss
|
||
import numpy as np
|
||
|
||
# Create sample data (1000 vectors, 128 dimensions)
|
||
d = 128
|
||
nb = 1000
|
||
vectors = np.random.random((nb, d)).astype('float32')
|
||
|
||
# Create index
|
||
index = faiss.IndexFlatL2(d) # L2 distance
|
||
index.add(vectors) # Add vectors
|
||
|
||
# Search
|
||
k = 5 # Find 5 nearest neighbors
|
||
query = np.random.random((1, d)).astype('float32')
|
||
distances, indices = index.search(query, k)
|
||
|
||
print(f"Nearest neighbors: {indices}")
|
||
print(f"Distances: {distances}")
|
||
```
|
||
|
||
## Index types
|
||
|
||
### 1. Flat (exact search)
|
||
|
||
```python
|
||
# L2 (Euclidean) distance
|
||
index = faiss.IndexFlatL2(d)
|
||
|
||
# Inner product (cosine similarity if normalized)
|
||
index = faiss.IndexFlatIP(d)
|
||
|
||
# Slowest, most accurate
|
||
```
|
||
|
||
### 2. IVF (inverted file) - Fast approximate
|
||
|
||
```python
|
||
# Create quantizer
|
||
quantizer = faiss.IndexFlatL2(d)
|
||
|
||
# IVF index with 100 clusters
|
||
nlist = 100
|
||
index = faiss.IndexIVFFlat(quantizer, d, nlist)
|
||
|
||
# Train on data
|
||
index.train(vectors)
|
||
|
||
# Add vectors
|
||
index.add(vectors)
|
||
|
||
# Search (nprobe = clusters to search)
|
||
index.nprobe = 10
|
||
distances, indices = index.search(query, k)
|
||
```
|
||
|
||
### 3. HNSW (Hierarchical NSW) - Best quality/speed
|
||
|
||
```python
|
||
# HNSW index
|
||
M = 32 # Number of connections per layer
|
||
index = faiss.IndexHNSWFlat(d, M)
|
||
|
||
# No training needed
|
||
index.add(vectors)
|
||
|
||
# Search
|
||
distances, indices = index.search(query, k)
|
||
```
|
||
|
||
### 4. Product Quantization - Memory efficient
|
||
|
||
```python
|
||
# PQ reduces memory by 16-32×
|
||
m = 8 # Number of subquantizers
|
||
nbits = 8
|
||
index = faiss.IndexPQ(d, m, nbits)
|
||
|
||
# Train and add
|
||
index.train(vectors)
|
||
index.add(vectors)
|
||
```
|
||
|
||
## Save and load
|
||
|
||
```python
|
||
# Save index
|
||
faiss.write_index(index, "large.index")
|
||
|
||
# Load index
|
||
index = faiss.read_index("large.index")
|
||
|
||
# Continue using
|
||
distances, indices = index.search(query, k)
|
||
```
|
||
|
||
## GPU acceleration
|
||
|
||
```python
|
||
# Single GPU
|
||
res = faiss.StandardGpuResources()
|
||
index_cpu = faiss.IndexFlatL2(d)
|
||
index_gpu = faiss.index_cpu_to_gpu(res, 0, index_cpu) # GPU 0
|
||
|
||
# Multi-GPU
|
||
index_gpu = faiss.index_cpu_to_all_gpus(index_cpu)
|
||
|
||
# 10-100× faster than CPU
|
||
```
|
||
|
||
## LangChain integration
|
||
|
||
```python
|
||
from langchain_community.vectorstores import FAISS
|
||
from langchain_openai import OpenAIEmbeddings
|
||
|
||
# Create FAISS vector store
|
||
vectorstore = FAISS.from_documents(docs, OpenAIEmbeddings())
|
||
|
||
# Save
|
||
vectorstore.save_local("faiss_index")
|
||
|
||
# Load
|
||
vectorstore = FAISS.load_local(
|
||
"faiss_index",
|
||
OpenAIEmbeddings(),
|
||
allow_dangerous_deserialization=True
|
||
)
|
||
|
||
# Search
|
||
results = vectorstore.similarity_search("query", k=5)
|
||
```
|
||
|
||
## LlamaIndex integration
|
||
|
||
```python
|
||
from llama_index.vector_stores.faiss import FaissVectorStore
|
||
import faiss
|
||
|
||
# Create FAISS index
|
||
d = 1536
|
||
faiss_index = faiss.IndexFlatL2(d)
|
||
|
||
vector_store = FaissVectorStore(faiss_index=faiss_index)
|
||
```
|
||
|
||
## Best practices
|
||
|
||
1. **Choose right index type** - Flat for <10K, IVF for 10K-1M, HNSW for quality
|
||
2. **Normalize for cosine** - Use IndexFlatIP with normalized vectors
|
||
3. **Use GPU for large datasets** - 10-100× faster
|
||
4. **Save trained indices** - Training is expensive
|
||
5. **Tune nprobe/ef_search** - Balance speed/accuracy
|
||
6. **Monitor memory** - PQ for large datasets
|
||
7. **Batch queries** - Better GPU utilization
|
||
|
||
## Performance
|
||
|
||
| Index Type | Build Time | Search Time | Memory | Accuracy |
|
||
|------------|------------|-------------|--------|----------|
|
||
| Flat | Fast | Slow | High | 100% |
|
||
| IVF | Medium | Fast | Medium | 95-99% |
|
||
| HNSW | Slow | Fastest | High | 99% |
|
||
| PQ | Medium | Fast | Low | 90-95% |
|
||
|
||
## Resources
|
||
|
||
- **GitHub**: https://github.com/facebookresearch/faiss ⭐ 31,700+
|
||
- **Wiki**: https://github.com/facebookresearch/faiss/wiki
|
||
- **License**: MIT
|
||
|
||
|