langchain-rag

Installation
Summary

Complete RAG pipeline for document ingestion, embedding, retrieval, and LLM-powered response generation.

  • Supports multiple document loaders (PDF, web pages, directories) and persistent vector stores (Chroma, FAISS, Pinecone) with configurable chunk size and overlap for optimal context preservation
  • Includes similarity search, MMR (Maximal Marginal Relevance) retrieval, and metadata filtering to balance relevance and diversity in results
  • Works with OpenAI embeddings and integrates seamlessly with LangChain agents and chat models for end-to-end RAG workflows
  • Provides Python and TypeScript implementations with best-practice guidance on chunk sizing (500–1500 characters), consistent embedding models, and persistent storage to avoid data loss
SKILL.md

Pipeline:

  1. Index: Load → Split → Embed → Store
  2. Retrieve: Query → Embed → Search → Return docs
  3. Generate: Docs + Query → LLM → Response

Key Components:

  • Document Loaders: Ingest data from files, web, databases
  • Text Splitters: Break documents into chunks
  • Embeddings: Convert text to vectors
  • Vector Stores: Store and search embeddings
Vector Store Use Case Persistence
Related skills
Installs
6.6K
GitHub Stars
678
First Seen
Mar 4, 2026