Getting Started
This guide helps you go from zero setup to a working local RAG pipeline in just a few minutes. RAGLight is designed to be simple, explicit, and fast to experiment with. You can start either from the CLI or directly from Python.Prerequisites
Before getting started, make sure you have:- Python 3.9+
-
A local or remote LLM provider
- Recommended for local use: Ollama
- Alternatives: LMStudio, vLLM, OpenAI, Mistral
Using Ollama (recommended)
Install Ollama and pull a model:Installation
Install RAGLight from PyPI:Option 1 — Instant RAG with the CLI
The fastest way to get started is using the interactive CLI wizard.- Selecting a local folder containing your documents
- Choosing an embedding model
- Choosing a vector store
- Selecting an LLM provider and model
- Configuring ignore folders (e.g.
.venv,node_modules)
- Ingest your documents
- Build the vector store
- Start an interactive chat session
Option 2 — Your First RAG Pipeline in Python
If you prefer explicit code, here is the minimal Python example.1. Define your knowledge sources
2. Create the RAG pipeline
3. Build the pipeline
- Parses your documents
- Generates embeddings
- Stores them in the vector database
4. Query your documents
What happens under the hood?
RAGLight keeps everything explicit:- Documents are ingested from your data sources
- Embeddings are generated using the selected model
- Vectors are stored in a vector database
- Relevant chunks are retrieved at query time
- The LLM generates an answer using retrieved context
Going further
Once you have a basic RAG running, you can explore:- Agentic RAG for multi-step retrieval and reasoning
- RAT (Retrieval-Augmented Thinking) with reflection loops
- Custom pipelines with the Builder API
- Multimodal document ingestion (PDFs with images)
- MCP integration for tool-augmented agents
Next steps
- Learn how each component works in the Core Concepts section
- Explore ready-to-run examples in the
examples/folder - Customize your pipeline step by step