A fully automated Retrieval-Augmented Generation (RAG) pipeline built with n8n, Ollama (local LLMs), and Pinecone Vector Database.
This project demonstrates how to ingest documents, generate embeddings, store them in a vector database, and query them using an AI Agent with real context.
- π Automated document ingestion from Google Drive
- βοΈ Intelligent document chunking
- π§ Embedding generation using local Ollama models
- π¦ Scalable vector storage with Pinecone
- π¬ Context-aware chat using n8n AI Agent
- π Runs locally with no external LLM dependency
n8n-rag-automation-ollama-pinecone/
β
βββ workflows/
β βββ file-ingestion-pipeline_rag-chat-automation.json
β
βββ screenshots/
β βββ file-ingestion-workflow.png
β βββ rag-chat-workflow.png
β
βββ .env.example
βββ .gitignore
βββ README.md
File Ingestion Pipeline
- Google Drive Trigger (file added/updated)
- File download
- Recursive Character Text Splitter
- Embeddings via
nomic-embed-text - Store vectors in Pinecone
RAG Chat Pipeline
- Chat trigger
- AI Agent (tool-enabled)
- Semantic search from Pinecone
- Context-aware responses using Llama 3.2
| Purpose | Model |
|---|---|
| Chat / Agent | llama3.2:latest |
| Embeddings | nomic-embed-text |
| Embedding Dimension | 768 |
| Similarity Metric | cosine |
- n8n (local or Docker)
- Ollama installed
- Pinecone account
- Google Drive credentials (for ingestion)
ollama pull llama3.2
ollama pull nomic-embed-text
This project was built to explore how automation, local LLMs, and vector databases come together to form real-world AI systems.
Everything here is designed to be practical, transparent, and extensible.
If this repository helps you learn, build, or experiment with RAG pipelines, feel free to fork it, adapt it, or improve it.
Contributions, suggestions, and discussions are always welcome.
β If you found this useful, consider starring the repo β it really helps!
