A No-Code, Scalable Document Intelligence System
Organizations today manage large volumes of unstructured dataβPDFs, spreadsheets, reports, and internal documents scattered across multiple systems. Retrieving relevant information from this fragmented data becomes increasingly complex as datasets grow.
To address this challenge, this publication presents a Retrieval-Augmented Generation (RAG) workflow built entirely in n8n (a no-code automation platform) and powered by Google Gemini AI. This workflow transforms unstructured documents into an interactive knowledge base that enables users to chat directly with their PDFs and CSV files using natural language.
By integrating document ingestion, Google Gemini embeddings, semantic vector storage, and context-aware response generation, this project delivers a powerful RAG system accessible even to non-developers. No custom backend coding or ML infrastructure is required β n8n orchestrates the entire workflow visually.
| Feature | Node Type |
|---|---|
| File Upload | n8n Form Trigger / Webhook |
| Text Extraction | File Read / PDF Extract (Community Node) |
| Embeddings | Google Gemini Embeddings Node |
| Vector Storage | Simple Vector Store (Core) |
| Chat Completion | Google Gemini Chat Node |
Tip: Install community nodes from:
Settings β Nodes β Install Community Node
first rag agent.jsonYour RAG Agent system is now ready.
Example queries:
βSummarize page 3.β
βList the important dates in the document.β
βββββββββββββββββββββ
β File Upload β
βββββββββββ¬ββββββββββ
β Extract Text
βββββββββββΌββββββββββ
β Gemini Embeddings β
βββββββββββ¬ββββββββββ
β Store vectors
βββββββββββΌββββββββββ
β Simple VectorStore β
βββββββββββ¬ββββββββββ
β Retrieve Relevant Chunks
βββββββββββΌββββββββββ
β Gemini Chat Model β
βββββββββββ¬ββββββββββ
β Response
βββββΌββββ
β User β
ββββββββ
Chunking determines retrieval quality. Poor chunking = poor results.
| Parameter | Value |
|---|---|
| Chunk Size | 800β1,000 characters |
| Chunk Overlap | ~200 characters (β20%) |
| Context Granularity | Paragraph-level chunks |
overlap = chunk_size * 0.20
The system uses two memory layers:
Gemini performs:
Testing with sample document: converted_text.pdf
| Metric | Result |
|---|---|
| Accuracy | Strong contextual extraction |
| Relevance | High-quality semantic chunk retrieval |
| Latency | Near real-time (<1s) responses |
| User Feedback | Intuitive and reliable conversational flow |
| Limitation | Vector store currently non-persistent |
| Query | System Response Summary |
|---|---|
| βWhat are the main topics?β | Summarized into key thematic sections |
| βExplain page 3.β | Retrieved relevant page content accurately |
| βList important dates.β | Extracted all date references with context |
The repository now includes:
β LICENSE (MIT)
β requirements.txt
β .gitignore with n8n patterns
β TECHNICAL_GUIDE.md
β ARCHITECTURE.md
β CHANGELOG.md
All files are structured for maintainability and enterprise-ready clarity.
This RAG workflow demonstrates how powerful AI-assisted document search can be achieved using pure no-code automation. The integration of n8n with Google Gemini enables organizations to convert their internal documents into a dynamic, conversational knowledge system without building complex infrastructure.