Automated Document Ingestion and Q&A with RAG AI

somdn_product_page

This n8n workflow automates the process of ingesting PDF documents into a vector database and leveraging retrieval-augmented generation (RAG) for answering questions. It monitors a specified Google Drive folder for new PDFs, downloads and processes these files by extracting text, splitting it into manageable chunks, and embedding the data using Cohere embeddings. The embeddings are stored in a Milvus vector database for efficient retrieval. When a chat message is received via webhook, the workflow retrieves relevant information from Milvus, supplies it to an OpenAI GPT-4 model, and responds accordingly. This setup is ideal for organizations needing an intelligent FAQ bot, knowledge base, or document-based AI assistant capable of handling large datasets and multilingual content.

Node Count

11 – 20 Nodes

Nodes Used

@n8n/n8n-nodes-langchain.agent, @n8n/n8n-nodes-langchain.chatTrigger, @n8n/n8n-nodes-langchain.documentDefaultDataLoader, @n8n/n8n-nodes-langchain.embeddingsCohere, @n8n/n8n-nodes-langchain.lmChatOpenAi, @n8n/n8n-nodes-langchain.memoryBufferWindow, @n8n/n8n-nodes-langchain.textSplitterRecursiveCharacterTextSplitter, @n8n/n8n-nodes-langchain.vectorStoreMilvus, extractFromFile, googleDrive, googleDriveTrigger, stickyNote

Reviews

There are no reviews yet.

Be the first to review “Automated Document Ingestion and Q&A with RAG AI”

Your email address will not be published. Required fields are marked *