n8nflow.net logo

Build & Query RAG System with Google Drive, OpenAI GPT-4o-mini, and Pinecone

by David Olusolaβ€’Updated: Last update 4 months agoβ€’Source: n8n.io
Loading workflow viewer...

Getting Started

πŸ” What This Workflow Does

This RAG Pipeline in n8n automates document ingestion from Google Drive, vectorizes it using OpenAI embeddings, stores it in Pinecone, and enables chat-based retrieval using LangChain agents.

Main Functions:

πŸ“‚ Auto-detects new files uploaded to a specific Google Drive folder.
🧠 Converts the file into embeddings using OpenAI.
πŸ“¦ Stores them in a Pinecone vector database.
πŸ’¬ Allows a user to query the knowledge base through a chat interface.
πŸ€– Uses a GPT-4o-mini model with LangChain to generate intelligent responses using retrieved context.
βš™οΈ Setup Instructions

  1. Connect Accounts
    Ensure these services are connected in n8n:

βœ… Google Drive (OAuth2)
βœ… OpenAI
βœ… Pinecone
You can do this in n8n > Credentials > New and use the matching names from the file:

Google Drive: "Google Drive account 2"
OpenAI: "OpenAi success"
Pinecone: "PineconeApi account 2"
2. Folder Setup
Upload your documents to this folder in Google Drive:

πŸ“ Power Folder

The workflow is triggered every minute when a new file is uploaded.

  1. Workflow Overview
    A. File Ingestion Path

Google Drive Trigger β€” detects new file.
Google Drive (Download) β€” downloads the new file.
Recursive Text Splitter β€” splits text into chunks.
Default Data Loader β€” loads content as LangChain documents.
OpenAI Embeddings β€” converts text chunks into embeddings.
Pinecone Vector Store β€” stores them in "ragfile" index.
B. Chat Retrieval Path

When chat message received β€”
AI Agent β€” LangChain agent managing tools.
OpenAI Chat Model (GPT-4o-mini) β€” generates replies.
Pinecone Vector Store (retrieval) β€” retrieves matching content.
Embeddings OpenAI1 β€” helps match queries to document chunks.