File Management n8n Workflows
n8n File Management workflow templates and automation recipes — 203 free workflow templates ready to import.
🛠️ Google Drive Tool MCP Server 💪 all 17 operations
Need help? Want access to this workflow + many more paid workflows + live Q&A sessions with a top verified n8n creator? Join the community Complete MCP server exposing all Google Drive Tool operations to AI agents. Zero configuration needed - all 17 operations pre-built. ⚡ Quick Setup Import this workflow into your n8n instance Activate the workflow to start your MCP server Copy the webhook URL from the MCP trigger node Connect AI agents using the MCP URL 🔧 How it Works • MCP Trigger: Serv
autor: David Ashby
Automate Screenshot Upload to Postimages.org with Airtop Browser Automation
Automating File (Image) Upload to Postimages.org Use Case Manually uploading screenshots or other image files to hosting platforms like Postimages.org can be tedious and time-consuming. This automation simplifies the process by automatically uploading an image to Postimages.org and validating the result, which is especially useful for repetitive QA tasks, reporting, or archiving visual web data. What This Automation Does This automation uploads a screenshot to Postimages.org using the followin
autor: Airtop
Upload & Rename Videos to Google Drive via Apps Script from URL
📄 Google Script Workflow: Upload File from URL to Google Drive (via n8n) 🔧 Purpose: This lightweight Google Apps Script acts as a server endpoint that receives a file URL (from n8n), downloads the file, uploads it to your specified Google Drive folder, and responds with the file’s metadata (like Drive file ID and URL). This is useful for large video/audio files that n8n cannot handle directly via HTTP Download nodes. 🚀 Setup Steps: 1. Create a New Script Project Go to https://script.google.c
autor: Joseph
Upload & Categorize Files with Supabase Storage and Secure URL Generation
Supabase Storage File Upload Workflow works with selfhosted Supabase ℹ️ How it works • Accepts file data (MIME type, filename, base64 content) from other workflows • Automatically routes files to appropriate storage buckets based on file type (images, audio, video, documents) • Uploads files to Supabase Storage using the REST API • Generates secure signed URLs for file access with 30-day expiration • Returns structured success/error responses for downstream processing 🏗️ Set up steps • Config
autor: Jan Willem Altink
Automatically Upload Scanned Documents to Nextcloud via ScanservJS
This workflow connects a USB scanner to Nextcloud via ScanservJS and the integrated API. It checks for new scans at a scheduled time (e.g., every 5 minutes). If there are any, they are automatically retrieved via HTTP request and then saved to a desired Nextcloud folder. Ideal for home offices, offices, or maker projects with Raspberry Pi and network scanners. Nodes used: Schedule Trigger – starts the flow cyclically HTTP Request – retrieves document data from ScanservJS Nextcloud Node –
autor: Joachim Hummel
🛠️ Microsoft OneDrive Tool MCP Server 💪 all 14 operations
Need help? Want access to this workflow + many more paid workflows + live Q&A sessions with a top verified n8n creator? Join the community Complete MCP server exposing all Microsoft OneDrive Tool operations to AI agents. Zero configuration needed - all 14 operations pre-built. ⚡ Quick Setup Import this workflow into your n8n instance Activate the workflow to start your MCP server Copy the webhook URL from the MCP trigger node Connect AI agents using the MCP URL 🔧 How it Works • MCP Trigger
autor: David Ashby
Airtable Base Backups to S3
This workflow exports every table in a base as its own CSV, saves the files in a time-stamped folder in Amazon S3, pings you on Slack, and optionally prunes older copies. You get an automated weekly backup that is easy to inspect or re-import as needed. You can easily swap the S3 node for the storage provider of your choice. ++How it works++ Weekly Backup Schedule trigger fires weekly Sets and formats the week ex. [2025-W12] Create a folder in S3 bucket with the week Loops through all tables i
autor: Autonomous Work
Auto-Cleanup of Cloudflare R2 Files Older Than 2 Weeks (+ Telegram Notifications)
This workflow automatically connects to a Cloudflare R2 bucket (via S3-compatible API), filters out files older than 14 days, deletes them, and then sends a Telegram notification for each deletion. It runs on a daily schedule. 🕘 Schedule Trigger Executes the workflow once a day at a specified hour (e.g., 9 AM). 📦 S3 Node – List Files Retrieves all objects from a specific folder (prefix) in a Cloudflare R2 bucket using the S3 API. 🔎 Code Node – Filter Files Older Than 2 Weeks Filters the retri
autor: Mobder
Process Multiple Files with Forms: A Tutorial on Binary Data and Loops
Let a user load multiple files with a Form node, and process the binary data. A very important workflow for many tools. This is a learning example of several core concepts that are hard to grasp in n8n: $binary data Loop and $runIndex Split Out The Save File deomonstrates how to access the binary data correctly, but could be swapped to POST the files to an AI, for example.
autor: Wyeth
Google Drive Workflow with Nested Folder Support
👤 Who’s it for Anyone* who needs to recursively fetch *all files** from a specific Google Drive folder, including files inside all its nested (child) folders. Ideal for automation builders, admins, and integrators who want to process, move, list, or report on all files in a complex folder tree. ⚙️ How it works / What it does This workflow will: Accept a Google Drive Folder ID as an input (from another workflow or trigger). Recursively discover all child folders starting from the parent, no ma
autor: Zain Ali
Copy Folder Structure Without Files in Google Drive
Why? Google Drive desktop lets you copy full folders—contents and all. But what if you only want to copy the folder names (not the files) from one drive location to another? Maybe you're reusing a folder structure or duplicating a template structure, but without this workflow each copy will hold different content. This n8n workflow helps you clone just the structure—no clutter, no file duplicates. How it works Load this workflow into your n8n instance Setup GDRIVE credentials if you have not
autor: Builds.Cool
Clone Nested Folder Structures in Google Drive with Custom Naming
Google Drive Nested Folder Structure Cloner How it works Takes any Google Drive folder structure and creates an exact copy with custom naming Perfect for agencies, consultants, or anyone who needs identical folder structures for multiple projects Creates folder hierarchy only (no files are copied) Handles nested folders automatically with level-by-level processing Set up steps Get your source and destination folder IDs from Google Drive URLs Edit the 3 variables in the CONFIG node (takes 30 se
autor: Jagged Frontiersman
Import E.ON W1000 Energy Meter Data to Home Assistant with Spook Integration
UPDATES: 2025-12-03 fix JS code in calculate hourly sum node E.ON W1000 → n8n → Home Assistant (Spook) “Integration” This workflow processes emails from the E.ON portal containing 15-minute +A -A (import/export) data and daily 1.8.0 2.8.0 meter readings. It extracts the required columns from the attached XLSX file, groups the 15-minute values by hour, then: updates the Spook/Recorder statistics under the IDs sensor.grid_energy_import and sensor.grid_energy_export, and sets the current me
autor: András Farkas
AI-Powered File Management Automation for Google Drive with MCP
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. 📁 Google Drive MCP Workflow – AI-Powered File Management Automation 🚀 🧠 Overview A secure and intelligent n8n workflow that connects with Google Drive via MCP (Model Control Protocol). Ideal for AI agent tasks, compliance-driven storage, and document automation. 🌟 Key Features 🔒 Built-In Safety Backs up files before edits (timestamped) Supports rollback using file history Validates file size
autor: David Olusola
Manage Google Cloud Storage with AI Image Generation using GPT-4 Mini
Beginner’s Tutorial: Manage Google Cloud Storage Buckets and Objects with n8n Watch the demo video below: Who’s it for Beginners who want to learn how to automate Google Cloud Storage (GCS) operations with n8n. Developers who want to combine AI image generation with cloud storage management. Anyone looking for a simple introduction to working with Buckets and Objects in GCS. How it works / What it does This workflow demonstrates end-to-end usage of Google Cloud Storage with AI integratio
autor: Trung Tran
Convert Multiple Files to Base64 with JavaScript Code
Base64 Encode Multiple Binary Files with a Code Node This template demonstrates how to handle multiple binary files in n8n by using a Code node to convert them into a Base64 encoded string. It's particularly useful when an API requires file uploads in this format and the standard 'Extract From File' node is not sufficient for batch processing. The workflow starts by downloading a ZIP file, unzipping it to get multiple binary files, and then uses a Code node with custom JavaScript to encode each
autor: Viktor Klepikovskyi
🛠️ Nextcloud Tool MCP Server 💪 all 17 operations
Need help? Want access to this workflow + many more paid workflows + live Q&A sessions with a top verified n8n creator? Join the community Complete MCP server exposing all Nextcloud Tool operations to AI agents. Zero configuration needed - all 17 operations pre-built. ⚡ Quick Setup Import this workflow into your n8n instance Activate the workflow to start your MCP server Copy the webhook URL from the MCP trigger node Connect AI agents using the MCP URL 🔧 How it Works • MCP Trigger: Serves
autor: David Ashby
Automated Project Kickoff with Dropbox MCP, OpenAI, Slack & Gmail
The Project starter bot takes the hassle out of launching projects by automatically creating a well-structured folder system in Dropbox and sending timely notifications through Slack and Gmail. By combining n8n's intelligent automation and seamless integration with Dropbox, Slack, and Gmail, this workflow streamlines project setup, saving you time and keeping everyone in the loop effortlessly. 💡 Why Use Project Starter Bot? Save Time: Cut down on the tedious manual folder creation by automatin
autor: Khairul Muhtadin
🛠️ Dropbox Tool MCP Server 💪 all 11 operations
Need help? Want access to this workflow + many more paid workflows + live Q&A sessions with a top verified n8n creator? Join the community Complete MCP server exposing all Dropbox Tool operations to AI agents. Zero configuration needed - all 11 operations pre-built. ⚡ Quick Setup Import this workflow into your n8n instance Activate the workflow to start your MCP server Copy the webhook URL from the MCP trigger node Connect AI agents using the MCP URL 🔧 How it Works • MCP Trigger: Serves as
autor: David Ashby
Automatic FTP File Backup to Google Drive with Scheduled Sync
What this workflow does This workflow automatically downloads files from an FTP folder and uploads them to Google Drive. It’s useful for backup, reporting, or syncing data between systems. How it works Cron Trigger runs on a schedule (default: every hour). FTP List node checks the specified folder on your FTP server. FTP Download node fetches each file found. Google Drive Upload node saves the file into your chosen Google Drive folder. Pre-conditions / Requirements An active FTP server
autor: Ziad Adel
Automatically Save & Organize Outlook Email Attachments in OneDrive Folders
Outlook to OneDrive This workflow automates the process of saving binary attachments from Outlook emails into newly created folders in OneDrive. It's ideal for users who regularly receive files and need them organized into separate folders without manual intervention. Each folder is automatically named based on the email subject and the current timestamp, allowing all attachments from that email to be stored inside the corresponding folder. This is particularly useful for streamlining document w
autor: Michael Gullo
🗜️ Automate image compression in Google Drive with Tinify API and email reports
Tags: Image Compression, Tinify API, TinyPNG, SEO Optimisation, E-commerce, Marketing Context Hi! I’m Samir Saci, Supply Chain Engineer, Data Scientist based in Paris, and founder of LogiGreen. I built this workflow for an agency specialising in e-commerce to automate the daily compression of their images stored in a Google Drive folder. This is particularly useful when managing large libraries of product photos, website assets or marketing visuals that need to stay lightweight for SEO, we
autor: Samir Saci
Salesforce to S3 File Migration & Cleanup
Salesforce to S3 File Migration & Cleanup Automate archiving old Salesforce files to Amazon S3, log them back in Salesforce, and free up org storage — all from a scheduled n8n workflow. 🔧 How It Works (High-Level) Schedule Trigger kicks off (e.g., daily). Query Salesforce for ContentDocument records older than 365 days. Loop Each File → download binary via REST. Upload to S3 with the original filename. Lookup Links (ContentDocumentLink) to keep the parent record reference. Filter Ou
autor: Le Nguyen
Sync SharePoint to Google Drive via Supabase (n8n Workflow)
Eliminate manual file transfers and data silos with this enterprise-grade synchronization pipeline. This n8n workflow automates the heavy lifting of mirroring your Microsoft SharePoint document libraries to Google Drive while maintaining a robust audit trail in Supabase or Postgres. Designed for high-reliability environments, the automation utilizes a sophisticated delta-check logic: it fetches SharePoint metadata, compares it against your existing database records using the 'Compare Datasets' node, and identifies only new or modified files for processing. By leveraging the 'Split in Batches' node, the system handles large volumes of data without hitting memory limits, ensuring every document is securely downloaded via HTTP and re-uploaded to its destination. Once the transfer is verified, the workflow updates your database status in real-time, providing a transparent log of your synchronization history. This is an essential solution for organizations transitioning between cloud ecosystems or those requiring a redundant, cross-platform backup strategy. It effectively bridges the gap between Microsoft and Google ecosystems using Supabase as a high-performance state-management layer. **Common Use Cases:** - Automated Multi-Cloud Disaster Recovery: Maintain a real-time mirror of critical SharePoint corporate assets in Google Drive to ensure business continuity during service outages. - Cross-Departmental Collaboration: Automatically sync project documentation from an executive SharePoint site to a creative team's Google Drive folder for seamless cross-platform access. - Centralized Document Auditing: Use Supabase to track every file movement between ecosystems, creating a searchable metadata index for compliance and regulatory reporting.
autor: DIGITAL BIZ TECH