Upload Large Files to Dropbox with Chunking & Web UI Progress Tracking — n8n ワークフロー

複雑度🔗 Webhook11個のノード🏷️ File Management作成者:Anthony

概要

Dropbox Large File Upload System

How It Works

This workflow enables uploading large files (300MB+) to Dropbox through a web interface with real-time progress tracking. It bypasses Dropbox's 150MB single-request limit by breaking files into 8MB chunks and uploading them sequentially using Dropbox's upload session API.

Upload Flow:

User accesses page - Visits /webhook/upload-page and sees HTML form with file picker and folder path input Selects file - Chooses file and clicks "Upload to Dropbox

使用ノード

HTTP Request

ワークフロープレビュー

S
Serve Upload Page
R
Respond with HTML
S
Start Session Webhook
Dropbox Start Session
R
Respond Session ID
A
Append Chunk Webhook
Dropbox Append Chunk
R
Respond Chunk OK
F
Finish Session Webhook
Dropbox Finish Session
R
Respond Complete
11 nodes7 edges

仕組み

  1. 1

    トリガー

    このワークフローは webhook トリガーで開始します。

  2. 2

    処理

    データは 11 個のノードを流れます, connecting httprequest, respondtowebhook, webhook。

  3. 3

    出力

    ワークフローは自動化を完了し、設定された宛先に結果を配信します。

ノード詳細 (11)

HT

HTTP Request

httpRequest

#1

このワークフローのインポート方法

  1. 1右側の JSONをダウンロード ボタンをクリックしてワークフローファイルを保存します。
  2. 2n8nインスタンスを開き、ワークフロー → 新規 → ファイルからインポート に進みます。
  3. 3ダウンロードした upload-large-files-to-dropbox-with-chunking-web-ui-progress-tracking ファイルを選択し、インポートをクリックします。
  4. 4各サービスノードの 認証情報(APIキー、OAuthなど)を設定します。
  5. 5ワークフローをテスト をクリックして動作確認し、有効化します。

またはn8nの JSONからインポート に直接貼り付け:

{ "name": "Upload Large Files to Dropbox with Chunking & Web UI Progress Tracking", "nodes": [...], ...}

インテグレーション

httprequestrespondtowebhookwebhook

このワークフローを取得

ワンクリックでダウンロード&インポート

JSONをダウンロードn8n.ioで見る
ノード11
複雑度medium
トリガーwebhook
カテゴリFile Management

作成者

Anthony

Anthony

@usascholar

タグ

httprequestrespondtowebhookwebhook

n8n初めてですか?

n8nは無料のオープンソースワークフロー自動化ツールです。セルフホストまたはクラウド版をご利用いただけます。

n8nを無料で始める →

Related File Management Workflows

COCOGOHT+5
medium

Sync SharePoint to Google Drive via Supabase (n8n Workflow)

Eliminate manual file transfers and data silos with this enterprise-grade synchronization pipeline. This n8n workflow automates the heavy lifting of mirroring your Microsoft SharePoint document libraries to Google Drive while maintaining a robust audit trail in Supabase or Postgres. Designed for high-reliability environments, the automation utilizes a sophisticated delta-check logic: it fetches SharePoint metadata, compares it against your existing database records using the 'Compare Datasets' node, and identifies only new or modified files for processing. By leveraging the 'Split in Batches' node, the system handles large volumes of data without hitting memory limits, ensuring every document is securely downloaded via HTTP and re-uploaded to its destination. Once the transfer is verified, the workflow updates your database status in real-time, providing a transparent log of your synchronization history. This is an essential solution for organizations transitioning between cloud ecosystems or those requiring a redundant, cross-platform backup strategy. It effectively bridges the gap between Microsoft and Google ecosystems using Supabase as a high-performance state-management layer. **Common Use Cases:** - Automated Multi-Cloud Disaster Recovery: Maintain a real-time mirror of critical SharePoint corporate assets in Google Drive to ensure business continuity during service outages. - Cross-Departmental Collaboration: Automatically sync project documentation from an executive SharePoint site to a creative team's Google Drive folder for seamless cross-platform access. - Centralized Document Auditing: Use Supabase to track every file movement between ecosystems, creating a searchable metadata index for compliance and regulatory reporting.

Scheduled·15 nodes
FOGOGOHT+2
medium

Automate Threads Video Backups to Google Drive & Sheets

Streamline your social media asset management with this robust n8n automation designed for digital marketers and content creators. This workflow eliminates the manual effort of downloading Threads content by providing a seamless bridge between Instagram's text-based platform and your cloud storage. Once triggered via a simple form input, the automation utilizes a RapidAPI connection to fetch high-quality video assets, processes the binary data, and securely uploads the file to a designated Google Drive folder. Simultaneously, the workflow maintains a centralized audit log in Google Sheets, capturing essential metadata such as timestamps and file IDs. Built-in logic gates and wait nodes ensure API rate limits are respected and file transfers are verified before logging. This is an essential tool for agencies managing high volumes of social content who need a reliable, automated pipeline for content repurposing, archival compliance, or collaborative review processes. By centralizing your Threads media assets, you ensure your creative team has instant access to raw footage without the friction of manual downloads. **Common Use Cases:** - Automated Social Media Content Archiving for Compliance - Centralized Asset Library Syncing for Creative Agencies - Competitor Research and Video Content Benchmarking

Trigger·9 nodes