Batch Document Processing¶
Process multiple documents from cloud storage on a schedule.
Workflow Overview¶
graph LR
A[Schedule Trigger] --> B[List Files]
B --> C[Loop]
C --> D[Download File]
D --> E[DeepTagger]
E --> F[Database]
F --> G[Archive File]
Use Case¶
Scenario: Process new documents uploaded to shared drive overnight
Automation: Every night at midnight, extract data from all new PDFs
Quick Setup¶
- Schedule Trigger
-
Cron:
0 0 * * *
(midnight daily) -
Google Drive / Dropbox
- List files in "/Invoices/New"
-
Filter: PDF files only
-
Loop Over Items
-
Google Drive Download
-
File ID:
{{$json["id"]}}
-
DeepTagger Node
- Project ID: (your project)
- Input Type: File
-
Binary Property: data
-
Database Insert
-
Insert extracted data
-
Move File
- Move to "/Invoices/Processed"
Performance Tips¶
- Add 1-second delay between iterations
- Enable "Continue on Fail" for fault tolerance
- Process max 100 files per run
- Log all processing to database
Expected Processing Time¶
- 50 documents × 5 seconds each = ~4 minutes total
Related Examples¶
- Invoice Processing - Single document workflow
- Email Receipts - Email-triggered processing