Skip to main content

Retries, Timeouts, and Large Files

Long-running jobs fail in real life. Good retry and timeout policy turns random failures into recoverable events.

Reliability Flags

FlagRole
--retriesRetry failed transfers
--retries-sleepBackoff interval
--low-level-retriesRetry backend API operations
--timeoutNetwork inactivity timeout
--contimeoutConnection setup timeout

Mapping

Hardened Transfer Example

resilient-sync.sh
rclone sync /srv/large-data remote-prod:archive/large-data \
--retries 8 \
--retries-sleep 10s \
--low-level-retries 20 \
--timeout 5m \
--contimeout 30s \
--log-file /var/log/rclone-resilient.log

Large File Guidance

ScenarioPractical adjustment
WAN with intermittent packet lossIncrease retries and timeout
Provider with strict API limitsLower transfers, add TPS limit
Multi-GB objectsRun during low-traffic windows
info

If large transfers repeatedly fail, test with smaller batches first to isolate network vs provider behavior.

Common Pitfalls

PitfallResultBetter approach
Tiny timeout valuesFrequent false failuresUse realistic timeout for link quality
Infinite retries without alertingHidden stuck jobsCap retries and monitor exit code
One giant monolithic jobHard restart and debuggingSegment path sets by domain

What's Next