diff --git a/src/storage/s3.js b/src/storage/s3.js index 48747a188..e2bd777aa 100644 --- a/src/storage/s3.js +++ b/src/storage/s3.js @@ -67,7 +67,7 @@ function getS3Config(apiConfig, callback) { customBackoff: () => 20000 // constant backoff - https://docs.aws.amazon.com/AWSJavaScriptSDK/latest/AWS/Config.html#retryDelayOptions-property }, httpOptions: { - connectTimeout: 20000, // https://github.com/aws/aws-sdk-js/pull/1446 + connectTimeout: 60000, // https://github.com/aws/aws-sdk-js/pull/1446 timeout: 0 // https://github.com/aws/aws-sdk-js/issues/1704 (allow unlimited time for chunk upload) } }; @@ -256,7 +256,7 @@ function copy(apiConfig, oldFilePath, newFilePath) { // S3 copyObject has a file size limit of 5GB so if we have larger files, we do a multipart copy // Exoscale and B2 take too long to copy 5GB - const largeFileLimit = (apiConfig.provider === 'exoscale-sos' || apiConfig.provider === 'backblaze-b2') ? 1024 * 1024 * 1024 : 5 * 1024 * 1024 * 1024; + const largeFileLimit = (apiConfig.provider === 'exoscale-sos' || apiConfig.provider === 'backblaze-b2' || apiConfig.provider === 'digitalocean-spaces') ? 1024 * 1024 * 1024 : 5 * 1024 * 1024 * 1024; if (entry.size < largeFileLimit) { events.emit('progress', `Copying ${relativePath || oldFilePath}`);