reduce DO spaces copy part size

This commit is contained in:
Girish Ramakrishnan
2020-12-15 14:37:18 -08:00
parent a68ddcbbc4
commit f55300eba5

View File

@@ -67,7 +67,7 @@ function getS3Config(apiConfig, callback) {
customBackoff: () => 20000 // constant backoff - https://docs.aws.amazon.com/AWSJavaScriptSDK/latest/AWS/Config.html#retryDelayOptions-property
},
httpOptions: {
connectTimeout: 20000, // https://github.com/aws/aws-sdk-js/pull/1446
connectTimeout: 60000, // https://github.com/aws/aws-sdk-js/pull/1446
timeout: 0 // https://github.com/aws/aws-sdk-js/issues/1704 (allow unlimited time for chunk upload)
}
};
@@ -256,7 +256,7 @@ function copy(apiConfig, oldFilePath, newFilePath) {
// S3 copyObject has a file size limit of 5GB so if we have larger files, we do a multipart copy
// Exoscale and B2 take too long to copy 5GB
const largeFileLimit = (apiConfig.provider === 'exoscale-sos' || apiConfig.provider === 'backblaze-b2') ? 1024 * 1024 * 1024 : 5 * 1024 * 1024 * 1024;
const largeFileLimit = (apiConfig.provider === 'exoscale-sos' || apiConfig.provider === 'backblaze-b2' || apiConfig.provider === 'digitalocean-spaces') ? 1024 * 1024 * 1024 : 5 * 1024 * 1024 * 1024;
if (entry.size < largeFileLimit) {
events.emit('progress', `Copying ${relativePath || oldFilePath}`);