"use strict"; var __defProp = Object.defineProperty; var __getOwnPropDesc = Object.getOwnPropertyDescriptor; var __getOwnPropNames = Object.getOwnPropertyNames; var __hasOwnProp = Object.prototype.hasOwnProperty; var __name = (target, value) => __defProp(target, "name", { value, configurable: true }); var __export = (target, all) => { for (var name in all) __defProp(target, name, { get: all[name], enumerable: true }); }; var __copyProps = (to, from, except, desc) => { if (from && typeof from === "object" || typeof from === "function") { for (let key of __getOwnPropNames(from)) if (!__hasOwnProp.call(to, key) && key !== except) __defProp(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable }); } return to; }; var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod); // src/index.ts var src_exports = {}; __export(src_exports, { Upload: () => Upload }); module.exports = __toCommonJS(src_exports); // src/Upload.ts var import_client_s3 = require("@aws-sdk/client-s3"); var import_abort_controller = require("@smithy/abort-controller"); var import_middleware_endpoint = require("@smithy/middleware-endpoint"); var import_smithy_client = require("@smithy/smithy-client"); var import_events = require("events"); // src/bytelength.ts var import_buffer = require("buffer"); var import_runtimeConfig = require("././runtimeConfig"); var byteLength = /* @__PURE__ */ __name((input) => { if (input === null || input === void 0) return 0; if (typeof input === "string") { return import_buffer.Buffer.byteLength(input); } if (typeof input.byteLength === "number") { return input.byteLength; } else if (typeof input.length === "number") { return input.length; } else if (typeof input.size === "number") { return input.size; } else if (typeof input.path === "string") { try { return import_runtimeConfig.ClientDefaultValues.lstatSync(input.path).size; } catch (error) { return void 0; } } return void 0; }, "byteLength"); // src/chunker.ts var import_stream = require("stream"); // src/chunks/getChunkStream.ts async function* getChunkStream(data, partSize, getNextData) { let partNumber = 1; const currentBuffer = { chunks: [], length: 0 }; for await (const datum of getNextData(data)) { currentBuffer.chunks.push(datum); currentBuffer.length += datum.byteLength; while (currentBuffer.length > partSize) { const dataChunk = currentBuffer.chunks.length > 1 ? import_buffer.Buffer.concat(currentBuffer.chunks) : currentBuffer.chunks[0]; yield { partNumber, data: dataChunk.subarray(0, partSize) }; currentBuffer.chunks = [dataChunk.subarray(partSize)]; currentBuffer.length = currentBuffer.chunks[0].byteLength; partNumber += 1; } } yield { partNumber, data: currentBuffer.chunks.length !== 1 ? import_buffer.Buffer.concat(currentBuffer.chunks) : currentBuffer.chunks[0], lastPart: true }; } __name(getChunkStream, "getChunkStream"); // src/chunks/getChunkUint8Array.ts async function* getChunkUint8Array(data, partSize) { let partNumber = 1; let startByte = 0; let endByte = partSize; while (endByte < data.byteLength) { yield { partNumber, data: data.subarray(startByte, endByte) }; partNumber += 1; startByte = endByte; endByte = startByte + partSize; } yield { partNumber, data: data.subarray(startByte), lastPart: true }; } __name(getChunkUint8Array, "getChunkUint8Array"); // src/chunks/getDataReadable.ts async function* getDataReadable(data) { for await (const chunk of data) { if (import_buffer.Buffer.isBuffer(chunk) || chunk instanceof Uint8Array) { yield chunk; } else { yield import_buffer.Buffer.from(chunk); } } } __name(getDataReadable, "getDataReadable"); // src/chunks/getDataReadableStream.ts async function* getDataReadableStream(data) { const reader = data.getReader(); try { while (true) { const { done, value } = await reader.read(); if (done) { return; } if (import_buffer.Buffer.isBuffer(value) || value instanceof Uint8Array) { yield value; } else { yield import_buffer.Buffer.from(value); } } } catch (e) { throw e; } finally { reader.releaseLock(); } } __name(getDataReadableStream, "getDataReadableStream"); // src/chunker.ts var getChunk = /* @__PURE__ */ __name((data, partSize) => { if (data instanceof Uint8Array) { return getChunkUint8Array(data, partSize); } if (data instanceof import_stream.Readable) { return getChunkStream(data, partSize, getDataReadable); } if (data instanceof String || typeof data === "string") { return getChunkUint8Array(import_buffer.Buffer.from(data), partSize); } if (typeof data.stream === "function") { return getChunkStream(data.stream(), partSize, getDataReadableStream); } if (data instanceof ReadableStream) { return getChunkStream(data, partSize, getDataReadableStream); } throw new Error( "Body Data is unsupported format, expected data to be one of: string | Uint8Array | Buffer | Readable | ReadableStream | Blob;." ); }, "getChunk"); // src/Upload.ts var MIN_PART_SIZE = 1024 * 1024 * 5; var _Upload = class _Upload extends import_events.EventEmitter { constructor(options) { super(); /** * S3 multipart upload does not allow more than 10,000 parts. */ this.MAX_PARTS = 1e4; // Defaults. this.queueSize = 4; this.partSize = MIN_PART_SIZE; this.leavePartsOnError = false; this.tags = []; this.concurrentUploaders = []; this.abortMultipartUploadCommand = null; this.uploadedParts = []; this.uploadEnqueuedPartsCount = 0; this.isMultiPart = true; this.sent = false; this.queueSize = options.queueSize || this.queueSize; this.partSize = options.partSize || this.partSize; this.leavePartsOnError = options.leavePartsOnError || this.leavePartsOnError; this.tags = options.tags || this.tags; this.client = options.client; this.params = options.params; this.__validateInput(); this.totalBytes = byteLength(this.params.Body); this.bytesUploadedSoFar = 0; this.abortController = options.abortController ?? new import_abort_controller.AbortController(); } async abort() { this.abortController.abort(); } async done() { if (this.sent) { throw new Error( "@aws-sdk/lib-storage: this instance of Upload has already executed .done(). Create a new instance." ); } this.sent = true; return await Promise.race([this.__doMultipartUpload(), this.__abortTimeout(this.abortController.signal)]); } on(event, listener) { this.uploadEvent = event; return super.on(event, listener); } async __uploadUsingPut(dataPart) { var _a; this.isMultiPart = false; const params = { ...this.params, Body: dataPart.data }; const clientConfig = this.client.config; const requestHandler = clientConfig.requestHandler; const eventEmitter = requestHandler instanceof import_events.EventEmitter ? requestHandler : null; const uploadEventListener = /* @__PURE__ */ __name((event) => { this.bytesUploadedSoFar = event.loaded; this.totalBytes = event.total; this.__notifyProgress({ loaded: this.bytesUploadedSoFar, total: this.totalBytes, part: dataPart.partNumber, Key: this.params.Key, Bucket: this.params.Bucket }); }, "uploadEventListener"); if (eventEmitter !== null) { eventEmitter.on("xhr.upload.progress", uploadEventListener); } const resolved = await Promise.all([this.client.send(new import_client_s3.PutObjectCommand(params)), (_a = clientConfig == null ? void 0 : clientConfig.endpoint) == null ? void 0 : _a.call(clientConfig)]); const putResult = resolved[0]; let endpoint = resolved[1]; if (!endpoint) { endpoint = (0, import_middleware_endpoint.toEndpointV1)( await (0, import_middleware_endpoint.getEndpointFromInstructions)(params, import_client_s3.PutObjectCommand, { ...clientConfig }) ); } if (!endpoint) { throw new Error('Could not resolve endpoint from S3 "client.config.endpoint()" nor EndpointsV2.'); } if (eventEmitter !== null) { eventEmitter.off("xhr.upload.progress", uploadEventListener); } const locationKey = this.params.Key.split("/").map((segment) => (0, import_smithy_client.extendedEncodeURIComponent)(segment)).join("/"); const locationBucket = (0, import_smithy_client.extendedEncodeURIComponent)(this.params.Bucket); const Location = (() => { const endpointHostnameIncludesBucket = endpoint.hostname.startsWith(`${locationBucket}.`); const forcePathStyle = this.client.config.forcePathStyle; const optionalPort = endpoint.port ? `:${endpoint.port}` : ``; if (forcePathStyle) { return `${endpoint.protocol}//${endpoint.hostname}${optionalPort}/${locationBucket}/${locationKey}`; } if (endpointHostnameIncludesBucket) { return `${endpoint.protocol}//${endpoint.hostname}${optionalPort}/${locationKey}`; } return `${endpoint.protocol}//${locationBucket}.${endpoint.hostname}${optionalPort}/${locationKey}`; })(); this.singleUploadResult = { ...putResult, Bucket: this.params.Bucket, Key: this.params.Key, Location }; const totalSize = byteLength(dataPart.data); this.__notifyProgress({ loaded: totalSize, total: totalSize, part: 1, Key: this.params.Key, Bucket: this.params.Bucket }); } async __createMultipartUpload() { if (!this.createMultiPartPromise) { const createCommandParams = { ...this.params, Body: void 0 }; this.createMultiPartPromise = this.client.send(new import_client_s3.CreateMultipartUploadCommand(createCommandParams)).then((createMpuResponse) => { this.abortMultipartUploadCommand = new import_client_s3.AbortMultipartUploadCommand({ Bucket: this.params.Bucket, Key: this.params.Key, UploadId: createMpuResponse.UploadId }); return createMpuResponse; }); } return this.createMultiPartPromise; } async __doConcurrentUpload(dataFeeder) { for await (const dataPart of dataFeeder) { if (this.uploadEnqueuedPartsCount > this.MAX_PARTS) { throw new Error( `Exceeded ${this.MAX_PARTS} parts in multipart upload to Bucket: ${this.params.Bucket} Key: ${this.params.Key}.` ); } if (this.abortController.signal.aborted) { return; } if (dataPart.partNumber === 1 && dataPart.lastPart) { return await this.__uploadUsingPut(dataPart); } if (!this.uploadId) { const { UploadId } = await this.__createMultipartUpload(); this.uploadId = UploadId; if (this.abortController.signal.aborted) { return; } } const partSize = byteLength(dataPart.data) || 0; const requestHandler = this.client.config.requestHandler; const eventEmitter = requestHandler instanceof import_events.EventEmitter ? requestHandler : null; let lastSeenBytes = 0; const uploadEventListener = /* @__PURE__ */ __name((event, request) => { const requestPartSize = Number(request.query["partNumber"]) || -1; if (requestPartSize !== dataPart.partNumber) { return; } if (event.total && partSize) { this.bytesUploadedSoFar += event.loaded - lastSeenBytes; lastSeenBytes = event.loaded; } this.__notifyProgress({ loaded: this.bytesUploadedSoFar, total: this.totalBytes, part: dataPart.partNumber, Key: this.params.Key, Bucket: this.params.Bucket }); }, "uploadEventListener"); if (eventEmitter !== null) { eventEmitter.on("xhr.upload.progress", uploadEventListener); } this.uploadEnqueuedPartsCount += 1; const partResult = await this.client.send( new import_client_s3.UploadPartCommand({ ...this.params, // dataPart.data is chunked into a non-streaming buffer // so the ContentLength from the input should not be used for MPU. ContentLength: void 0, UploadId: this.uploadId, Body: dataPart.data, PartNumber: dataPart.partNumber }) ); if (eventEmitter !== null) { eventEmitter.off("xhr.upload.progress", uploadEventListener); } if (this.abortController.signal.aborted) { return; } if (!partResult.ETag) { throw new Error( `Part ${dataPart.partNumber} is missing ETag in UploadPart response. Missing Bucket CORS configuration for ETag header?` ); } this.uploadedParts.push({ PartNumber: dataPart.partNumber, ETag: partResult.ETag, ...partResult.ChecksumCRC32 && { ChecksumCRC32: partResult.ChecksumCRC32 }, ...partResult.ChecksumCRC32C && { ChecksumCRC32C: partResult.ChecksumCRC32C }, ...partResult.ChecksumSHA1 && { ChecksumSHA1: partResult.ChecksumSHA1 }, ...partResult.ChecksumSHA256 && { ChecksumSHA256: partResult.ChecksumSHA256 } }); if (eventEmitter === null) { this.bytesUploadedSoFar += partSize; } this.__notifyProgress({ loaded: this.bytesUploadedSoFar, total: this.totalBytes, part: dataPart.partNumber, Key: this.params.Key, Bucket: this.params.Bucket }); } } async __doMultipartUpload() { const dataFeeder = getChunk(this.params.Body, this.partSize); const concurrentUploaderFailures = []; for (let index = 0; index < this.queueSize; index++) { const currentUpload = this.__doConcurrentUpload(dataFeeder).catch((err) => { concurrentUploaderFailures.push(err); }); this.concurrentUploaders.push(currentUpload); } await Promise.all(this.concurrentUploaders); if (concurrentUploaderFailures.length >= 1) { await this.markUploadAsAborted(); throw concurrentUploaderFailures[0]; } if (this.abortController.signal.aborted) { await this.markUploadAsAborted(); throw Object.assign(new Error("Upload aborted."), { name: "AbortError" }); } let result; if (this.isMultiPart) { this.uploadedParts.sort((a, b) => a.PartNumber - b.PartNumber); const uploadCompleteParams = { ...this.params, Body: void 0, UploadId: this.uploadId, MultipartUpload: { Parts: this.uploadedParts } }; result = await this.client.send(new import_client_s3.CompleteMultipartUploadCommand(uploadCompleteParams)); if (typeof (result == null ? void 0 : result.Location) === "string" && result.Location.includes("%2F")) { result.Location = result.Location.replace(/%2F/g, "/"); } } else { result = this.singleUploadResult; } this.abortMultipartUploadCommand = null; if (this.tags.length) { await this.client.send( new import_client_s3.PutObjectTaggingCommand({ ...this.params, Tagging: { TagSet: this.tags } }) ); } return result; } /** * Abort the last multipart upload in progress * if we know the upload id, the user did not specify to leave the parts, and * we have a prepared AbortMultipartUpload command. */ async markUploadAsAborted() { if (this.uploadId && !this.leavePartsOnError && null !== this.abortMultipartUploadCommand) { await this.client.send(this.abortMultipartUploadCommand); this.abortMultipartUploadCommand = null; } } __notifyProgress(progress) { if (this.uploadEvent) { this.emit(this.uploadEvent, progress); } } async __abortTimeout(abortSignal) { return new Promise((resolve, reject) => { abortSignal.onabort = () => { const abortError = new Error("Upload aborted."); abortError.name = "AbortError"; reject(abortError); }; }); } __validateInput() { if (!this.params) { throw new Error(`InputError: Upload requires params to be passed to upload.`); } if (!this.client) { throw new Error(`InputError: Upload requires a AWS client to do uploads with.`); } if (this.partSize < MIN_PART_SIZE) { throw new Error( `EntityTooSmall: Your proposed upload partsize [${this.partSize}] is smaller than the minimum allowed size [${MIN_PART_SIZE}] (5MB)` ); } if (this.queueSize < 1) { throw new Error(`Queue size: Must have at least one uploading queue.`); } } }; __name(_Upload, "Upload"); var Upload = _Upload; // Annotate the CommonJS export names for ESM import in node: 0 && (module.exports = { Upload });