From 74f8731231a022b82ded212310998bc9ddeeb30f Mon Sep 17 00:00:00 2001 From: Jarred Sumner Date: Sat, 25 Jan 2025 07:36:20 -0800 Subject: [PATCH] Update s3.test.ts --- test/js/bun/s3/s3.test.ts | 16 ++++++++-------- 1 file changed, 8 insertions(+), 8 deletions(-) diff --git a/test/js/bun/s3/s3.test.ts b/test/js/bun/s3/s3.test.ts index 4d72254a933bbc..0476c672fb0481 100644 --- a/test/js/bun/s3/s3.test.ts +++ b/test/js/bun/s3/s3.test.ts @@ -23,13 +23,13 @@ function isDockerEnabled(): boolean { } const allCredentials = [ - // { - // accessKeyId: getSecret("S3_R2_ACCESS_KEY"), - // secretAccessKey: getSecret("S3_R2_SECRET_KEY"), - // endpoint: getSecret("S3_R2_ENDPOINT"), - // bucket: getSecret("S3_R2_BUCKET"), - // service: "R2" as string, - // }, + { + accessKeyId: getSecret("S3_R2_ACCESS_KEY"), + secretAccessKey: getSecret("S3_R2_SECRET_KEY"), + endpoint: getSecret("S3_R2_ENDPOINT"), + bucket: getSecret("S3_R2_BUCKET"), + service: "R2" as string, + }, ]; if (isDockerEnabled()) { @@ -543,7 +543,7 @@ for (let credentials of allCredentials) { expect(bytes).toBe(10); expect(Buffer.concat(chunks)).toEqual(Buffer.from("Hello Bun!")); }); - it.only("should work with large files ", async () => { + it("should work with large files ", async () => { const s3file = s3(tmp_filename + "-readable-stream-big", options); await s3file.write(bigishPayload); const stream = s3file.stream();