summaryrefslogtreecommitdiffstats
path: root/testing/web-platform/tests/compression/compression-multiple-chunks.tentative.any.js
diff options
context:
space:
mode:
authorDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-07 09:22:09 +0000
committerDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-07 09:22:09 +0000
commit43a97878ce14b72f0981164f87f2e35e14151312 (patch)
tree620249daf56c0258faa40cbdcf9cfba06de2a846 /testing/web-platform/tests/compression/compression-multiple-chunks.tentative.any.js
parentInitial commit. (diff)
downloadfirefox-43a97878ce14b72f0981164f87f2e35e14151312.tar.xz
firefox-43a97878ce14b72f0981164f87f2e35e14151312.zip
Adding upstream version 110.0.1.upstream/110.0.1upstream
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'testing/web-platform/tests/compression/compression-multiple-chunks.tentative.any.js')
-rw-r--r--testing/web-platform/tests/compression/compression-multiple-chunks.tentative.any.js67
1 files changed, 67 insertions, 0 deletions
diff --git a/testing/web-platform/tests/compression/compression-multiple-chunks.tentative.any.js b/testing/web-platform/tests/compression/compression-multiple-chunks.tentative.any.js
new file mode 100644
index 0000000000..f43f13b541
--- /dev/null
+++ b/testing/web-platform/tests/compression/compression-multiple-chunks.tentative.any.js
@@ -0,0 +1,67 @@
+// META: global=window,worker
+// META: script=third_party/pako/pako_inflate.min.js
+// META: timeout=long
+
+'use strict';
+
+// This test asserts that compressing multiple chunks should work.
+
+// Example: ('Hello', 3) => TextEncoder().encode('HelloHelloHello')
+function makeExpectedChunk(input, numberOfChunks) {
+ const expectedChunk = input.repeat(numberOfChunks);
+ return new TextEncoder().encode(expectedChunk);
+}
+
+// Example: ('Hello', 3, 'deflate') => compress ['Hello', 'Hello', Hello']
+async function compressMultipleChunks(input, numberOfChunks, format) {
+ const cs = new CompressionStream(format);
+ const writer = cs.writable.getWriter();
+ const chunk = new TextEncoder().encode(input);
+ for (let i = 0; i < numberOfChunks; ++i) {
+ writer.write(chunk);
+ }
+ const closePromise = writer.close();
+ const out = [];
+ const reader = cs.readable.getReader();
+ let totalSize = 0;
+ while (true) {
+ const { value, done } = await reader.read();
+ if (done)
+ break;
+ out.push(value);
+ totalSize += value.byteLength;
+ }
+ await closePromise;
+ const concatenated = new Uint8Array(totalSize);
+ let offset = 0;
+ for (const array of out) {
+ concatenated.set(array, offset);
+ offset += array.byteLength;
+ }
+ return concatenated;
+}
+
+const hello = 'Hello';
+
+for (let numberOfChunks = 2; numberOfChunks <= 16; ++numberOfChunks) {
+ promise_test(async t => {
+ const compressedData = await compressMultipleChunks(hello, numberOfChunks, 'deflate');
+ const expectedValue = makeExpectedChunk(hello, numberOfChunks);
+ // decompress with pako, and check that we got the same result as our original string
+ assert_array_equals(expectedValue, pako.inflate(compressedData), 'value should match');
+ }, `compressing ${numberOfChunks} chunks with deflate should work`);
+
+ promise_test(async t => {
+ const compressedData = await compressMultipleChunks(hello, numberOfChunks, 'gzip');
+ const expectedValue = makeExpectedChunk(hello, numberOfChunks);
+ // decompress with pako, and check that we got the same result as our original string
+ assert_array_equals(expectedValue, pako.inflate(compressedData), 'value should match');
+ }, `compressing ${numberOfChunks} chunks with gzip should work`);
+
+ promise_test(async t => {
+ const compressedData = await compressMultipleChunks(hello, numberOfChunks, 'deflate-raw');
+ const expectedValue = makeExpectedChunk(hello, numberOfChunks);
+ // decompress with pako, and check that we got the same result as our original string
+ assert_array_equals(expectedValue, pako.inflateRaw(compressedData), 'value should match');
+ }, `compressing ${numberOfChunks} chunks with deflate-raw should work`);
+}