node/test/parallel/test-http-no-read-no-dump.js
Robert Nagy 1abff07392
stream: bump default highWaterMark
This should give a performance boost accross the board.

Given that the old limit is a decod old and memory capacity has
doubled many times since I think it is appropriate to slightly bump
the default limit.

PR-URL: https://github.com/nodejs/node/pull/52037
Refs: https://github.com/nodejs/node/pull/46608
Refs: https://github.com/nodejs/node/pull/50120
Reviewed-By: Rafael Gonzaga <rafael.nunu@hotmail.com>
Reviewed-By: Matteo Collina <matteo.collina@gmail.com>
Reviewed-By: Yagiz Nizipli <yagiz.nizipli@sentry.io>
Reviewed-By: Chengzhong Wu <legendecas@gmail.com>
Reviewed-By: Moshe Atlow <moshe@atlow.co.il>
Reviewed-By: Mohammed Keyvanzadeh <mohammadkeyvanzade94@gmail.com>
Reviewed-By: Trivikram Kamat <trivikr.dev@gmail.com>
Reviewed-By: Ruben Bridgewater <ruben@bridgewater.de>
2024-03-13 19:02:14 +00:00

57 lines
1.1 KiB
JavaScript

'use strict';
const common = require('../common');
const http = require('http');
let onPause = null;
const server = http.createServer((req, res) => {
if (req.method === 'GET')
return res.end();
res.writeHead(200);
res.flushHeaders();
req.on('close', common.mustCall(() => {
req.on('end', common.mustNotCall());
}));
req.connection.on('pause', () => {
res.end();
onPause();
});
}).listen(0, common.mustCall(() => {
const agent = new http.Agent({
maxSockets: 1,
keepAlive: true
});
const port = server.address().port;
const post = http.request({
agent,
method: 'POST',
port,
}, common.mustCall((res) => {
res.resume();
post.write(Buffer.alloc(64 * 1024).fill('X'));
onPause = () => {
post.end('something');
};
}));
// What happens here is that the server `end`s the response before we send
// `something`, and the client thought that this is a green light for sending
// next GET request
post.write('initial');
http.request({
agent,
method: 'GET',
port,
}, common.mustCall((res) => {
server.close();
res.connection.end();
})).end();
}));