mirror of
https://github.com/nodejs/node.git
synced 2024-11-24 20:29:23 +01:00
7889416b8a
Some benchmarks' results are small values, so keeping decimals when running them manually (not comparing) can be helpful. PR-URL: https://github.com/nodejs/node/pull/10559 Reviewed-By: James M Snell <jasnell@gmail.com>
237 lines
6.9 KiB
JavaScript
237 lines
6.9 KiB
JavaScript
'use strict';
|
|
|
|
const child_process = require('child_process');
|
|
const http_benchmarkers = require('./_http-benchmarkers.js');
|
|
|
|
exports.createBenchmark = function(fn, configs, options) {
|
|
return new Benchmark(fn, configs, options);
|
|
};
|
|
|
|
function Benchmark(fn, configs, options) {
|
|
// Use the file name as the name of the benchmark
|
|
this.name = require.main.filename.slice(__dirname.length + 1);
|
|
// Parse job-specific configuration from the command line arguments
|
|
const parsed_args = this._parseArgs(process.argv.slice(2), configs);
|
|
this.options = parsed_args.cli;
|
|
this.extra_options = parsed_args.extra;
|
|
// The configuration list as a queue of jobs
|
|
this.queue = this._queue(this.options);
|
|
// The configuration of the current job, head of the queue
|
|
this.config = this.queue[0];
|
|
// Execution arguments i.e. flags used to run the jobs
|
|
this.flags = [];
|
|
if (options && options.flags) {
|
|
this.flags = this.flags.concat(options.flags);
|
|
}
|
|
// Holds process.hrtime value
|
|
this._time = [0, 0];
|
|
// Used to make sure a benchmark only start a timer once
|
|
this._started = false;
|
|
|
|
// this._run will use fork() to create a new process for each configuration
|
|
// combination.
|
|
if (process.env.hasOwnProperty('NODE_RUN_BENCHMARK_FN')) {
|
|
process.nextTick(() => fn(this.config));
|
|
} else {
|
|
process.nextTick(() => this._run());
|
|
}
|
|
}
|
|
|
|
Benchmark.prototype._parseArgs = function(argv, configs) {
|
|
const cliOptions = Object.assign({}, configs);
|
|
const extraOptions = {};
|
|
// Parse configuration arguments
|
|
for (const arg of argv) {
|
|
const match = arg.match(/^(.+?)=([\s\S]*)$/);
|
|
if (!match || !match[1]) {
|
|
console.error('bad argument: ' + arg);
|
|
process.exit(1);
|
|
}
|
|
|
|
if (configs[match[1]]) {
|
|
// Infer the type from the config object and parse accordingly
|
|
const isNumber = typeof configs[match[1]][0] === 'number';
|
|
const value = isNumber ? +match[2] : match[2];
|
|
cliOptions[match[1]] = [value];
|
|
} else {
|
|
extraOptions[match[1]] = match[2];
|
|
}
|
|
}
|
|
return { cli: cliOptions, extra: extraOptions };
|
|
};
|
|
|
|
Benchmark.prototype._queue = function(options) {
|
|
const queue = [];
|
|
const keys = Object.keys(options);
|
|
|
|
// Perform a depth-first walk though all options to generate a
|
|
// configuration list that contains all combinations.
|
|
function recursive(keyIndex, prevConfig) {
|
|
const key = keys[keyIndex];
|
|
const values = options[key];
|
|
const type = typeof values[0];
|
|
|
|
for (const value of values) {
|
|
if (typeof value !== 'number' && typeof value !== 'string') {
|
|
throw new TypeError(`configuration "${key}" had type ${typeof value}`);
|
|
}
|
|
if (typeof value !== type) {
|
|
// This is a requirement for being able to consistently and predictably
|
|
// parse CLI provided configuration values.
|
|
throw new TypeError(`configuration "${key}" has mixed types`);
|
|
}
|
|
|
|
const currConfig = Object.assign({ [key]: value }, prevConfig);
|
|
|
|
if (keyIndex + 1 < keys.length) {
|
|
recursive(keyIndex + 1, currConfig);
|
|
} else {
|
|
queue.push(currConfig);
|
|
}
|
|
}
|
|
}
|
|
|
|
if (keys.length > 0) {
|
|
recursive(0, {});
|
|
} else {
|
|
queue.push({});
|
|
}
|
|
|
|
return queue;
|
|
};
|
|
|
|
// Benchmark an http server.
|
|
exports.default_http_benchmarker =
|
|
http_benchmarkers.default_http_benchmarker;
|
|
exports.PORT = http_benchmarkers.PORT;
|
|
|
|
Benchmark.prototype.http = function(options, cb) {
|
|
const self = this;
|
|
const http_options = Object.assign({ }, options);
|
|
http_options.benchmarker = http_options.benchmarker ||
|
|
self.config.benchmarker ||
|
|
self.extra_options.benchmarker ||
|
|
exports.default_http_benchmarker;
|
|
http_benchmarkers.run(http_options, function(error, code, used_benchmarker,
|
|
result, elapsed) {
|
|
if (cb) {
|
|
cb(code);
|
|
}
|
|
if (error) {
|
|
console.error(error);
|
|
process.exit(code || 1);
|
|
}
|
|
self.config.benchmarker = used_benchmarker;
|
|
self.report(result, elapsed);
|
|
});
|
|
};
|
|
|
|
Benchmark.prototype._run = function() {
|
|
const self = this;
|
|
|
|
(function recursive(queueIndex) {
|
|
const config = self.queue[queueIndex];
|
|
|
|
// set NODE_RUN_BENCHMARK_FN to indicate that the child shouldn't construct
|
|
// a configuration queue, but just execute the benchmark function.
|
|
const childEnv = Object.assign({}, process.env);
|
|
childEnv.NODE_RUN_BENCHMARK_FN = '';
|
|
|
|
// Create configuration arguments
|
|
const childArgs = [];
|
|
for (const key of Object.keys(config)) {
|
|
childArgs.push(`${key}=${config[key]}`);
|
|
}
|
|
for (const key of Object.keys(self.extra_options)) {
|
|
childArgs.push(`${key}=${self.extra_options[key]}`);
|
|
}
|
|
|
|
const child = child_process.fork(require.main.filename, childArgs, {
|
|
env: childEnv,
|
|
execArgv: self.flags.concat(process.execArgv)
|
|
});
|
|
child.on('message', sendResult);
|
|
child.on('close', function(code) {
|
|
if (code) {
|
|
process.exit(code);
|
|
return;
|
|
}
|
|
|
|
if (queueIndex + 1 < self.queue.length) {
|
|
recursive(queueIndex + 1);
|
|
}
|
|
});
|
|
})(0);
|
|
};
|
|
|
|
Benchmark.prototype.start = function() {
|
|
if (this._started) {
|
|
throw new Error('Called start more than once in a single benchmark');
|
|
}
|
|
this._started = true;
|
|
this._time = process.hrtime();
|
|
};
|
|
|
|
Benchmark.prototype.end = function(operations) {
|
|
// get elapsed time now and do error checking later for accuracy.
|
|
const elapsed = process.hrtime(this._time);
|
|
|
|
if (!this._started) {
|
|
throw new Error('called end without start');
|
|
}
|
|
if (typeof operations !== 'number') {
|
|
throw new Error('called end() without specifying operation count');
|
|
}
|
|
|
|
const time = elapsed[0] + elapsed[1] / 1e9;
|
|
const rate = operations / time;
|
|
this.report(rate, elapsed);
|
|
};
|
|
|
|
function formatResult(data) {
|
|
// Construct configuration string, " A=a, B=b, ..."
|
|
let conf = '';
|
|
for (const key of Object.keys(data.conf)) {
|
|
conf += ' ' + key + '=' + JSON.stringify(data.conf[key]);
|
|
}
|
|
|
|
var rate = data.rate.toString().split('.');
|
|
rate[0] = rate[0].replace(/(\d)(?=(\d\d\d)+(?!\d))/g, '$1,');
|
|
rate = (rate[1] ? rate.join('.') : rate[0]);
|
|
return `${data.name}${conf}: ${rate}`;
|
|
}
|
|
|
|
function sendResult(data) {
|
|
if (process.send) {
|
|
// If forked, report by process send
|
|
process.send(data);
|
|
} else {
|
|
// Otherwise report by stdout
|
|
console.log(formatResult(data));
|
|
}
|
|
}
|
|
exports.sendResult = sendResult;
|
|
|
|
Benchmark.prototype.report = function(rate, elapsed) {
|
|
sendResult({
|
|
name: this.name,
|
|
conf: this.config,
|
|
rate: rate,
|
|
time: elapsed[0] + elapsed[1] / 1e9
|
|
});
|
|
};
|
|
|
|
exports.v8ForceOptimization = function(method) {
|
|
if (typeof method !== 'function')
|
|
return;
|
|
|
|
const v8 = require('v8');
|
|
v8.setFlagsFromString('--allow_natives_syntax');
|
|
|
|
const args = Array.prototype.slice.call(arguments, 1);
|
|
method.apply(null, args);
|
|
eval('%OptimizeFunctionOnNextCall(method)');
|
|
method.apply(null, args);
|
|
return eval('%GetOptimizationStatus(method)');
|
|
};
|