mirror of
https://github.com/nushell/nushell.git
synced 2024-11-23 00:43:33 +01:00
215ed141e7
# Description Broken after #7415 We currently don't try to build the benchmarks in the CI thus this slipped through the cracks. # User-Facing Changes None # Tests + Formatting Compile check is currently missing, but working towards that
192 lines
6.4 KiB
Rust
192 lines
6.4 KiB
Rust
use criterion::{criterion_group, criterion_main, BatchSize, Criterion};
|
|
use nu_cli::eval_source;
|
|
use nu_parser::parse;
|
|
use nu_plugin::{EncodingType, PluginResponse};
|
|
use nu_protocol::{PipelineData, Span, Value};
|
|
use nu_utils::{get_default_config, get_default_env};
|
|
|
|
// FIXME: All benchmarks live in this 1 file to speed up build times when benchmarking.
|
|
// When the *_benchmarks functions were in different files, `cargo bench` would build
|
|
// an executable for every single one - incredibly slowly. Would be nice to figure out
|
|
// a way to split things up again.
|
|
|
|
fn parser_benchmarks(c: &mut Criterion) {
|
|
let mut engine_state = nu_command::create_default_context();
|
|
// parsing config.nu breaks without PWD set
|
|
engine_state.add_env_var(
|
|
"PWD".into(),
|
|
Value::string("/some/dir".to_string(), Span::test_data()),
|
|
);
|
|
|
|
let default_env = get_default_env().as_bytes();
|
|
c.bench_function("parse_default_env_file", |b| {
|
|
b.iter_batched(
|
|
|| nu_protocol::engine::StateWorkingSet::new(&engine_state),
|
|
|mut working_set| parse(&mut working_set, None, default_env, false, &[]),
|
|
BatchSize::SmallInput,
|
|
)
|
|
});
|
|
|
|
let default_config = get_default_config().as_bytes();
|
|
c.bench_function("parse_default_config_file", |b| {
|
|
b.iter_batched(
|
|
|| nu_protocol::engine::StateWorkingSet::new(&engine_state),
|
|
|mut working_set| parse(&mut working_set, None, default_config, false, &[]),
|
|
BatchSize::SmallInput,
|
|
)
|
|
});
|
|
|
|
c.bench_function("eval default_env.nu", |b| {
|
|
b.iter(|| {
|
|
let mut engine_state = nu_command::create_default_context();
|
|
let mut stack = nu_protocol::engine::Stack::new();
|
|
eval_source(
|
|
&mut engine_state,
|
|
&mut stack,
|
|
get_default_env().as_bytes(),
|
|
"default_env.nu",
|
|
PipelineData::empty(),
|
|
false,
|
|
)
|
|
})
|
|
});
|
|
|
|
c.bench_function("eval default_config.nu", |b| {
|
|
b.iter(|| {
|
|
let mut engine_state = nu_command::create_default_context();
|
|
// parsing config.nu breaks without PWD set
|
|
engine_state.add_env_var(
|
|
"PWD".into(),
|
|
Value::string("/some/dir".to_string(), Span::test_data()),
|
|
);
|
|
let mut stack = nu_protocol::engine::Stack::new();
|
|
eval_source(
|
|
&mut engine_state,
|
|
&mut stack,
|
|
get_default_config().as_bytes(),
|
|
"default_config.nu",
|
|
PipelineData::empty(),
|
|
false,
|
|
)
|
|
})
|
|
});
|
|
}
|
|
|
|
fn eval_benchmarks(c: &mut Criterion) {
|
|
c.bench_function("eval default_env.nu", |b| {
|
|
b.iter(|| {
|
|
let mut engine_state = nu_command::create_default_context();
|
|
let mut stack = nu_protocol::engine::Stack::new();
|
|
eval_source(
|
|
&mut engine_state,
|
|
&mut stack,
|
|
get_default_env().as_bytes(),
|
|
"default_env.nu",
|
|
PipelineData::empty(),
|
|
false,
|
|
)
|
|
})
|
|
});
|
|
|
|
c.bench_function("eval default_config.nu", |b| {
|
|
b.iter(|| {
|
|
let mut engine_state = nu_command::create_default_context();
|
|
// parsing config.nu breaks without PWD set
|
|
engine_state.add_env_var(
|
|
"PWD".into(),
|
|
Value::string("/some/dir".to_string(), Span::test_data()),
|
|
);
|
|
let mut stack = nu_protocol::engine::Stack::new();
|
|
eval_source(
|
|
&mut engine_state,
|
|
&mut stack,
|
|
get_default_config().as_bytes(),
|
|
"default_config.nu",
|
|
PipelineData::empty(),
|
|
false,
|
|
)
|
|
})
|
|
});
|
|
}
|
|
|
|
// generate a new table data with `row_cnt` rows, `col_cnt` columns.
|
|
fn encoding_test_data(row_cnt: usize, col_cnt: usize) -> Value {
|
|
let columns: Vec<String> = (0..col_cnt).map(|x| format!("col_{x}")).collect();
|
|
let vals: Vec<Value> = (0..col_cnt as i64).map(Value::test_int).collect();
|
|
|
|
Value::List {
|
|
vals: (0..row_cnt)
|
|
.map(|_| Value::test_record(columns.clone(), vals.clone()))
|
|
.collect(),
|
|
span: Span::test_data(),
|
|
}
|
|
}
|
|
|
|
fn encoding_benchmarks(c: &mut Criterion) {
|
|
let mut group = c.benchmark_group("Encoding");
|
|
let test_cnt_pairs = [
|
|
(100, 5),
|
|
(100, 10),
|
|
(100, 15),
|
|
(1000, 5),
|
|
(1000, 10),
|
|
(1000, 15),
|
|
(10000, 5),
|
|
(10000, 10),
|
|
(10000, 15),
|
|
];
|
|
for (row_cnt, col_cnt) in test_cnt_pairs.into_iter() {
|
|
for fmt in ["json", "msgpack"] {
|
|
group.bench_function(&format!("{fmt} encode {row_cnt} * {col_cnt}"), |b| {
|
|
let mut res = vec![];
|
|
let test_data =
|
|
PluginResponse::Value(Box::new(encoding_test_data(row_cnt, col_cnt)));
|
|
let encoder = EncodingType::try_from_bytes(fmt.as_bytes()).unwrap();
|
|
b.iter(|| encoder.encode_response(&test_data, &mut res))
|
|
});
|
|
}
|
|
}
|
|
group.finish();
|
|
}
|
|
|
|
fn decoding_benchmarks(c: &mut Criterion) {
|
|
let mut group = c.benchmark_group("Decoding");
|
|
let test_cnt_pairs = [
|
|
(100, 5),
|
|
(100, 10),
|
|
(100, 15),
|
|
(1000, 5),
|
|
(1000, 10),
|
|
(1000, 15),
|
|
(10000, 5),
|
|
(10000, 10),
|
|
(10000, 15),
|
|
];
|
|
for (row_cnt, col_cnt) in test_cnt_pairs.into_iter() {
|
|
for fmt in ["json", "msgpack"] {
|
|
group.bench_function(&format!("{fmt} decode for {row_cnt} * {col_cnt}"), |b| {
|
|
let mut res = vec![];
|
|
let test_data =
|
|
PluginResponse::Value(Box::new(encoding_test_data(row_cnt, col_cnt)));
|
|
let encoder = EncodingType::try_from_bytes(fmt.as_bytes()).unwrap();
|
|
encoder.encode_response(&test_data, &mut res).unwrap();
|
|
let mut binary_data = std::io::Cursor::new(res);
|
|
b.iter(|| {
|
|
binary_data.set_position(0);
|
|
encoder.decode_response(&mut binary_data)
|
|
})
|
|
});
|
|
}
|
|
}
|
|
group.finish();
|
|
}
|
|
|
|
criterion_group!(
|
|
benches,
|
|
parser_benchmarks,
|
|
eval_benchmarks,
|
|
encoding_benchmarks,
|
|
decoding_benchmarks
|
|
);
|
|
criterion_main!(benches);
|