Skip to main content
Module

x/deno/cli/bench/main.rs

A modern runtime for JavaScript and TypeScript.
Go to Latest
File
// Copyright 2018-2022 the Deno authors. All rights reserved. MIT license.
use deno_core::error::AnyError;use deno_core::serde_json;use deno_core::serde_json::Value;use serde::Serialize;use std::collections::HashMap;use std::convert::From;use std::env;use std::fs;use std::path::Path;use std::path::PathBuf;use std::process::Command;use std::process::Stdio;use std::time::SystemTime;
mod http;mod lsp;
fn read_json(filename: &str) -> Result<Value> { let f = fs::File::open(filename)?; Ok(serde_json::from_reader(f)?)}
fn write_json(filename: &str, value: &Value) -> Result<()> { let f = fs::File::create(filename)?; serde_json::to_writer(f, value)?; Ok(())}
/// The list of the tuples of the benchmark name, arguments and return codeconst EXEC_TIME_BENCHMARKS: &[(&str, &[&str], Option<i32>)] = &[ // we need to run the cold_* benchmarks before the _warm_ ones as they ensure // the cache is properly populated, instead of other tests possibly // invalidating that cache. ( "cold_hello", &["run", "--reload", "cli/tests/testdata/002_hello.ts"], None, ), ( "cold_relative_import", &[ "run", "--reload", "cli/tests/testdata/003_relative_import.ts", ], None, ), ("hello", &["run", "cli/tests/testdata/002_hello.ts"], None), ( "relative_import", &["run", "cli/tests/testdata/003_relative_import.ts"], None, ), ( "error_001", &["run", "cli/tests/testdata/error_001.ts"], Some(1), ), ( "no_check_hello", &[ "run", "--reload", "--no-check", "cli/tests/testdata/002_hello.ts", ], None, ), ( "workers_startup", &[ "run", "--allow-read", "cli/tests/testdata/workers/bench_startup.ts", ], None, ), ( "workers_round_robin", &[ "run", "--allow-read", "cli/tests/testdata/workers/bench_round_robin.ts", ], None, ), ( "workers_large_message", &[ "run", "--allow-read", "cli/tests/testdata/workers/bench_large_message.ts", ], None, ), ( "text_decoder", &["run", "cli/tests/testdata/text_decoder_perf.js"], None, ), ( "text_encoder", &["run", "cli/tests/testdata/text_encoder_perf.js"], None, ), ( "text_encoder_into", &["run", "cli/tests/testdata/text_encoder_into_perf.js"], None, ), ( "response_string", &["run", "cli/tests/testdata/response_string_perf.js"], None, ), ( "check", &[ "cache", "--reload", "test_util/std/examples/chat/server_test.ts", ], None, ), ( "no_check", &[ "cache", "--reload", "--no-check", "test_util/std/examples/chat/server_test.ts", ], None, ), ( "bundle", &["bundle", "test_util/std/examples/chat/server_test.ts"], None, ), ( "bundle_no_check", &[ "bundle", "--no-check", "test_util/std/examples/chat/server_test.ts", ], None, ),];
const RESULT_KEYS: &[&str] = &["mean", "stddev", "user", "system", "min", "max"];fn run_exec_time( deno_exe: &Path, target_dir: &Path,) -> Result<HashMap<String, HashMap<String, f64>>> { let hyperfine_exe = test_util::prebuilt_tool_path("hyperfine");
let benchmark_file = target_dir.join("hyperfine_results.json"); let benchmark_file = benchmark_file.to_str().unwrap();
let mut command = [ hyperfine_exe.to_str().unwrap(), "--export-json", benchmark_file, "--warmup", "3", ] .iter() .map(|s| s.to_string()) .collect::<Vec<_>>();
for (_, args, return_code) in EXEC_TIME_BENCHMARKS { let ret_code_test = if let Some(code) = return_code { // Bash test which asserts the return code value of the previous command // $? contains the return code of the previous command format!("; test $? -eq {}", code) } else { "".to_string() }; command.push(format!( "{} {} {}", deno_exe.to_str().unwrap(), args.join(" "), ret_code_test )); }
test_util::run( &command.iter().map(|s| s.as_ref()).collect::<Vec<_>>(), None, None, None, true, );
let mut results = HashMap::<String, HashMap<String, f64>>::new(); let hyperfine_results = read_json(benchmark_file)?; for ((name, _, _), data) in EXEC_TIME_BENCHMARKS.iter().zip( hyperfine_results .as_object() .unwrap() .get("results") .unwrap() .as_array() .unwrap(), ) { let data = data.as_object().unwrap().clone(); results.insert( name.to_string(), data .into_iter() .filter(|(key, _)| RESULT_KEYS.contains(&key.as_str())) .map(|(key, val)| (key, val.as_f64().unwrap())) .collect(), ); }
Ok(results)}
fn rlib_size(target_dir: &std::path::Path, prefix: &str) -> u64 { let mut size = 0; let mut seen = std::collections::HashSet::new(); for entry in std::fs::read_dir(target_dir.join("deps")).unwrap() { let entry = entry.unwrap(); let os_str = entry.file_name(); let name = os_str.to_str().unwrap(); if name.starts_with(prefix) && name.ends_with(".rlib") { let start = name.split('-').next().unwrap().to_string(); if seen.contains(&start) { println!("skip {}", name); } else { seen.insert(start); size += entry.metadata().unwrap().len(); println!("check size {} {}", name, size); } } } assert!(size > 0); size}
const BINARY_TARGET_FILES: &[&str] = &["CLI_SNAPSHOT.bin", "COMPILER_SNAPSHOT.bin"];fn get_binary_sizes(target_dir: &Path) -> Result<HashMap<String, u64>> { let mut sizes = HashMap::<String, u64>::new(); let mut mtimes = HashMap::<String, SystemTime>::new();
sizes.insert( "deno".to_string(), test_util::deno_exe_path().metadata()?.len(), );
// add up size for everything in target/release/deps/libswc* let swc_size = rlib_size(target_dir, "libswc"); println!("swc {} bytes", swc_size); sizes.insert("swc_rlib".to_string(), swc_size);
let v8_size = rlib_size(target_dir, "libv8"); println!("v8 {} bytes", v8_size); sizes.insert("rusty_v8_rlib".to_string(), v8_size);
// Because cargo's OUT_DIR is not predictable, search the build tree for // snapshot related files. for file in walkdir::WalkDir::new(target_dir) { let file = match file { Ok(file) => file, Err(_) => continue, }; let filename = file.file_name().to_str().unwrap().to_string();
if !BINARY_TARGET_FILES.contains(&filename.as_str()) { continue; }
let meta = file.metadata()?; let file_mtime = meta.modified()?;
// If multiple copies of a file are found, use the most recent one. if let Some(stored_mtime) = mtimes.get(&filename) { if *stored_mtime > file_mtime { continue; } }
mtimes.insert(filename.clone(), file_mtime); sizes.insert(filename, meta.len()); }
Ok(sizes)}
const BUNDLES: &[(&str, &str)] = &[ ("file_server", "./test_util/std/http/file_server.ts"), ("gist", "./test_util/std/examples/gist.ts"),];fn bundle_benchmark(deno_exe: &Path) -> Result<HashMap<String, u64>> { let mut sizes = HashMap::<String, u64>::new();
for (name, url) in BUNDLES { let path = format!("{}.bundle.js", name); test_util::run( &[ deno_exe.to_str().unwrap(), "bundle", "--unstable", url, &path, ], None, None, None, true, );
let file = PathBuf::from(path); assert!(file.is_file()); sizes.insert(name.to_string(), file.metadata()?.len()); let _ = fs::remove_file(file); }
Ok(sizes)}
fn run_http(target_dir: &Path, new_data: &mut BenchResult) -> Result<()> { let stats = http::benchmark(target_dir)?;
new_data.req_per_sec = stats .iter() .map(|(name, result)| (name.clone(), result.requests)) .collect();
new_data.max_latency = stats .iter() .map(|(name, result)| (name.clone(), result.latency)) .collect();
Ok(())}
fn run_strace_benchmarks( deno_exe: &Path, new_data: &mut BenchResult,) -> Result<()> { use std::io::Read;
let mut thread_count = HashMap::<String, u64>::new(); let mut syscall_count = HashMap::<String, u64>::new();
for (name, args, expected_exit_code) in EXEC_TIME_BENCHMARKS { let mut file = secure_tempfile::NamedTempFile::new()?;
let exit_status = Command::new("strace") .args(&[ "-c", "-f", "-o", file.path().to_str().unwrap(), deno_exe.to_str().unwrap(), ]) .args(args.iter()) .stdout(Stdio::null()) .spawn()? .wait()?; let expected_exit_code = expected_exit_code.unwrap_or(0); assert_eq!(exit_status.code(), Some(expected_exit_code));
let mut output = String::new(); file.as_file_mut().read_to_string(&mut output)?;
let strace_result = test_util::parse_strace_output(&output); let clone = strace_result.get("clone").map(|d| d.calls).unwrap_or(0) + 1; let total = strace_result.get("total").unwrap().calls; thread_count.insert(name.to_string(), clone); syscall_count.insert(name.to_string(), total); }
new_data.thread_count = thread_count; new_data.syscall_count = syscall_count;
Ok(())}
fn run_max_mem_benchmark(deno_exe: &Path) -> Result<HashMap<String, u64>> { let mut results = HashMap::<String, u64>::new();
for (name, args, return_code) in EXEC_TIME_BENCHMARKS { let proc = Command::new("time") .args(&["-v", deno_exe.to_str().unwrap()]) .args(args.iter()) .stdout(Stdio::null()) .stderr(Stdio::piped()) .spawn()?;
let proc_result = proc.wait_with_output()?; if let Some(code) = return_code { assert_eq!(proc_result.status.code().unwrap(), *code); } let out = String::from_utf8(proc_result.stderr)?;
results.insert(name.to_string(), test_util::parse_max_mem(&out).unwrap()); }
Ok(results)}
fn cargo_deps() -> usize { let cargo_lock = test_util::root_path().join("Cargo.lock"); let mut count = 0; let file = std::fs::File::open(cargo_lock).unwrap(); use std::io::BufRead; for line in std::io::BufReader::new(file).lines() { if line.unwrap().starts_with("[[package]]") { count += 1 } } println!("cargo_deps {}", count); assert!(count > 10); // Sanity check. count}
#[derive(Default, Serialize)]struct BenchResult { created_at: String, sha1: String,
// TODO(ry) The "benchmark" benchmark should actually be called "exec_time". // When this is changed, the historical data in gh-pages branch needs to be // changed too. benchmark: HashMap<String, HashMap<String, f64>>, binary_size: HashMap<String, u64>, bundle_size: HashMap<String, u64>, cargo_deps: usize, max_latency: HashMap<String, f64>, max_memory: HashMap<String, u64>, lsp_exec_time: HashMap<String, u64>, req_per_sec: HashMap<String, u64>, syscall_count: HashMap<String, u64>, thread_count: HashMap<String, u64>,}
/* TODO(SyrupThinker) Switch to the #[bench] attribute once it is stabilized. Before that the #[test] tests won't be run because we replace the harness with our own runner here.*/fn main() -> Result<()> { if !env::args().any(|s| s == "--bench") { return Ok(()); }
println!("Starting Deno benchmark");
let target_dir = test_util::target_dir(); let deno_exe = test_util::deno_exe_path();
env::set_current_dir(&test_util::root_path())?;
let mut new_data = BenchResult { created_at: chrono::Utc::now() .to_rfc3339_opts(chrono::SecondsFormat::Secs, true), sha1: test_util::run_collect( &["git", "rev-parse", "HEAD"], None, None, None, true, ) .0 .trim() .to_string(), benchmark: run_exec_time(&deno_exe, &target_dir)?, binary_size: get_binary_sizes(&target_dir)?, bundle_size: bundle_benchmark(&deno_exe)?, cargo_deps: cargo_deps(), lsp_exec_time: lsp::benchmarks(&deno_exe)?, ..Default::default() };
if cfg!(not(target_os = "windows")) { run_http(&target_dir, &mut new_data)?; }
if cfg!(target_os = "linux") { run_strace_benchmarks(&deno_exe, &mut new_data)?; new_data.max_memory = run_max_mem_benchmark(&deno_exe)?; }
println!("===== <BENCHMARK RESULTS>"); serde_json::to_writer_pretty(std::io::stdout(), &new_data)?; println!("\n===== </BENCHMARK RESULTS>");
if let Some(filename) = target_dir.join("bench.json").to_str() { write_json(filename, &serde_json::to_value(&new_data)?)?; } else { eprintln!("Cannot write bench.json, path is invalid"); }
Ok(())}
pub type Result<T> = std::result::Result<T, AnyError>;