pub(crate) mod data; use std::{ collections::{BTreeMap, hash_map::Entry}, fs::File, hash::BuildHasherDefault, path::{Path, PathBuf}, time::{Duration, UNIX_EPOCH}, }; use anyhow::{Context, anyhow}; use chrono::{DateTime, Utc}; use indexmap::IndexSet; use rustc_hash::{FxHashMap, FxHasher}; use walkdir::WalkDir; use self::data::Benchmark; use crate::summarize_bench::data::{BaseBenchmarks, CStats}; type FxIndexSet = IndexSet>; #[derive(Debug)] struct BenchDataFile { path: PathBuf, sha: String, timestamp: u64, system: String, } impl<'a> TryFrom<&'a Path> for BenchDataFile { type Error = &'static str; fn try_from(path: &'a Path) -> Result { let key_dir = path.parent().ok_or("invalid structure")?; let sha_dir = key_dir.parent().ok_or("invalid structure")?; let system_dir = sha_dir.parent().ok_or("invalid structure")?; let (timestamp, sha) = sha_dir .file_name() .ok_or("invalid filename")? .to_str() .ok_or("invalid chars in file name")? .split_once('-') .ok_or("missing dash in timestamp-sha directory")?; Ok(Self { path: path.to_path_buf(), sha: sha.to_string(), timestamp: timestamp.parse().map_err(|_| "unable to parse timestamp")?, system: system_dir .file_name() .ok_or("invalid filename")? .to_str() .ok_or("invalid chars in file name")? .to_string(), }) } } pub fn process_all(path: PathBuf) { let mut map = FxHashMap::default(); for entry in WalkDir::new(&path) .into_iter() .filter_map(Result::ok) .filter(|e| { e.file_type().is_file() && e.file_name().to_str().map(|n| n.ends_with("raw.json")) == Some(true) }) { let data_file_result: Result = entry.path().try_into(); if let Ok(data_file) = data_file_result { let inner_map: &mut FxHashMap<_, _> = map.entry(data_file.system.clone()).or_default(); let items: &mut Vec = inner_map.entry(data_file.sha.clone()).or_default(); items.push(data_file); } } for (system, inner_map) in map { let mut latest: FxHashMap = FxHashMap::default(); for (sha, data_files) in inner_map { let min_ts = data_files.iter().map(|d| d.timestamp).min().unwrap(); let max_ts = data_files.iter().map(|d| d.timestamp).max().unwrap(); let mut items = data_files .iter() .map(|data_file| { let file = File::open(&data_file.path).unwrap(); let reader = std::io::BufReader::new(file); let data: BaseBenchmarks = serde_json::from_reader(reader) .with_context(|| anyhow!("unable to read {}", data_file.path.display())) .unwrap(); data }) .collect::>(); let mut by_name = FxHashMap::default(); for (i, data) in items.iter().enumerate() { for (name, bench) in data.benchmarks.iter() { let list: &mut Vec<_> = by_name.entry(name).or_default(); list.push((i, bench)); } } let mut normalization_state = items.iter().map(|_| (0.0, 0)).collect::>(); for list in by_name.values().filter(|l| l.len() > 1) { let avg = list .iter() .map(|(_, b)| b.estimates.mean.point_estimate) .sum::() / list.len() as f64; for (i, b) in list { let correction = avg / b.estimates.mean.point_estimate; let (sum, count) = &mut normalization_state[*i]; *sum += correction; *count += 1; } } for (i, benches) in items.iter_mut().enumerate() { let (sum, count) = normalization_state[i]; if count <= 1 { continue; } let correction = sum / count as f64; for bench in benches.benchmarks.values_mut() { fn correct(s: &mut CStats, f: f64) { s.point_estimate *= f; s.standard_error *= f; s.confidence_interval.lower_bound *= f; s.confidence_interval.upper_bound *= f; } correct(&mut bench.estimates.mean, correction); correct(&mut bench.estimates.median, correction); correct(&mut bench.estimates.median_abs_dev, correction); if let Some(slope) = bench.estimates.slope.as_mut() { correct(slope, correction); } correct(&mut bench.estimates.std_dev, correction); } // let data_file = &data_files[i]; // benches.name = format!("{}-{sha}", data_file.timestamp); // let normalized = // data_file.path.parent().unwrap().join("normalized.json"); // let file = File::create(&normalized).unwrap(); // let writer = std::io::BufWriter::new(file); // serde_json::to_writer_pretty(writer, benches).unwrap(); } let baseline = &sha[..7]; let mut latest_for_sha = BaseBenchmarks { name: baseline.to_string(), benchmarks: BTreeMap::new(), }; let all_benchmark_keys = items .iter() .flat_map(|b| b.benchmarks.keys()) .collect::>(); for key in all_benchmark_keys { let (ts, bench) = items .iter() .enumerate() .map(|(i, b)| (data_files[i].timestamp, b)) .filter_map(|(ts, benches)| benches.benchmarks.get(key).map(|b| (ts, b))) .max_by_key(|(ts, _)| *ts) .unwrap(); let ts = UNIX_EPOCH + Duration::from_secs(ts); let ts = DateTime::::from(ts); let ts = ts.format("%Y-%m-%d %H:%M"); let key_ts = format!("{key} ({ts})"); let key_ts_sha = format!("{key} ({ts}, {})", &sha[..7]); match latest.entry(key.to_string()) { Entry::Occupied(mut e) => { if e.get().0 < min_ts { e.insert((min_ts, key_ts_sha, bench.clone())); } } Entry::Vacant(e) => { e.insert((min_ts, key_ts_sha, bench.clone())); } } latest_for_sha.benchmarks.insert(key_ts, bench.clone()); } let latest_for_sha_path = data_files .first() .unwrap() .path .parent() .unwrap() .parent() .unwrap() .parent() .unwrap() .join(format!("{min_ts}-{max_ts}-{sha}.json")); let file = File::create(&latest_for_sha_path).unwrap(); let writer = std::io::BufWriter::new(file); serde_json::to_writer_pretty(writer, &latest_for_sha).unwrap(); println!("{}", latest_for_sha_path.display()); } let latest_for_system = BaseBenchmarks { name: "latest".to_string(), benchmarks: latest .into_values() .map(|(_, key, bench)| (key, bench)) .collect(), }; let latest_path = path.join(format!("{system}.json")); let file = File::create(&latest_path).unwrap(); let writer = std::io::BufWriter::new(file); serde_json::to_writer_pretty(writer, &latest_for_system).unwrap(); println!("{}", latest_path.display()); } }