Azren/moved to lib (#40)
This commit is contained in:
@@ -23,7 +23,7 @@
|
||||
//!
|
||||
//! This produces graphs that look roughly like, for two levels:
|
||||
//!
|
||||
//! ```
|
||||
//! ``` ignored
|
||||
//! L2 <-------------------- L2 <---------- ...
|
||||
//! ^--- L1 <--- L1 <--- L1 ^--- L1 <--- L1 <--- L1
|
||||
//! ```
|
||||
|
||||
393
src/lib.rs
Normal file
393
src/lib.rs
Normal file
@@ -0,0 +1,393 @@
|
||||
// Copyright 2018 New Vector Ltd
|
||||
//
|
||||
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||
// you may not use this file except in compliance with the License.
|
||||
// You may obtain a copy of the License at
|
||||
//
|
||||
// http://www.apache.org/licenses/LICENSE-2.0
|
||||
//
|
||||
// Unless required by applicable law or agreed to in writing, software
|
||||
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
// See the License for the specific language governing permissions and
|
||||
// limitations under the License.
|
||||
|
||||
//! This is a tool that attempts to further compress state maps within a
|
||||
//! Synapse instance's database. Specifically, it aims to reduce the number of
|
||||
//! rows that a given room takes up in the `state_groups_state` table.
|
||||
|
||||
mod compressor;
|
||||
mod database;
|
||||
|
||||
#[global_allocator]
|
||||
static GLOBAL: jemallocator::Jemalloc = jemallocator::Jemalloc;
|
||||
|
||||
use compressor::Compressor;
|
||||
use database::PGEscape;
|
||||
|
||||
use clap::{
|
||||
crate_authors, crate_description, crate_name, crate_version, value_t_or_exit, App, Arg,
|
||||
};
|
||||
use indicatif::{ProgressBar, ProgressStyle};
|
||||
use rayon::prelude::*;
|
||||
use state_map::StateMap;
|
||||
use std::{collections::BTreeMap, fs::File, io::Write, str::FromStr};
|
||||
use string_cache::DefaultAtom as Atom;
|
||||
|
||||
/// An entry for a state group. Consists of an (optional) previous group and the
|
||||
/// delta from that previous group (or the full state if no previous group)
|
||||
#[derive(Default, Debug, Clone, PartialEq, Eq)]
|
||||
pub struct StateGroupEntry {
|
||||
prev_state_group: Option<i64>,
|
||||
state_map: StateMap<Atom>,
|
||||
}
|
||||
|
||||
/// Gets the full state for a given group from the map (of deltas)
|
||||
fn collapse_state_maps(map: &BTreeMap<i64, StateGroupEntry>, state_group: i64) -> StateMap<Atom> {
|
||||
let mut entry = &map[&state_group];
|
||||
let mut state_map = StateMap::new();
|
||||
|
||||
let mut stack = vec![state_group];
|
||||
|
||||
while let Some(prev_state_group) = entry.prev_state_group {
|
||||
stack.push(prev_state_group);
|
||||
if !map.contains_key(&prev_state_group) {
|
||||
panic!("Missing {}", prev_state_group);
|
||||
}
|
||||
entry = &map[&prev_state_group];
|
||||
}
|
||||
|
||||
for sg in stack.iter().rev() {
|
||||
state_map.extend(
|
||||
map[&sg]
|
||||
.state_map
|
||||
.iter()
|
||||
.map(|((t, s), e)| ((t, s), e.clone())),
|
||||
);
|
||||
}
|
||||
|
||||
state_map
|
||||
}
|
||||
|
||||
/// Helper struct for parsing the `level_sizes` argument.
|
||||
struct LevelSizes(Vec<usize>);
|
||||
|
||||
impl FromStr for LevelSizes {
|
||||
type Err = &'static str;
|
||||
|
||||
fn from_str(s: &str) -> Result<Self, Self::Err> {
|
||||
let mut sizes = Vec::new();
|
||||
|
||||
for size_str in s.split(',') {
|
||||
let size: usize = size_str
|
||||
.parse()
|
||||
.map_err(|_| "Not a comma separated list of numbers")?;
|
||||
sizes.push(size);
|
||||
}
|
||||
|
||||
Ok(LevelSizes(sizes))
|
||||
}
|
||||
}
|
||||
|
||||
pub struct Config {
|
||||
db_url: String,
|
||||
output_file: Option<File>,
|
||||
room_id: String,
|
||||
max_state_group: Option<i64>,
|
||||
min_saved_rows: Option<i32>,
|
||||
transactions: bool,
|
||||
level_sizes: LevelSizes,
|
||||
}
|
||||
|
||||
impl Config {
|
||||
pub fn parse_arguments() -> Config {
|
||||
let matches = App::new(crate_name!())
|
||||
.version(crate_version!())
|
||||
.author(crate_authors!("\n"))
|
||||
.about(crate_description!())
|
||||
.arg(
|
||||
Arg::with_name("postgres-url")
|
||||
.short("p")
|
||||
.value_name("URL")
|
||||
.help("The url for connecting to the postgres database")
|
||||
.takes_value(true)
|
||||
.required(true),
|
||||
).arg(
|
||||
Arg::with_name("room_id")
|
||||
.short("r")
|
||||
.value_name("ROOM_ID")
|
||||
.help("The room to process")
|
||||
.takes_value(true)
|
||||
.required(true),
|
||||
).arg(
|
||||
Arg::with_name("max_state_group")
|
||||
.short("s")
|
||||
.value_name("MAX_STATE_GROUP")
|
||||
.help("The maximum state group to process up to")
|
||||
.takes_value(true)
|
||||
.required(false),
|
||||
).arg(
|
||||
Arg::with_name("min_saved_rows")
|
||||
.short("m")
|
||||
.value_name("COUNT")
|
||||
.help("Suppress output if fewer than COUNT rows would be saved")
|
||||
.takes_value(true)
|
||||
.required(false),
|
||||
).arg(
|
||||
Arg::with_name("output_file")
|
||||
.short("o")
|
||||
.value_name("FILE")
|
||||
.help("File to output the changes to in SQL")
|
||||
.takes_value(true),
|
||||
).arg(
|
||||
Arg::with_name("transactions")
|
||||
.short("t")
|
||||
.help("Whether to wrap each state group change in a transaction")
|
||||
.requires("output_file"),
|
||||
).arg(
|
||||
Arg::with_name("level_sizes")
|
||||
.short("l")
|
||||
.value_name("LEVELS")
|
||||
.help("Sizes of each new level in the compression algorithm, as a comma separated list.")
|
||||
.long_help(concat!(
|
||||
"Sizes of each new level in the compression algorithm, as a comma separated list.",
|
||||
" The first entry in the list is for the lowest, most granular level,",
|
||||
" with each subsequent entry being for the next highest level.",
|
||||
" The number of entries in the list determines the number of levels",
|
||||
" that will be used.",
|
||||
"\nThe sum of the sizes of the levels effect the performance of fetching the state",
|
||||
" from the database, as the sum of the sizes is the upper bound on number of",
|
||||
" iterations needed to fetch a given set of state.",
|
||||
))
|
||||
.default_value("100,50,25")
|
||||
.takes_value(true),
|
||||
).get_matches();
|
||||
|
||||
let db_url = matches
|
||||
.value_of("postgres-url")
|
||||
.expect("db url should be required");
|
||||
|
||||
let output_file = matches
|
||||
.value_of("output_file")
|
||||
.map(|path| File::create(path).unwrap());
|
||||
|
||||
let room_id = matches
|
||||
.value_of("room_id")
|
||||
.expect("room_id should be required since no file");
|
||||
|
||||
let max_state_group = matches
|
||||
.value_of("max_state_group")
|
||||
.map(|s| s.parse().expect("max_state_group must be an integer"));
|
||||
|
||||
let min_saved_rows = matches
|
||||
.value_of("min_saved_rows")
|
||||
.map(|v| v.parse().expect("COUNT must be an integer"));
|
||||
|
||||
let transactions = matches.is_present("transactions");
|
||||
|
||||
let level_sizes = value_t_or_exit!(matches, "level_sizes", LevelSizes);
|
||||
|
||||
Config {
|
||||
db_url: String::from(db_url),
|
||||
output_file,
|
||||
room_id: String::from(room_id),
|
||||
max_state_group,
|
||||
min_saved_rows,
|
||||
transactions,
|
||||
level_sizes,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub fn run(mut config: Config) {
|
||||
// let mut config = Config::parse_arguments();
|
||||
|
||||
// First we need to get the current state groups
|
||||
println!("Fetching state from DB for room '{}'...", config.room_id);
|
||||
|
||||
let state_group_map =
|
||||
database::get_data_from_db(&config.db_url, &config.room_id, config.max_state_group);
|
||||
|
||||
println!("Number of state groups: {}", state_group_map.len());
|
||||
|
||||
let original_summed_size = state_group_map
|
||||
.iter()
|
||||
.fold(0, |acc, (_, v)| acc + v.state_map.len());
|
||||
|
||||
println!("Number of rows in current table: {}", original_summed_size);
|
||||
|
||||
// Now we actually call the compression algorithm.
|
||||
|
||||
println!("Compressing state...");
|
||||
|
||||
let compressor = Compressor::compress(&state_group_map, &config.level_sizes.0);
|
||||
|
||||
let new_state_group_map = compressor.new_state_group_map;
|
||||
|
||||
// Done! Now to print a bunch of stats.
|
||||
|
||||
let compressed_summed_size = new_state_group_map
|
||||
.iter()
|
||||
.fold(0, |acc, (_, v)| acc + v.state_map.len());
|
||||
|
||||
let ratio = (compressed_summed_size as f64) / (original_summed_size as f64);
|
||||
|
||||
println!(
|
||||
"Number of rows after compression: {} ({:.2}%)",
|
||||
compressed_summed_size,
|
||||
ratio * 100.
|
||||
);
|
||||
|
||||
println!("Compression Statistics:");
|
||||
println!(
|
||||
" Number of forced resets due to lacking prev: {}",
|
||||
compressor.stats.resets_no_suitable_prev
|
||||
);
|
||||
println!(
|
||||
" Number of compressed rows caused by the above: {}",
|
||||
compressor.stats.resets_no_suitable_prev_size
|
||||
);
|
||||
println!(
|
||||
" Number of state groups changed: {}",
|
||||
compressor.stats.state_groups_changed
|
||||
);
|
||||
|
||||
if let Some(min) = config.min_saved_rows {
|
||||
let saving = (original_summed_size - compressed_summed_size) as i32;
|
||||
if saving < min {
|
||||
println!(
|
||||
"Only {} rows would be saved by this compression. Skipping output.",
|
||||
saving
|
||||
);
|
||||
return;
|
||||
}
|
||||
}
|
||||
|
||||
check_that_maps_match(&state_group_map, &new_state_group_map);
|
||||
|
||||
// If we are given an output file, we output the changes as SQL. If the
|
||||
// `transactions` argument is set we wrap each change to a state group in a
|
||||
// transaction.
|
||||
|
||||
output_sql(&mut config, &state_group_map, &new_state_group_map);
|
||||
}
|
||||
|
||||
fn output_sql(
|
||||
config: &mut Config,
|
||||
old_map: &BTreeMap<i64, StateGroupEntry>,
|
||||
new_map: &BTreeMap<i64, StateGroupEntry>,
|
||||
) {
|
||||
if config.output_file.is_none() {
|
||||
return;
|
||||
}
|
||||
|
||||
println!("Writing changes...");
|
||||
|
||||
let pb = ProgressBar::new(old_map.len() as u64);
|
||||
pb.set_style(
|
||||
ProgressStyle::default_bar().template("[{elapsed_precise}] {bar} {pos}/{len} {msg}"),
|
||||
);
|
||||
pb.set_message("state groups");
|
||||
pb.enable_steady_tick(100);
|
||||
|
||||
if let Some(output) = &mut config.output_file {
|
||||
for (sg, old_entry) in old_map {
|
||||
let new_entry = &new_map[sg];
|
||||
|
||||
if old_entry != new_entry {
|
||||
if config.transactions {
|
||||
writeln!(output, "BEGIN;").unwrap();
|
||||
}
|
||||
|
||||
writeln!(
|
||||
output,
|
||||
"DELETE FROM state_group_edges WHERE state_group = {};",
|
||||
sg
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
if let Some(prev_sg) = new_entry.prev_state_group {
|
||||
writeln!(output, "INSERT INTO state_group_edges (state_group, prev_state_group) VALUES ({}, {});", sg, prev_sg).unwrap();
|
||||
}
|
||||
|
||||
writeln!(
|
||||
output,
|
||||
"DELETE FROM state_groups_state WHERE state_group = {};",
|
||||
sg
|
||||
)
|
||||
.unwrap();
|
||||
if !new_entry.state_map.is_empty() {
|
||||
writeln!(output, "INSERT INTO state_groups_state (state_group, room_id, type, state_key, event_id) VALUES").unwrap();
|
||||
let mut first = true;
|
||||
for ((t, s), e) in new_entry.state_map.iter() {
|
||||
if first {
|
||||
write!(output, " ").unwrap();
|
||||
first = false;
|
||||
} else {
|
||||
write!(output, " ,").unwrap();
|
||||
}
|
||||
writeln!(
|
||||
output,
|
||||
"({}, {}, {}, {}, {})",
|
||||
sg,
|
||||
PGEscape(&config.room_id),
|
||||
PGEscape(t),
|
||||
PGEscape(s),
|
||||
PGEscape(e)
|
||||
)
|
||||
.unwrap();
|
||||
}
|
||||
writeln!(output, ";").unwrap();
|
||||
}
|
||||
|
||||
if config.transactions {
|
||||
writeln!(output, "COMMIT;").unwrap();
|
||||
}
|
||||
writeln!(output).unwrap();
|
||||
}
|
||||
|
||||
pb.inc(1);
|
||||
}
|
||||
}
|
||||
|
||||
pb.finish();
|
||||
}
|
||||
|
||||
fn check_that_maps_match(
|
||||
old_map: &BTreeMap<i64, StateGroupEntry>,
|
||||
new_map: &BTreeMap<i64, StateGroupEntry>,
|
||||
) {
|
||||
println!("Checking that state maps match...");
|
||||
|
||||
let pb = ProgressBar::new(old_map.len() as u64);
|
||||
pb.set_style(
|
||||
ProgressStyle::default_bar().template("[{elapsed_precise}] {bar} {pos}/{len} {msg}"),
|
||||
);
|
||||
pb.set_message("state groups");
|
||||
pb.enable_steady_tick(100);
|
||||
|
||||
// Now let's iterate through and assert that the state for each group
|
||||
// matches between the two versions.
|
||||
old_map
|
||||
.par_iter() // This uses rayon to run the checks in parallel
|
||||
.try_for_each(|(sg, _)| {
|
||||
let expected = collapse_state_maps(&old_map, *sg);
|
||||
let actual = collapse_state_maps(&new_map, *sg);
|
||||
|
||||
pb.inc(1);
|
||||
|
||||
if expected != actual {
|
||||
println!("State Group: {}", sg);
|
||||
println!("Expected: {:#?}", expected);
|
||||
println!("actual: {:#?}", actual);
|
||||
Err(format!("State for group {} do not match", sg))
|
||||
} else {
|
||||
Ok(())
|
||||
}
|
||||
})
|
||||
.expect("expected state to match");
|
||||
|
||||
pb.finish();
|
||||
|
||||
println!("New state map matches old one");
|
||||
}
|
||||
377
src/main.rs
377
src/main.rs
@@ -16,381 +16,8 @@
|
||||
//! Synapse instance's database. Specifically, it aims to reduce the number of
|
||||
//! rows that a given room takes up in the `state_groups_state` table.
|
||||
|
||||
mod compressor;
|
||||
mod database;
|
||||
|
||||
#[global_allocator]
|
||||
static GLOBAL: jemallocator::Jemalloc = jemallocator::Jemalloc;
|
||||
|
||||
use compressor::Compressor;
|
||||
use database::PGEscape;
|
||||
|
||||
use clap::{
|
||||
crate_authors, crate_description, crate_name, crate_version, value_t_or_exit, App, Arg,
|
||||
};
|
||||
use indicatif::{ProgressBar, ProgressStyle};
|
||||
use rayon::prelude::*;
|
||||
use state_map::StateMap;
|
||||
use std::{collections::BTreeMap, fs::File, io::Write, str::FromStr};
|
||||
use string_cache::DefaultAtom as Atom;
|
||||
|
||||
/// An entry for a state group. Consists of an (optional) previous group and the
|
||||
/// delta from that previous group (or the full state if no previous group)
|
||||
#[derive(Default, Debug, Clone, PartialEq, Eq)]
|
||||
pub struct StateGroupEntry {
|
||||
prev_state_group: Option<i64>,
|
||||
state_map: StateMap<Atom>,
|
||||
}
|
||||
|
||||
/// Gets the full state for a given group from the map (of deltas)
|
||||
pub fn collapse_state_maps(
|
||||
map: &BTreeMap<i64, StateGroupEntry>,
|
||||
state_group: i64,
|
||||
) -> StateMap<Atom> {
|
||||
let mut entry = &map[&state_group];
|
||||
let mut state_map = StateMap::new();
|
||||
|
||||
let mut stack = vec![state_group];
|
||||
|
||||
while let Some(prev_state_group) = entry.prev_state_group {
|
||||
stack.push(prev_state_group);
|
||||
if !map.contains_key(&prev_state_group) {
|
||||
panic!("Missing {}", prev_state_group);
|
||||
}
|
||||
entry = &map[&prev_state_group];
|
||||
}
|
||||
|
||||
for sg in stack.iter().rev() {
|
||||
state_map.extend(
|
||||
map[&sg]
|
||||
.state_map
|
||||
.iter()
|
||||
.map(|((t, s), e)| ((t, s), e.clone())),
|
||||
);
|
||||
}
|
||||
|
||||
state_map
|
||||
}
|
||||
|
||||
/// Helper struct for parsing the `level_sizes` argument.
|
||||
struct LevelSizes(Vec<usize>);
|
||||
|
||||
impl FromStr for LevelSizes {
|
||||
type Err = &'static str;
|
||||
|
||||
fn from_str(s: &str) -> Result<Self, Self::Err> {
|
||||
let mut sizes = Vec::new();
|
||||
|
||||
for size_str in s.split(',') {
|
||||
let size: usize = size_str
|
||||
.parse()
|
||||
.map_err(|_| "Not a comma separated list of numbers")?;
|
||||
sizes.push(size);
|
||||
}
|
||||
|
||||
Ok(LevelSizes(sizes))
|
||||
}
|
||||
}
|
||||
|
||||
struct Config {
|
||||
db_url: String,
|
||||
output_file: Option<File>,
|
||||
room_id: String,
|
||||
max_state_group: Option<i64>,
|
||||
min_saved_rows: Option<i32>,
|
||||
transactions: bool,
|
||||
level_sizes: LevelSizes,
|
||||
}
|
||||
|
||||
impl Config {
|
||||
fn parse_arguments() -> Config {
|
||||
let matches = App::new(crate_name!())
|
||||
.version(crate_version!())
|
||||
.author(crate_authors!("\n"))
|
||||
.about(crate_description!())
|
||||
.arg(
|
||||
Arg::with_name("postgres-url")
|
||||
.short("p")
|
||||
.value_name("URL")
|
||||
.help("The url for connecting to the postgres database")
|
||||
.takes_value(true)
|
||||
.required(true),
|
||||
).arg(
|
||||
Arg::with_name("room_id")
|
||||
.short("r")
|
||||
.value_name("ROOM_ID")
|
||||
.help("The room to process")
|
||||
.takes_value(true)
|
||||
.required(true),
|
||||
).arg(
|
||||
Arg::with_name("max_state_group")
|
||||
.short("s")
|
||||
.value_name("MAX_STATE_GROUP")
|
||||
.help("The maximum state group to process up to")
|
||||
.takes_value(true)
|
||||
.required(false),
|
||||
).arg(
|
||||
Arg::with_name("min_saved_rows")
|
||||
.short("m")
|
||||
.value_name("COUNT")
|
||||
.help("Suppress output if fewer than COUNT rows would be saved")
|
||||
.takes_value(true)
|
||||
.required(false),
|
||||
).arg(
|
||||
Arg::with_name("output_file")
|
||||
.short("o")
|
||||
.value_name("FILE")
|
||||
.help("File to output the changes to in SQL")
|
||||
.takes_value(true),
|
||||
).arg(
|
||||
Arg::with_name("transactions")
|
||||
.short("t")
|
||||
.help("Whether to wrap each state group change in a transaction")
|
||||
.requires("output_file"),
|
||||
).arg(
|
||||
Arg::with_name("level_sizes")
|
||||
.short("l")
|
||||
.value_name("LEVELS")
|
||||
.help("Sizes of each new level in the compression algorithm, as a comma separated list.")
|
||||
.long_help(concat!(
|
||||
"Sizes of each new level in the compression algorithm, as a comma separated list.",
|
||||
" The first entry in the list is for the lowest, most granular level,",
|
||||
" with each subsequent entry being for the next highest level.",
|
||||
" The number of entries in the list determines the number of levels",
|
||||
" that will be used.",
|
||||
"\nThe sum of the sizes of the levels effect the performance of fetching the state",
|
||||
" from the database, as the sum of the sizes is the upper bound on number of",
|
||||
" iterations needed to fetch a given set of state.",
|
||||
))
|
||||
.default_value("100,50,25")
|
||||
.takes_value(true),
|
||||
).get_matches();
|
||||
|
||||
let db_url = matches
|
||||
.value_of("postgres-url")
|
||||
.expect("db url should be required");
|
||||
|
||||
let output_file = matches
|
||||
.value_of("output_file")
|
||||
.map(|path| File::create(path).unwrap());
|
||||
|
||||
let room_id = matches
|
||||
.value_of("room_id")
|
||||
.expect("room_id should be required since no file");
|
||||
|
||||
let max_state_group = matches
|
||||
.value_of("max_state_group")
|
||||
.map(|s| s.parse().expect("max_state_group must be an integer"));
|
||||
|
||||
let min_saved_rows = matches
|
||||
.value_of("min_saved_rows")
|
||||
.map(|v| v.parse().expect("COUNT must be an integer"));
|
||||
|
||||
let transactions = matches.is_present("transactions");
|
||||
|
||||
let level_sizes = value_t_or_exit!(matches, "level_sizes", LevelSizes);
|
||||
|
||||
Config {
|
||||
db_url: String::from(db_url),
|
||||
output_file,
|
||||
room_id: String::from(room_id),
|
||||
max_state_group,
|
||||
min_saved_rows,
|
||||
transactions,
|
||||
level_sizes,
|
||||
}
|
||||
}
|
||||
}
|
||||
use synapse_compress_state as comp_state;
|
||||
|
||||
fn main() {
|
||||
let mut config = Config::parse_arguments();
|
||||
|
||||
// First we need to get the current state groups
|
||||
println!("Fetching state from DB for room '{}'...", config.room_id);
|
||||
|
||||
let state_group_map =
|
||||
database::get_data_from_db(&config.db_url, &config.room_id, config.max_state_group);
|
||||
|
||||
println!("Number of state groups: {}", state_group_map.len());
|
||||
|
||||
let original_summed_size = state_group_map
|
||||
.iter()
|
||||
.fold(0, |acc, (_, v)| acc + v.state_map.len());
|
||||
|
||||
println!("Number of rows in current table: {}", original_summed_size);
|
||||
|
||||
// Now we actually call the compression algorithm.
|
||||
|
||||
println!("Compressing state...");
|
||||
|
||||
let compressor = Compressor::compress(&state_group_map, &config.level_sizes.0);
|
||||
|
||||
let new_state_group_map = compressor.new_state_group_map;
|
||||
|
||||
// Done! Now to print a bunch of stats.
|
||||
|
||||
let compressed_summed_size = new_state_group_map
|
||||
.iter()
|
||||
.fold(0, |acc, (_, v)| acc + v.state_map.len());
|
||||
|
||||
let ratio = (compressed_summed_size as f64) / (original_summed_size as f64);
|
||||
|
||||
println!(
|
||||
"Number of rows after compression: {} ({:.2}%)",
|
||||
compressed_summed_size,
|
||||
ratio * 100.
|
||||
);
|
||||
|
||||
println!("Compression Statistics:");
|
||||
println!(
|
||||
" Number of forced resets due to lacking prev: {}",
|
||||
compressor.stats.resets_no_suitable_prev
|
||||
);
|
||||
println!(
|
||||
" Number of compressed rows caused by the above: {}",
|
||||
compressor.stats.resets_no_suitable_prev_size
|
||||
);
|
||||
println!(
|
||||
" Number of state groups changed: {}",
|
||||
compressor.stats.state_groups_changed
|
||||
);
|
||||
|
||||
if let Some(min) = config.min_saved_rows {
|
||||
let saving = (original_summed_size - compressed_summed_size) as i32;
|
||||
if saving < min {
|
||||
println!(
|
||||
"Only {} rows would be saved by this compression. Skipping output.",
|
||||
saving
|
||||
);
|
||||
return;
|
||||
}
|
||||
}
|
||||
|
||||
check_that_maps_match(&state_group_map, &new_state_group_map);
|
||||
|
||||
// If we are given an output file, we output the changes as SQL. If the
|
||||
// `transactions` argument is set we wrap each change to a state group in a
|
||||
// transaction.
|
||||
|
||||
output_sql(&mut config, &state_group_map, &new_state_group_map);
|
||||
}
|
||||
|
||||
fn output_sql(
|
||||
config: &mut Config,
|
||||
old_map: &BTreeMap<i64, StateGroupEntry>,
|
||||
new_map: &BTreeMap<i64, StateGroupEntry>,
|
||||
) {
|
||||
if config.output_file.is_none() {
|
||||
return;
|
||||
}
|
||||
|
||||
println!("Writing changes...");
|
||||
|
||||
let pb = ProgressBar::new(old_map.len() as u64);
|
||||
pb.set_style(
|
||||
ProgressStyle::default_bar().template("[{elapsed_precise}] {bar} {pos}/{len} {msg}"),
|
||||
);
|
||||
pb.set_message("state groups");
|
||||
pb.enable_steady_tick(100);
|
||||
|
||||
if let Some(output) = &mut config.output_file {
|
||||
for (sg, old_entry) in old_map {
|
||||
let new_entry = &new_map[sg];
|
||||
|
||||
if old_entry != new_entry {
|
||||
if config.transactions {
|
||||
writeln!(output, "BEGIN;").unwrap();
|
||||
}
|
||||
|
||||
writeln!(
|
||||
output,
|
||||
"DELETE FROM state_group_edges WHERE state_group = {};",
|
||||
sg
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
if let Some(prev_sg) = new_entry.prev_state_group {
|
||||
writeln!(output, "INSERT INTO state_group_edges (state_group, prev_state_group) VALUES ({}, {});", sg, prev_sg).unwrap();
|
||||
}
|
||||
|
||||
writeln!(
|
||||
output,
|
||||
"DELETE FROM state_groups_state WHERE state_group = {};",
|
||||
sg
|
||||
)
|
||||
.unwrap();
|
||||
if !new_entry.state_map.is_empty() {
|
||||
writeln!(output, "INSERT INTO state_groups_state (state_group, room_id, type, state_key, event_id) VALUES").unwrap();
|
||||
let mut first = true;
|
||||
for ((t, s), e) in new_entry.state_map.iter() {
|
||||
if first {
|
||||
write!(output, " ").unwrap();
|
||||
first = false;
|
||||
} else {
|
||||
write!(output, " ,").unwrap();
|
||||
}
|
||||
writeln!(
|
||||
output,
|
||||
"({}, {}, {}, {}, {})",
|
||||
sg,
|
||||
PGEscape(&config.room_id),
|
||||
PGEscape(t),
|
||||
PGEscape(s),
|
||||
PGEscape(e)
|
||||
)
|
||||
.unwrap();
|
||||
}
|
||||
writeln!(output, ";").unwrap();
|
||||
}
|
||||
|
||||
if config.transactions {
|
||||
writeln!(output, "COMMIT;").unwrap();
|
||||
}
|
||||
writeln!(output).unwrap();
|
||||
}
|
||||
|
||||
pb.inc(1);
|
||||
}
|
||||
}
|
||||
|
||||
pb.finish();
|
||||
}
|
||||
|
||||
fn check_that_maps_match(
|
||||
old_map: &BTreeMap<i64, StateGroupEntry>,
|
||||
new_map: &BTreeMap<i64, StateGroupEntry>,
|
||||
) {
|
||||
println!("Checking that state maps match...");
|
||||
|
||||
let pb = ProgressBar::new(old_map.len() as u64);
|
||||
pb.set_style(
|
||||
ProgressStyle::default_bar().template("[{elapsed_precise}] {bar} {pos}/{len} {msg}"),
|
||||
);
|
||||
pb.set_message("state groups");
|
||||
pb.enable_steady_tick(100);
|
||||
|
||||
// Now let's iterate through and assert that the state for each group
|
||||
// matches between the two versions.
|
||||
old_map
|
||||
.par_iter() // This uses rayon to run the checks in parallel
|
||||
.try_for_each(|(sg, _)| {
|
||||
let expected = collapse_state_maps(&old_map, *sg);
|
||||
let actual = collapse_state_maps(&new_map, *sg);
|
||||
|
||||
pb.inc(1);
|
||||
|
||||
if expected != actual {
|
||||
println!("State Group: {}", sg);
|
||||
println!("Expected: {:#?}", expected);
|
||||
println!("actual: {:#?}", actual);
|
||||
Err(format!("State for group {} do not match", sg))
|
||||
} else {
|
||||
Ok(())
|
||||
}
|
||||
})
|
||||
.expect("expected state to match");
|
||||
|
||||
pb.finish();
|
||||
|
||||
println!("New state map matches old one");
|
||||
comp_state::run(comp_state::Config::parse_arguments());
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user