// Copyright (c) 2022-2023 Niko Bonnieure, Par le Peuple, NextGraph.org developers // All rights reserved. // This code is partly derived from work written by TG x Thoth from P2Pcollab. // Copyright 2022 TG x Thoth // Licensed under the Apache License, Version 2.0 // // or the MIT license , // at your option. All files in the project carrying such // notice may not be copied, modified, or distributed except // according to those terms. //! Merkle hash tree of Objects use std::collections::{HashMap, HashSet}; use debug_print::*; use chacha20::cipher::{KeyIvInit, StreamCipher}; use chacha20::ChaCha20; use crate::store::*; use crate::types::*; /// Size of a serialized empty Block const EMPTY_BLOCK_SIZE: usize = 12; /// Size of a serialized BlockId const BLOCK_ID_SIZE: usize = 33; /// Size of serialized SymKey const BLOCK_KEY_SIZE: usize = 33; /// Size of serialized Object with deps reference. const EMPTY_ROOT_SIZE_DEPSREF: usize = 77; /// Extra size needed if depsRef used instead of deps list. const DEPSREF_OVERLOAD: usize = EMPTY_ROOT_SIZE_DEPSREF - EMPTY_BLOCK_SIZE; /// Varint extra bytes when reaching the maximum value we will ever use const BIG_VARINT_EXTRA: usize = 3; /// Varint extra bytes when reaching the maximum size of data byte arrays. const DATA_VARINT_EXTRA: usize = 4; /// Max extra space used by the deps list const MAX_DEPS_SIZE: usize = 8 * BLOCK_ID_SIZE; #[derive(Debug)] pub struct Object { /// Blocks of the Object (nodes of the tree) blocks: Vec, /// Dependencies deps: Vec, } /// Object parsing errors #[derive(Debug)] pub enum ObjectParseError { /// Missing blocks MissingBlocks(Vec), /// Missing root key MissingRootKey, /// Invalid BlockId encountered in the tree InvalidBlockId, /// Too many or too few children of a block InvalidChildren, /// Number of keys does not match number of children of a block InvalidKeys, /// Invalid DepList object content InvalidDeps, /// Error deserializing content of a block BlockDeserializeError, /// Error deserializing content of the object ObjectDeserializeError, } /// Object copy error #[derive(Debug)] pub enum ObjectCopyError { NotFound, ParseError, } impl Object { fn convergence_key(repo_pubkey: PubKey, repo_secret: SymKey) -> [u8; blake3::OUT_LEN] { let key_material = match (repo_pubkey, repo_secret) { (PubKey::Ed25519PubKey(pubkey), SymKey::ChaCha20Key(secret)) => { [pubkey, secret].concat() } }; blake3::derive_key("NextGraph Data BLAKE3 key", key_material.as_slice()) } fn make_block( content: &[u8], conv_key: &[u8; blake3::OUT_LEN], children: Vec, deps: ObjectDeps, expiry: Option, ) -> Block { let key_hash = blake3::keyed_hash(conv_key, content); let nonce = [0u8; 12]; let key = key_hash.as_bytes(); let mut cipher = ChaCha20::new(key.into(), &nonce.into()); let mut content_enc = Vec::from(content); let mut content_enc_slice = &mut content_enc.as_mut_slice(); cipher.apply_keystream(&mut content_enc_slice); let key = SymKey::ChaCha20Key(key.clone()); let block = Block::new(children, deps, expiry, content_enc, Some(key)); //debug_println!(">>> make_block:"); //debug_println!("!! id: {:?}", obj.id()); //debug_println!("!! children: ({}) {:?}", children.len(), children); block } fn make_deps( deps_vec: Vec, object_size: usize, repo_pubkey: PubKey, repo_secret: SymKey, ) -> ObjectDeps { if deps_vec.len() <= 8 { ObjectDeps::ObjectIdList(deps_vec) } else { let dep_list = DepList::V0(deps_vec); let dep_obj = Object::new( ObjectContent::DepList(dep_list), vec![], None, object_size, repo_pubkey, repo_secret, ); let dep_ref = ObjectRef { id: dep_obj.id(), key: dep_obj.key().unwrap(), }; ObjectDeps::DepListRef(dep_ref) } } /// Build tree from leaves, returns parent nodes fn make_tree( leaves: &[Block], conv_key: &ChaCha20Key, root_deps: &ObjectDeps, expiry: Option, arity: usize, ) -> Vec { let mut parents = vec![]; let chunks = leaves.chunks(arity); let mut it = chunks.peekable(); while let Some(nodes) = it.next() { let keys = nodes.iter().map(|block| block.key().unwrap()).collect(); let children = nodes.iter().map(|block| block.id()).collect(); let content = BlockContentV0::InternalNode(keys); let content_ser = serde_bare::to_vec(&content).unwrap(); let child_deps = ObjectDeps::ObjectIdList(vec![]); let deps = if parents.is_empty() && it.peek().is_none() { root_deps.clone() } else { child_deps }; parents.push(Self::make_block( content_ser.as_slice(), conv_key, children, deps, expiry, )); } //debug_println!("parents += {}", parents.len()); if 1 < parents.len() { let mut great_parents = Self::make_tree(parents.as_slice(), conv_key, root_deps, expiry, arity); parents.append(&mut great_parents); } parents } /// Create new Object from given content /// /// The Object is chunked and stored in a Merkle tree /// The arity of the Merkle tree is the maximum that fits in the given `max_object_size` /// /// Arguments: /// * `content`: Object content /// * `deps`: Dependencies of the object /// * `block_size`: Desired block size for chunking content, rounded up to nearest valid block size /// * `repo_pubkey`: Repository public key /// * `repo_secret`: Repository secret pub fn new( content: ObjectContent, deps: Vec, expiry: Option, block_size: usize, repo_pubkey: PubKey, repo_secret: SymKey, ) -> Object { // create blocks by chunking + encrypting content let valid_block_size = store_valid_value_size(block_size); let data_chunk_size = valid_block_size - EMPTY_BLOCK_SIZE - DATA_VARINT_EXTRA; let mut blocks: Vec = vec![]; let conv_key = Self::convergence_key(repo_pubkey, repo_secret); let obj_deps = Self::make_deps(deps.clone(), valid_block_size, repo_pubkey, repo_secret); let content_ser = serde_bare::to_vec(&content).unwrap(); if EMPTY_BLOCK_SIZE + DATA_VARINT_EXTRA + BLOCK_ID_SIZE * deps.len() + content_ser.len() <= valid_block_size { // content fits in root node let data_chunk = BlockContentV0::DataChunk(content_ser.clone()); let content_ser = serde_bare::to_vec(&data_chunk).unwrap(); blocks.push(Self::make_block( content_ser.as_slice(), &conv_key, vec![], obj_deps, expiry, )); } else { // chunk content and create leaf nodes for chunk in content_ser.chunks(data_chunk_size) { let data_chunk = BlockContentV0::DataChunk(chunk.to_vec()); let content_ser = serde_bare::to_vec(&data_chunk).unwrap(); blocks.push(Self::make_block( content_ser.as_slice(), &conv_key, vec![], ObjectDeps::ObjectIdList(vec![]), expiry, )); } // internal nodes // arity: max number of ObjectRefs that fit inside an InternalNode Object within the object_size limit let arity: usize = (valid_block_size - EMPTY_BLOCK_SIZE - BIG_VARINT_EXTRA * 2 - MAX_DEPS_SIZE) / (BLOCK_ID_SIZE + BLOCK_KEY_SIZE); let mut parents = Self::make_tree(blocks.as_slice(), &conv_key, &obj_deps, expiry, arity); blocks.append(&mut parents); } Object { blocks, deps } } pub fn copy( &self, expiry: Option, repo_pubkey: PubKey, repo_secret: SymKey, ) -> Result { // getting the old object from store let leaves: Vec = self.leaves().map_err(|_e| ObjectCopyError::ParseError)?; let conv_key = Self::convergence_key(repo_pubkey, repo_secret); let block_size = leaves.first().unwrap().content().len(); let valid_block_size = store_valid_value_size(block_size); let mut blocks: Vec = vec![]; for block in leaves { let mut copy = block.clone(); copy.set_expiry(expiry); blocks.push(copy); } // internal nodes // arity: max number of ObjectRefs that fit inside an InternalNode Object within the object_size limit let arity: usize = (valid_block_size - EMPTY_BLOCK_SIZE - BIG_VARINT_EXTRA * 2 - MAX_DEPS_SIZE) / (BLOCK_ID_SIZE + BLOCK_KEY_SIZE); let mut parents = Self::make_tree( blocks.as_slice(), &conv_key, self.root().deps(), expiry, arity, ); blocks.append(&mut parents); Ok(Object { blocks, deps: self.deps().clone(), }) } /// Load an Object from RepoStore /// /// Returns Ok(Object) or an Err(Vec) of missing BlockIds pub fn load( id: ObjectId, key: Option, store: &impl RepoStore, ) -> Result { fn load_tree( parents: Vec, store: &impl RepoStore, blocks: &mut Vec, missing: &mut Vec, ) { let mut children: Vec = vec![]; for id in parents { match store.get(&id) { Ok(block) => { blocks.insert(0, block.clone()); match block { Block::V0(o) => { children.extend(o.children.iter().rev()); } } } Err(_) => missing.push(id.clone()), } } if !children.is_empty() { load_tree(children, store, blocks, missing); } } let mut blocks: Vec = vec![]; let mut missing: Vec = vec![]; load_tree(vec![id], store, &mut blocks, &mut missing); if !missing.is_empty() { return Err(ObjectParseError::MissingBlocks(missing)); } let root = blocks.last_mut().unwrap(); if key.is_some() { root.set_key(key); } let deps = match root.deps().clone() { ObjectDeps::ObjectIdList(deps_vec) => deps_vec, ObjectDeps::DepListRef(deps_ref) => { let obj = Object::load(deps_ref.id, Some(deps_ref.key), store)?; match obj.content()? { ObjectContent::DepList(DepList::V0(deps_vec)) => deps_vec, _ => return Err(ObjectParseError::InvalidDeps), } } }; Ok(Object { blocks, deps }) } /// Save blocks of the object in the store pub fn save(&self, store: &mut impl RepoStore) -> Result<(), StorageError> { let mut deduplicated: HashSet = HashSet::new(); for block in &self.blocks { let id = block.id(); if deduplicated.get(&id).is_none() { store.put(block)?; deduplicated.insert(id); } } Ok(()) } /// Get the ID of the Object pub fn id(&self) -> ObjectId { self.blocks.last().unwrap().id() } /// Get the key for the Object pub fn key(&self) -> Option { self.blocks.last().unwrap().key() } /// Get an `ObjectRef` for the root object pub fn reference(&self) -> Option { if self.key().is_some() { Some(ObjectRef { id: self.id(), key: self.key().unwrap(), }) } else { None } } pub fn is_root(&self) -> bool { self.deps().len() == 0 //TODO: add && sdeps().len() == 0 && self.acks().len() == 0 && self.nacks().len() == 0 } pub fn root(&self) -> &Block { self.blocks.last().unwrap() } pub fn expiry(&self) -> Option { self.blocks.last().unwrap().expiry() } pub fn deps(&self) -> &Vec { &self.deps } pub fn blocks(&self) -> &Vec { &self.blocks } pub fn to_hashmap(&self) -> HashMap { let mut map: HashMap = HashMap::new(); for block in &self.blocks { map.insert(block.id(), block.clone()); } map } /// Collect leaves from the tree fn collect_leaves( blocks: &Vec, parents: &Vec<(ObjectId, SymKey)>, parent_index: usize, leaves: &mut Option<&mut Vec>, obj_content: &mut Option<&mut Vec>, ) -> Result<(), ObjectParseError> { /*debug_println!( ">>> collect_leaves: #{}..{}", parent_index, parent_index + parents.len() - 1 );*/ let mut children: Vec<(ObjectId, SymKey)> = vec![]; let mut i = parent_index; for (id, key) in parents { //debug_println!("!!! parent: #{}", i); let block = &blocks[i]; i += 1; // verify object ID if *id != block.id() { debug_println!("Invalid ObjectId.\nExp: {:?}\nGot: {:?}", *id, block.id()); return Err(ObjectParseError::InvalidBlockId); } match block { Block::V0(b) => { // decrypt content let mut content_dec = b.content.clone(); match key { SymKey::ChaCha20Key(key) => { let nonce = [0u8; 12]; let mut cipher = ChaCha20::new(key.into(), &nonce.into()); let mut content_dec_slice = &mut content_dec.as_mut_slice(); cipher.apply_keystream(&mut content_dec_slice); } } // deserialize content let content: BlockContentV0; match serde_bare::from_slice(content_dec.as_slice()) { Ok(c) => content = c, Err(e) => { debug_println!("Block deserialize error: {}", e); return Err(ObjectParseError::BlockDeserializeError); } } // parse content match content { BlockContentV0::InternalNode(keys) => { if keys.len() != b.children.len() { debug_println!( "Invalid keys length: got {}, expected {}", keys.len(), b.children.len() ); debug_println!("!!! children: {:?}", b.children); debug_println!("!!! keys: {:?}", keys); return Err(ObjectParseError::InvalidKeys); } for (id, key) in b.children.iter().zip(keys.iter()) { children.push((id.clone(), key.clone())); } } BlockContentV0::DataChunk(chunk) => { if leaves.is_some() { let mut leaf = block.clone(); leaf.set_key(Some(*key)); let l = &mut **leaves.as_mut().unwrap(); l.push(leaf); } if obj_content.is_some() { let c = &mut **obj_content.as_mut().unwrap(); c.extend_from_slice(chunk.as_slice()); } } } } } } if !children.is_empty() { if parent_index < children.len() { return Err(ObjectParseError::InvalidChildren); } match Self::collect_leaves( blocks, &children, parent_index - children.len(), leaves, obj_content, ) { Ok(_) => (), Err(e) => return Err(e), } } Ok(()) } /// Parse the Object and return the leaf Blocks with decryption key set pub fn leaves(&self) -> Result, ObjectParseError> { let mut leaves: Vec = vec![]; let parents = vec![(self.id(), self.key().unwrap())]; match Self::collect_leaves( &self.blocks, &parents, self.blocks.len() - 1, &mut Some(&mut leaves), &mut None, ) { Ok(_) => Ok(leaves), Err(e) => Err(e), } } /// Parse the Object and return the decrypted content assembled from Blocks pub fn content(&self) -> Result { if self.key().is_none() { return Err(ObjectParseError::MissingRootKey); } let mut obj_content: Vec = vec![]; let parents = vec![(self.id(), self.key().unwrap())]; match Self::collect_leaves( &self.blocks, &parents, self.blocks.len() - 1, &mut None, &mut Some(&mut obj_content), ) { Ok(_) => { let content: ObjectContent; match serde_bare::from_slice(obj_content.as_slice()) { Ok(c) => Ok(c), Err(e) => { debug_println!("Object deserialize error: {}", e); Err(ObjectParseError::ObjectDeserializeError) } } } Err(e) => Err(e), } } } #[cfg(test)] mod test { use crate::object::*; use crate::store::*; use crate::types::*; // Those constants are calculated with RepoStore::get_max_value_size /// Maximum arity of branch containing max number of leaves const MAX_ARITY_LEAVES: usize = 31774; /// Maximum arity of root branch const MAX_ARITY_ROOT: usize = 31770; /// Maximum data that can fit in object.content const MAX_DATA_PAYLOAD_SIZE: usize = 2097112; /// Test tree API #[test] pub fn test_object() { let file = File::V0(FileV0 { content_type: Vec::from("file/test"), metadata: Vec::from("some meta data here"), content: [(0..255).collect::>().as_slice(); 320].concat(), }); let content = ObjectContent::File(file); let deps: Vec = vec![Digest::Blake3Digest32([9; 32])]; let exp = Some(2u32.pow(31)); let max_object_size = 0; let repo_secret = SymKey::ChaCha20Key([0; 32]); let repo_pubkey = PubKey::Ed25519PubKey([1; 32]); let obj = Object::new( content.clone(), deps.clone(), exp, max_object_size, repo_pubkey, repo_secret, ); println!("obj.id: {:?}", obj.id()); println!("obj.key: {:?}", obj.key()); println!("obj.deps: {:?}", obj.deps()); println!("obj.blocks.len: {:?}", obj.blocks().len()); let mut i = 0; for node in obj.blocks() { println!("#{}: {:?}", i, node.id()); i += 1; } assert_eq!(*obj.deps(), deps); match obj.content() { Ok(cnt) => { assert_eq!(content, cnt); } Err(e) => panic!("Object parse error: {:?}", e), } let mut store = HashMapRepoStore::new(); obj.save(&mut store).expect("Object save error"); let obj2 = Object::load(obj.id(), obj.key(), &store).unwrap(); println!("obj2.id: {:?}", obj2.id()); println!("obj2.key: {:?}", obj2.key()); println!("obj2.deps: {:?}", obj2.deps()); println!("obj2.blocks.len: {:?}", obj2.blocks().len()); let mut i = 0; for node in obj2.blocks() { println!("#{}: {:?}", i, node.id()); i += 1; } assert_eq!(*obj2.deps(), deps); assert_eq!(*obj2.deps(), deps); match obj2.content() { Ok(cnt) => { assert_eq!(content, cnt); } Err(e) => panic!("Object2 parse error: {:?}", e), } let obj3 = Object::load(obj.id(), None, &store).unwrap(); println!("obj3.id: {:?}", obj3.id()); println!("obj3.key: {:?}", obj3.key()); println!("obj3.deps: {:?}", obj3.deps()); println!("obj3.blocks.len: {:?}", obj3.blocks().len()); let mut i = 0; for node in obj3.blocks() { println!("#{}: {:?}", i, node.id()); i += 1; } assert_eq!(*obj3.deps(), deps); match obj3.content() { Err(ObjectParseError::MissingRootKey) => (), Err(e) => panic!("Object3 parse error: {:?}", e), Ok(_) => panic!("Object3 should not return content"), } let exp4 = Some(2342); let obj4 = obj.copy(exp4, repo_pubkey, repo_secret).unwrap(); obj4.save(&mut store).unwrap(); assert_eq!(obj4.expiry(), exp4); assert_eq!(*obj.deps(), deps); match obj4.content() { Ok(cnt) => { assert_eq!(content, cnt); } Err(e) => panic!("Object3 parse error: {:?}", e), } } /// Checks that a content that fits the root node, will not be chunked into children nodes #[test] pub fn test_depth_1() { let deps: Vec = vec![Digest::Blake3Digest32([9; 32])]; let empty_file = ObjectContent::File(File::V0(FileV0 { content_type: vec![], metadata: vec![], content: vec![], })); let empty_file_ser = serde_bare::to_vec(&empty_file).unwrap(); println!("empty file size: {}", empty_file_ser.len()); let size = store_max_value_size() - EMPTY_BLOCK_SIZE - DATA_VARINT_EXTRA - BLOCK_ID_SIZE * deps.len() - empty_file_ser.len() - DATA_VARINT_EXTRA; println!("file size: {}", size); let content = ObjectContent::File(File::V0(FileV0 { content_type: vec![], metadata: vec![], content: vec![99; size], })); let content_ser = serde_bare::to_vec(&content).unwrap(); println!("content len: {}", content_ser.len()); let expiry = Some(2u32.pow(31)); let max_object_size = store_max_value_size(); let repo_secret = SymKey::ChaCha20Key([0; 32]); let repo_pubkey = PubKey::Ed25519PubKey([1; 32]); let object = Object::new( content, deps, expiry, max_object_size, repo_pubkey, repo_secret, ); println!("root_id: {:?}", object.id()); println!("root_key: {:?}", object.key().unwrap()); println!("nodes.len: {:?}", object.blocks().len()); //println!("root: {:?}", tree.root()); //println!("nodes: {:?}", object.blocks); assert_eq!(object.blocks.len(), 1); } #[test] pub fn test_block_size() { let max_block_size = store_max_value_size(); println!("max_object_size: {}", max_block_size); let id = Digest::Blake3Digest32([0u8; 32]); let key = SymKey::ChaCha20Key([0u8; 32]); let one_key = BlockContentV0::InternalNode(vec![key]); let one_key_ser = serde_bare::to_vec(&one_key).unwrap(); let two_keys = BlockContentV0::InternalNode(vec![key, key]); let two_keys_ser = serde_bare::to_vec(&two_keys).unwrap(); let max_keys = BlockContentV0::InternalNode(vec![key; MAX_ARITY_LEAVES]); let max_keys_ser = serde_bare::to_vec(&max_keys).unwrap(); let data = BlockContentV0::DataChunk(vec![]); let data_ser = serde_bare::to_vec(&data).unwrap(); let data_full = BlockContentV0::DataChunk(vec![0; MAX_DATA_PAYLOAD_SIZE]); let data_full_ser = serde_bare::to_vec(&data_full).unwrap(); let leaf_empty = Block::new( vec![], ObjectDeps::ObjectIdList(vec![]), Some(2342), data_ser.clone(), None, ); let leaf_empty_ser = serde_bare::to_vec(&leaf_empty).unwrap(); let leaf_full_data = Block::new( vec![], ObjectDeps::ObjectIdList(vec![]), Some(2342), data_full_ser.clone(), None, ); let leaf_full_data_ser = serde_bare::to_vec(&leaf_full_data).unwrap(); let root_depsref = Block::new( vec![], ObjectDeps::DepListRef(ObjectRef { id: id, key: key }), Some(2342), data_ser.clone(), None, ); let root_depsref_ser = serde_bare::to_vec(&root_depsref).unwrap(); let internal_max = Block::new( vec![id; MAX_ARITY_LEAVES], ObjectDeps::ObjectIdList(vec![]), Some(2342), max_keys_ser.clone(), None, ); let internal_max_ser = serde_bare::to_vec(&internal_max).unwrap(); let internal_one = Block::new( vec![id; 1], ObjectDeps::ObjectIdList(vec![]), Some(2342), one_key_ser.clone(), None, ); let internal_one_ser = serde_bare::to_vec(&internal_one).unwrap(); let internal_two = Block::new( vec![id; 2], ObjectDeps::ObjectIdList(vec![]), Some(2342), two_keys_ser.clone(), None, ); let internal_two_ser = serde_bare::to_vec(&internal_two).unwrap(); let root_one = Block::new( vec![id; 1], ObjectDeps::ObjectIdList(vec![id; 8]), Some(2342), one_key_ser.clone(), None, ); let root_one_ser = serde_bare::to_vec(&root_one).unwrap(); let root_two = Block::new( vec![id; 2], ObjectDeps::ObjectIdList(vec![id; 8]), Some(2342), two_keys_ser.clone(), None, ); let root_two_ser = serde_bare::to_vec(&root_two).unwrap(); println!( "range of valid value sizes {} {}", store_valid_value_size(0), store_max_value_size() ); println!( "max_data_payload_of_object: {}", max_block_size - EMPTY_BLOCK_SIZE - DATA_VARINT_EXTRA ); println!( "max_data_payload_depth_1: {}", max_block_size - EMPTY_BLOCK_SIZE - DATA_VARINT_EXTRA - MAX_DEPS_SIZE ); println!( "max_data_payload_depth_2: {}", MAX_ARITY_ROOT * MAX_DATA_PAYLOAD_SIZE ); println!( "max_data_payload_depth_3: {}", MAX_ARITY_ROOT * MAX_ARITY_LEAVES * MAX_DATA_PAYLOAD_SIZE ); let max_arity_leaves = (max_block_size - EMPTY_BLOCK_SIZE - BIG_VARINT_EXTRA * 2) / (BLOCK_ID_SIZE + BLOCK_KEY_SIZE); println!("max_arity_leaves: {}", max_arity_leaves); assert_eq!(max_arity_leaves, MAX_ARITY_LEAVES); assert_eq!( max_block_size - EMPTY_BLOCK_SIZE - DATA_VARINT_EXTRA, MAX_DATA_PAYLOAD_SIZE ); let max_arity_root = (max_block_size - EMPTY_BLOCK_SIZE - MAX_DEPS_SIZE - BIG_VARINT_EXTRA * 2) / (BLOCK_ID_SIZE + BLOCK_KEY_SIZE); println!("max_arity_root: {}", max_arity_root); assert_eq!(max_arity_root, MAX_ARITY_ROOT); println!("store_max_value_size: {}", leaf_full_data_ser.len()); assert_eq!(leaf_full_data_ser.len(), max_block_size); println!("leaf_empty: {}", leaf_empty_ser.len()); assert_eq!(leaf_empty_ser.len(), EMPTY_BLOCK_SIZE); println!("root_depsref: {}", root_depsref_ser.len()); assert_eq!(root_depsref_ser.len(), EMPTY_ROOT_SIZE_DEPSREF); println!("internal_max: {}", internal_max_ser.len()); assert_eq!( internal_max_ser.len(), EMPTY_BLOCK_SIZE + BIG_VARINT_EXTRA * 2 + MAX_ARITY_LEAVES * (BLOCK_ID_SIZE + BLOCK_KEY_SIZE) ); assert!(internal_max_ser.len() < max_block_size); println!("internal_one: {}", internal_one_ser.len()); assert_eq!( internal_one_ser.len(), EMPTY_BLOCK_SIZE + 1 * BLOCK_ID_SIZE + 1 * BLOCK_KEY_SIZE ); println!("internal_two: {}", internal_two_ser.len()); assert_eq!( internal_two_ser.len(), EMPTY_BLOCK_SIZE + 2 * BLOCK_ID_SIZE + 2 * BLOCK_KEY_SIZE ); println!("root_one: {}", root_one_ser.len()); assert_eq!( root_one_ser.len(), EMPTY_BLOCK_SIZE + 8 * BLOCK_ID_SIZE + 1 * BLOCK_ID_SIZE + 1 * BLOCK_KEY_SIZE ); println!("root_two: {}", root_two_ser.len()); assert_eq!( root_two_ser.len(), EMPTY_BLOCK_SIZE + 8 * BLOCK_ID_SIZE + 2 * BLOCK_ID_SIZE + 2 * BLOCK_KEY_SIZE ); // let object_size_1 = 4096 * 1 - VALUE_HEADER_SIZE; // let object_size_512 = 4096 * MAX_PAGES_PER_VALUE - VALUE_HEADER_SIZE; // let arity_1: usize = // (object_size_1 - 8 * OBJECT_ID_SIZE) / (OBJECT_ID_SIZE + OBJECT_KEY_SIZE); // let arity_512: usize = // (object_size_512 - 8 * OBJECT_ID_SIZE) / (OBJECT_ID_SIZE + OBJECT_KEY_SIZE); // println!("1-page object_size: {}", object_size_1); // println!("512-page object_size: {}", object_size_512); // println!("max arity of 1-page object: {}", arity_1); // println!("max arity of 512-page object: {}", arity_512); } }