mirror of
https://github.com/quickwit-oss/tantivy.git
synced 2026-01-06 09:12:55 +00:00
246 lines
6.8 KiB
Rust
246 lines
6.8 KiB
Rust
//! 32-bits Memory arena for types implementing `Copy`.
|
|
//! This Memory arena has been implemented to fit the use of tantivy's indexer
|
|
//! and has *twisted specifications*.
|
|
//!
|
|
//! - It works on stable rust.
|
|
//! - One can get an accurate figure of the memory usage of the arena.
|
|
//! - Allocation are very cheap.
|
|
//! - Allocation happening consecutively are very likely to have great locality.
|
|
//! - Addresses (`Addr`) are 32bits.
|
|
//! - Dropping the whole `MemoryArena` is cheap.
|
|
//!
|
|
//! # Limitations
|
|
//!
|
|
//! - Your object shall not implement `Drop`.
|
|
//! - `Addr` to the `Arena` are 32-bits. The maximum capacity of the arena
|
|
//! is 4GB. *(Tantivy's indexer uses one arena per indexing thread.)*
|
|
//! - The arena only works for objects much smaller than `1MB`.
|
|
//! Allocating more than `1MB` at a time will result in a panic,
|
|
//! and allocating a lot of large object (> 500KB) will result in a fragmentation.
|
|
//! - Your objects are store in an unaligned fashion. For this reason,
|
|
//! the API does not let you access them as references.
|
|
//!
|
|
//! Instead, you store and access your data via `.write(...)` and `.read(...)`, which under the hood
|
|
//! stores your object using `ptr::write_unaligned` and `ptr::read_unaligned`.
|
|
use std::{mem, ptr};
|
|
|
|
const NUM_BITS_PAGE_ADDR: usize = 20;
|
|
const PAGE_SIZE: usize = 1 << NUM_BITS_PAGE_ADDR; // pages are 1 MB large
|
|
|
|
/// Represents a pointer into the `MemoryArena`
|
|
/// .
|
|
/// Pointer are 32-bits and are split into
|
|
/// two parts.
|
|
///
|
|
/// The first 12 bits represent the id of a
|
|
/// page of memory.
|
|
///
|
|
/// The last 20 bits are an address within this page of memory.
|
|
#[derive(Copy, Clone, Debug)]
|
|
pub struct Addr(u32);
|
|
|
|
impl Addr {
|
|
/// Creates a null pointer.
|
|
pub fn null_pointer() -> Addr {
|
|
Addr(u32::MAX)
|
|
}
|
|
|
|
/// Returns the `Addr` object for `addr + offset`
|
|
pub fn offset(self, offset: u32) -> Addr {
|
|
Addr(self.0.wrapping_add(offset))
|
|
}
|
|
|
|
fn new(page_id: usize, local_addr: usize) -> Addr {
|
|
Addr((page_id << NUM_BITS_PAGE_ADDR | local_addr) as u32)
|
|
}
|
|
|
|
fn page_id(self) -> usize {
|
|
(self.0 as usize) >> NUM_BITS_PAGE_ADDR
|
|
}
|
|
|
|
fn page_local_addr(self) -> usize {
|
|
(self.0 as usize) & (PAGE_SIZE - 1)
|
|
}
|
|
|
|
/// Returns true if and only if the `Addr` is null.
|
|
pub fn is_null(self) -> bool {
|
|
self.0 == u32::MAX
|
|
}
|
|
}
|
|
|
|
pub fn store<Item: Copy + 'static>(dest: &mut [u8], val: Item) {
|
|
assert_eq!(dest.len(), std::mem::size_of::<Item>());
|
|
unsafe {
|
|
ptr::write_unaligned(dest.as_mut_ptr() as *mut Item, val);
|
|
}
|
|
}
|
|
|
|
pub fn load<Item: Copy + 'static>(data: &[u8]) -> Item {
|
|
assert_eq!(data.len(), std::mem::size_of::<Item>());
|
|
unsafe { ptr::read_unaligned(data.as_ptr() as *const Item) }
|
|
}
|
|
|
|
/// The `MemoryArena`
|
|
#[allow(clippy::new_without_default)]
|
|
pub struct MemoryArena {
|
|
pages: Vec<Page>,
|
|
}
|
|
|
|
impl MemoryArena {
|
|
/// Creates a new memory arena.
|
|
pub fn new() -> MemoryArena {
|
|
let first_page = Page::new(0);
|
|
MemoryArena {
|
|
pages: vec![first_page],
|
|
}
|
|
}
|
|
|
|
fn add_page(&mut self) -> &mut Page {
|
|
let new_page_id = self.pages.len();
|
|
self.pages.push(Page::new(new_page_id));
|
|
&mut self.pages[new_page_id]
|
|
}
|
|
|
|
/// Returns an estimate in number of bytes
|
|
/// of resident memory consumed by the `MemoryArena`.
|
|
///
|
|
/// Internally, it counts a number of `1MB` pages
|
|
/// and therefore delivers an upperbound.
|
|
pub fn mem_usage(&self) -> usize {
|
|
self.pages.len() * PAGE_SIZE
|
|
}
|
|
|
|
pub fn write_at<Item: Copy + 'static>(&mut self, addr: Addr, val: Item) {
|
|
let dest = self.slice_mut(addr, std::mem::size_of::<Item>());
|
|
store(dest, val);
|
|
}
|
|
|
|
/// Read an item in the memory arena at the given `address`.
|
|
///
|
|
/// # Panics
|
|
///
|
|
/// If the address is erroneous
|
|
pub fn read<Item: Copy + 'static>(&self, addr: Addr) -> Item {
|
|
load(self.slice(addr, mem::size_of::<Item>()))
|
|
}
|
|
|
|
pub fn slice(&self, addr: Addr, len: usize) -> &[u8] {
|
|
self.pages[addr.page_id()].slice(addr.page_local_addr(), len)
|
|
}
|
|
|
|
pub fn slice_from(&self, addr: Addr) -> &[u8] {
|
|
self.pages[addr.page_id()].slice_from(addr.page_local_addr())
|
|
}
|
|
|
|
#[inline]
|
|
pub fn slice_mut(&mut self, addr: Addr, len: usize) -> &mut [u8] {
|
|
self.pages[addr.page_id()].slice_mut(addr.page_local_addr(), len)
|
|
}
|
|
|
|
/// Allocates `len` bytes and returns the allocated address.
|
|
pub fn allocate_space(&mut self, len: usize) -> Addr {
|
|
let page_id = self.pages.len() - 1;
|
|
if let Some(addr) = self.pages[page_id].allocate_space(len) {
|
|
return addr;
|
|
}
|
|
self.add_page().allocate_space(len).unwrap()
|
|
}
|
|
}
|
|
|
|
struct Page {
|
|
page_id: usize,
|
|
len: usize,
|
|
data: Box<[u8]>,
|
|
}
|
|
|
|
impl Page {
|
|
fn new(page_id: usize) -> Page {
|
|
Page {
|
|
page_id,
|
|
len: 0,
|
|
data: vec![0u8; PAGE_SIZE].into_boxed_slice(),
|
|
}
|
|
}
|
|
|
|
#[inline]
|
|
fn is_available(&self, len: usize) -> bool {
|
|
len + self.len <= PAGE_SIZE
|
|
}
|
|
|
|
fn slice(&self, local_addr: usize, len: usize) -> &[u8] {
|
|
&self.slice_from(local_addr)[..len]
|
|
}
|
|
|
|
fn slice_from(&self, local_addr: usize) -> &[u8] {
|
|
&self.data[local_addr..]
|
|
}
|
|
|
|
fn slice_mut(&mut self, local_addr: usize, len: usize) -> &mut [u8] {
|
|
&mut self.data[local_addr..][..len]
|
|
}
|
|
|
|
fn allocate_space(&mut self, len: usize) -> Option<Addr> {
|
|
if self.is_available(len) {
|
|
let addr = Addr::new(self.page_id, self.len);
|
|
self.len += len;
|
|
Some(addr)
|
|
} else {
|
|
None
|
|
}
|
|
}
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod tests {
|
|
|
|
use super::MemoryArena;
|
|
|
|
#[test]
|
|
fn test_arena_allocate_slice() {
|
|
let mut arena = MemoryArena::new();
|
|
let a = b"hello";
|
|
let b = b"happy tax payer";
|
|
|
|
let addr_a = arena.allocate_space(a.len());
|
|
arena.slice_mut(addr_a, a.len()).copy_from_slice(a);
|
|
|
|
let addr_b = arena.allocate_space(b.len());
|
|
arena.slice_mut(addr_b, b.len()).copy_from_slice(b);
|
|
|
|
assert_eq!(arena.slice(addr_a, a.len()), a);
|
|
assert_eq!(arena.slice(addr_b, b.len()), b);
|
|
}
|
|
|
|
#[derive(Clone, Copy, Debug, Eq, PartialEq)]
|
|
struct MyTest {
|
|
pub a: usize,
|
|
pub b: u8,
|
|
pub c: u32,
|
|
}
|
|
|
|
#[test]
|
|
fn test_store_object() {
|
|
let mut arena = MemoryArena::new();
|
|
let a = MyTest {
|
|
a: 143,
|
|
b: 21,
|
|
c: 32,
|
|
};
|
|
let b = MyTest {
|
|
a: 113,
|
|
b: 221,
|
|
c: 12,
|
|
};
|
|
|
|
let num_bytes = std::mem::size_of::<MyTest>();
|
|
let addr_a = arena.allocate_space(num_bytes);
|
|
arena.write_at(addr_a, a);
|
|
|
|
let addr_b = arena.allocate_space(num_bytes);
|
|
arena.write_at(addr_b, b);
|
|
|
|
assert_eq!(arena.read::<MyTest>(addr_a), a);
|
|
assert_eq!(arena.read::<MyTest>(addr_b), b);
|
|
}
|
|
}
|