//! Platform independent, software paged memory implementation use core::mem::size_of; pub mod lookup; pub mod paging; #[cfg(feature = "alloc")] pub mod mapping; use { crate::{LoadError, Memory, MemoryAccessReason, StoreError}, core::slice::SliceIndex, lookup::{AddrPageLookupError, AddrPageLookupOk, AddrPageLookuper}, paging::{PageTable, Permission}, }; /// HoleyBytes software paged memory #[deprecated = "Use platform-specific memory implementation"] #[derive(Clone, Debug)] pub struct SoftPagedMem<'p, PfH> { /// Root page table pub root_pt: *mut PageTable, /// Page fault handler pub pf_handler: PfH, /// Program memory segment pub program: &'p [u8], } impl<'p, PfH: HandlePageFault> Memory for SoftPagedMem<'p, PfH> { /// Load value from an address /// /// # Safety /// Applies same conditions as for [`core::ptr::copy_nonoverlapping`] unsafe fn load(&mut self, addr: u64, target: *mut u8, count: usize) -> Result<(), LoadError> { self.memory_access( MemoryAccessReason::Load, addr, target, count, perm_check::readable, |src, dst, count| core::ptr::copy_nonoverlapping(src, dst, count), ) .map_err(LoadError) } /// Store value to an address /// /// # Safety /// Applies same conditions as for [`core::ptr::copy_nonoverlapping`] unsafe fn store( &mut self, addr: u64, source: *const u8, count: usize, ) -> Result<(), StoreError> { self.memory_access( MemoryAccessReason::Store, addr, source.cast_mut(), count, perm_check::writable, |dst, src, count| core::ptr::copy_nonoverlapping(src, dst, count), ) .map_err(StoreError) } unsafe fn prog_read(&mut self, addr: u64) -> Option { let addr = addr as usize; self.program .get(addr..addr + size_of::()) .map(|x| x.as_ptr().cast::().read()) } unsafe fn prog_read_unchecked(&mut self, addr: u64) -> T { self.program.as_ptr().add(addr as _).cast::().read() } } impl<'p, PfH: HandlePageFault> SoftPagedMem<'p, PfH> { // Everyone behold, the holy function, the god of HBVM memory accesses! /// Split address to pages, check their permissions and feed pointers with offset /// to a specified function. /// /// If page is not found, execute page fault trap handler. #[allow(clippy::too_many_arguments)] // Silence peasant fn memory_access( &mut self, reason: MemoryAccessReason, src: u64, mut dst: *mut u8, len: usize, permission_check: fn(Permission) -> bool, action: fn(*mut u8, *mut u8, usize), ) -> Result<(), u64> { // Memory load from program section let (src, len) = if src < self.program.len() as _ { // Allow only loads if reason != MemoryAccessReason::Load { return Err(src); } // Determine how much data to copy from here let to_copy = len.clamp(0, self.program.len().saturating_sub(src as _)); // Perform action action( unsafe { self.program.as_ptr().add(src as _).cast_mut() }, dst, to_copy, ); // Return shifted from what we've already copied ( src.saturating_add(to_copy as _), len.saturating_sub(to_copy), ) } else { (src, len) // Nothing weird! }; // Nothing to copy? Don't bother doing anything, bail. if len == 0 { return Ok(()); } // Create new splitter let mut pspl = AddrPageLookuper::new(src, len, self.root_pt); loop { match pspl.next() { // Page is found Some(Ok(AddrPageLookupOk { vaddr, ptr, size, perm, })) => { if !permission_check(perm) { return Err(vaddr); } // Perform specified memory action and bump destination pointer action(ptr, dst, size); dst = unsafe { dst.add(size) }; } // No page found Some(Err(AddrPageLookupError { addr, size })) => { // Attempt to execute page fault handler if self.pf_handler.page_fault( reason, unsafe { &mut *self.root_pt }, addr, size, dst, ) { // Shift the splitter address pspl.bump(size); // Bump dst pointer dst = unsafe { dst.add(size as _) }; } else { return Err(addr); // Unhandleable, VM will yield. } } // No remaining pages, we are done! None => return Ok(()), } } } } /// Extract index in page table on specified level /// /// The level shall not be larger than 4, otherwise /// the output of the function is unspecified (yes, it can also panic :) pub fn addr_extract_index(addr: u64, lvl: u8) -> usize { debug_assert!(lvl <= 4); usize::try_from((addr >> (lvl * 8 + 12)) & ((1 << 8) - 1)).expect("?conradluget a better CPU") } /// Page size #[derive(Clone, Copy, Debug, PartialEq, Eq)] pub enum PageSize { /// 4 KiB page (on level 0) Size4K = 4096, /// 2 MiB page (on level 1) Size2M = 1024 * 1024 * 2, /// 1 GiB page (on level 2) Size1G = 1024 * 1024 * 1024, } impl PageSize { /// Convert page table level to size of page const fn from_lvl(lvl: u8) -> Option { match lvl { 0 => Some(PageSize::Size4K), 1 => Some(PageSize::Size2M), 2 => Some(PageSize::Size1G), _ => None, } } } /// Permisison checks pub mod perm_check { use super::paging::Permission; /// Page is readable #[inline(always)] pub const fn readable(perm: Permission) -> bool { matches!( perm, Permission::Readonly | Permission::Write | Permission::Exec ) } /// Page is writable #[inline(always)] pub const fn writable(perm: Permission) -> bool { matches!(perm, Permission::Write) } /// Page is executable #[inline(always)] pub const fn executable(perm: Permission) -> bool { matches!(perm, Permission::Exec) } } /// Handle VM traps pub trait HandlePageFault { /// Handle page fault /// /// Return true if handling was sucessful, /// otherwise the program will be interrupted and will /// yield an error. fn page_fault( &mut self, reason: MemoryAccessReason, pagetable: &mut PageTable, vaddr: u64, size: PageSize, dataptr: *mut u8, ) -> bool where Self: Sized; }