diff --git a/Cargo.lock b/Cargo.lock index 810c4a1..4ffaadb 100644 --- a/Cargo.lock +++ b/Cargo.lock @@ -373,7 +373,6 @@ version = "0.2.0" dependencies = [ "crossbeam-queue", "limine", - "linked_list_allocator", "log", "slab", "spin", @@ -401,15 +400,6 @@ version = "0.1.8" source = "registry+https://github.com/rust-lang/crates.io-index" checksum = "eac64618e505bab2387986e5b1014f5d344130d609685bbd8b71a10af3ee599d" -[[package]] -name = "linked_list_allocator" -version = "0.9.1" -source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "549ce1740e46b291953c4340adcd74c59bcf4308f4cac050fd33ba91b7168f4a" -dependencies = [ - "spinning_top", -] - [[package]] name = "lock_api" version = "0.4.7" @@ -755,15 +745,6 @@ dependencies = [ "lock_api", ] -[[package]] -name = "spinning_top" -version = "0.2.4" -source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "75adad84ee84b521fb2cca2d4fd0f1dab1d8d026bda3c5bea4ca63b5f9f9293c" -dependencies = [ - "lock_api", -] - [[package]] name = "static_assertions" version = "1.1.0" diff --git a/kernel/Cargo.toml b/kernel/Cargo.toml index 00afab0..b5eca17 100644 --- a/kernel/Cargo.toml +++ b/kernel/Cargo.toml @@ -4,7 +4,6 @@ name = "kernel" version = "0.2.0" [dependencies] -linked_list_allocator = "0.9" slab = { version = "0.4", default-features = false } spin = "0.9" versioning = { git = "https://git.ablecorp.us/able/aos_userland" } diff --git a/kernel/lds/x86_64.ld b/kernel/lds/x86_64.ld index 499b8fc..e430d28 100644 --- a/kernel/lds/x86_64.ld +++ b/kernel/lds/x86_64.ld @@ -26,15 +26,15 @@ SECTIONS *(.text .text.*) } :text - /* Move to the next memory page for .rodata */ - . += CONSTANT(MAXPAGESIZE); + /* Align .rodata to page boundary */ + . = ALIGN(4K); .rodata : { *(.rodata .rodata.*) } :rodata - /* Move to the next memory page for .data */ - . += CONSTANT(MAXPAGESIZE); + /* Align .data to page boundary */ + . = ALIGN(4K); .data : { *(.data .data.*) @@ -43,5 +43,11 @@ SECTIONS .bss : { *(COMMON) *(.bss .bss.*) + + /* Align initial kernel heap to page boundary */ + . = ALIGN(4K); + PROVIDE(_initial_kernel_heap_start = .); + PROVIDE(_initial_kernel_heap_size = 1024 * 1024); + . += _initial_kernel_heap_size; } :data } diff --git a/kernel/src/allocator.rs b/kernel/src/allocator.rs index 670b1a0..5b357b7 100644 --- a/kernel/src/allocator.rs +++ b/kernel/src/allocator.rs @@ -1,16 +1,342 @@ //! Memory allocator +/* + * This file incorporates work covered by the following license notice: + * + * Copyright (c) 2020, the SerenityOS developers. + * All rights reserved. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions are met: + * + * 1. Redistributions of source code must retain the above copyright notice, this + * list of conditions and the following disclaimer. + * + * 2. Redistributions in binary form must reproduce the above copyright notice, + * this list of conditions and the following disclaimer in the documentation + * and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" + * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE + * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE + * DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE + * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL + * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR + * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER + * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, + * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE + * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + */ -/// -pub const HEAP_START: usize = 0x_4444_4444_0000; +use core::{ + alloc::{GlobalAlloc, Layout}, + mem, + ptr::{self, NonNull}, +}; -/// -pub const HEAP_MULTIPLIER: usize = 100000; +use spin::Mutex; -/// -pub const HEAP_BASE: usize = 100; +extern "C" { + fn _initial_kernel_heap_start(); + fn _initial_kernel_heap_size(); +} -/// -pub const HEAP_SIZE: usize = HEAP_BASE * HEAP_MULTIPLIER; +const INITIAL_KERNEL_HEAP_START: *mut u8 = _initial_kernel_heap_start as _; +const INITIAL_KERNEL_HEAP_SIZE: *const () = _initial_kernel_heap_size as _; + +struct Allocator(Mutex>); + +unsafe impl GlobalAlloc for Allocator { + unsafe fn alloc(&self, layout: Layout) -> *mut u8 { + let mut lock = self.0.lock(); + let allocator = lock.as_mut().expect("heap allocator should be initialized"); + + match allocator.allocate(layout.size(), layout.align()) { + Some(ptr) => ptr.as_ptr(), + None => ptr::null_mut(), + } + } + + unsafe fn dealloc(&self, ptr: *mut u8, _: Layout) { + let mut lock = self.0.lock(); + let allocator = lock.as_mut().expect("heap allocator should be initialized"); + allocator.deallocate(ptr); + } +} + +#[global_allocator] +static ALLOCATOR: Allocator = Allocator(Mutex::new(None)); + +pub fn init() { + log::info!("Initialising kernel heap allocator"); + let memory_size = unsafe { mem::transmute(INITIAL_KERNEL_HEAP_SIZE) }; + *ALLOCATOR.0.lock() = Some(unsafe { Heap::new(INITIAL_KERNEL_HEAP_START, memory_size) }); +} + +// FIXME: these are arch-specific +const CHUNK_SIZE: usize = 16; +const MINIMUM_ALIGNMENT: usize = 8; + +struct Header { + size_in_chunks: usize, +} + +// compile-time assertions to make sure that AllocationHeader's size is a power of two +// and CHUNK_SIZE is bigger than AllocationHeader's size +const _: () = { + assert!(mem::size_of::
().is_power_of_two()); + assert!(CHUNK_SIZE >= mem::size_of::
()); +}; + +/// A first-fit heap allocator, with CHUNK_SIZE chunks and a set size +/// In the future these will become subheaps and the actual heap allocator will create more +/// subheaps as needed +struct Heap { + total_chunks: usize, + allocated_chunks: usize, + chunks: *mut u8, + bitmap: *mut u8, +} + +impl Heap { + /// # Safety + /// This function assumes that the pointer given points at a valid memory address + unsafe fn new(memory: *mut u8, memory_size: usize) -> Self { + let total_chunks = Self::calculate_chunks(memory_size); + assert!(total_chunks * CHUNK_SIZE + (total_chunks + 7) / 8 <= memory_size); + Self { + total_chunks, + allocated_chunks: 0, + chunks: memory, + bitmap: unsafe { memory.add(total_chunks * CHUNK_SIZE) }, + } + } + + fn allocate(&mut self, size: usize, alignment: usize) -> Option> { + assert!(alignment.is_power_of_two()); + let alignment = if alignment < MINIMUM_ALIGNMENT { + MINIMUM_ALIGNMENT + } else { + alignment + }; + + // We need space for the header as well + let size = size + mem::size_of::
(); + let chunks_needed = (size + CHUNK_SIZE - 1) / CHUNK_SIZE; + let chunk_alignment = (alignment + CHUNK_SIZE - 1) / CHUNK_SIZE; + log::info!("size: {size} chunks: {chunks_needed} align: {chunk_alignment}"); + + if chunks_needed + chunk_alignment > self.free_chunks() { + return None; + } + + // FIXME: should utilize the alignment directly instead of trying to allocate `size + alignment` + let first_chunk = self.find_first_fit(chunks_needed + chunk_alignment)?; + let chunks_addr = self.chunks as usize; + let addr_unaligned = chunks_addr + first_chunk * CHUNK_SIZE; + + // Align the starting address and verify that we haven't gone outside the calculated free area + let addr = + addr_unaligned + alignment - (addr_unaligned + mem::size_of::
()) % alignment; + log::info!( + "Addr unaligned: 0x{addr_unaligned:x} (offset: 0x{:x})", + addr_unaligned - chunks_addr + ); + log::info!("Addr: 0x{addr:x} (offset: 0x{:x})", addr - chunks_addr); + let aligned_first_chunk = (addr - chunks_addr) / CHUNK_SIZE; + assert!(first_chunk <= aligned_first_chunk); + assert!( + aligned_first_chunk + chunks_needed <= first_chunk + chunks_needed + chunk_alignment + ); + + let header: *mut Header = unsafe { mem::transmute(addr) }; + unsafe { + (*header).size_in_chunks = chunks_needed; + } + + self.bitmap_set_range(aligned_first_chunk, chunks_needed, true); + + self.allocated_chunks += chunks_needed; + + let ptr: *mut u8 = unsafe { mem::transmute(header.add(1)) }; + log::info!("{ptr:p}"); + // FIXME: zero or scrub memory? + assert!(ptr.is_aligned_to(alignment)); + NonNull::new(ptr) + } + + fn deallocate(&mut self, ptr: *mut u8) { + let header = Self::allocation_header(ptr); + let start = (header as usize - self.chunks as usize) / CHUNK_SIZE; + assert!(self.bitmap_get(start)); + let size = unsafe { (*header).size_in_chunks }; + self.bitmap_set_range(start, size, false); + self.allocated_chunks -= size; + // FIXME: zero or scrub memory? + } + + /// Finds first hole that can fit an allocation of `size` chunks, returns the start of the + /// found free chunks + fn find_first_fit(&self, size: usize) -> Option { + let mut start_of_free_chunks = 0; + let mut free_chunks = 0; + for i in 0..self.total_chunks / usize::BITS as usize { + if free_chunks >= size { + return Some(start_of_free_chunks); + } + + let mut bucket = unsafe { *self.bitmap.cast::().add(i) }; + if bucket == usize::MAX { + // Skip over completely full bucket + free_chunks = 0; + continue; + } + if bucket == 0 { + // Skip over completely empty bucket + if free_chunks == 0 { + start_of_free_chunks = i * usize::BITS as usize; + } + + free_chunks += usize::BITS as usize; + continue; + } + + let mut viewed_bits = 0; + while viewed_bits < usize::BITS as usize { + if bucket == 0 { + if free_chunks == 0 { + start_of_free_chunks = i * usize::BITS as usize + viewed_bits; + } + + free_chunks += usize::BITS as usize - viewed_bits; + viewed_bits = usize::BITS as usize; + } else { + let trailing_zeros = bucket.trailing_zeros() as usize; + bucket >>= trailing_zeros; + + if free_chunks == 0 { + start_of_free_chunks = i * usize::BITS as usize + viewed_bits; + } + + free_chunks += trailing_zeros; + viewed_bits += trailing_zeros; + + if free_chunks >= size { + return Some(start_of_free_chunks); + } + + let trailing_ones = bucket.trailing_ones() as usize; + bucket >>= trailing_ones; + viewed_bits += trailing_ones; + free_chunks = 0; + } + } + } + + if free_chunks >= size { + return Some(start_of_free_chunks); + } + + let first_trailing_bit = (self.total_chunks / usize::BITS as usize) * usize::BITS as usize; + let trailing_bits = self.total_chunks % usize::BITS as usize; + for i in 0..trailing_bits { + if self.bitmap_get(first_trailing_bit + i) { + free_chunks = 0; + continue; + } + + if free_chunks == 0 { + start_of_free_chunks = first_trailing_bit + i; + } + + free_chunks += 1; + if free_chunks >= size { + return Some(start_of_free_chunks); + } + } + + None + } + + fn bitmap_set_range(&mut self, start: usize, length: usize, value: bool) { + assert!(start + length <= self.total_chunks); + if length == 0 { + return; + } + + const BITMASK_FIRST_BYTE: [u8; 8] = [0xFF, 0xFE, 0xFC, 0xF8, 0xF0, 0xE0, 0xC0, 0x80]; + const BITMASK_LAST_BYTE: [u8; 8] = [0, 1, 3, 7, 0xF, 0x1F, 0x3F, 0x7F]; + let first = unsafe { self.bitmap.add(start / 8) }; + let last = unsafe { self.bitmap.add((start + length) / 8) }; + let mut byte_mask = BITMASK_FIRST_BYTE[start % 8]; + if first == last { + byte_mask &= BITMASK_LAST_BYTE[(start + length) % 8]; + if value { + unsafe { + *first |= byte_mask; + } + } else { + unsafe { + *first &= !byte_mask; + } + } + } else { + if value { + unsafe { + *first |= byte_mask; + } + } else { + unsafe { + *first &= !byte_mask; + } + } + + byte_mask = BITMASK_LAST_BYTE[(start + length) % 8]; + if value { + unsafe { + *last |= byte_mask; + } + } else { + unsafe { + *last &= !byte_mask; + } + } + + let first = unsafe { first.add(1) }; + if first >= last { + return; + } + + if value { + unsafe { + first.write_bytes(0xFF, last.sub_ptr(first)); + } + } else { + unsafe { + first.write_bytes(0, last.sub_ptr(first)); + } + } + } + } + + fn bitmap_get(&self, index: usize) -> bool { + assert!(index < self.total_chunks); + (unsafe { *self.bitmap.add(index / 8) } & (1 << (index % 8))) != 0 + } + + const fn free_chunks(&self) -> usize { + self.total_chunks - self.allocated_chunks + } + + fn allocation_header(ptr: *mut u8) -> *mut Header { + unsafe { mem::transmute::<_, *mut Header>(ptr).sub(1) } + } + + const fn calculate_chunks(memory_size: usize) -> usize { + memory_size / (CHUNK_SIZE + 1) + } +} + +unsafe impl Send for Heap {} #[alloc_error_handler] fn alloc_error_handler(layout: alloc::alloc::Layout) -> ! { diff --git a/kernel/src/arch/x86_64/allocator.rs b/kernel/src/arch/x86_64/allocator.rs deleted file mode 100644 index 8cf17c1..0000000 --- a/kernel/src/arch/x86_64/allocator.rs +++ /dev/null @@ -1,46 +0,0 @@ -use linked_list_allocator::LockedHeap; -use x86_64::{ - structures::paging::{ - mapper::MapToError, FrameAllocator, Mapper, Page, PageTableFlags, Size4KiB, - }, - VirtAddr, -}; - -use crate::allocator::{HEAP_SIZE, HEAP_START}; - -#[global_allocator] -static ALLOCATOR: LockedHeap = LockedHeap::empty(); - -pub unsafe fn init_alloc() -> Result<(), MapToError> { - log::info!("Initialising kernel heap allocator"); - let page_range = Page::range_inclusive( - Page::containing_address(VirtAddr::new(HEAP_START as u64)), - Page::containing_address(VirtAddr::new(HEAP_START as u64) + HEAP_SIZE - 1u64), - ); - - let mut frame_allocator = super::memory::FRAME_ALLOC - .get() - .expect("frame allocator is not initialised") - .lock(); - - let mut mapper = super::memory::PAGE_TABLE - .get() - .expect("page table is not initialised") - .lock(); - - for page in page_range { - let frame = frame_allocator - .allocate_frame() - .ok_or(MapToError::FrameAllocationFailed)?; - let flags = PageTableFlags::PRESENT | PageTableFlags::WRITABLE; - mapper - .map_to(page, frame, flags, &mut *frame_allocator)? - .flush(); - } - - ALLOCATOR - .lock() - .init(crate::allocator::HEAP_START, crate::allocator::HEAP_SIZE); - - Ok(()) -} diff --git a/kernel/src/arch/x86_64/mod.rs b/kernel/src/arch/x86_64/mod.rs index f094ed0..fa6e153 100644 --- a/kernel/src/arch/x86_64/mod.rs +++ b/kernel/src/arch/x86_64/mod.rs @@ -3,7 +3,8 @@ use spin::Mutex; use uart_16550::SerialPort; use x86_64::VirtAddr; -mod allocator; +use crate::allocator; + mod gdt; mod interrupts; mod memory; @@ -37,7 +38,7 @@ unsafe extern "C" fn _kernel_start() -> ! { .expect("tried to get memory map from Limine"), ); - allocator::init_alloc().expect("tried to initialise allocator"); + allocator::init(); gdt::init(); interrupts::init(); diff --git a/kernel/src/kmain.rs b/kernel/src/kmain.rs index a61ebef..99e318e 100644 --- a/kernel/src/kmain.rs +++ b/kernel/src/kmain.rs @@ -4,6 +4,6 @@ pub fn kmain(cmdline: &str, initrd: Option<&'static [u8]>) -> ! { log::debug!("Entered kmain"); log::info!("Cmdline: \"{cmdline}\""); let initrd = initrd.expect("no initrd found"); - + crate::arch::sloop() } diff --git a/kernel/src/lib.rs b/kernel/src/lib.rs index 6c2e901..f88219c 100644 --- a/kernel/src/lib.rs +++ b/kernel/src/lib.rs @@ -4,7 +4,9 @@ abi_x86_interrupt, alloc_error_handler, panic_info_message, - prelude_import + pointer_is_aligned, + prelude_import, + ptr_sub_ptr )] #![no_std] diff --git a/repbuild/src/main.rs b/repbuild/src/main.rs index 48d25dc..7d5801c 100644 --- a/repbuild/src/main.rs +++ b/repbuild/src/main.rs @@ -83,11 +83,11 @@ fn main() -> Result<(), Box> { // Create disk directory fs::create_dir("./disk").unwrap(); - // Clone limine 3.X binaries + // Clone limine 4.x binaries Command::new("git") .arg("clone") .arg("https://github.com/limine-bootloader/limine.git") - .arg("--branch=v3.0-branch-binary") + .arg("--branch=v4.x-branch-binary") .arg("--depth=1") .status() .unwrap(); @@ -108,7 +108,10 @@ fn main() -> Result<(), Box> { let dbus_conn = zbus::blocking::Connection::system()?; // Setup loop device - let disk_img = File::options().read(true).write(true).open("./target/disk.img")?; + let disk_img = File::options() + .read(true) + .write(true) + .open("./target/disk.img")?; let loopdev = udisks::manager::UDisks2ManagerProxyBlocking::new(&dbus_conn)? .loop_setup( disk_img.as_raw_fd().into(), @@ -147,12 +150,12 @@ fn main() -> Result<(), Box> { .path(&filesystem)? .build()?; - // Obtain mountpoint - let mountpoint = loop { - if let Some(m) = fsproxy.mount_points()?.get(0) { - break m.to_string(); - } - }; + // Mount the filesystem + let mountpoint = fsproxy.mount(MountOptions { + no_user_interaction: true, + fs_type: String::new(), + mount_options: String::new(), + })?; // copy ./base/* over to ./disk Command::new("sh") @@ -261,7 +264,10 @@ fn main() -> Result<(), Box> { // Setup loopback device for disk.img, with partitions // FIXME: don't do ths if running without changes // Setup loop device - let disk_img = File::options().read(true).write(true).open("./target/disk.img")?; + let disk_img = File::options() + .read(true) + .write(true) + .open("./target/disk.img")?; let dbus_conn = zbus::blocking::Connection::system()?; let loopdev = udisks::manager::UDisks2ManagerProxyBlocking::new(&dbus_conn)? .loop_setup( @@ -339,7 +345,10 @@ fn main() -> Result<(), Box> { // Setup loopback device for disk.img, with partitions // FIXME: don't do ths if running without changes - let disk_img = File::options().read(true).write(true).open("./target/disk.img")?; + let disk_img = File::options() + .read(true) + .write(true) + .open("./target/disk.img")?; let dbus_conn = zbus::blocking::Connection::system()?; let loopdev = udisks::manager::UDisks2ManagerProxyBlocking::new(&dbus_conn)? .loop_setup( @@ -363,12 +372,12 @@ fn main() -> Result<(), Box> { .path(&parts[0])? .build()?; - // Obtain mountpoint - let mountpoint = loop { - if let Some(m) = fsproxy.mount_points()?.get(0) { - break m.to_string(); - } - }; + // Mount the filesystem + let mountpoint = fsproxy.mount(MountOptions { + no_user_interaction: true, + fs_type: String::new(), + mount_options: String::new(), + })?; // copy the kernel over to ./disk/boot/kernel Command::new("cp")