Compare commits

...

19 Commits

Author SHA1 Message Date
Philipp Oppermann
3fc7bf6aa1 Fix register typo: rpb -> rbp
Fixes #746
2020-02-16 18:02:24 +01:00
Philipp Oppermann
b337f65abb Add a yield_now function and use it in idle thread 2020-01-28 12:28:50 +01:00
Philipp Oppermann
22c6bd5aa7 Add idle thread and begin support for blocking 2020-01-28 12:22:37 +01:00
Philipp Oppermann
87719f2260 Switch current_thread_id before context switch 2020-01-28 11:29:21 +01:00
Philipp Oppermann
0caf5c351e Run cargo fmt 2020-01-23 14:24:13 +01:00
Philipp Oppermann
cb7bb0ddef Refactor and rewrite 2020-01-23 14:22:29 +01:00
Philipp Oppermann
7ad30651fb Rename allocator.rs to allocator/mod.rs 2020-01-23 11:46:45 +01:00
Philipp Oppermann
49923acb3f Print thread id instead of hardcoding numbers 2020-01-23 11:03:15 +01:00
Philipp Oppermann
f2b1f3a593 Fix handling of current thread id 2020-01-23 10:49:51 +01:00
Philipp Oppermann
5e2e0b629e Refactor threading code 2020-01-23 10:42:37 +01:00
Philipp Oppermann
35379c90e6 Force unlock writer on panic to avoid deadlocks 2020-01-23 10:42:19 +01:00
Philipp Oppermann
e5d10fcaec Increase stack size to avoid stack overflow 2020-01-23 10:41:59 +01:00
Philipp Oppermann
e1242a867f Move global_asm inline in threads module 2020-01-23 09:20:17 +01:00
Philipp Oppermann
cd138a3a1b Rename multitasking module to threads 2020-01-23 09:19:38 +01:00
Philipp Oppermann
11a0eb679c Fix race condition
The first timer interrupt might occur before the heap is initialized. With lazy_static, this causes an allocation failure since the VecDeque is allocated when it's accessed the first time. This commit fixes this by only initializing VecDeque in `add_thread`.
2020-01-23 08:50:35 +01:00
Philipp Oppermann
241c1ab2c9 Add support for closures 2020-01-23 08:24:48 +01:00
Philipp Oppermann
b75406b37e Add new modules 2020-01-22 17:24:17 +01:00
Philipp Oppermann
c3450b6df7 Refactor a bit 2020-01-22 16:33:23 +01:00
Philipp Oppermann
ce1fdcf768 Wip 2020-01-22 16:15:23 +01:00
10 changed files with 460 additions and 3 deletions

View File

@@ -33,3 +33,6 @@ test-args = [
"-display", "none"
]
test-success-exit-code = 33 # (0x10 << 1) | 1
[profile.release]
lto = true

View File

@@ -77,6 +77,7 @@ extern "x86-interrupt" fn timer_interrupt_handler(_stack_frame: &mut InterruptSt
PICS.lock()
.notify_end_of_interrupt(InterruptIndex::Timer.as_u8());
}
crate::multitasking::invoke_scheduler();
}
extern "x86-interrupt" fn keyboard_interrupt_handler(_stack_frame: &mut InterruptStackFrame) {

View File

@@ -6,6 +6,12 @@
#![feature(const_fn)]
#![feature(alloc_layout_extra)]
#![feature(const_in_array_repeat_expressions)]
#![feature(global_asm)]
#![feature(asm)]
#![feature(raw)]
#![feature(never_type)]
#![feature(naked_functions)]
#![feature(option_expect_none)]
#![test_runner(crate::test_runner)]
#![reexport_test_harness_main = "test_main"]
@@ -17,6 +23,7 @@ pub mod allocator;
pub mod gdt;
pub mod interrupts;
pub mod memory;
pub mod multitasking;
pub mod serial;
pub mod vga_buffer;

View File

@@ -7,7 +7,8 @@
extern crate alloc;
use alloc::{boxed::Box, rc::Rc, vec, vec::Vec};
use blog_os::println;
use blog_os::multitasking::{self, thread::Thread, with_scheduler};
use blog_os::{print, println};
use bootloader::{entry_point, BootInfo};
use core::panic::PanicInfo;
@@ -54,14 +55,45 @@ fn kernel_main(boot_info: &'static BootInfo) -> ! {
#[cfg(test)]
test_main();
let idle_thread = Thread::create(idle_thread, 2, &mut mapper, &mut frame_allocator).unwrap();
with_scheduler(|s| s.set_idle_thread(idle_thread));
for _ in 0..10 {
let thread = Thread::create(thread_entry, 2, &mut mapper, &mut frame_allocator).unwrap();
with_scheduler(|s| s.add_new_thread(thread));
}
let thread =
Thread::create_from_closure(|| thread_entry(), 2, &mut mapper, &mut frame_allocator)
.unwrap();
with_scheduler(|s| s.add_new_thread(thread));
println!("It did not crash!");
blog_os::hlt_loop();
thread_entry();
}
fn idle_thread() -> ! {
loop {
x86_64::instructions::hlt();
multitasking::yield_now();
}
}
fn thread_entry() -> ! {
let thread_id = with_scheduler(|s| s.current_thread_id()).as_u64();
for _ in 0..=thread_id {
print!("{}", thread_id);
x86_64::instructions::hlt();
}
multitasking::exit_thread();
}
/// This function is called on panic.
#[cfg(not(test))]
#[panic_handler]
fn panic(info: &PanicInfo) -> ! {
unsafe {
blog_os::vga_buffer::WRITER.force_unlock();
}
println!("{}", info);
blog_os::hlt_loop();
}

View File

@@ -1,7 +1,7 @@
use bootloader::bootinfo::{MemoryMap, MemoryRegionType};
use x86_64::{
structures::paging::{
FrameAllocator, Mapper, OffsetPageTable, Page, PageTable, PhysFrame, Size4KiB,
mapper, FrameAllocator, Mapper, OffsetPageTable, Page, PageTable, PhysFrame, Size4KiB,
UnusedPhysFrame,
},
PhysAddr, VirtAddr,
@@ -36,6 +36,54 @@ unsafe fn active_level_4_table(physical_memory_offset: VirtAddr) -> &'static mut
&mut *page_table_ptr // unsafe
}
#[derive(Debug, Clone, Copy, PartialEq, Eq)]
pub struct StackBounds {
start: VirtAddr,
end: VirtAddr,
}
impl StackBounds {
pub fn start(&self) -> VirtAddr {
self.start
}
pub fn end(&self) -> VirtAddr {
self.end
}
}
pub fn alloc_stack(
size_in_pages: u64,
mapper: &mut impl Mapper<Size4KiB>,
frame_allocator: &mut impl FrameAllocator<Size4KiB>,
) -> Result<StackBounds, mapper::MapToError> {
use core::sync::atomic::{AtomicU64, Ordering};
use x86_64::structures::paging::PageTableFlags as Flags;
static STACK_ALLOC_NEXT: AtomicU64 = AtomicU64::new(0x_5555_5555_0000);
let guard_page_start = STACK_ALLOC_NEXT.fetch_add(
(size_in_pages + 1) * Page::<Size4KiB>::SIZE,
Ordering::SeqCst,
);
let guard_page = Page::from_start_address(VirtAddr::new(guard_page_start))
.expect("`STACK_ALLOC_NEXT` not page aligned");
let stack_start = guard_page + 1;
let stack_end = stack_start + size_in_pages;
let flags = Flags::PRESENT | Flags::WRITABLE;
for page in Page::range(stack_start, stack_end) {
let frame = frame_allocator
.allocate_frame()
.ok_or(mapper::MapToError::FrameAllocationFailed)?;
mapper.map_to(page, frame, flags, frame_allocator)?.flush();
}
Ok(StackBounds {
start: stack_start.start_address(),
end: stack_end.start_address(),
})
}
/// Creates an example mapping for the given page to frame `0xb8000`.
pub fn create_example_mapping(
page: Page,

View File

@@ -0,0 +1,105 @@
use super::{with_scheduler, SwitchReason};
use crate::multitasking::thread::ThreadId;
use alloc::boxed::Box;
use core::mem;
use core::raw::TraitObject;
use x86_64::VirtAddr;
pub struct Stack {
pointer: VirtAddr,
}
impl Stack {
pub unsafe fn new(stack_pointer: VirtAddr) -> Self {
Stack {
pointer: stack_pointer,
}
}
pub fn get_stack_pointer(self) -> VirtAddr {
self.pointer
}
pub fn set_up_for_closure(&mut self, closure: Box<dyn FnOnce() -> !>) {
let trait_object: TraitObject = unsafe { mem::transmute(closure) };
unsafe { self.push(trait_object.data) };
unsafe { self.push(trait_object.vtable) };
self.set_up_for_entry_point(call_closure_entry);
}
pub fn set_up_for_entry_point(&mut self, entry_point: fn() -> !) {
unsafe { self.push(entry_point) };
let rflags: u64 = 0x200;
unsafe { self.push(rflags) };
}
unsafe fn push<T>(&mut self, value: T) {
self.pointer -= core::mem::size_of::<T>();
let ptr: *mut T = self.pointer.as_mut_ptr();
ptr.write(value);
}
}
pub unsafe fn context_switch_to(
new_stack_pointer: VirtAddr,
prev_thread_id: ThreadId,
switch_reason: SwitchReason,
) {
asm!(
"call asm_context_switch"
:
: "{rdi}"(new_stack_pointer), "{rsi}"(prev_thread_id), "{rdx}"(switch_reason as u64)
: "rax", "rbx", "rcx", "rdx", "rsi", "rdi", "rbp", "r8", "r9", "r10",
"r11", "r12", "r13", "r14", "r15", "rflags", "memory"
: "intel", "volatile"
);
}
global_asm!(
"
.intel_syntax noprefix
// asm_context_switch(stack_pointer: u64, thread_id: u64)
asm_context_switch:
pushfq
mov rax, rsp
mov rsp, rdi
mov rdi, rax
call add_paused_thread
popfq
ret
"
);
#[no_mangle]
pub extern "C" fn add_paused_thread(
paused_stack_pointer: VirtAddr,
paused_thread_id: ThreadId,
switch_reason: SwitchReason,
) {
with_scheduler(|s| s.add_paused_thread(paused_stack_pointer, paused_thread_id, switch_reason));
}
#[naked]
fn call_closure_entry() -> ! {
unsafe {
asm!("
pop rsi
pop rdi
call call_closure
" ::: "mem" : "intel", "volatile")
};
unreachable!();
}
// no_mangle required because of https://github.com/rust-lang/rust/issues/68136
#[no_mangle]
extern "C" fn call_closure(data: *mut (), vtable: *mut ()) -> ! {
let trait_object = TraitObject { data, vtable };
let f: Box<dyn FnOnce() -> !> = unsafe { mem::transmute(trait_object) };
f()
}

57
src/multitasking/mod.rs Normal file
View File

@@ -0,0 +1,57 @@
use scheduler::Scheduler;
pub mod context_switch;
pub mod scheduler;
pub mod thread;
static SCHEDULER: spin::Mutex<Option<Scheduler>> = spin::Mutex::new(None);
#[repr(u64)]
pub enum SwitchReason {
Paused,
Yield,
Blocked,
Exit,
}
pub fn invoke_scheduler() {
let next = SCHEDULER
.try_lock()
.and_then(|mut scheduler| scheduler.as_mut().and_then(|s| s.schedule()));
if let Some((next_stack_pointer, prev_thread_id)) = next {
unsafe {
context_switch::context_switch_to(
next_stack_pointer,
prev_thread_id,
SwitchReason::Paused,
)
};
}
}
pub fn exit_thread() -> ! {
synchronous_context_switch(SwitchReason::Exit).expect("can't exit last thread");
unreachable!("finished thread continued");
}
pub fn yield_now() {
let _ = synchronous_context_switch(SwitchReason::Yield);
}
fn synchronous_context_switch(reason: SwitchReason) -> Result<(), ()> {
let next = with_scheduler(|s| s.schedule());
match next {
Some((next_stack_pointer, prev_thread_id)) => unsafe {
context_switch::context_switch_to(next_stack_pointer, prev_thread_id, reason);
Ok(())
},
None => Err(()),
}
}
pub fn with_scheduler<F, T>(f: F) -> T
where
F: FnOnce(&mut Scheduler) -> T,
{
f(SCHEDULER.lock().get_or_insert_with(Scheduler::new))
}

View File

@@ -0,0 +1,122 @@
use super::SwitchReason;
use crate::multitasking::thread::{Thread, ThreadId};
use alloc::collections::{BTreeMap, BTreeSet, VecDeque};
use core::mem;
use x86_64::VirtAddr;
pub struct Scheduler {
threads: BTreeMap<ThreadId, Thread>,
idle_thread_id: Option<ThreadId>,
current_thread_id: ThreadId,
paused_threads: VecDeque<ThreadId>,
blocked_threads: BTreeSet<ThreadId>,
wakeups: BTreeSet<ThreadId>,
}
impl Scheduler {
pub fn new() -> Self {
let root_thread = Thread::create_root_thread();
let root_id = root_thread.id();
let mut threads = BTreeMap::new();
threads
.insert(root_id, root_thread)
.expect_none("map is not empty after creation");
Scheduler {
threads,
current_thread_id: root_id,
paused_threads: VecDeque::new(),
blocked_threads: BTreeSet::new(),
wakeups: BTreeSet::new(),
idle_thread_id: None,
}
}
fn next_thread(&mut self) -> Option<ThreadId> {
self.paused_threads.pop_front()
}
pub fn schedule(&mut self) -> Option<(VirtAddr, ThreadId)> {
let mut next_thread_id = self.next_thread();
if next_thread_id.is_none() && Some(self.current_thread_id) != self.idle_thread_id {
next_thread_id = self.idle_thread_id
}
if let Some(next_id) = next_thread_id {
let next_thread = self
.threads
.get_mut(&next_id)
.expect("next thread does not exist");
let next_stack_pointer = next_thread
.stack_pointer()
.take()
.expect("paused thread has no stack pointer");
let prev_thread_id = mem::replace(&mut self.current_thread_id, next_thread.id());
Some((next_stack_pointer, prev_thread_id))
} else {
None
}
}
pub(super) fn add_paused_thread(
&mut self,
paused_stack_pointer: VirtAddr,
paused_thread_id: ThreadId,
switch_reason: SwitchReason,
) {
let paused_thread = self
.threads
.get_mut(&paused_thread_id)
.expect("paused thread does not exist");
paused_thread
.stack_pointer()
.replace(paused_stack_pointer)
.expect_none("running thread should have stack pointer set to None");
if Some(paused_thread_id) == self.idle_thread_id {
return; // do nothing
}
match switch_reason {
SwitchReason::Paused | SwitchReason::Yield => {
self.paused_threads.push_back(paused_thread_id)
}
SwitchReason::Blocked => {
self.blocked_threads.insert(paused_thread_id);
self.check_for_wakeup(paused_thread_id);
}
SwitchReason::Exit => {
let thread = self
.threads
.remove(&paused_thread_id)
.expect("thread not found");
// TODO: free stack memory again
}
}
}
pub fn add_new_thread(&mut self, thread: Thread) {
let thread_id = thread.id();
self.threads
.insert(thread_id, thread)
.expect_none("thread already exists");
self.paused_threads.push_back(thread_id);
}
pub fn set_idle_thread(&mut self, thread: Thread) {
let thread_id = thread.id();
self.threads
.insert(thread_id, thread)
.expect_none("thread already exists");
self.idle_thread_id
.replace(thread_id)
.expect_none("idle thread should be set only once");
}
pub fn current_thread_id(&self) -> ThreadId {
self.current_thread_id
}
fn check_for_wakeup(&mut self, thread_id: ThreadId) {
if self.wakeups.remove(&thread_id) {
assert!(self.blocked_threads.remove(&thread_id));
self.paused_threads.push_back(thread_id);
}
}
}

View File

@@ -0,0 +1,82 @@
use crate::memory::{alloc_stack, StackBounds};
use crate::multitasking::context_switch::Stack;
use alloc::boxed::Box;
use x86_64::{
structures::paging::{mapper, FrameAllocator, Mapper, Size4KiB},
VirtAddr,
};
#[derive(Debug, Clone, Copy, PartialEq, Eq, PartialOrd, Ord)]
pub struct ThreadId(u64);
impl ThreadId {
pub fn as_u64(&self) -> u64 {
self.0
}
fn new() -> Self {
use core::sync::atomic::{AtomicU64, Ordering};
static NEXT_THREAD_ID: AtomicU64 = AtomicU64::new(1);
ThreadId(NEXT_THREAD_ID.fetch_add(1, Ordering::SeqCst))
}
}
#[derive(Debug)]
pub struct Thread {
id: ThreadId,
stack_pointer: Option<VirtAddr>,
stack_bounds: Option<StackBounds>,
}
impl Thread {
pub fn create(
entry_point: fn() -> !,
stack_size: u64,
mapper: &mut impl Mapper<Size4KiB>,
frame_allocator: &mut impl FrameAllocator<Size4KiB>,
) -> Result<Self, mapper::MapToError> {
let stack_bounds = alloc_stack(stack_size, mapper, frame_allocator)?;
let mut stack = unsafe { Stack::new(stack_bounds.end()) };
stack.set_up_for_entry_point(entry_point);
Ok(Self::new(stack.get_stack_pointer(), stack_bounds))
}
pub fn create_from_closure<F>(
closure: F,
stack_size: u64,
mapper: &mut impl Mapper<Size4KiB>,
frame_allocator: &mut impl FrameAllocator<Size4KiB>,
) -> Result<Self, mapper::MapToError>
where
F: FnOnce() -> ! + 'static + Send + Sync,
{
let stack_bounds = alloc_stack(stack_size, mapper, frame_allocator)?;
let mut stack = unsafe { Stack::new(stack_bounds.end()) };
stack.set_up_for_closure(Box::new(closure));
Ok(Self::new(stack.get_stack_pointer(), stack_bounds))
}
fn new(stack_pointer: VirtAddr, stack_bounds: StackBounds) -> Self {
Thread {
id: ThreadId::new(),
stack_pointer: Some(stack_pointer),
stack_bounds: Some(stack_bounds),
}
}
pub(super) fn create_root_thread() -> Self {
Thread {
id: ThreadId(0),
stack_pointer: None,
stack_bounds: None,
}
}
pub fn id(&self) -> ThreadId {
self.id
}
pub(super) fn stack_pointer(&mut self) -> &mut Option<VirtAddr> {
&mut self.stack_pointer
}
}