1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151
//! Implementation of the prologue/epilogue modell.
#![allow(unused)]
use core::cell::UnsafeCell;
use core::ops::{Deref, DerefMut};
use core::sync::atomic::AtomicBool;
use core::sync::atomic::Ordering::Relaxed;
use arraydeque::ArrayDeque;
use super::epilogue::Epilogue;
use crate::arch::cpu;
use crate::util::{Ticket, TicketGuard};
use crate::MAX_CPUS;
/// Size of the epilogue queue
const LEN: usize = 8;
/// Synchronizes the kernel with interrupts using the Prologue/Epilogue Model
pub static GUARD: Guard = Guard::new();
/// Synchronizes the kernel with interrupts using the Prologue/Epilogue Model
///
/// The [Guard] is used to synchronize between "normal" core activities (currently
/// just the text output, later system calls) and interrupt handling routines.
/// For this purpose, [Guard] has to contain one or more "queues",
/// in which [Epilogue] objects can be added. This is necessary if the critical
/// section is occupied at the time when an interrupt occurs, and the [Epilogue::run]
/// method cannot be executed immediately. The queued epilogues are processed
/// when leaving the critical section.
///
/// The [Guard] protects and synchronizes various global kernel objects.
/// These objects are only accessible when the guard is locked, by the control
/// flow holding the lock. This prevents concurrent access and race conditions.
///
/// ## Hints
///
/// The epilogue queue is a central data structure, whose consistency
/// must be ensured. The implementation provided by the [ArrayDeque] is not
/// safe against concurrency, i.e. there must never be accesses by two cores
/// at the same time. You need to disable interrupts during operations on the queue.
///
/// For SMP, you need a separate epilogue queue for each core,
/// in which each processor serializes *its* epilogues. However, epilogues
/// on different cores could then be executed in parallel, since the
/// critical section is managed separately on a per-core basis. This must be
/// prevented by using a global [Ticket] lock to avoid concurrent
/// execution of epilogues -- there must never be more than one epilogue
/// active on the whole system at the same time!
///
/// > *Please note:* This [giant lock](https://en.wikipedia.org/wiki/Giant_lock)
/// > (synchronizing all cores) should not be confused with the (core-specific)
/// > flag variable that marks only the entry to the epilogue level on the
/// > corresponding core!
///
/// Interrupts should be disabled for as short as possible. Due to this
/// reason, the prologue/epilogue model allows epilogues to be interrupted
/// by prologues. This means that interrupts should be
/// [enabled][crate::arch::int::enable] again before the epilogue is
/// executed (this includes notifying the APIC about the "End-Of-Interrupt")
pub struct Guard {
guarded: Ticket<Guarded>,
/// Indicates whether the epilogue layer is active on the corresponding core
active: [AtomicBool; MAX_CPUS],
/// CPU-local queues for epilogues
epilogues: UnsafeCell<[ArrayDeque<Epilogue, LEN>; MAX_CPUS]>,
}
unsafe impl Sync for Guard {}
/// Protected and synchronized kernel objects
pub struct Guarded {
// TODO: BSB A4 - Scheduler
// TODO: BSB A6 - BellRinger, KeyBuffer
}
impl Guard {
pub const fn new() -> Self {
Self {
guarded: Ticket::new(Guarded {}),
active: [const { AtomicBool::new(false) }; MAX_CPUS],
epilogues: UnsafeCell::new([const { ArrayDeque::new() }; MAX_CPUS]),
}
}
fn epilogues(&self) -> &mut ArrayDeque<Epilogue, LEN> {
// TODO: Check if this is safe! When can this be accessed?
&mut (unsafe { &mut *self.epilogues.get() })[cpu::id()]
}
/// Temporary read-only access.
///
/// Safety: beware race conditions!
pub unsafe fn read(&self) -> &Guarded {
self.guarded.raw()
}
/// Enter the epilogue layer or wait synchronously if it is already occupied.
///
/// A guard object is returned that unlocks the epilogue layer when it falls out of scope.
#[track_caller]
pub fn enter(&self) -> GuardedGuard<'_> {
todo!("BSB A3")
}
/// Register the given epilogue, which is either executed directly if possible
/// or it is enqueued for later execution.
pub fn relay(&self, mut _epilogue: Epilogue) {
// We have to secure locking the CPU to prevent a timer interrupt scheduling
// us onto a different core after reading the CPUID (but before setting "epi_flag")
todo!("BSB A3")
}
/// Leave the epilogue layer.
pub fn leave(&self) {
todo!("BSB A3")
}
/// Returns wether the layer 1/2 is active
pub fn active(&self) -> bool {
self.active[cpu::id()].load(Relaxed)
}
}
/// Provides access to protected objects and unlocks the guard
/// when it goes out of scope.
pub struct GuardedGuard<'a> {
inner: Option<TicketGuard<'a, Guarded>>,
}
impl Deref for GuardedGuard<'_> {
type Target = Guarded;
fn deref(&self) -> &Self::Target {
self.inner.as_ref().unwrap()
}
}
impl DerefMut for GuardedGuard<'_> {
fn deref_mut(&mut self) -> &mut Self::Target {
self.inner.as_mut().unwrap()
}
}
impl Drop for GuardedGuard<'_> {
fn drop(&mut self) {
// already unlocked as part of leave
core::mem::forget(self.inner.take());
GUARD.leave();
}
}