2021-03-17 01:47:45 +01:00
|
|
|
use atomic_polyfill::{AtomicPtr, Ordering};
|
2020-12-26 16:42:44 +01:00
|
|
|
use core::ptr;
|
2021-02-02 05:14:52 +01:00
|
|
|
use core::ptr::NonNull;
|
2020-12-26 16:42:44 +01:00
|
|
|
|
2021-08-25 00:20:29 +02:00
|
|
|
use super::TaskHeader;
|
2020-12-26 16:42:44 +01:00
|
|
|
|
|
|
|
pub(crate) struct RunQueueItem {
|
2021-03-18 00:20:02 +01:00
|
|
|
next: AtomicPtr<TaskHeader>,
|
2020-12-26 16:42:44 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
impl RunQueueItem {
|
|
|
|
pub const fn new() -> Self {
|
|
|
|
Self {
|
|
|
|
next: AtomicPtr::new(ptr::null_mut()),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Atomic task queue using a very, very simple lock-free linked-list queue:
|
|
|
|
///
|
|
|
|
/// To enqueue a task, task.next is set to the old head, and head is atomically set to task.
|
|
|
|
///
|
|
|
|
/// Dequeuing is done in batches: the queue is emptied by atomically replacing head with
|
|
|
|
/// null. Then the batch is iterated following the next pointers until null is reached.
|
|
|
|
///
|
|
|
|
/// Note that batches will be iterated in the reverse order as they were enqueued. This is OK
|
|
|
|
/// for our purposes: it can't crate fairness problems since the next batch won't run until the
|
|
|
|
/// current batch is completely processed, so even if a task enqueues itself instantly (for example
|
|
|
|
/// by waking its own waker) can't prevent other tasks from running.
|
|
|
|
pub(crate) struct RunQueue {
|
2021-03-18 00:20:02 +01:00
|
|
|
head: AtomicPtr<TaskHeader>,
|
2020-12-26 16:42:44 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
impl RunQueue {
|
|
|
|
pub const fn new() -> Self {
|
|
|
|
Self {
|
|
|
|
head: AtomicPtr::new(ptr::null_mut()),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Enqueues an item. Returns true if the queue was empty.
|
2021-03-18 00:20:02 +01:00
|
|
|
pub(crate) unsafe fn enqueue(&self, item: *mut TaskHeader) -> bool {
|
2020-12-26 16:42:44 +01:00
|
|
|
let mut prev = self.head.load(Ordering::Acquire);
|
|
|
|
loop {
|
|
|
|
(*item).run_queue_item.next.store(prev, Ordering::Relaxed);
|
|
|
|
match self
|
|
|
|
.head
|
|
|
|
.compare_exchange_weak(prev, item, Ordering::AcqRel, Ordering::Acquire)
|
|
|
|
{
|
|
|
|
Ok(_) => break,
|
|
|
|
Err(next_prev) => prev = next_prev,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
prev.is_null()
|
|
|
|
}
|
|
|
|
|
2021-03-18 00:20:02 +01:00
|
|
|
pub(crate) unsafe fn dequeue_all(&self, on_task: impl Fn(NonNull<TaskHeader>)) {
|
2020-12-26 16:42:44 +01:00
|
|
|
let mut task = self.head.swap(ptr::null_mut(), Ordering::AcqRel);
|
|
|
|
|
|
|
|
while !task.is_null() {
|
|
|
|
// If the task re-enqueues itself, the `next` pointer will get overwritten.
|
|
|
|
// Therefore, first read the next pointer, and only then process the task.
|
|
|
|
let next = (*task).run_queue_item.next.load(Ordering::Relaxed);
|
|
|
|
|
2021-02-02 05:14:52 +01:00
|
|
|
on_task(NonNull::new_unchecked(task));
|
2020-12-26 16:42:44 +01:00
|
|
|
|
|
|
|
task = next
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|