//! A deferred forced unwinding should occur when a drop handler function //! is active in the call stack and it further calls a function that overflows //! the stack, and when the task does not enable dynamic stack extension. #![no_std] #![no_main] extern crate alloc; use core::{ mem::MaybeUninit, sync::atomic::{AtomicUsize, Ordering}, }; use hopter::{ debug::semihosting::{self, dbg_println}, task, task::main, }; #[main] fn main(_: cortex_m::Peripherals) { task::build() .set_entry(test_task) .disable_dynamic_stack() .set_stack_limit(2048) .spawn_restartable() .unwrap(); } fn test_task() { // A persistent counter. static CNT: AtomicUsize = AtomicUsize::new(0); // Every time the task runs we increment it by 1. let cnt = CNT.fetch_add(1, Ordering::SeqCst); // When the task is executed for the first time, run the drop function. // Even if this drop function uses a large stack frame and will overflow // the task's stack while dynamic stack extension is not enabled, the // segmented stack runtime should still allow the drop function to proceed // because we cannot initiate an unwinding inside a drop function. A // deferred forced unwinding will be executed after the drop handler // finishes. if cnt == 0 { core::mem::drop(HasDrop); } if cnt == 0 { // The task should have been unwound so this print should not be // reachable. dbg_println!("Should not print this."); } if cnt > 0 { dbg_println!("Task successfully restarted after a deferred forced unwinding."); #[cfg(feature = "qemu")] semihosting::terminate(true); #[cfg(not(feature = "qemu"))] { dbg_println!("test complete!"); loop {} } } else { #[cfg(feature = "qemu")] semihosting::terminate(false); #[cfg(not(feature = "qemu"))] { dbg_println!("test complete!"); loop {} } } } #[inline(never)] fn large_func() { let _padding = StackFramePadding::new(); dbg_println!("Large function executed."); } struct HasDrop; // A drop function that uses a large stack frame. impl Drop for HasDrop { #[inline(never)] fn drop(&mut self) { large_func(); dbg_println!("Drop executed."); } } /// A padding that causes large stack frame. struct StackFramePadding { _padding: [u8; 4096], } impl StackFramePadding { /// Use volatile write to prevent the compiler from optimizing away the /// padding. fn new() -> Self { let mut padding = MaybeUninit::<[u8; 4096]>::uninit(); let mut ptr = unsafe { (*padding.as_mut_ptr()).as_mut_ptr() }; for _ in 0..4096 { unsafe { ptr.write_volatile(0); ptr = ptr.offset(1); } } let padding = unsafe { padding.assume_init() }; Self { _padding: padding } } }