/* * Copyright (c) 2023-2024 Ian Marco Moffett and the Osmora Team. * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are met: * * 1. Redistributions of source code must retain the above copyright notice, * this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. Neither the name of Hyra nor the names of its * contributors may be used to endorse or promote products derived from * this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE * POSSIBILITY OF SUCH DAMAGE. */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include /* * Thread ready queue - all threads ready to be * scheduled should be added to this queue. */ static TAILQ_HEAD(, proc) td_queue; static size_t nthread = 0; /* * Thread queue lock - all operations to `td_queue' * must be done with this lock acquired. */ static struct spinlock tdq_lock = {0}; /* * Perform timer oneshot * * @now: True for shortest timeslice. */ static inline void sched_oneshot(bool now) { struct timer timer; size_t usec = (now) ? SHORT_TIMESLICE_USEC : DEFAULT_TIMESLICE_USEC; tmrr_status_t tmr_status; tmr_status = req_timer(TIMER_SCHED, &timer); __assert(tmr_status == TMRR_SUCCESS); timer.oneshot_us(usec); } /* * Push a thread into the thread ready queue * allowing it to be eventually dequeued * and ran. */ static void sched_enqueue_td(struct proc *td) { /* Sanity check */ if (td == NULL) return; spinlock_acquire(&tdq_lock); td->pid = nthread++; TAILQ_INSERT_TAIL(&td_queue, td, link); spinlock_release(&tdq_lock); } /* * Dequeue the first thread in the thread ready * queue. */ static struct proc * sched_dequeue_td(void) { struct proc *td = NULL; spinlock_acquire(&tdq_lock); if (!TAILQ_EMPTY(&td_queue)) { td = TAILQ_FIRST(&td_queue); TAILQ_REMOVE(&td_queue, td, link); } spinlock_release(&tdq_lock); return td; } __noreturn static void sched_idle(void) { for (;;) { hint_spinwait(); } } /* * Processor awaiting tasks to be assigned will be here spinning. */ __noreturn static void sched_enter(void) { sched_oneshot(false); sched_idle(); __builtin_unreachable(); } static uintptr_t sched_create_stack(struct vas vas, bool user, struct exec_args args, struct proc *td) { int status; uintptr_t stack; const vm_prot_t USER_STACK_PROT = PROT_WRITE | PROT_USER; struct vm_range *stack_range = &td->addr_range[ADDR_RANGE_STACK]; if (!user) { stack = (uintptr_t)dynalloc(PROC_STACK_SIZE); stack_range->start = (uintptr_t)stack; stack_range->end = (uintptr_t)stack + PROC_STACK_SIZE; return loader_init_stack((void *)(stack + PROC_STACK_SIZE), args); } stack = vm_alloc_pageframe(PROC_STACK_PAGES); stack_range->start = stack; stack_range->end = stack + PROC_STACK_SIZE; status = vm_map_create(vas, stack, stack, USER_STACK_PROT, PROC_STACK_SIZE); if (status != 0) { return 0; } memset(USER_TO_KERN(stack), 0, PROC_STACK_SIZE); stack = loader_init_stack((void *)USER_TO_KERN(stack + PROC_STACK_SIZE), args); return stack; } static struct proc * sched_create_td(uintptr_t rip, struct exec_args args, bool is_user, struct vm_range *prog_range) { struct proc *td; struct vm_range *exec_range; uintptr_t stack; struct trapframe *tf; tf = dynalloc(sizeof(struct trapframe)); if (tf == NULL) { return NULL; } td = dynalloc(sizeof(struct proc)); if (td == NULL) { /* TODO: Free stack */ dynfree(tf); return NULL; } memset(td, 0, sizeof(struct proc)); stack = sched_create_stack(args.vas, is_user, args, td); if (stack == 0) { dynfree(tf); dynfree(td); return NULL; } memset(tf, 0, sizeof(struct trapframe)); /* Setup process itself */ exec_range = &td->addr_range[ADDR_RANGE_EXEC]; td->pid = 0; /* Don't assign PID until enqueued */ td->cpu = NULL; /* Not yet assigned a core */ td->tf = tf; td->addrsp = args.vas; td->is_user = is_user; for (size_t i = 0; i < MTAB_ENTRIES; ++i) { /* Init the memory mapping table */ TAILQ_INIT(&td->mapspace.mtab[i]); } if (prog_range != NULL) { memcpy(exec_range, prog_range, sizeof(struct vm_range)); } processor_init_pcb(td); /* Allocate standard file descriptors */ __assert(fd_alloc(td, NULL) == 0); /* STDIN */ __assert(fd_alloc(td, NULL) == 0); /* STDOUT */ __assert(fd_alloc(td, NULL) == 0); /* STDERR */ /* Setup trapframe */ if (!is_user) { init_frame(tf, rip, (uintptr_t)stack); } else { init_frame_user(tf, rip, KERN_TO_USER(stack)); } return td; } static void sched_destroy_td(struct proc *td) { const struct vm_range *stack_range = &td->addr_range[ADDR_RANGE_STACK]; struct vm_range *exec_range = &td->addr_range[ADDR_RANGE_EXEC]; vm_mapq_t *mapq; processor_free_pcb(td); /* * User stacks are allocated with vm_alloc_pageframe(), * while kernel stacks are allocated with dynalloc(). * We want to check if we are a user program or kernel * program to perform the proper deallocation method. */ if (td->is_user) { vm_free_pageframe(stack_range->start, PROC_STACK_PAGES); } else { dynfree((void *)stack_range->start); } /* Close all of the file descriptors */ for (size_t i = 0; i < PROC_MAX_FDS; ++i) { fd_close_fdnum(td, i); } for (size_t i = 0; i < MTAB_ENTRIES; ++i) { mapq = &td->mapspace.mtab[i]; vm_free_mapq(mapq); } loader_unload(td->addrsp, exec_range); pmap_free_vas(vm_get_ctx(), td->addrsp); dynfree(td); } /* * Create the idle thread. */ static void sched_make_idletd(void) { struct exec_args exec_args = {0}; struct vm_range range; struct proc *td; char *argv[] = {NULL}; char *envp[] = {NULL}; exec_args.argp = argv; exec_args.envp = envp; exec_args.vas = pmap_read_vas(); td = sched_create_td((uintptr_t)sched_idle, exec_args, false, &range); sched_enqueue_td(td); } /* * Cause an early preemption and lets * the next thread run. */ void sched_rest(void) { sched_oneshot(true); } void sched_exit(void) { struct proc *td; struct vas kvas = vm_get_kvas(); intr_mask(); td = this_td(); spinlock_acquire(&td->lock); /* Never release */ /* Switch back to the kernel address space and destroy ourself */ pmap_switch_vas(vm_get_ctx(), kvas); TAILQ_REMOVE(&td_queue, td, link); sched_destroy_td(td); intr_unmask(); sched_enter(); } /* * Get the current running thread. */ struct proc * this_td(void) { struct sched_state *state; struct cpu_info *ci; ci = this_cpu(); state = &ci->sched_state; return state->td; } /* * Thread context switch routine */ void sched_context_switch(struct trapframe *tf) { struct cpu_info *ci = this_cpu(); struct sched_state *state = &ci->sched_state; struct proc *td, *next_td; if (state->td != NULL) { signal_handle(state->td); } /* * If we have no threads, we should not * preempt at all. */ if (nthread == 0 || (next_td = sched_dequeue_td()) == NULL) { sched_oneshot(false); return; } /* * If we have a thread currently running and we are switching * to another, we shall save our current register state * by copying the trapframe. */ if (state->td != NULL) { td = state->td; memcpy(td->tf, tf, sizeof(struct trapframe)); } /* Copy over the next thread's register state to us */ memcpy(tf, next_td->tf, sizeof(struct trapframe)); td = state->td; state->td = next_td; /* Re-enqueue the previous thread if it exists */ if (td != NULL) { sched_enqueue_td(td); } /* Do architecture specific context switch logic */ processor_switch_to(td, next_td); /* Done, switch out our vas and oneshot */ pmap_switch_vas(vm_get_ctx(), next_td->addrsp); sched_oneshot(false); } uint64_t sys_exit(struct syscall_args *args) { sched_exit(); __builtin_unreachable(); } void sched_init(void) { struct proc *init; struct vm_range init_range; const char *init_bin; struct exec_args exec_args = {0}; struct auxval *auxvp = &exec_args.auxv; struct vas vas; char *argv[] = {"/usr/sbin/init", NULL}; char *envp[] = {NULL}; vas = pmap_create_vas(vm_get_ctx()); exec_args.vas = vas; exec_args.argp = argv; exec_args.envp = envp; TAILQ_INIT(&td_queue); sched_make_idletd(); if ((init_bin = initramfs_open("/usr/sbin/init")) == NULL) { panic("Could not open /usr/sbin/init\n"); } if (loader_load(vas, init_bin, auxvp, 0, NULL, &init_range) != 0) { panic("Could not load init\n"); } init = sched_create_td((uintptr_t)auxvp->at_entry, exec_args, true, &init_range); if (init == NULL) { panic("Failed to create thread for init\n"); } sched_enqueue_td(init); } /* * Setup scheduler related things and enqueue AP. */ void sched_init_processor(struct cpu_info *ci) { struct sched_state *sched_state = &ci->sched_state; (void)sched_state; /* TODO */ sched_enter(); __builtin_unreachable(); }