diff options
author | Ian Moffett <ian@osmora.org> | 2024-04-20 16:23:59 -0400 |
---|---|---|
committer | Ian Moffett <ian@osmora.org> | 2024-04-20 16:25:49 -0400 |
commit | 11976acd80c0c360de66a99647f58dbb8e5ecf3b (patch) | |
tree | 83491b21087b4078bc54d78c5f66edff09430863 /sys/vm | |
parent | 852b00cbc75b83390984a77048e443a43dd0ae48 (diff) |
kernel: vm: Add initial demand paging support
Signed-off-by: Ian Moffett <ian@osmora.org>
Diffstat (limited to 'sys/vm')
-rw-r--r-- | sys/vm/vm_fault.c | 114 | ||||
-rw-r--r-- | sys/vm/vm_map.c | 67 |
2 files changed, 170 insertions, 11 deletions
diff --git a/sys/vm/vm_fault.c b/sys/vm/vm_fault.c new file mode 100644 index 0000000..ec733e8 --- /dev/null +++ b/sys/vm/vm_fault.c @@ -0,0 +1,114 @@ +/* + * Copyright (c) 2023-2024 Ian Marco Moffett and the Osmora Team. + * All rights reserved. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions are met: + * + * 1. Redistributions of source code must retain the above copyright notice, + * this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * 3. Neither the name of Hyra nor the names of its + * contributors may be used to endorse or promote products derived from + * this software without specific prior written permission. + * + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" + * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE + * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE + * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE + * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS + * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN + * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) + * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE + * POSSIBILITY OF SUCH DAMAGE. + */ + +#include <sys/types.h> +#include <sys/sched.h> +#include <vm/fault.h> +#include <vm/map.h> +#include <vm/pmap.h> +#include <vm/vm.h> +#include <vm/physseg.h> + +static struct vm_mapping * +vm_mapq_search(vm_mapq_t *mq, vaddr_t addr) +{ + struct vm_mapping *mapping; + const struct vm_range *range; + + TAILQ_FOREACH(mapping, mq, link) { + range = &mapping->range; + if (addr >= range->start && addr <= range->end) { + return mapping; + } + } + + return NULL; +} + +static struct vm_mapping * +vm_find_mapping(vaddr_t addr) +{ + struct vm_mapping *mapping; + struct proc *td; + vm_mapq_t *mapq; + + mapping = vm_mapping_fetch(&td->mapspace, addr); + if (mapping != NULL) + return mapping; + + /* Need to search other maps */ + td = this_td(); + for (size_t i = 0; i < MTAB_ENTRIES; ++i) { + mapq = &td->mapspace.mtab[i]; + mapping = vm_mapq_search(mapq, addr); + if (mapping != NULL) + return mapping; + } + + return NULL; +} + +int +vm_fault(vaddr_t va, vm_prot_t access_type) +{ + struct proc *td = this_td(); + struct vm_mapping *mapping; + struct vm_object *vmobj; + + size_t granule = vm_get_page_size(); + vaddr_t va_base = va &= ~(granule - 1); + + int s; + paddr_t pa_base; + + mapping = vm_find_mapping(va_base); + if (mapping == NULL) + return -1; + + if ((vmobj = mapping->vmobj) == NULL) + /* Virtual memory object non-existent */ + return -1; + if (!vmobj->demand) + /* Demand paging not enabled for this object */ + return -1; + if ((access_type & ~mapping->prot) != 0) + /* Invalid access type */ + return -1; + + /* Allocate physical memory if needed */ + if (mapping->physmem_base == 0) { + pa_base = vm_alloc_pageframe(1); + mapping->physmem_base = pa_base; + } else { + pa_base = mapping->physmem_base; + } + + s = vm_map_create(td->addrsp, va_base, pa_base, access_type, granule); + return s; +} diff --git a/sys/vm/vm_map.c b/sys/vm/vm_map.c index 8f7accc..7e67ebd 100644 --- a/sys/vm/vm_map.c +++ b/sys/vm/vm_map.c @@ -140,7 +140,7 @@ vm_mapspace_insert(struct vm_mapspace *ms, struct vm_mapping *mapping) } /* - * Create an anonymous mapping. + * Create a mapping (internal helper) * * @addr: Address to map. * @physmem: Physical address, set to 0 to alloc one here @@ -149,7 +149,7 @@ vm_mapspace_insert(struct vm_mapspace *ms, struct vm_mapping *mapping) * Returns zero on failure. */ static paddr_t -vm_anon_map(void *addr, paddr_t physmem, vm_prot_t prot, size_t len) +vm_map(void *addr, paddr_t physmem, vm_prot_t prot, size_t len) { struct proc *td = this_td(); const size_t GRANULE = vm_get_page_size(); @@ -187,7 +187,8 @@ vm_anon_map(void *addr, paddr_t physmem, vm_prot_t prot, size_t len) * @fd: File descriptor. */ static paddr_t -vm_fd_map(void *addr, vm_prot_t prot, size_t len, off_t off, int fd) +vm_fd_map(void *addr, vm_prot_t prot, size_t len, off_t off, int fd, + struct vm_mapping *mapping) { paddr_t physmem = 0; @@ -216,6 +217,9 @@ vm_fd_map(void *addr, vm_prot_t prot, size_t len, off_t off, int fd) if (vm_obj_init(&vp->vmobj, vp) != 0) return 0; + mapping->vmobj = vp->vmobj; + vm_object_ref(vp->vmobj); + /* Try to fetch a physical address */ if (vm_pager_paddr(vp->vmobj, &physmem, prot) != 0) { vm_obj_destroy(vp->vmobj); @@ -228,7 +232,7 @@ vm_fd_map(void *addr, vm_prot_t prot, size_t len, off_t off, int fd) * then connect it to the physical address (creates a shared mapping) */ if (physmem != 0) { - vm_anon_map(addr, physmem, prot, len); + vm_map(addr, physmem, prot, len); return physmem; } @@ -238,7 +242,7 @@ vm_fd_map(void *addr, vm_prot_t prot, size_t len, off_t off, int fd) * anonymous mapping then page-in from whatever filesystem * (creates a shared mapping) */ - physmem = vm_anon_map(addr, 0, prot, len); + physmem = vm_map(addr, 0, prot, len); pg.physaddr = physmem; if (vm_pager_get(vp->vmobj, off, len, &pg) != 0) { @@ -255,6 +259,9 @@ munmap(void *addr, size_t len) struct proc *td = this_td(); struct vm_mapping *mapping; + struct vm_object *obj; + struct vnode *vp; + struct vm_mapspace *ms; size_t map_len, granule; vaddr_t map_start, map_end; @@ -272,6 +279,23 @@ munmap(void *addr, size_t len) map_end = mapping->range.end; map_len = map_end - map_start; + /* Try to release any virtual memory objects */ + if ((obj = mapping->vmobj) != NULL) { + spinlock_acquire(&obj->lock); + /* + * Drop our ref and try to cleanup. If the refcount + * is > 1, something is still holding it and we can't + * do much. + */ + vm_object_unref(obj); + vp = obj->vnode; + if (vp != NULL && obj->ref == 1) { + vp->vmobj = NULL; + vm_obj_destroy(obj); + } + spinlock_release(&obj->lock); + } + /* Release the mapping */ vm_map_destroy(td->addrsp, map_start, map_len); vm_free_pageframe(mapping->range.start, map_len / granule); @@ -292,10 +316,13 @@ mmap(void *addr, size_t len, int prot, int flags, int fildes, off_t off) struct proc *td = this_td(); struct vm_mapping *mapping = ALLOC_MAPPING(); + struct vm_object *vmobj; size_t misalign = ((vaddr_t)addr) & (GRANULE - 1); paddr_t physmem = 0; + mapping->prot = prot | PROT_USER; + /* Ensure of valid prot flags */ if ((prot & ~PROT_MASK) != 0) return MAP_FAILED; @@ -315,15 +342,33 @@ mmap(void *addr, size_t len, int prot, int flags, int fildes, off_t off) * this is. */ if (__TEST(flags, MAP_ANONYMOUS)) { - /* Handle an anonymous map request */ - physmem = vm_anon_map(addr, 0, prot, len); + /* Try to create a virtual memory object */ + if (vm_obj_init(&vmobj, NULL) != 0) + return 0; + + /* + * Enable demand paging for this object if + * `addr` is not NULL. + */ + if (addr != NULL) { + vmobj->is_anon = 1; + vmobj->demand = 1; + + mapping->vmobj = vmobj; + mapping->physmem_base = 0; + } else { + physmem = vm_map(addr, 0, prot, len); + } + + /* Did this work? */ + if (physmem == 0 && addr == NULL) + return MAP_FAILED; } else if (__TEST(flags, MAP_SHARED)) { - physmem = vm_fd_map(addr, prot, len, off, fildes); + physmem = vm_fd_map(addr, prot, len, off, fildes, mapping); + if (physmem == 0) + return MAP_FAILED; } - if (physmem == 0) { - return MAP_FAILED; - } map_start = __ALIGN_DOWN((vaddr_t)addr, GRANULE); map_end = map_start + __ALIGN_UP(len + misalign, GRANULE); |