/* * Copyright (c) 2010, Stefan Lankes, RWTH Aachen University * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are met: * * Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * * Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * Neither the name of the University nor the names of its contributors * may be used to endorse or promote products derived from this * software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE * DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE FOR ANY * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #include #include #include #include #include #include #include #include #include #include extern uint64_t base; extern uint64_t limit; typedef struct free_list { size_t start, end; struct free_list* next; struct free_list* prev; } free_list_t; /* * Note that linker symbols are not variables, they have no memory allocated for * maintaining a value, rather their address is their value. */ extern const void kernel_start; // defined in entry.S extern size_t image_size; static spinlock_t list_lock = SPINLOCK_INIT; static free_list_t init_list = {0, 0, NULL, NULL}; static free_list_t* free_start = (free_list_t*) &init_list; atomic_int64_t total_pages = ATOMIC_INIT(0); atomic_int64_t total_allocated_pages = ATOMIC_INIT(0); atomic_int64_t total_available_pages = ATOMIC_INIT(0); free_list_t *get_free_list(void) { return free_start; } size_t get_pages(size_t npages) { size_t i, ret = 0; free_list_t* curr = free_start; if (BUILTIN_EXPECT(!npages, 0)) return 0; if (BUILTIN_EXPECT(npages > atomic_int64_read(&total_available_pages), 0)) return 0; spinlock_lock(&list_lock); while(curr) { i = (curr->end - curr->start) / PAGE_SIZE; if (i > npages) { ret = curr->start; curr->start += npages * PAGE_SIZE; goto out; } else if (i == npages) { ret = curr->start; if (curr->prev) curr->prev = curr->next; else free_start = curr->next; if (curr != &init_list) kfree(curr); goto out; } curr = curr->next; } out: LOG_DEBUG("get_pages: ret 0x%llx, npages %zd, curr->start 0x%llx, curr->end 0x%llx\n", ret, npages, curr->start, curr->end); spinlock_unlock(&list_lock); if (ret) { atomic_int64_add(&total_allocated_pages, npages); atomic_int64_sub(&total_available_pages, npages); } return ret; } DEFINE_PER_CORE(size_t, ztmp_addr, 0); size_t get_zeroed_page(void) { size_t phyaddr = get_page(); size_t viraddr; uint8_t flags; if (BUILTIN_EXPECT(!phyaddr, 0)) return 0; flags = irq_nested_disable(); viraddr = per_core(ztmp_addr); if (BUILTIN_EXPECT(!viraddr, 0)) { viraddr = vma_alloc(PAGE_SIZE, VMA_READ|VMA_WRITE|VMA_CACHEABLE); if (BUILTIN_EXPECT(!viraddr, 0)) goto novaddr; LOG_DEBUG("Core %d uses 0x%zx as temporary address\n", CORE_ID, viraddr); set_per_core(ztmp_addr, viraddr); } __page_map(viraddr, phyaddr, 1, PG_GLOBAL|PG_RW|PG_PRESENT); memset((void*) viraddr, 0x00, PAGE_SIZE); novaddr: irq_nested_enable(flags); return phyaddr; } /* TODO: reunion of elements is still missing */ int put_pages(size_t phyaddr, size_t npages) { free_list_t* curr = free_start; if (BUILTIN_EXPECT(!phyaddr, 0)) return -EINVAL; if (BUILTIN_EXPECT(!npages, 0)) return -EINVAL; spinlock_lock(&list_lock); while(curr) { if (phyaddr+npages*PAGE_SIZE == curr->start) { curr->start = phyaddr; goto out; } else if (phyaddr == curr->end) { curr->end += npages*PAGE_SIZE; goto out; } if (phyaddr > curr->end) { free_list_t* n = kmalloc(sizeof(free_list_t)); if (BUILTIN_EXPECT(!n, 0)) goto out_err; /* add new element */ n->start = phyaddr; n->end = phyaddr + npages * PAGE_SIZE; n->prev = curr; n->next = curr->next; curr->next = n; } curr = curr->next; } out: spinlock_unlock(&list_lock); atomic_int64_sub(&total_allocated_pages, npages); atomic_int64_add(&total_available_pages, npages); return 0; out_err: spinlock_unlock(&list_lock); return -ENOMEM; } void* page_alloc(size_t sz, uint32_t flags) { size_t viraddr = 0; size_t phyaddr; uint32_t npages = PAGE_FLOOR(sz) >> PAGE_BITS; size_t pflags = PG_PRESENT|PG_GLOBAL; //|PG_XD; if (BUILTIN_EXPECT(!npages, 0)) goto oom; viraddr = vma_alloc(PAGE_FLOOR(sz), flags); if (BUILTIN_EXPECT(!viraddr, 0)) goto oom; phyaddr = get_pages(npages); if (BUILTIN_EXPECT(!phyaddr, 0)) { vma_free(viraddr, viraddr+npages*PAGE_SIZE); viraddr = 0; goto oom; } if (flags & VMA_WRITE) pflags |= PG_RW; if (!(flags & VMA_CACHEABLE)) pflags |= PG_PCD; int ret = page_map(viraddr, phyaddr, npages, pflags); if (BUILTIN_EXPECT(ret, 0)) { vma_free(viraddr, viraddr+npages*PAGE_SIZE); put_pages(phyaddr, npages); viraddr = 0; } oom: return (void*) viraddr; } void page_free(void* viraddr, size_t sz) { size_t phyaddr; if (BUILTIN_EXPECT(!viraddr || !sz, 0)) return; phyaddr = virt_to_phys((size_t)viraddr); vma_free((size_t) viraddr, (size_t) viraddr + PAGE_FLOOR(sz)); if (phyaddr) put_pages(phyaddr, PAGE_FLOOR(sz) >> PAGE_BITS); } int memory_init(void) { int ret = 0; // enable paging and map Multiboot modules etc. ret = page_init(); if (BUILTIN_EXPECT(ret, 0)) { LOG_ERROR("Failed to initialize paging!\n"); return ret; } LOG_INFO("memory_init: base 0x%zx, image_size 0x%zx, limit 0x%zx\n", base, image_size, limit); // determine available memory atomic_int64_add(&total_pages, (limit-base) >> PAGE_BITS); atomic_int64_add(&total_available_pages, (limit-base) >> PAGE_BITS); //initialize free list init_list.start = PAGE_FLOOR((size_t) &kernel_start + image_size + (16+511)*PAGE_SIZE); if (limit < GICD_BASE) init_list.end = limit; else init_list.end = GICD_BASE; // determine allocated memory atomic_int64_add(&total_allocated_pages, PAGE_FLOOR((size_t) &kernel_start + image_size + 511*PAGE_SIZE) >> PAGE_BITS); atomic_int64_sub(&total_available_pages, PAGE_FLOOR((size_t) &kernel_start + image_size + 511*PAGE_SIZE) >> PAGE_BITS); LOG_INFO("free list starts at 0x%zx, limit 0x%zx\n", init_list.start, init_list.end); ret = vma_init(); if (BUILTIN_EXPECT(ret, 0)) LOG_WARNING("Failed to initialize VMA regions: %d\n", ret); if (limit > GICD_BASE + GICD_SIZE + GICC_SIZE) { init_list.next = kmalloc(sizeof(free_list_t)); if (BUILTIN_EXPECT(!init_list.next, 0)) { LOG_ERROR("Unable to allocate new element for the free list\n"); goto oom; } LOG_INFO("Add region 0x%zx - 0x%zx\n", GICD_BASE + GICD_SIZE + GICC_SIZE, limit); init_list.next->prev = &init_list; init_list.next->next = NULL; init_list.start = GICD_BASE + GICD_SIZE + GICC_SIZE; init_list.end = limit; } oom: return ret; }