mirror of
https://github.com/hermitcore/libhermit.git
synced 2025-03-09 00:00:03 +01:00

Add an interface for sharing the free list with the hypervisor. This can be used to accelerate the cold migration of guests.
304 lines
7.7 KiB
C
304 lines
7.7 KiB
C
/*
|
|
* Copyright (c) 2010, Stefan Lankes, RWTH Aachen University
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are met:
|
|
* * Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* * Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* * Neither the name of the University nor the names of its contributors
|
|
* may be used to endorse or promote products derived from this
|
|
* software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
|
|
* WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
|
* DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE FOR ANY
|
|
* DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
|
|
* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
|
|
* LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
|
|
* ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
|
|
* SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#include <hermit/stddef.h>
|
|
#include <hermit/stdlib.h>
|
|
#include <hermit/stdio.h>
|
|
#include <hermit/string.h>
|
|
#include <hermit/spinlock.h>
|
|
#include <hermit/memory.h>
|
|
#include <hermit/logging.h>
|
|
|
|
#include <asm/atomic.h>
|
|
#include <asm/page.h>
|
|
#include <asm/irq.h>
|
|
|
|
extern uint64_t base;
|
|
extern uint64_t limit;
|
|
|
|
typedef struct free_list {
|
|
size_t start, end;
|
|
struct free_list* next;
|
|
struct free_list* prev;
|
|
} free_list_t;
|
|
|
|
/*
|
|
* Note that linker symbols are not variables, they have no memory allocated for
|
|
* maintaining a value, rather their address is their value.
|
|
*/
|
|
extern const void kernel_start;
|
|
|
|
// defined in entry.S
|
|
extern size_t image_size;
|
|
|
|
static spinlock_t list_lock = SPINLOCK_INIT;
|
|
|
|
static free_list_t init_list = {0, 0, NULL, NULL};
|
|
static free_list_t* free_start = (free_list_t*) &init_list;
|
|
|
|
atomic_int64_t total_pages = ATOMIC_INIT(0);
|
|
atomic_int64_t total_allocated_pages = ATOMIC_INIT(0);
|
|
atomic_int64_t total_available_pages = ATOMIC_INIT(0);
|
|
|
|
free_list_t *get_free_list(void)
|
|
{
|
|
return free_start;
|
|
}
|
|
|
|
size_t get_pages(size_t npages)
|
|
{
|
|
size_t i, ret = 0;
|
|
free_list_t* curr = free_start;
|
|
|
|
if (BUILTIN_EXPECT(!npages, 0))
|
|
return 0;
|
|
if (BUILTIN_EXPECT(npages > atomic_int64_read(&total_available_pages), 0))
|
|
return 0;
|
|
|
|
spinlock_lock(&list_lock);
|
|
|
|
while(curr) {
|
|
i = (curr->end - curr->start) / PAGE_SIZE;
|
|
if (i > npages) {
|
|
ret = curr->start;
|
|
curr->start += npages * PAGE_SIZE;
|
|
goto out;
|
|
} else if (i == npages) {
|
|
ret = curr->start;
|
|
if (curr->prev)
|
|
curr->prev = curr->next;
|
|
else
|
|
free_start = curr->next;
|
|
if (curr != &init_list)
|
|
kfree(curr);
|
|
goto out;
|
|
}
|
|
|
|
curr = curr->next;
|
|
}
|
|
out:
|
|
LOG_DEBUG("get_pages: ret 0x%llx, npages %zd, curr->start 0x%llx, curr->end 0x%llx\n", ret, npages, curr->start, curr->end);
|
|
|
|
spinlock_unlock(&list_lock);
|
|
|
|
if (ret) {
|
|
atomic_int64_add(&total_allocated_pages, npages);
|
|
atomic_int64_sub(&total_available_pages, npages);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
DEFINE_PER_CORE(size_t, ztmp_addr, 0);
|
|
|
|
size_t get_zeroed_page(void)
|
|
{
|
|
size_t phyaddr = get_page();
|
|
size_t viraddr;
|
|
uint8_t flags;
|
|
|
|
if (BUILTIN_EXPECT(!phyaddr, 0))
|
|
return 0;
|
|
|
|
flags = irq_nested_disable();
|
|
|
|
viraddr = per_core(ztmp_addr);
|
|
if (BUILTIN_EXPECT(!viraddr, 0))
|
|
{
|
|
viraddr = vma_alloc(PAGE_SIZE, VMA_READ|VMA_WRITE|VMA_CACHEABLE);
|
|
if (BUILTIN_EXPECT(!viraddr, 0))
|
|
goto novaddr;
|
|
|
|
LOG_DEBUG("Core %d uses 0x%zx as temporary address\n", CORE_ID, viraddr);
|
|
set_per_core(ztmp_addr, viraddr);
|
|
}
|
|
|
|
__page_map(viraddr, phyaddr, 1, PG_GLOBAL|PG_RW|PG_PRESENT);
|
|
|
|
memset((void*) viraddr, 0x00, PAGE_SIZE);
|
|
|
|
novaddr:
|
|
irq_nested_enable(flags);
|
|
|
|
return phyaddr;
|
|
}
|
|
|
|
/* TODO: reunion of elements is still missing */
|
|
int put_pages(size_t phyaddr, size_t npages)
|
|
{
|
|
free_list_t* curr = free_start;
|
|
|
|
if (BUILTIN_EXPECT(!phyaddr, 0))
|
|
return -EINVAL;
|
|
if (BUILTIN_EXPECT(!npages, 0))
|
|
return -EINVAL;
|
|
|
|
spinlock_lock(&list_lock);
|
|
|
|
while(curr) {
|
|
if (phyaddr+npages*PAGE_SIZE == curr->start) {
|
|
curr->start = phyaddr;
|
|
goto out;
|
|
} else if (phyaddr == curr->end) {
|
|
curr->end += npages*PAGE_SIZE;
|
|
goto out;
|
|
} if (phyaddr > curr->end) {
|
|
free_list_t* n = kmalloc(sizeof(free_list_t));
|
|
|
|
if (BUILTIN_EXPECT(!n, 0))
|
|
goto out_err;
|
|
|
|
/* add new element */
|
|
n->start = phyaddr;
|
|
n->end = phyaddr + npages * PAGE_SIZE;
|
|
n->prev = curr;
|
|
n->next = curr->next;
|
|
curr->next = n;
|
|
}
|
|
|
|
curr = curr->next;
|
|
}
|
|
out:
|
|
spinlock_unlock(&list_lock);
|
|
|
|
atomic_int64_sub(&total_allocated_pages, npages);
|
|
atomic_int64_add(&total_available_pages, npages);
|
|
|
|
return 0;
|
|
|
|
out_err:
|
|
spinlock_unlock(&list_lock);
|
|
|
|
return -ENOMEM;
|
|
}
|
|
|
|
void* page_alloc(size_t sz, uint32_t flags)
|
|
{
|
|
size_t viraddr = 0;
|
|
size_t phyaddr;
|
|
uint32_t npages = PAGE_FLOOR(sz) >> PAGE_BITS;
|
|
size_t pflags = PG_PRESENT|PG_GLOBAL; //|PG_XD;
|
|
|
|
if (BUILTIN_EXPECT(!npages, 0))
|
|
goto oom;
|
|
|
|
viraddr = vma_alloc(PAGE_FLOOR(sz), flags);
|
|
if (BUILTIN_EXPECT(!viraddr, 0))
|
|
goto oom;
|
|
|
|
phyaddr = get_pages(npages);
|
|
if (BUILTIN_EXPECT(!phyaddr, 0))
|
|
{
|
|
vma_free(viraddr, viraddr+npages*PAGE_SIZE);
|
|
viraddr = 0;
|
|
goto oom;
|
|
}
|
|
|
|
if (flags & VMA_WRITE)
|
|
pflags |= PG_RW;
|
|
if (!(flags & VMA_CACHEABLE))
|
|
pflags |= PG_PCD;
|
|
|
|
int ret = page_map(viraddr, phyaddr, npages, pflags);
|
|
if (BUILTIN_EXPECT(ret, 0))
|
|
{
|
|
vma_free(viraddr, viraddr+npages*PAGE_SIZE);
|
|
put_pages(phyaddr, npages);
|
|
viraddr = 0;
|
|
}
|
|
|
|
oom:
|
|
return (void*) viraddr;
|
|
}
|
|
|
|
void page_free(void* viraddr, size_t sz)
|
|
{
|
|
size_t phyaddr;
|
|
|
|
if (BUILTIN_EXPECT(!viraddr || !sz, 0))
|
|
return;
|
|
|
|
phyaddr = virt_to_phys((size_t)viraddr);
|
|
|
|
vma_free((size_t) viraddr, (size_t) viraddr + PAGE_FLOOR(sz));
|
|
|
|
if (phyaddr)
|
|
put_pages(phyaddr, PAGE_FLOOR(sz) >> PAGE_BITS);
|
|
}
|
|
|
|
int memory_init(void)
|
|
{
|
|
int ret = 0;
|
|
|
|
// enable paging and map Multiboot modules etc.
|
|
ret = page_init();
|
|
if (BUILTIN_EXPECT(ret, 0)) {
|
|
LOG_ERROR("Failed to initialize paging!\n");
|
|
return ret;
|
|
}
|
|
|
|
LOG_INFO("memory_init: base 0x%zx, image_size 0x%zx, limit 0x%zx\n", base, image_size, limit);
|
|
|
|
// determine available memory
|
|
atomic_int64_add(&total_pages, (limit-base) >> PAGE_BITS);
|
|
atomic_int64_add(&total_available_pages, (limit-base) >> PAGE_BITS);
|
|
|
|
//initialize free list
|
|
init_list.start = PAGE_FLOOR((size_t) &kernel_start + image_size + (16+511)*PAGE_SIZE);
|
|
if (limit < GICD_BASE)
|
|
init_list.end = limit;
|
|
else
|
|
init_list.end = GICD_BASE;
|
|
|
|
// determine allocated memory
|
|
atomic_int64_add(&total_allocated_pages, PAGE_FLOOR((size_t) &kernel_start + image_size + 511*PAGE_SIZE) >> PAGE_BITS);
|
|
atomic_int64_sub(&total_available_pages, PAGE_FLOOR((size_t) &kernel_start + image_size + 511*PAGE_SIZE) >> PAGE_BITS);
|
|
|
|
LOG_INFO("free list starts at 0x%zx, limit 0x%zx\n", init_list.start, init_list.end);
|
|
|
|
ret = vma_init();
|
|
if (BUILTIN_EXPECT(ret, 0))
|
|
LOG_WARNING("Failed to initialize VMA regions: %d\n", ret);
|
|
|
|
if (limit > GICD_BASE + GICD_SIZE + GICC_SIZE) {
|
|
init_list.next = kmalloc(sizeof(free_list_t));
|
|
if (BUILTIN_EXPECT(!init_list.next, 0)) {
|
|
LOG_ERROR("Unable to allocate new element for the free list\n");
|
|
goto oom;
|
|
}
|
|
|
|
LOG_INFO("Add region 0x%zx - 0x%zx\n", GICD_BASE + GICD_SIZE + GICC_SIZE, limit);
|
|
|
|
init_list.next->prev = &init_list;
|
|
init_list.next->next = NULL;
|
|
init_list.start = GICD_BASE + GICD_SIZE + GICC_SIZE;
|
|
init_list.end = limit;
|
|
}
|
|
|
|
oom:
|
|
return ret;
|
|
}
|