genode/repos/base-pistachio/src/core/platform.cc

651 lines
17 KiB
C++

/*
* \brief Pistachio platform interface implementation
* \author Christian Helmuth
* \date 2006-04-11
*/
/*
* Copyright (C) 2006-2017 Genode Labs GmbH
*
* This file is part of the Genode OS framework, which is distributed
* under the terms of the GNU Affero General Public License version 3.
*/
/* Genode includes */
#include <base/log.h>
#include <base/allocator_avl.h>
#include <base/sleep.h>
#include <base/capability.h>
#include <util/misc_math.h>
/* base-internal includes */
#include <base/internal/crt0.h>
#include <base/internal/stack_area.h>
#include <base/internal/capability_space_tpl.h>
#include <base/internal/globals.h>
/* core includes */
#include <boot_modules.h>
#include <core_log.h>
#include <map_local.h>
#include <platform.h>
#include <platform_thread.h>
#include <platform_pd.h>
#include <util.h>
#include <kip.h>
#include <print_l4_thread_id.h>
/* Pistachio includes */
namespace Pistachio {
#include <l4/kip.h>
#include <l4/sigma0.h>
#include <l4/space.h>
#include <l4/bootinfo.h>
#include <l4/schedule.h>
}
using namespace Genode;
static const bool verbose = true;
static const bool verbose_core_pf = true;
/***********************************
** Core address space management **
***********************************/
static Synced_range_allocator<Allocator_avl> &_core_address_ranges()
{
static Synced_range_allocator<Allocator_avl> _core_address_ranges(nullptr);
return _core_address_ranges;
}
enum { PAGER_STACK_ELEMENTS = 512 };
static unsigned long _core_pager_stack[PAGER_STACK_ELEMENTS];
static inline bool write_fault(Pistachio::L4_Word_t flags) {
return (flags & 2); }
static bool wait_for_page_fault(Pistachio::L4_ThreadId_t &from,
Pistachio::L4_Word_t &pf_addr,
Pistachio::L4_Word_t &pf_ip,
Pistachio::L4_Word_t &flags)
{
using namespace Pistachio;
L4_Accept(L4_UntypedWordsAcceptor);
L4_MsgTag_t res = L4_Wait(&from);
L4_Msg_t msg;
enum { EXPECT = 2 };
if (L4_IpcFailed(res) || (L4_UntypedWords(res)) != EXPECT) {
error(__func__, ": got ", L4_UntypedWords(res), " words, "
"expected ", (int)EXPECT);
return false;
}
L4_Store(res, &msg);
pf_addr = L4_Get(&msg, 0);
pf_ip = L4_Get(&msg, 1);
flags = res.X.flags;
return true;
}
static bool reply_and_wait_for_page_fault(Pistachio::L4_ThreadId_t to,
Pistachio::L4_MapItem_t item,
Pistachio::L4_ThreadId_t &from,
Pistachio::L4_Word_t &pf_addr,
Pistachio::L4_Word_t &pf_ip,
Pistachio::L4_Word_t &flags)
{
using namespace Pistachio;
L4_Msg_t msg;
L4_Clear(&msg);
L4_Append(&msg, item);
L4_Accept(L4_UntypedWordsAcceptor);
L4_MsgLoad(&msg);
L4_MsgTag_t res = L4_ReplyWait(to, &from);
enum { EXPECT = 2 };
if (L4_IpcFailed(res) || (L4_UntypedWords(res)) != EXPECT) {
error(__func__, ": got ", L4_UntypedWords(res), " words, "
"expected ", (int)EXPECT);
return wait_for_page_fault(from, pf_addr, pf_ip, flags);
}
L4_Store(res, &msg);
pf_addr = L4_Get(&msg, 0);
pf_ip = L4_Get(&msg, 1);
flags = res.X.flags;
return true;
}
/****************
** Core pager **
****************/
static void _core_pager_loop()
{
using namespace Pistachio;
L4_ThreadId_t t;
L4_Word_t pf_addr, pf_ip;
L4_Word_t page_size = Genode::get_page_size();
L4_Word_t flags;
L4_MapItem_t item;
bool send_reply = false;
while (1) {
if (send_reply)
reply_and_wait_for_page_fault(t, item, t, pf_addr, pf_ip, flags);
else
wait_for_page_fault(t, pf_addr, pf_ip, flags);
/* XXX Ignore fault messages from non-core tasks */
/*
* Check for local echo mapping request. To request a local
* mappings, a core thread may send an IPC to the core pager with
* the message word 1 (which normally carries the pf_ip) set to 0.
* The message word 0 contains a pointer to a map item to be used
* for the echo reply.
*/
if (pf_ip == 0) {
item = *(L4_MapItem_t *)pf_addr;
send_reply = true;
continue;
}
log(L4_GlobalId(t));
/* check for NULL pointer */
if (pf_addr < page_size) {
error("possible null pointer ",
write_fault(flags) ? "WRITE" : "READ/EXEC", " at "
"address ", Hex(pf_addr), " at EIP ", Hex(pf_ip), " in ", t);
/* do not unblock faulter */
break;
} else if (!_core_address_ranges().valid_addr(pf_addr)) {
/* page-fault address is not in RAM */
error(write_fault(flags) ? "WRITE" : "READ", " access outside of RAM "
"at ", Hex(pf_addr), " IP ", Hex(pf_ip), " by ", t);
/* do not unblock faulter */
break;
} else if (verbose_core_pf) {
log(__func__, ": pfa=", Hex(pf_addr), " ip=", Hex(pf_ip), " by ", t);
}
/* my pf handler is sigma0 - just touch the appropriate page */
L4_Fpage_t res = L4_Sigma0_GetPage(get_sigma0(),
L4_Fpage(trunc_page(pf_addr), page_size));
if (L4_IsNilFpage(res)) {
panic("Unhandled page fault");
}
/* answer pagefault */
L4_Fpage_t fpage = L4_Fpage(pf_addr, page_size);
fpage += L4_FullyAccessible;
item = L4_MapItem(fpage, pf_addr);
send_reply = true;
}
}
Platform::Sigma0::Sigma0()
:
Pager_object(Cpu_session_capability(), Thread_capability(),
0, Affinity::Location(),
Session_label(), Cpu_session::Name("sigma0"))
{
cap(Capability_space::import(Pistachio::get_sigma0(), Rpc_obj_key()));
}
Platform::Sigma0 &Platform::sigma0()
{
static Sigma0 _sigma0;
return _sigma0;
}
Platform::Core_pager::Core_pager(Platform_pd &core_pd)
:
Platform_thread("core.pager"),
Pager_object(Cpu_session_capability(), Thread_capability(),
0, Affinity::Location(),
Session_label(), Cpu_session::Name(name()))
{
Platform_thread::pager(sigma0());
core_pd.bind_thread(*this);
cap(Capability_space::import(native_thread_id(), Rpc_obj_key()));
/* stack begins at the top end of the '_core_pager_stack' array */
void *sp = (void *)&_core_pager_stack[PAGER_STACK_ELEMENTS - 1];
start((void *)_core_pager_loop, sp);
/* pager0 receives pagefaults from me - for NULL pointer detection */
L4_Set_Pager(native_thread_id());
}
Platform::Core_pager &Platform::core_pager()
{
static Core_pager _core_pager(core_pd());
return _core_pager;
}
/***********************************
** Helper for L4 region handling **
***********************************/
struct Region
{
addr_t start;
addr_t end;
Region() : start(0), end(0) { }
Region(addr_t s, addr_t e) : start(s), end(e) { }
/**
* Returns true if the specified range intersects with the region
*/
bool intersects(addr_t base, size_t size) const
{
return (((base + size) > start) && (base < end));
}
void print(Genode::Output &out) const
{
size_t const size = end - start;
Genode::print(out, Hex_range<addr_t>(start, size), " ",
"size: ", Hex(size, Hex::PREFIX, Hex::PAD));
}
};
/**
* Add region to allocator
*/
static inline void add_region(Region r, Range_allocator &alloc)
{
if (r.start >= r.end) {
error(__func__, ": bad range ", r);
panic("add_region called with bogus parameters.");
}
/* adjust region */
addr_t start = trunc_page(r.start);
addr_t end = round_page(r.end);
alloc.add_range(start, end - start);
}
/**
* Remove region from allocator
*/
static inline void remove_region(Region r, Range_allocator &alloc)
{
if (r.start >= r.end)
panic("remove_region called with bogus parameters.");
/* adjust region */
addr_t start = trunc_page(r.start);
addr_t end = round_page(r.end);
alloc.remove_range(start, end - start);
}
static bool intersects_kip_archdep(Pistachio::L4_KernelInterfacePage_t *kip,
addr_t start, size_t size)
{
using namespace Pistachio;
L4_Word_t num_desc = L4_NumMemoryDescriptors(kip);
for (L4_Word_t i = 0; i < num_desc; i++) {
L4_MemoryDesc_t *d = L4_MemoryDesc(kip, i);
enum { ARCH_DEPENDENT_MEM = 15 };
if (!L4_IsVirtual(d) && ((L4_Type(d) & 0xF) == ARCH_DEPENDENT_MEM)) {
if (L4_Low(d) <= start && start <= L4_High(d))
return true;
if (L4_Low(d) <= start+size-1 && start+size-1 <= L4_High(d))
return true;
if (L4_Low(d) <= start && start+size-1 <= L4_High(d))
return true;
}
}
return false;
}
static void dump_kip_memdesc(Pistachio::L4_KernelInterfacePage_t *kip)
{
using namespace Pistachio;
L4_Word_t num_desc = L4_NumMemoryDescriptors(kip);
static const char *types[16] =
{
"Undefined", "Conventional", "Reserved by kernel",
"Dedicated", "Shared", "?", "?", "?", "?", "?",
"?", "?", "?", "?", "Boot loader",
"Architecture-dependent"
};
for (L4_Word_t i = 0; i < num_desc; i++) {
L4_MemoryDesc_t *d = L4_MemoryDesc(kip, i);
log("mem ", i, ": ",
Hex_range<addr_t>(L4_Low(d), L4_High(d) - L4_Low(d) + 1), " "
"type=", Hex(L4_Type(d)), " (", types[L4_Type(d) & 0xf], ") ",
L4_IsVirtual(d) ? "Virtual" : "Non-Virtual");
}
}
/**
* Request any RAM page from Sigma0
*/
bool sigma0_req_region(addr_t *addr, unsigned log2size)
{
using namespace Pistachio;
L4_Fpage_t fpage = L4_Sigma0_GetAny(get_sigma0(), log2size,
L4_CompleteAddressSpace);
if (L4_IsNilFpage(fpage))
return false;
*addr = L4_Address(fpage);
return true;
}
void Platform::_setup_mem_alloc()
{
Pistachio::L4_KernelInterfacePage_t *kip = Pistachio::get_kip();
/*
* Completely map program image by touching all pages read-only to
* prevent sigma0 from handing out those page as anonymous memory.
*/
volatile const char *beg, *end;
beg = (const char *)(((unsigned)&_prog_img_beg) & get_page_mask());
end = (const char *)&_prog_img_end;
for ( ; beg < end; beg += get_page_size()) (void)(*beg);
Pistachio::L4_Word_t page_size_mask = Pistachio::L4_PageSizeMask(kip);
unsigned int size_log2;
/*
* Allocate all memory from sigma0 in descending page sizes. Only
* try page sizes that are hardware supported.
*/
for ( size_log2 = 31; page_size_mask != 0; size_log2-- ) {
unsigned int size = 1 << size_log2;
/* if this page size is not supported try next */
if ((page_size_mask & size) == 0)
continue;
/* mask out that size bit */
page_size_mask &= ~size;
log("Trying to allocate ", size >> 10, "K pages from sigma0.");
/*
* Suck out sigma0. The spec says that we get only "conventional
* memory". Let's hope this is true.
*/
bool succ = true;
unsigned int bytes_got = 0;
while (succ) {
addr_t addr;
Region region;
succ = sigma0_req_region(&addr, size_log2);
if (!succ)
continue;
/* XXX do not allocate page0 */
if (addr == 0) {
// L4_Fpage_t f = L4_FpageLog2(0, pslog2);
// f += L4_FullyAccessible;
// L4_Flush(f);
continue;
}
/* exclude phys ram which is unaccessible - 1:1 mappings! */
if (addr >= _vm_start + _vm_size)
continue;
if ((addr + size > _vm_start + _vm_size) || ( addr + size <= addr))
size = _vm_start + _vm_size - addr;
region.start = addr; region.end = addr + size;
if (region.intersects(stack_area_virtual_base(),
stack_area_virtual_size()) ||
intersects_kip_archdep(kip, addr, size)) {
unmap_local(region.start, size >> get_page_size_log2());
} else {
add_region(region, _ram_alloc);
add_region(region, _core_address_ranges());
}
remove_region(region, _io_mem_alloc);
remove_region(region, _region_alloc);
bytes_got += size;
}
log("Got ", bytes_got >> 10, "K in ", size >> 10, "K pieces.");
}
}
void Platform::_setup_irq_alloc() { _irq_alloc.add_range(0, 0x10); }
void Platform::_setup_preemption()
{
/*
* The roottask has the maximum priority
*/
L4_Set_Priority(Pistachio::L4_Myself(),
Platform_thread::DEFAULT_PRIORITY);
}
void Platform::_setup_basics()
{
using namespace Pistachio;
/* store mapping base from received mapping */
L4_KernelInterfacePage_t *kip = get_kip();
if (kip->magic != L4_MAGIC)
panic("we got something but not the KIP");
if (verbose) {
log("");
log("KIP @ ", kip);
log(" magic: ", Hex(kip->magic, Hex::PREFIX, Hex::PAD));
log(" version: ", Hex(kip->ApiVersion.raw, Hex::PREFIX, Hex::PAD));
log(" BootInfo: ", Hex(kip->BootInfo, Hex::PREFIX, Hex::PAD));
}
dump_kip_memdesc(kip);
/* add KIP as ROM module */
_rom_fs.insert(&_kip_rom);
L4_Fpage_t bipage = L4_Sigma0_GetPage(get_sigma0(),
L4_Fpage(kip->BootInfo,
get_page_size()));
if (L4_IsNilFpage(bipage))
panic("Could not map BootInfo.");
/* done magic */
/* get UTCB memory */
Platform_pd::touch_utcb_space();
/* I/O memory could be the whole user address space */
_io_mem_alloc.add_range(0, ~0);
unsigned int kip_size = sizeof(L4_KernelInterfacePage_t);
_vm_start = 0x0;
_vm_size = 0x0;
/*
* Determine the valid virtual address range by iterating through the
* memory descriptors provided by the KIP. We expect only one
* virtual-memory descriptor.
*/
for (unsigned i = 0; i < L4_NumMemoryDescriptors(kip); i++) {
L4_MemoryDesc_t *md = L4_MemoryDesc(kip, i);
if (!L4_IsVirtual(md)) continue;
if (_vm_start != 0x0 || _vm_size != 0x0) {
warning("KIP has multiple virtual-memory descriptors. Taking only the first.");
break;
}
/*
* Exclude the zero page so that we are able to see null-pointer
* dereference bugs.
*/
_vm_start = max((L4_Word_t)0x1000, L4_MemoryDescLow(md));
_vm_size = L4_MemoryDescHigh(md) - _vm_start + 1;
log("KIP reports virtual memory region at ",
Hex_range<addr_t>(L4_MemoryDescLow(md),
L4_MemoryDescHigh(md) - L4_MemoryDescLow(md)));
}
/* configure core's virtual memory, exclude KIP, stack area */
_region_alloc.add_range(_vm_start, _vm_size);
_region_alloc.remove_range((addr_t)kip, kip_size);
_region_alloc.remove_range(stack_area_virtual_base(),
stack_area_virtual_size());
/* remove KIP area from region and IO_MEM allocator */
remove_region(Region((addr_t)kip, (addr_t)kip + kip_size), _region_alloc);
/* remove utcb area */
addr_t utcb_ptr = (addr_t)Platform_pd::_core_utcb_ptr;
remove_region(Region(utcb_ptr, utcb_ptr + L4_UtcbAreaSize (kip)), _region_alloc);
/* remove core program image memory from region allocator */
addr_t img_start = (addr_t) &_prog_img_beg;
addr_t img_end = (addr_t) &_prog_img_end;
remove_region(Region(img_start, img_end), _region_alloc);
remove_region(Region(img_start, img_end), _io_mem_alloc);
/* image is accessible by core */
add_region(Region(img_start, img_end), _core_address_ranges());
}
Platform_pd &Platform::core_pd()
{
/* on first call, setup task object for core task */
static Platform_pd _core_pd(true);
return _core_pd;
}
Platform::Platform()
:
_ram_alloc(nullptr), _io_mem_alloc(&core_mem_alloc()),
_io_port_alloc(&core_mem_alloc()), _irq_alloc(&core_mem_alloc()),
_region_alloc(&core_mem_alloc()),
_kip_rom((addr_t)Pistachio::get_kip(),
sizeof(Pistachio::L4_KernelInterfacePage_t), "pistachio_kip")
{
/*
* We must be single-threaded at this stage and so this is safe.
*/
static bool initialized = 0;
if (initialized) panic("Platform constructed twice!");
initialized = true;
_setup_basics();
_setup_preemption();
_setup_mem_alloc();
_setup_io_port_alloc();
_setup_irq_alloc();
_init_rom_modules();
log(_rom_fs);
/*
* We setup the thread object for thread0 in core task using a
* special interface that allows us to specify the thread
* ID. For core, this creates the situation that task_id ==
* thread_id of first task. But since we do not destroy this
* task, it should be no problem.
*/
static Platform_thread core_thread("core.main");
core_thread.set_l4_thread_id(Pistachio::L4_MyGlobalId());
core_thread.pager(sigma0());
core_pd().bind_thread(core_thread);
/* core log as ROM module */
{
void * phys_ptr = nullptr;
unsigned const pages = 1;
size_t const log_size = pages << get_page_size_log2();
ram_alloc().alloc_aligned(log_size, &phys_ptr, get_page_size_log2());
addr_t const phys_addr = reinterpret_cast<addr_t>(phys_ptr);
void * const core_local_ptr = phys_ptr;
addr_t const core_local_addr = phys_addr;
/* let one page free after the log buffer */
region_alloc().remove_range(core_local_addr, log_size + get_page_size());
memset(core_local_ptr, 0, log_size);
_rom_fs.insert(new (core_mem_alloc()) Rom_module(phys_addr, log_size,
"core_log"));
init_core_log(Core_log_range { core_local_addr, log_size } );
}
}
/********************************
** Generic platform interface **
********************************/
void Platform::wait_for_exit()
{
/*
* On Pistachio, core never exits. So let us sleep forever.
*/
sleep_forever();
}