/* * Copyright (c) 2017, ARM Limited and Contributors. All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are met: * * Redistributions of source code must retain the above copyright notice, this * list of conditions and the following disclaimer. * * Redistributions in binary form must reproduce the above copyright notice, * this list of conditions and the following disclaimer in the documentation * and/or other materials provided with the distribution. * * Neither the name of ARM nor the names of its contributors may be used * to endorse or promote products derived from this software without specific * prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE * POSSIBILITY OF SUCH DAMAGE. */ #include #include #include #include #include #include #include #include #include #include #include #include #ifdef AARCH32 # include "aarch32/xlat_tables_arch.h" #else # include "aarch64/xlat_tables_arch.h" #endif #include "xlat_tables_private.h" /* * Private variables used by the TF */ static mmap_region_t tf_mmap[MAX_MMAP_REGIONS + 1]; static uint64_t tf_xlat_tables[MAX_XLAT_TABLES][XLAT_TABLE_ENTRIES] __aligned(XLAT_TABLE_SIZE) __section("xlat_table"); static uint64_t tf_base_xlat_table[NUM_BASE_LEVEL_ENTRIES] __aligned(NUM_BASE_LEVEL_ENTRIES * sizeof(uint64_t)); static mmap_region_t tf_mmap[MAX_MMAP_REGIONS + 1]; #if PLAT_XLAT_TABLES_DYNAMIC static int xlat_tables_mapped_regions[MAX_XLAT_TABLES]; #endif /* PLAT_XLAT_TABLES_DYNAMIC */ xlat_ctx_t tf_xlat_ctx = { .pa_max_address = PLAT_PHY_ADDR_SPACE_SIZE - 1, .va_max_address = PLAT_VIRT_ADDR_SPACE_SIZE - 1, .mmap = tf_mmap, .mmap_num = MAX_MMAP_REGIONS, .tables = tf_xlat_tables, .tables_num = MAX_XLAT_TABLES, #if PLAT_XLAT_TABLES_DYNAMIC .tables_mapped_regions = xlat_tables_mapped_regions, #endif /* PLAT_XLAT_TABLES_DYNAMIC */ .base_table = tf_base_xlat_table, .base_table_entries = NUM_BASE_LEVEL_ENTRIES, .max_pa = 0, .max_va = 0, .next_table = 0, .base_level = XLAT_TABLE_LEVEL_BASE, .initialized = 0 }; void mmap_add_region(unsigned long long base_pa, uintptr_t base_va, size_t size, mmap_attr_t attr) { mmap_region_t mm = { .base_va = base_va, .base_pa = base_pa, .size = size, .attr = attr, }; mmap_add_region_ctx(&tf_xlat_ctx, (mmap_region_t *)&mm); } void mmap_add(const mmap_region_t *mm) { while (mm->size) { mmap_add_region_ctx(&tf_xlat_ctx, (mmap_region_t *)mm); mm++; } } #if PLAT_XLAT_TABLES_DYNAMIC int mmap_add_dynamic_region(unsigned long long base_pa, uintptr_t base_va, size_t size, mmap_attr_t attr) { mmap_region_t mm = { .base_va = base_va, .base_pa = base_pa, .size = size, .attr = attr, }; return mmap_add_dynamic_region_ctx(&tf_xlat_ctx, &mm); } int mmap_remove_dynamic_region(uintptr_t base_va, size_t size) { return mmap_remove_dynamic_region_ctx(&tf_xlat_ctx, base_va, size); } #endif /* PLAT_XLAT_TABLES_DYNAMIC */ void init_xlat_tables(void) { assert(!is_mmu_enabled()); assert(!tf_xlat_ctx.initialized); print_mmap(tf_xlat_ctx.mmap); tf_xlat_ctx.execute_never_mask = xlat_arch_get_xn_desc(xlat_arch_current_el()); init_xlation_table(&tf_xlat_ctx); xlat_tables_print(&tf_xlat_ctx); assert(tf_xlat_ctx.max_va <= PLAT_VIRT_ADDR_SPACE_SIZE - 1); assert(tf_xlat_ctx.max_pa <= PLAT_PHY_ADDR_SPACE_SIZE - 1); init_xlat_tables_arch(tf_xlat_ctx.max_pa); } #ifdef AARCH32 void enable_mmu_secure(unsigned int flags) { enable_mmu_arch(flags, tf_xlat_ctx.base_table); } #else void enable_mmu_el1(unsigned int flags) { enable_mmu_arch(flags, tf_xlat_ctx.base_table); } void enable_mmu_el3(unsigned int flags) { enable_mmu_arch(flags, tf_xlat_ctx.base_table); } #endif /* AARCH32 */