Commit ef27980d authored by Andrew Thoelke's avatar Andrew Thoelke
Browse files

Merge pull request #69 from sandrine-bailleux:sb/split-mmu-fcts-per-el

parents 19ea62d3 b793e431
...@@ -203,7 +203,7 @@ func tsp_cpu_on_entry ...@@ -203,7 +203,7 @@ func tsp_cpu_on_entry
* Initialise the MMU * Initialise the MMU
* --------------------------------------------- * ---------------------------------------------
*/ */
bl enable_mmu bl enable_mmu_el1
/* --------------------------------------------- /* ---------------------------------------------
* Give ourselves a stack allocated in Normal * Give ourselves a stack allocated in Normal
......
...@@ -106,9 +106,7 @@ void __dead2 raise_el(aapcs64_params_t *args) ...@@ -106,9 +106,7 @@ void __dead2 raise_el(aapcs64_params_t *args)
*/ */
void __dead2 change_el(el_change_info_t *info) void __dead2 change_el(el_change_info_t *info)
{ {
unsigned long current_el = read_current_el(); if (IS_IN_EL3()) {
if (GET_EL(current_el) == MODE_EL3) {
/* /*
* We can go anywhere from EL3. So find where. * We can go anywhere from EL3. So find where.
* TODO: Lots to do if we are going non-secure. * TODO: Lots to do if we are going non-secure.
...@@ -551,7 +549,6 @@ void __dead2 run_image(unsigned long entrypoint, ...@@ -551,7 +549,6 @@ void __dead2 run_image(unsigned long entrypoint,
void *second_arg) void *second_arg)
{ {
el_change_info_t run_image_info; el_change_info_t run_image_info;
unsigned long current_el = read_current_el();
/* Tell next EL what we want done */ /* Tell next EL what we want done */
run_image_info.args.arg0 = RUN_IMAGE; run_image_info.args.arg0 = RUN_IMAGE;
...@@ -565,7 +562,7 @@ void __dead2 run_image(unsigned long entrypoint, ...@@ -565,7 +562,7 @@ void __dead2 run_image(unsigned long entrypoint,
* to jump to a higher EL and issue an SMC. Contents of argY * to jump to a higher EL and issue an SMC. Contents of argY
* will go into the general purpose register xY e.g. arg0->x0 * will go into the general purpose register xY e.g. arg0->x0
*/ */
if (GET_EL(current_el) == MODE_EL3) { if (IS_IN_EL3()) {
run_image_info.args.arg1 = (unsigned long) first_arg; run_image_info.args.arg1 = (unsigned long) first_arg;
run_image_info.args.arg2 = (unsigned long) second_arg; run_image_info.args.arg2 = (unsigned long) second_arg;
} else { } else {
......
...@@ -264,5 +264,10 @@ extern void write_cpuectlr(unsigned long); ...@@ -264,5 +264,10 @@ extern void write_cpuectlr(unsigned long);
extern void write_cptr_el2(unsigned long); extern void write_cptr_el2(unsigned long);
extern void write_cptr_el3(unsigned long); extern void write_cptr_el3(unsigned long);
#define IS_IN_EL(x) \
(GET_EL(read_current_el()) == MODE_EL##x)
#define IS_IN_EL1() IS_IN_EL(1)
#define IS_IN_EL3() IS_IN_EL(3)
#endif /* __ARCH_HELPERS_H__ */ #endif /* __ARCH_HELPERS_H__ */
...@@ -47,81 +47,68 @@ ...@@ -47,81 +47,68 @@
static unsigned long platform_config[CONFIG_LIMIT]; static unsigned long platform_config[CONFIG_LIMIT];
/******************************************************************************* /*******************************************************************************
* Enable the MMU assuming that the pagetables have already been created * Macro generating the code for the function enabling the MMU in the given
*******************************************************************************/ * exception level, assuming that the pagetables have already been created.
void enable_mmu() *
{ * _el: Exception level at which the function will run
unsigned long mair, tcr, ttbr, sctlr; * _tcr_extra: Extra bits to set in the TCR register. This mask will
unsigned long current_el = read_current_el(); * be OR'ed with the default TCR value.
* _tlbi_fct: Function to invalidate the TLBs at the current
/* Set the attributes in the right indices of the MAIR */ * exception level
mair = MAIR_ATTR_SET(ATTR_DEVICE, ATTR_DEVICE_INDEX); ******************************************************************************/
mair |= MAIR_ATTR_SET(ATTR_IWBWA_OWBWA_NTR, #define DEFINE_ENABLE_MMU_EL(_el, _tcr_extra, _tlbi_fct) \
ATTR_IWBWA_OWBWA_NTR_INDEX); void enable_mmu_el##_el(void) \
{ \
/* uint64_t mair, tcr, ttbr; \
* Set TCR bits as well. Inner & outer WBWA & shareable + T0SZ = 32 uint32_t sctlr; \
*/ \
tcr = TCR_SH_INNER_SHAREABLE | TCR_RGN_OUTER_WBA | assert(IS_IN_EL(_el)); \
TCR_RGN_INNER_WBA | TCR_T0SZ_4GB; assert((read_sctlr_el##_el() & SCTLR_M_BIT) == 0); \
\
/* Set TTBR bits as well */ /* Set attributes in the right indices of the MAIR */ \
ttbr = (unsigned long) l1_xlation_table; mair = MAIR_ATTR_SET(ATTR_DEVICE, ATTR_DEVICE_INDEX); \
mair |= MAIR_ATTR_SET(ATTR_IWBWA_OWBWA_NTR, \
if (GET_EL(current_el) == MODE_EL3) { ATTR_IWBWA_OWBWA_NTR_INDEX); \
assert((read_sctlr_el3() & SCTLR_M_BIT) == 0); write_mair_el##_el(mair); \
\
write_mair_el3(mair); /* Invalidate TLBs at the current exception level */ \
tcr |= TCR_EL3_RES1; _tlbi_fct(); \
/* Invalidate EL3 TLBs */ \
tlbialle3(); /* Set TCR bits as well. */ \
/* Inner & outer WBWA & shareable + T0SZ = 32 */ \
write_tcr_el3(tcr); tcr = TCR_SH_INNER_SHAREABLE | TCR_RGN_OUTER_WBA | \
write_ttbr0_el3(ttbr); TCR_RGN_INNER_WBA | TCR_T0SZ_4GB; \
tcr |= _tcr_extra; \
/* ensure all translation table writes have drained into memory, write_tcr_el##_el(tcr); \
* the TLB invalidation is complete, and translation register \
* writes are committed before enabling the MMU /* Set TTBR bits as well */ \
*/ ttbr = (uint64_t) l1_xlation_table; \
dsb(); write_ttbr0_el##_el(ttbr); \
isb(); \
/* Ensure all translation table writes have drained */ \
sctlr = read_sctlr_el3(); /* into memory, the TLB invalidation is complete, */ \
sctlr |= SCTLR_WXN_BIT | SCTLR_M_BIT | SCTLR_I_BIT; /* and translation register writes are committed */ \
sctlr |= SCTLR_A_BIT | SCTLR_C_BIT; /* before enabling the MMU */ \
write_sctlr_el3(sctlr); dsb(); \
} else { isb(); \
assert((read_sctlr_el1() & SCTLR_M_BIT) == 0); \
sctlr = read_sctlr_el##_el(); \
write_mair_el1(mair); sctlr |= SCTLR_WXN_BIT | SCTLR_M_BIT | SCTLR_I_BIT; \
/* Invalidate EL1 TLBs */ sctlr |= SCTLR_A_BIT | SCTLR_C_BIT; \
tlbivmalle1(); write_sctlr_el##_el(sctlr); \
\
write_tcr_el1(tcr); /* Ensure the MMU enable takes effect immediately */ \
write_ttbr0_el1(ttbr); isb(); \
/* ensure all translation table writes have drained into memory,
* the TLB invalidation is complete, and translation register
* writes are committed before enabling the MMU
*/
dsb();
isb();
sctlr = read_sctlr_el1();
sctlr |= SCTLR_WXN_BIT | SCTLR_M_BIT | SCTLR_I_BIT;
sctlr |= SCTLR_A_BIT | SCTLR_C_BIT;
write_sctlr_el1(sctlr);
} }
/* ensure the MMU enable takes effect immediately */
isb();
return; /* Define EL1 and EL3 variants of the function enabling the MMU */
} DEFINE_ENABLE_MMU_EL(1, 0, tlbivmalle1)
DEFINE_ENABLE_MMU_EL(3, TCR_EL3_RES1, tlbialle3)
/* /*
* Table of regions to map using the MMU. * Table of regions to map using the MMU.
* This doesn't include TZRAM as the 'mem_layout' argument passed to to * This doesn't include TZRAM as the 'mem_layout' argument passed to
* configure_mmu() will give the available subset of that, * configure_mmu_elx() will give the available subset of that,
*/ */
const mmap_region_t fvp_mmap[] = { const mmap_region_t fvp_mmap[] = {
{ TZROM_BASE, TZROM_SIZE, MT_MEMORY | MT_RO | MT_SECURE }, { TZROM_BASE, TZROM_SIZE, MT_MEMORY | MT_RO | MT_SECURE },
...@@ -139,28 +126,32 @@ const mmap_region_t fvp_mmap[] = { ...@@ -139,28 +126,32 @@ const mmap_region_t fvp_mmap[] = {
}; };
/******************************************************************************* /*******************************************************************************
* Setup the pagetables as per the platform memory map & initialize the mmu * Macro generating the code for the function setting up the pagetables as per
*******************************************************************************/ * the platform memory map & initialize the mmu, for the given exception level
void configure_mmu(meminfo_t *mem_layout, ******************************************************************************/
unsigned long ro_start, #define DEFINE_CONFIGURE_MMU_EL(_el) \
unsigned long ro_limit, void configure_mmu_el##_el(meminfo_t *mem_layout, \
unsigned long coh_start, unsigned long ro_start, \
unsigned long coh_limit) unsigned long ro_limit, \
{ unsigned long coh_start, \
mmap_add_region(mem_layout->total_base, mem_layout->total_size, unsigned long coh_limit) \
MT_MEMORY | MT_RW | MT_SECURE); { \
mmap_add_region(ro_start, ro_limit - ro_start, mmap_add_region(mem_layout->total_base, \
MT_MEMORY | MT_RO | MT_SECURE); mem_layout->total_size, \
mmap_add_region(coh_start, coh_limit - coh_start, MT_MEMORY | MT_RW | MT_SECURE); \
MT_DEVICE | MT_RW | MT_SECURE); mmap_add_region(ro_start, ro_limit - ro_start, \
MT_MEMORY | MT_RO | MT_SECURE); \
mmap_add(fvp_mmap); mmap_add_region(coh_start, coh_limit - coh_start, \
MT_DEVICE | MT_RW | MT_SECURE); \
init_xlat_tables(); mmap_add(fvp_mmap); \
init_xlat_tables(); \
\
enable_mmu_el##_el(); \
}
enable_mmu(); /* Define EL1 and EL3 variants of the function initialising the MMU */
return; DEFINE_CONFIGURE_MMU_EL(1)
} DEFINE_CONFIGURE_MMU_EL(3)
/* Simple routine which returns a configuration variable value */ /* Simple routine which returns a configuration variable value */
unsigned long platform_get_cfgvar(unsigned int var_id) unsigned long platform_get_cfgvar(unsigned int var_id)
......
...@@ -138,9 +138,9 @@ void bl1_plat_arch_setup(void) ...@@ -138,9 +138,9 @@ void bl1_plat_arch_setup(void)
cci_enable_coherency(read_mpidr()); cci_enable_coherency(read_mpidr());
} }
configure_mmu(&bl1_tzram_layout, configure_mmu_el3(&bl1_tzram_layout,
TZROM_BASE, TZROM_BASE,
TZROM_BASE + TZROM_SIZE, TZROM_BASE + TZROM_SIZE,
BL1_COHERENT_RAM_BASE, BL1_COHERENT_RAM_BASE,
BL1_COHERENT_RAM_LIMIT); BL1_COHERENT_RAM_LIMIT);
} }
...@@ -172,9 +172,9 @@ void bl2_platform_setup() ...@@ -172,9 +172,9 @@ void bl2_platform_setup()
******************************************************************************/ ******************************************************************************/
void bl2_plat_arch_setup() void bl2_plat_arch_setup()
{ {
configure_mmu(&bl2_tzram_layout, configure_mmu_el1(&bl2_tzram_layout,
BL2_RO_BASE, BL2_RO_BASE,
BL2_RO_LIMIT, BL2_RO_LIMIT,
BL2_COHERENT_RAM_BASE, BL2_COHERENT_RAM_BASE,
BL2_COHERENT_RAM_LIMIT); BL2_COHERENT_RAM_LIMIT);
} }
...@@ -172,9 +172,9 @@ void bl31_platform_setup() ...@@ -172,9 +172,9 @@ void bl31_platform_setup()
******************************************************************************/ ******************************************************************************/
void bl31_plat_arch_setup() void bl31_plat_arch_setup()
{ {
configure_mmu(&bl2_to_bl31_args->bl31_meminfo, configure_mmu_el3(&bl2_to_bl31_args->bl31_meminfo,
BL31_RO_BASE, BL31_RO_BASE,
BL31_RO_LIMIT, BL31_RO_LIMIT,
BL31_COHERENT_RAM_BASE, BL31_COHERENT_RAM_BASE,
BL31_COHERENT_RAM_LIMIT); BL31_COHERENT_RAM_LIMIT);
} }
...@@ -111,9 +111,9 @@ void bl32_platform_setup() ...@@ -111,9 +111,9 @@ void bl32_platform_setup()
******************************************************************************/ ******************************************************************************/
void bl32_plat_arch_setup() void bl32_plat_arch_setup()
{ {
configure_mmu(&bl32_tzdram_layout, configure_mmu_el1(&bl32_tzdram_layout,
BL32_RO_BASE, BL32_RO_BASE,
BL32_RO_LIMIT, BL32_RO_LIMIT,
BL32_COHERENT_RAM_BASE, BL32_COHERENT_RAM_BASE,
BL32_COHERENT_RAM_LIMIT); BL32_COHERENT_RAM_LIMIT);
} }
...@@ -373,12 +373,18 @@ extern void bl2_plat_arch_setup(void); ...@@ -373,12 +373,18 @@ extern void bl2_plat_arch_setup(void);
extern void bl31_plat_arch_setup(void); extern void bl31_plat_arch_setup(void);
extern int platform_setup_pm(const struct plat_pm_ops **); extern int platform_setup_pm(const struct plat_pm_ops **);
extern unsigned int platform_get_core_pos(unsigned long mpidr); extern unsigned int platform_get_core_pos(unsigned long mpidr);
extern void enable_mmu(void); extern void enable_mmu_el1(void);
extern void configure_mmu(struct meminfo *, extern void enable_mmu_el3(void);
unsigned long, extern void configure_mmu_el1(struct meminfo *mem_layout,
unsigned long, unsigned long ro_start,
unsigned long, unsigned long ro_limit,
unsigned long); unsigned long coh_start,
unsigned long coh_limit);
extern void configure_mmu_el3(struct meminfo *mem_layout,
unsigned long ro_start,
unsigned long ro_limit,
unsigned long coh_start,
unsigned long coh_limit);
extern unsigned long platform_get_cfgvar(unsigned int); extern unsigned long platform_get_cfgvar(unsigned int);
extern int platform_config_setup(void); extern int platform_config_setup(void);
extern void plat_report_exception(unsigned long); extern void plat_report_exception(unsigned long);
......
...@@ -362,7 +362,7 @@ static unsigned int psci_afflvl0_on_finish(unsigned long mpidr, ...@@ -362,7 +362,7 @@ static unsigned int psci_afflvl0_on_finish(unsigned long mpidr,
/* /*
* Arch. management: Turn on mmu & restore architectural state * Arch. management: Turn on mmu & restore architectural state
*/ */
enable_mmu(); enable_mmu_el3();
/* /*
* All the platform specific actions for turning this cpu * All the platform specific actions for turning this cpu
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment