1*91f16700Schasinglulu /* 2*91f16700Schasinglulu * Copyright (c) 2021, Arm Limited and Contributors. All rights reserved. 3*91f16700Schasinglulu * 4*91f16700Schasinglulu * SPDX-License-Identifier: BSD-3-Clause 5*91f16700Schasinglulu */ 6*91f16700Schasinglulu 7*91f16700Schasinglulu #include <assert.h> 8*91f16700Schasinglulu #include <errno.h> 9*91f16700Schasinglulu #include <stdbool.h> 10*91f16700Schasinglulu #include <stdint.h> 11*91f16700Schasinglulu #include <string.h> 12*91f16700Schasinglulu 13*91f16700Schasinglulu #include <arch_features.h> 14*91f16700Schasinglulu #include <common/debug.h> 15*91f16700Schasinglulu #include <lib/utils_def.h> 16*91f16700Schasinglulu #include <lib/xlat_tables/xlat_tables_defs.h> 17*91f16700Schasinglulu #include <lib/xlat_tables/xlat_tables_v2.h> 18*91f16700Schasinglulu #include "xlat_mpu_private.h" 19*91f16700Schasinglulu 20*91f16700Schasinglulu #include <fvp_r_arch_helpers.h> 21*91f16700Schasinglulu #include <platform_def.h> 22*91f16700Schasinglulu 23*91f16700Schasinglulu #warning "xlat_mpu library is currently experimental and its API may change in future." 24*91f16700Schasinglulu 25*91f16700Schasinglulu 26*91f16700Schasinglulu /* Helper function that cleans the data cache only if it is enabled. */ 27*91f16700Schasinglulu static inline __attribute__((unused)) 28*91f16700Schasinglulu void xlat_clean_dcache_range(uintptr_t addr, size_t size) 29*91f16700Schasinglulu { 30*91f16700Schasinglulu if (is_dcache_enabled()) { 31*91f16700Schasinglulu clean_dcache_range(addr, size); 32*91f16700Schasinglulu } 33*91f16700Schasinglulu } 34*91f16700Schasinglulu 35*91f16700Schasinglulu 36*91f16700Schasinglulu 37*91f16700Schasinglulu /* Calculate region-attributes byte for PRBAR part of MPU-region descriptor: */ 38*91f16700Schasinglulu uint64_t prbar_attr_value(uint32_t attr) 39*91f16700Schasinglulu { 40*91f16700Schasinglulu uint64_t retValue = UL(0); 41*91f16700Schasinglulu uint64_t extract; /* temp var holding bit extracted from attr */ 42*91f16700Schasinglulu 43*91f16700Schasinglulu /* Extract and stuff SH: */ 44*91f16700Schasinglulu extract = (uint64_t) ((attr >> MT_SHAREABILITY_SHIFT) 45*91f16700Schasinglulu & MT_SHAREABILITY_MASK); 46*91f16700Schasinglulu retValue |= (extract << PRBAR_SH_SHIFT); 47*91f16700Schasinglulu 48*91f16700Schasinglulu /* Extract and stuff AP: */ 49*91f16700Schasinglulu extract = (uint64_t) ((attr >> MT_PERM_SHIFT) & MT_PERM_MASK); 50*91f16700Schasinglulu if (extract == 0U) { 51*91f16700Schasinglulu retValue |= (UL(2) << PRBAR_AP_SHIFT); 52*91f16700Schasinglulu } else /* extract == 1 */ { 53*91f16700Schasinglulu retValue |= (UL(0) << PRBAR_AP_SHIFT); 54*91f16700Schasinglulu } 55*91f16700Schasinglulu 56*91f16700Schasinglulu /* Extract and stuff XN: */ 57*91f16700Schasinglulu extract = (uint64_t) ((attr >> MT_EXECUTE_SHIFT) & MT_EXECUTE_MASK); 58*91f16700Schasinglulu retValue |= (extract << PRBAR_XN_SHIFT); 59*91f16700Schasinglulu /* However, also don't execute in peripheral space: */ 60*91f16700Schasinglulu extract = (uint64_t) ((attr >> MT_TYPE_SHIFT) & MT_TYPE_MASK); 61*91f16700Schasinglulu if (extract == 0U) { 62*91f16700Schasinglulu retValue |= (UL(1) << PRBAR_XN_SHIFT); 63*91f16700Schasinglulu } 64*91f16700Schasinglulu return retValue; 65*91f16700Schasinglulu } 66*91f16700Schasinglulu 67*91f16700Schasinglulu /* Calculate region-attributes byte for PRLAR part of MPU-region descriptor: */ 68*91f16700Schasinglulu uint64_t prlar_attr_value(uint32_t attr) 69*91f16700Schasinglulu { 70*91f16700Schasinglulu uint64_t retValue = UL(0); 71*91f16700Schasinglulu uint64_t extract; /* temp var holding bit extracted from attr */ 72*91f16700Schasinglulu 73*91f16700Schasinglulu /* Extract and stuff AttrIndx: */ 74*91f16700Schasinglulu extract = (uint64_t) ((attr >> MT_TYPE_SHIFT) 75*91f16700Schasinglulu & MT_TYPE_MASK); 76*91f16700Schasinglulu switch (extract) { 77*91f16700Schasinglulu case UL(0): 78*91f16700Schasinglulu retValue |= (UL(1) << PRLAR_ATTR_SHIFT); 79*91f16700Schasinglulu break; 80*91f16700Schasinglulu case UL(2): 81*91f16700Schasinglulu /* 0, so OR in nothing */ 82*91f16700Schasinglulu break; 83*91f16700Schasinglulu case UL(3): 84*91f16700Schasinglulu retValue |= (UL(2) << PRLAR_ATTR_SHIFT); 85*91f16700Schasinglulu break; 86*91f16700Schasinglulu default: 87*91f16700Schasinglulu retValue |= (extract << PRLAR_ATTR_SHIFT); 88*91f16700Schasinglulu break; 89*91f16700Schasinglulu } 90*91f16700Schasinglulu 91*91f16700Schasinglulu /* Stuff EN: */ 92*91f16700Schasinglulu retValue |= (UL(1) << PRLAR_EN_SHIFT); 93*91f16700Schasinglulu 94*91f16700Schasinglulu /* Force NS to 0 (Secure); v8-R64 only supports Secure: */ 95*91f16700Schasinglulu extract = ~(1U << PRLAR_NS_SHIFT); 96*91f16700Schasinglulu retValue &= extract; 97*91f16700Schasinglulu 98*91f16700Schasinglulu return retValue; 99*91f16700Schasinglulu } 100*91f16700Schasinglulu 101*91f16700Schasinglulu /* 102*91f16700Schasinglulu * Function that writes an MPU "translation" into the MPU registers. If not 103*91f16700Schasinglulu * possible (e.g., if no more MPU regions available) boot is aborted. 104*91f16700Schasinglulu */ 105*91f16700Schasinglulu static void mpu_map_region(mmap_region_t *mm) 106*91f16700Schasinglulu { 107*91f16700Schasinglulu uint64_t prenr_el2_value = 0UL; 108*91f16700Schasinglulu uint64_t prbar_attrs = 0UL; 109*91f16700Schasinglulu uint64_t prlar_attrs = 0UL; 110*91f16700Schasinglulu int region_to_use = 0; 111*91f16700Schasinglulu 112*91f16700Schasinglulu /* If all MPU regions in use, then abort boot: */ 113*91f16700Schasinglulu prenr_el2_value = read_prenr_el2(); 114*91f16700Schasinglulu assert(prenr_el2_value != 0xffffffff); 115*91f16700Schasinglulu 116*91f16700Schasinglulu /* Find and select first-available MPU region (PRENR has an enable bit 117*91f16700Schasinglulu * for each MPU region, 1 for in-use or 0 for unused): 118*91f16700Schasinglulu */ 119*91f16700Schasinglulu for (region_to_use = 0; region_to_use < N_MPU_REGIONS; 120*91f16700Schasinglulu region_to_use++) { 121*91f16700Schasinglulu if (((prenr_el2_value >> region_to_use) & 1) == 0) { 122*91f16700Schasinglulu break; 123*91f16700Schasinglulu } 124*91f16700Schasinglulu } 125*91f16700Schasinglulu write_prselr_el2((uint64_t) (region_to_use)); 126*91f16700Schasinglulu isb(); 127*91f16700Schasinglulu 128*91f16700Schasinglulu /* Set base and limit addresses: */ 129*91f16700Schasinglulu write_prbar_el2(mm->base_pa & PRBAR_PRLAR_ADDR_MASK); 130*91f16700Schasinglulu write_prlar_el2((mm->base_pa + mm->size - 1UL) 131*91f16700Schasinglulu & PRBAR_PRLAR_ADDR_MASK); 132*91f16700Schasinglulu dsbsy(); 133*91f16700Schasinglulu isb(); 134*91f16700Schasinglulu 135*91f16700Schasinglulu /* Set attributes: */ 136*91f16700Schasinglulu prbar_attrs = prbar_attr_value(mm->attr); 137*91f16700Schasinglulu write_prbar_el2(read_prbar_el2() | prbar_attrs); 138*91f16700Schasinglulu prlar_attrs = prlar_attr_value(mm->attr); 139*91f16700Schasinglulu write_prlar_el2(read_prlar_el2() | prlar_attrs); 140*91f16700Schasinglulu dsbsy(); 141*91f16700Schasinglulu isb(); 142*91f16700Schasinglulu 143*91f16700Schasinglulu /* Mark this MPU region as used: */ 144*91f16700Schasinglulu prenr_el2_value |= (1 << region_to_use); 145*91f16700Schasinglulu write_prenr_el2(prenr_el2_value); 146*91f16700Schasinglulu isb(); 147*91f16700Schasinglulu } 148*91f16700Schasinglulu 149*91f16700Schasinglulu /* 150*91f16700Schasinglulu * Function that verifies that a region can be mapped. 151*91f16700Schasinglulu * Returns: 152*91f16700Schasinglulu * 0: Success, the mapping is allowed. 153*91f16700Schasinglulu * EINVAL: Invalid values were used as arguments. 154*91f16700Schasinglulu * ERANGE: The memory limits were surpassed. 155*91f16700Schasinglulu * ENOMEM: There is not enough memory in the mmap array. 156*91f16700Schasinglulu * EPERM: Region overlaps another one in an invalid way. 157*91f16700Schasinglulu */ 158*91f16700Schasinglulu static int mmap_add_region_check(const xlat_ctx_t *ctx, const mmap_region_t *mm) 159*91f16700Schasinglulu { 160*91f16700Schasinglulu unsigned long long base_pa = mm->base_pa; 161*91f16700Schasinglulu uintptr_t base_va = mm->base_va; 162*91f16700Schasinglulu size_t size = mm->size; 163*91f16700Schasinglulu 164*91f16700Schasinglulu unsigned long long end_pa = base_pa + size - 1U; 165*91f16700Schasinglulu uintptr_t end_va = base_va + size - 1U; 166*91f16700Schasinglulu 167*91f16700Schasinglulu if (base_pa != base_va) { 168*91f16700Schasinglulu return -EINVAL; /* MPU does not perform address translation */ 169*91f16700Schasinglulu } 170*91f16700Schasinglulu if ((base_pa % 64ULL) != 0ULL) { 171*91f16700Schasinglulu return -EINVAL; /* MPU requires 64-byte alignment */ 172*91f16700Schasinglulu } 173*91f16700Schasinglulu /* Check for overflows */ 174*91f16700Schasinglulu if ((base_pa > end_pa) || (base_va > end_va)) { 175*91f16700Schasinglulu return -ERANGE; 176*91f16700Schasinglulu } 177*91f16700Schasinglulu if (end_pa > ctx->pa_max_address) { 178*91f16700Schasinglulu return -ERANGE; 179*91f16700Schasinglulu } 180*91f16700Schasinglulu /* Check that there is space in the ctx->mmap array */ 181*91f16700Schasinglulu if (ctx->mmap[ctx->mmap_num - 1].size != 0U) { 182*91f16700Schasinglulu return -ENOMEM; 183*91f16700Schasinglulu } 184*91f16700Schasinglulu /* Check for PAs and VAs overlaps with all other regions */ 185*91f16700Schasinglulu for (const mmap_region_t *mm_cursor = ctx->mmap; 186*91f16700Schasinglulu mm_cursor->size != 0U; ++mm_cursor) { 187*91f16700Schasinglulu 188*91f16700Schasinglulu uintptr_t mm_cursor_end_va = 189*91f16700Schasinglulu mm_cursor->base_va + mm_cursor->size - 1U; 190*91f16700Schasinglulu 191*91f16700Schasinglulu /* 192*91f16700Schasinglulu * Check if one of the regions is completely inside the other 193*91f16700Schasinglulu * one. 194*91f16700Schasinglulu */ 195*91f16700Schasinglulu bool fully_overlapped_va = 196*91f16700Schasinglulu ((base_va >= mm_cursor->base_va) && 197*91f16700Schasinglulu (end_va <= mm_cursor_end_va)) || 198*91f16700Schasinglulu ((mm_cursor->base_va >= base_va) && 199*91f16700Schasinglulu (mm_cursor_end_va <= end_va)); 200*91f16700Schasinglulu 201*91f16700Schasinglulu /* 202*91f16700Schasinglulu * Full VA overlaps are only allowed if both regions are 203*91f16700Schasinglulu * identity mapped (zero offset) or have the same VA to PA 204*91f16700Schasinglulu * offset. Also, make sure that it's not the exact same area. 205*91f16700Schasinglulu * This can only be done with static regions. 206*91f16700Schasinglulu */ 207*91f16700Schasinglulu if (fully_overlapped_va) { 208*91f16700Schasinglulu 209*91f16700Schasinglulu #if PLAT_XLAT_TABLES_DYNAMIC 210*91f16700Schasinglulu if (((mm->attr & MT_DYNAMIC) != 0U) || 211*91f16700Schasinglulu ((mm_cursor->attr & MT_DYNAMIC) != 0U)) { 212*91f16700Schasinglulu return -EPERM; 213*91f16700Schasinglulu } 214*91f16700Schasinglulu #endif /* PLAT_XLAT_TABLES_DYNAMIC */ 215*91f16700Schasinglulu if ((mm_cursor->base_va - mm_cursor->base_pa) 216*91f16700Schasinglulu != (base_va - base_pa)) { 217*91f16700Schasinglulu return -EPERM; 218*91f16700Schasinglulu } 219*91f16700Schasinglulu if ((base_va == mm_cursor->base_va) && 220*91f16700Schasinglulu (size == mm_cursor->size)) { 221*91f16700Schasinglulu return -EPERM; 222*91f16700Schasinglulu } 223*91f16700Schasinglulu } else { 224*91f16700Schasinglulu /* 225*91f16700Schasinglulu * If the regions do not have fully overlapping VAs, 226*91f16700Schasinglulu * then they must have fully separated VAs and PAs. 227*91f16700Schasinglulu * Partial overlaps are not allowed 228*91f16700Schasinglulu */ 229*91f16700Schasinglulu 230*91f16700Schasinglulu unsigned long long mm_cursor_end_pa = 231*91f16700Schasinglulu mm_cursor->base_pa + mm_cursor->size - 1U; 232*91f16700Schasinglulu 233*91f16700Schasinglulu bool separated_pa = (end_pa < mm_cursor->base_pa) || 234*91f16700Schasinglulu (base_pa > mm_cursor_end_pa); 235*91f16700Schasinglulu bool separated_va = (end_va < mm_cursor->base_va) || 236*91f16700Schasinglulu (base_va > mm_cursor_end_va); 237*91f16700Schasinglulu 238*91f16700Schasinglulu if (!separated_va || !separated_pa) { 239*91f16700Schasinglulu return -EPERM; 240*91f16700Schasinglulu } 241*91f16700Schasinglulu } 242*91f16700Schasinglulu } 243*91f16700Schasinglulu 244*91f16700Schasinglulu return 0; 245*91f16700Schasinglulu } 246*91f16700Schasinglulu 247*91f16700Schasinglulu void mmap_add_region_ctx(xlat_ctx_t *ctx, const mmap_region_t *mm) 248*91f16700Schasinglulu { 249*91f16700Schasinglulu mmap_region_t *mm_cursor = ctx->mmap, *mm_destination; 250*91f16700Schasinglulu const mmap_region_t *mm_end = ctx->mmap + ctx->mmap_num; 251*91f16700Schasinglulu const mmap_region_t *mm_last; 252*91f16700Schasinglulu unsigned long long end_pa = mm->base_pa + mm->size - 1U; 253*91f16700Schasinglulu uintptr_t end_va = mm->base_va + mm->size - 1U; 254*91f16700Schasinglulu int ret; 255*91f16700Schasinglulu 256*91f16700Schasinglulu /* Ignore empty regions */ 257*91f16700Schasinglulu if (mm->size == 0U) { 258*91f16700Schasinglulu return; 259*91f16700Schasinglulu } 260*91f16700Schasinglulu 261*91f16700Schasinglulu /* Static regions must be added before initializing the xlat tables. */ 262*91f16700Schasinglulu assert(!ctx->initialized); 263*91f16700Schasinglulu 264*91f16700Schasinglulu ret = mmap_add_region_check(ctx, mm); 265*91f16700Schasinglulu if (ret != 0) { 266*91f16700Schasinglulu ERROR("mmap_add_region_check() failed. error %d\n", ret); 267*91f16700Schasinglulu assert(false); 268*91f16700Schasinglulu return; 269*91f16700Schasinglulu } 270*91f16700Schasinglulu 271*91f16700Schasinglulu /* 272*91f16700Schasinglulu * Find the last entry marker in the mmap 273*91f16700Schasinglulu */ 274*91f16700Schasinglulu mm_last = ctx->mmap; 275*91f16700Schasinglulu while ((mm_last->size != 0U) && (mm_last < mm_end)) { 276*91f16700Schasinglulu ++mm_last; 277*91f16700Schasinglulu } 278*91f16700Schasinglulu 279*91f16700Schasinglulu /* 280*91f16700Schasinglulu * Check if we have enough space in the memory mapping table. 281*91f16700Schasinglulu * This shouldn't happen as we have checked in mmap_add_region_check 282*91f16700Schasinglulu * that there is free space. 283*91f16700Schasinglulu */ 284*91f16700Schasinglulu assert(mm_last->size == 0U); 285*91f16700Schasinglulu 286*91f16700Schasinglulu /* Make room for new region by moving other regions up by one place */ 287*91f16700Schasinglulu mm_destination = mm_cursor + 1; 288*91f16700Schasinglulu (void)memmove(mm_destination, mm_cursor, 289*91f16700Schasinglulu (uintptr_t)mm_last - (uintptr_t)mm_cursor); 290*91f16700Schasinglulu 291*91f16700Schasinglulu /* 292*91f16700Schasinglulu * Check we haven't lost the empty sentinel from the end of the array. 293*91f16700Schasinglulu * This shouldn't happen as we have checked in mmap_add_region_check 294*91f16700Schasinglulu * that there is free space. 295*91f16700Schasinglulu */ 296*91f16700Schasinglulu assert(mm_end->size == 0U); 297*91f16700Schasinglulu 298*91f16700Schasinglulu *mm_cursor = *mm; 299*91f16700Schasinglulu 300*91f16700Schasinglulu if (end_pa > ctx->max_pa) { 301*91f16700Schasinglulu ctx->max_pa = end_pa; 302*91f16700Schasinglulu } 303*91f16700Schasinglulu if (end_va > ctx->max_va) { 304*91f16700Schasinglulu ctx->max_va = end_va; 305*91f16700Schasinglulu } 306*91f16700Schasinglulu } 307*91f16700Schasinglulu 308*91f16700Schasinglulu void mmap_add_ctx(xlat_ctx_t *ctx, const mmap_region_t *mm) 309*91f16700Schasinglulu { 310*91f16700Schasinglulu const mmap_region_t *mm_cursor = mm; 311*91f16700Schasinglulu 312*91f16700Schasinglulu while (mm_cursor->granularity != 0U) { 313*91f16700Schasinglulu mmap_add_region_ctx(ctx, mm_cursor); 314*91f16700Schasinglulu mm_cursor++; 315*91f16700Schasinglulu } 316*91f16700Schasinglulu } 317*91f16700Schasinglulu 318*91f16700Schasinglulu void __init init_xlat_tables_ctx(xlat_ctx_t *ctx) 319*91f16700Schasinglulu { 320*91f16700Schasinglulu uint64_t mair = UL(0); 321*91f16700Schasinglulu 322*91f16700Schasinglulu assert(ctx != NULL); 323*91f16700Schasinglulu assert(!ctx->initialized); 324*91f16700Schasinglulu assert((ctx->xlat_regime == EL2_REGIME) || 325*91f16700Schasinglulu (ctx->xlat_regime == EL1_EL0_REGIME)); 326*91f16700Schasinglulu /* Note: Add EL3_REGIME if EL3 is supported in future v8-R64 cores. */ 327*91f16700Schasinglulu assert(!is_mpu_enabled_ctx(ctx)); 328*91f16700Schasinglulu 329*91f16700Schasinglulu mmap_region_t *mm = ctx->mmap; 330*91f16700Schasinglulu 331*91f16700Schasinglulu assert(ctx->va_max_address >= 332*91f16700Schasinglulu (xlat_get_min_virt_addr_space_size() - 1U)); 333*91f16700Schasinglulu assert(ctx->va_max_address <= (MAX_VIRT_ADDR_SPACE_SIZE - 1U)); 334*91f16700Schasinglulu assert(IS_POWER_OF_TWO(ctx->va_max_address + 1U)); 335*91f16700Schasinglulu 336*91f16700Schasinglulu xlat_mmap_print(mm); 337*91f16700Schasinglulu 338*91f16700Schasinglulu /* All tables must be zeroed before mapping any region. */ 339*91f16700Schasinglulu 340*91f16700Schasinglulu for (unsigned int i = 0U; i < ctx->base_table_entries; i++) 341*91f16700Schasinglulu ctx->base_table[i] = INVALID_DESC; 342*91f16700Schasinglulu 343*91f16700Schasinglulu /* Also mark all MPU regions as invalid in the MPU hardware itself: */ 344*91f16700Schasinglulu write_prenr_el2(0); 345*91f16700Schasinglulu /* Sufficient for current, max-32-region implementations. */ 346*91f16700Schasinglulu dsbsy(); 347*91f16700Schasinglulu isb(); 348*91f16700Schasinglulu while (mm->size != 0U) { 349*91f16700Schasinglulu if (read_prenr_el2() == ALL_MPU_EL2_REGIONS_USED) { 350*91f16700Schasinglulu ERROR("Not enough MPU regions to map region:\n" 351*91f16700Schasinglulu " VA:0x%lx PA:0x%llx size:0x%zx attr:0x%x\n", 352*91f16700Schasinglulu mm->base_va, mm->base_pa, mm->size, mm->attr); 353*91f16700Schasinglulu panic(); 354*91f16700Schasinglulu } else { 355*91f16700Schasinglulu #if !(HW_ASSISTED_COHERENCY || WARMBOOT_ENABLE_DCACHE_EARLY) 356*91f16700Schasinglulu xlat_clean_dcache_range((uintptr_t)mm->base_va, 357*91f16700Schasinglulu mm->size); 358*91f16700Schasinglulu #endif 359*91f16700Schasinglulu mpu_map_region(mm); 360*91f16700Schasinglulu } 361*91f16700Schasinglulu mm++; 362*91f16700Schasinglulu } 363*91f16700Schasinglulu 364*91f16700Schasinglulu ctx->initialized = true; 365*91f16700Schasinglulu 366*91f16700Schasinglulu xlat_tables_print(ctx); 367*91f16700Schasinglulu 368*91f16700Schasinglulu /* Set attributes in the right indices of the MAIR */ 369*91f16700Schasinglulu mair = MAIR_ATTR_SET(ATTR_DEVICE, ATTR_DEVICE_INDEX); 370*91f16700Schasinglulu mair |= MAIR_ATTR_SET(ATTR_IWBWA_OWBWA_NTR, 371*91f16700Schasinglulu ATTR_IWBWA_OWBWA_NTR_INDEX); 372*91f16700Schasinglulu mair |= MAIR_ATTR_SET(ATTR_NON_CACHEABLE, 373*91f16700Schasinglulu ATTR_NON_CACHEABLE_INDEX); 374*91f16700Schasinglulu write_mair_el2(mair); 375*91f16700Schasinglulu dsbsy(); 376*91f16700Schasinglulu isb(); 377*91f16700Schasinglulu } 378*91f16700Schasinglulu 379*91f16700Schasinglulu /* 380*91f16700Schasinglulu * Function to wipe clean and disable all MPU regions. This function expects 381*91f16700Schasinglulu * that the MPU has already been turned off, and caching concerns addressed, 382*91f16700Schasinglulu * but it nevertheless also explicitly turns off the MPU. 383*91f16700Schasinglulu */ 384*91f16700Schasinglulu void clear_all_mpu_regions(void) 385*91f16700Schasinglulu { 386*91f16700Schasinglulu uint64_t sctlr_el2_value = 0UL; 387*91f16700Schasinglulu uint64_t region_n = 0UL; 388*91f16700Schasinglulu 389*91f16700Schasinglulu /* 390*91f16700Schasinglulu * MPU should already be disabled, but explicitly disable it 391*91f16700Schasinglulu * nevertheless: 392*91f16700Schasinglulu */ 393*91f16700Schasinglulu sctlr_el2_value = read_sctlr_el2() & ~(1UL); 394*91f16700Schasinglulu write_sctlr_el2(sctlr_el2_value); 395*91f16700Schasinglulu 396*91f16700Schasinglulu /* Disable all regions: */ 397*91f16700Schasinglulu write_prenr_el2(0UL); 398*91f16700Schasinglulu 399*91f16700Schasinglulu /* Sequence through all regions, zeroing them out and turning off: */ 400*91f16700Schasinglulu for (region_n = 0UL; region_n < N_MPU_REGIONS; region_n++) { 401*91f16700Schasinglulu write_prselr_el2(region_n); 402*91f16700Schasinglulu isb(); 403*91f16700Schasinglulu write_prbar_el2((uint64_t) 0); 404*91f16700Schasinglulu write_prlar_el2((uint64_t) 0); 405*91f16700Schasinglulu dsbsy(); 406*91f16700Schasinglulu isb(); 407*91f16700Schasinglulu } 408*91f16700Schasinglulu } 409