| /* |
| * This file is part of the coreboot project. |
| * |
| * Copyright 2013 Google Inc. |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions |
| * are met: |
| * 1. Redistributions of source code must retain the above copyright |
| * notice, this list of conditions and the following disclaimer. |
| * 2. Redistributions in binary form must reproduce the above copyright |
| * notice, this list of conditions and the following disclaimer in the |
| * documentation and/or other materials provided with the distribution. |
| * 3. The name of the author may not be used to endorse or promote products |
| * derived from this software without specific prior written permission. |
| * |
| * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND |
| * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
| * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE |
| * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE |
| * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL |
| * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS |
| * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) |
| * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT |
| * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY |
| * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF |
| * SUCH DAMAGE. |
| * |
| * cache.h: Cache maintenance API for ARMv7 |
| */ |
| |
| #ifndef ARMV7_CACHE_H |
| #define ARMV7_CACHE_H |
| |
| #include <stddef.h> |
| #include <stdint.h> |
| |
| /* SCTLR bits */ |
| #define SCTLR_M (1 << 0) /* MMU enable */ |
| #define SCTLR_A (1 << 1) /* Alignment check enable */ |
| #define SCTLR_C (1 << 2) /* Data/unified cache enable */ |
| /* Bits 4:3 are reserved */ |
| #define SCTLR_CP15BEN (1 << 5) /* CP15 barrier enable */ |
| /* Bit 6 is reserved */ |
| #define SCTLR_B (1 << 7) /* Endianness */ |
| /* Bits 9:8 */ |
| #define SCTLR_SW (1 << 10) /* SWP and SWPB enable */ |
| #define SCTLR_Z (1 << 11) /* Branch prediction enable */ |
| #define SCTLR_I (1 << 12) /* Instruction cache enable */ |
| #define SCTLR_V (1 << 13) /* Low/high exception vectors */ |
| #define SCTLR_RR (1 << 14) /* Round Robin select */ |
| /* Bits 16:15 are reserved */ |
| #define SCTLR_HA (1 << 17) /* Hardware Access flag enable */ |
| /* Bit 18 is reserved */ |
| /* Bits 20:19 reserved virtualization not supported */ |
| #define SCTLR_WXN (1 << 19) /* Write permission implies XN */ |
| #define SCTLR_UWXN (1 << 20) /* Unprivileged write permission |
| implies PL1 XN */ |
| #define SCTLR_FI (1 << 21) /* Fast interrupt config enable */ |
| #define SCTLR_U (1 << 22) /* Unaligned access behavior */ |
| #define SCTLR_VE (1 << 24) /* Interrupt vectors enable */ |
| #define SCTLR_EE (1 << 25) /* Exception endianness */ |
| /* Bit 26 is reserved */ |
| #define SCTLR_NMFI (1 << 27) /* Non-maskable FIQ support */ |
| #define SCTLR_TRE (1 << 28) /* TEX remap enable */ |
| #define SCTLR_AFE (1 << 29) /* Access flag enable */ |
| #define SCTLR_TE (1 << 30) /* Thumb exception enable */ |
| /* Bit 31 is reserved */ |
| |
| /* |
| * Sync primitives |
| */ |
| |
| /* data memory barrier */ |
| static inline void dmb(void) |
| { |
| asm volatile ("dmb" : : : "memory"); |
| } |
| |
| /* data sync barrier */ |
| static inline void dsb(void) |
| { |
| asm volatile ("dsb" : : : "memory"); |
| } |
| |
| /* instruction sync barrier */ |
| static inline void isb(void) |
| { |
| asm volatile ("isb" : : : "memory"); |
| } |
| |
| /* |
| * Low-level TLB maintenance operations |
| */ |
| |
| /* invalidate entire data TLB */ |
| static inline void dtlbiall(void) |
| { |
| asm volatile ("mcr p15, 0, %0, c8, c6, 0" : : "r" (0) : "memory"); |
| } |
| |
| /* invalidate entire instruction TLB */ |
| static inline void itlbiall(void) |
| { |
| asm volatile ("mcr p15, 0, %0, c8, c5, 0" : : "r" (0)); |
| } |
| |
| /* invalidate entire unified TLB */ |
| static inline void tlbiall(void) |
| { |
| asm volatile ("mcr p15, 0, %0, c8, c7, 0" : : "r" (0) : "memory"); |
| } |
| |
| /* invalidate unified TLB by MVA, all ASID */ |
| static inline void tlbimvaa(unsigned long mva) |
| { |
| asm volatile ("mcr p15, 0, %0, c8, c7, 3" : : "r" (mva) : "memory"); |
| } |
| |
| /* write data access control register (DACR) */ |
| static inline void write_dacr(uint32_t val) |
| { |
| asm volatile ("mcr p15, 0, %0, c3, c0, 0" : : "r" (val)); |
| } |
| |
| /* write translation table base register 0 (TTBR0) */ |
| static inline void write_ttbr0(uint32_t val) |
| { |
| asm volatile ("mcr p15, 0, %0, c2, c0, 0" : : "r" (val) : "memory"); |
| } |
| |
| /* read translation table base control register (TTBCR) */ |
| static inline uint32_t read_ttbcr(void) |
| { |
| uint32_t val = 0; |
| asm volatile ("mrc p15, 0, %0, c2, c0, 2" : "=r" (val)); |
| return val; |
| } |
| |
| /* write translation table base control register (TTBCR) */ |
| static inline void write_ttbcr(uint32_t val) |
| { |
| asm volatile ("mcr p15, 0, %0, c2, c0, 2" : : "r" (val) : "memory"); |
| } |
| |
| /* |
| * Low-level cache maintenance operations |
| */ |
| |
| /* branch predictor invalidate all */ |
| static inline void bpiall(void) |
| { |
| asm volatile ("mcr p15, 0, %0, c7, c5, 6" : : "r" (0)); |
| } |
| |
| /* data cache clean and invalidate by MVA to PoC */ |
| static inline void dccimvac(unsigned long mva) |
| { |
| asm volatile ("mcr p15, 0, %0, c7, c14, 1" : : "r" (mva) : "memory"); |
| } |
| |
| /* data cache invalidate by set/way */ |
| static inline void dccisw(uint32_t val) |
| { |
| asm volatile ("mcr p15, 0, %0, c7, c14, 2" : : "r" (val) : "memory"); |
| } |
| |
| /* data cache clean by MVA to PoC */ |
| static inline void dccmvac(unsigned long mva) |
| { |
| asm volatile ("mcr p15, 0, %0, c7, c10, 1" : : "r" (mva) : "memory"); |
| } |
| |
| /* data cache clean by set/way */ |
| static inline void dccsw(uint32_t val) |
| { |
| asm volatile ("mcr p15, 0, %0, c7, c10, 2" : : "r" (val) : "memory"); |
| } |
| |
| /* data cache invalidate by MVA to PoC */ |
| static inline void dcimvac(unsigned long mva) |
| { |
| asm volatile ("mcr p15, 0, %0, c7, c6, 1" : : "r" (mva) : "memory"); |
| } |
| |
| /* data cache invalidate by set/way */ |
| static inline void dcisw(uint32_t val) |
| { |
| asm volatile ("mcr p15, 0, %0, c7, c6, 2" : : "r" (val) : "memory"); |
| } |
| |
| /* instruction cache invalidate all by PoU */ |
| static inline void iciallu(void) |
| { |
| asm volatile ("mcr p15, 0, %0, c7, c5, 0" : : "r" (0)); |
| } |
| |
| /* |
| * Cache co-processor (CP15) access functions |
| */ |
| |
| /* read cache level ID register (CLIDR) */ |
| static inline uint32_t read_clidr(void) |
| { |
| uint32_t val = 0; |
| asm volatile ("mrc p15, 1, %0, c0, c0, 1" : "=r" (val)); |
| return val; |
| } |
| |
| /* read cache size ID register register (CCSIDR) */ |
| static inline uint32_t read_ccsidr(void) |
| { |
| uint32_t val = 0; |
| asm volatile ("mrc p15, 1, %0, c0, c0, 0" : "=r" (val)); |
| return val; |
| } |
| |
| /* read cache size selection register (CSSELR) */ |
| static inline uint32_t read_csselr(void) |
| { |
| uint32_t val = 0; |
| asm volatile ("mrc p15, 2, %0, c0, c0, 0" : "=r" (val)); |
| return val; |
| } |
| |
| /* write to cache size selection register (CSSELR) */ |
| static inline void write_csselr(uint32_t val) |
| { |
| /* |
| * Bits [3:1] - Cache level + 1 (0b000 = L1, 0b110 = L7, 0b111 is rsvd) |
| * Bit 0 - 0 = data or unified cache, 1 = instruction cache |
| */ |
| asm volatile ("mcr p15, 2, %0, c0, c0, 0" : : "r" (val)); |
| isb(); /* ISB to sync the change to CCSIDR */ |
| } |
| |
| /* read L2 control register (L2CTLR) */ |
| static inline uint32_t read_l2ctlr(void) |
| { |
| uint32_t val = 0; |
| asm volatile ("mrc p15, 1, %0, c9, c0, 2" : "=r" (val)); |
| return val; |
| } |
| |
| /* write L2 control register (L2CTLR) */ |
| static inline void write_l2ctlr(uint32_t val) |
| { |
| /* |
| * Note: L2CTLR can only be written when the L2 memory system |
| * is idle, ie before the MMU is enabled. |
| */ |
| asm volatile("mcr p15, 1, %0, c9, c0, 2" : : "r" (val) : "memory" ); |
| isb(); |
| } |
| |
| /* read L2 Auxiliary Control Register (L2ACTLR) */ |
| static inline uint32_t read_l2actlr(void) |
| { |
| uint32_t val = 0; |
| asm volatile ("mrc p15, 1, %0, c15, c0, 0" : "=r" (val)); |
| return val; |
| } |
| |
| /* write L2 Auxiliary Control Register (L2ACTLR) */ |
| static inline void write_l2actlr(uint32_t val) |
| { |
| asm volatile ("mcr p15, 1, %0, c15, c0, 0" : : "r" (val) : "memory" ); |
| isb(); |
| } |
| |
| /* read system control register (SCTLR) */ |
| static inline uint32_t read_sctlr(void) |
| { |
| uint32_t val; |
| asm volatile ("mrc p15, 0, %0, c1, c0, 0" : "=r" (val)); |
| return val; |
| } |
| |
| /* write system control register (SCTLR) */ |
| static inline void write_sctlr(uint32_t val) |
| { |
| asm volatile ("mcr p15, 0, %0, c1, c0, 0" : : "r" (val) : "cc"); |
| isb(); |
| } |
| |
| /* |
| * Cache maintenance API |
| */ |
| |
| /* dcache clean and invalidate all (on current level given by CCSELR) */ |
| void dcache_clean_invalidate_all(void); |
| |
| /* dcache clean by modified virtual address to PoC */ |
| void dcache_clean_by_mva(void const *addr, size_t len); |
| |
| /* dcache clean and invalidate by modified virtual address to PoC */ |
| void dcache_clean_invalidate_by_mva(void const *addr, size_t len); |
| |
| /* dcache invalidate by modified virtual address to PoC */ |
| void dcache_invalidate_by_mva(void const *addr, size_t len); |
| |
| void dcache_clean_all(void); |
| |
| /* dcache invalidate all (on current level given by CCSELR) */ |
| void dcache_invalidate_all(void); |
| |
| /* dcache and MMU disable */ |
| void dcache_mmu_disable(void); |
| |
| /* dcache and MMU enable */ |
| void dcache_mmu_enable(void); |
| |
| /* icache invalidate all (on current level given by CSSELR) */ |
| void icache_invalidate_all(void); |
| |
| /* tlb invalidate all */ |
| void tlb_invalidate_all(void); |
| |
| /* |
| * Generalized setup/init functions |
| */ |
| |
| /* invalidate all caches on ARMv7 */ |
| void armv7_invalidate_caches(void); |
| |
| /* mmu initialization (set page table address, set permissions, etc) */ |
| void mmu_init(void); |
| |
| enum dcache_policy { |
| DCACHE_OFF, |
| DCACHE_WRITEBACK, |
| DCACHE_WRITETHROUGH, |
| }; |
| |
| /* disable the mmu for a range. Primarily useful to lock out address 0. */ |
| void mmu_disable_range(unsigned long start_mb, unsigned long size_mb); |
| /* mmu range configuration (set dcache policy) */ |
| void mmu_config_range(unsigned long start_mb, unsigned long size_mb, |
| enum dcache_policy policy); |
| |
| #endif /* ARMV7_CACHE_H */ |