123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538 |
- /*
- * Copyright (c) 2017-2024, Arm Limited and Contributors. All rights reserved.
- *
- * SPDX-License-Identifier: BSD-3-Clause
- */
- #include <arch.h>
- #include <asm_macros.S>
- #include <common/bl_common.h>
- #include <cortex_a76.h>
- #include <cpu_macros.S>
- #include <plat_macros.S>
- #include <services/arm_arch_svc.h>
- #include "wa_cve_2022_23960_bhb.S"
- /* Hardware handled coherency */
- #if HW_ASSISTED_COHERENCY == 0
- #error "Cortex-A76 must be compiled with HW_ASSISTED_COHERENCY enabled"
- #endif
- .globl cortex_a76_reset_func
- .globl cortex_a76_core_pwr_dwn
- .globl cortex_a76_disable_wa_cve_2018_3639
- /* 64-bit only core */
- #if CTX_INCLUDE_AARCH32_REGS == 1
- #error "Cortex-A76 supports only AArch64. Compile with CTX_INCLUDE_AARCH32_REGS=0"
- #endif
- #define ESR_EL3_A64_SMC0 0x5e000000
- #define ESR_EL3_A32_SMC0 0x4e000000
- #if DYNAMIC_WORKAROUND_CVE_2018_3639
- /*
- * This macro applies the mitigation for CVE-2018-3639.
- * It implements a fast path where `SMCCC_ARCH_WORKAROUND_2`
- * SMC calls from a lower EL running in AArch32 or AArch64
- * will go through the fast and return early.
- *
- * The macro saves x2-x3 to the context. In the fast path
- * x0-x3 registers do not need to be restored as the calling
- * context will have saved them. The macro also saves
- * x29-x30 to the context in the sync_exception path.
- */
- .macro apply_cve_2018_3639_wa _is_sync_exception _esr_el3_val
- stp x2, x3, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X2]
- .if \_is_sync_exception
- stp x29, x30, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X29]
- mov_imm w2, \_esr_el3_val
- bl apply_cve_2018_3639_sync_wa
- ldp x29, x30, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X29]
- .endif
- /*
- * Always enable v4 mitigation during EL3 execution. This is not
- * required for the fast path above because it does not perform any
- * memory loads.
- */
- mrs x2, CORTEX_A76_CPUACTLR2_EL1
- orr x2, x2, #CORTEX_A76_CPUACTLR2_EL1_DISABLE_LOAD_PASS_STORE
- msr CORTEX_A76_CPUACTLR2_EL1, x2
- isb
- /*
- * The caller may have passed arguments to EL3 via x2-x3.
- * Restore these registers from the context before jumping to the
- * main runtime vector table entry.
- */
- ldp x2, x3, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X2]
- .endm
- #endif /* DYNAMIC_WORKAROUND_CVE_2018_3639 */
- #if DYNAMIC_WORKAROUND_CVE_2018_3639 || WORKAROUND_CVE_2022_23960
- vector_base cortex_a76_wa_cve_vbar
- /* ---------------------------------------------------------------------
- * Current EL with SP_EL0 : 0x0 - 0x200
- * ---------------------------------------------------------------------
- */
- vector_entry cortex_a76_sync_exception_sp_el0
- b sync_exception_sp_el0
- end_vector_entry cortex_a76_sync_exception_sp_el0
- vector_entry cortex_a76_irq_sp_el0
- b irq_sp_el0
- end_vector_entry cortex_a76_irq_sp_el0
- vector_entry cortex_a76_fiq_sp_el0
- b fiq_sp_el0
- end_vector_entry cortex_a76_fiq_sp_el0
- vector_entry cortex_a76_serror_sp_el0
- b serror_sp_el0
- end_vector_entry cortex_a76_serror_sp_el0
- /* ---------------------------------------------------------------------
- * Current EL with SP_ELx: 0x200 - 0x400
- * ---------------------------------------------------------------------
- */
- vector_entry cortex_a76_sync_exception_sp_elx
- b sync_exception_sp_elx
- end_vector_entry cortex_a76_sync_exception_sp_elx
- vector_entry cortex_a76_irq_sp_elx
- b irq_sp_elx
- end_vector_entry cortex_a76_irq_sp_elx
- vector_entry cortex_a76_fiq_sp_elx
- b fiq_sp_elx
- end_vector_entry cortex_a76_fiq_sp_elx
- vector_entry cortex_a76_serror_sp_elx
- b serror_sp_elx
- end_vector_entry cortex_a76_serror_sp_elx
- /* ---------------------------------------------------------------------
- * Lower EL using AArch64 : 0x400 - 0x600
- * ---------------------------------------------------------------------
- */
- vector_entry cortex_a76_sync_exception_aarch64
- #if WORKAROUND_CVE_2022_23960
- apply_cve_2022_23960_bhb_wa CORTEX_A76_BHB_LOOP_COUNT
- #endif /* WORKAROUND_CVE_2022_23960 */
- #if DYNAMIC_WORKAROUND_CVE_2018_3639
- apply_cve_2018_3639_wa _is_sync_exception=1 _esr_el3_val=ESR_EL3_A64_SMC0
- #endif /* DYNAMIC_WORKAROUND_CVE_2018_3639*/
- b sync_exception_aarch64
- end_vector_entry cortex_a76_sync_exception_aarch64
- vector_entry cortex_a76_irq_aarch64
- #if WORKAROUND_CVE_2022_23960
- apply_cve_2022_23960_bhb_wa CORTEX_A76_BHB_LOOP_COUNT
- #endif /* WORKAROUND_CVE_2022_23960 */
- #if DYNAMIC_WORKAROUND_CVE_2018_3639
- apply_cve_2018_3639_wa _is_sync_exception=0 _esr_el3_val=ESR_EL3_A64_SMC0
- #endif /* DYNAMIC_WORKAROUND_CVE_2018_3639*/
- b irq_aarch64
- end_vector_entry cortex_a76_irq_aarch64
- vector_entry cortex_a76_fiq_aarch64
- #if WORKAROUND_CVE_2022_23960
- apply_cve_2022_23960_bhb_wa CORTEX_A76_BHB_LOOP_COUNT
- #endif /* WORKAROUND_CVE_2022_23960 */
- #if DYNAMIC_WORKAROUND_CVE_2018_3639
- apply_cve_2018_3639_wa _is_sync_exception=0 _esr_el3_val=ESR_EL3_A64_SMC0
- #endif /* DYNAMIC_WORKAROUND_CVE_2018_3639*/
- b fiq_aarch64
- end_vector_entry cortex_a76_fiq_aarch64
- vector_entry cortex_a76_serror_aarch64
- #if WORKAROUND_CVE_2022_23960
- apply_cve_2022_23960_bhb_wa CORTEX_A76_BHB_LOOP_COUNT
- #endif /* WORKAROUND_CVE_2022_23960 */
- #if DYNAMIC_WORKAROUND_CVE_2018_3639
- apply_cve_2018_3639_wa _is_sync_exception=0 _esr_el3_val=ESR_EL3_A64_SMC0
- #endif /* DYNAMIC_WORKAROUND_CVE_2018_3639*/
- b serror_aarch64
- end_vector_entry cortex_a76_serror_aarch64
- /* ---------------------------------------------------------------------
- * Lower EL using AArch32 : 0x600 - 0x800
- * ---------------------------------------------------------------------
- */
- vector_entry cortex_a76_sync_exception_aarch32
- #if WORKAROUND_CVE_2022_23960
- apply_cve_2022_23960_bhb_wa CORTEX_A76_BHB_LOOP_COUNT
- #endif /* WORKAROUND_CVE_2022_23960 */
- #if DYNAMIC_WORKAROUND_CVE_2018_3639
- apply_cve_2018_3639_wa _is_sync_exception=1 _esr_el3_val=ESR_EL3_A32_SMC0
- #endif /* DYNAMIC_WORKAROUND_CVE_2018_3639*/
- b sync_exception_aarch32
- end_vector_entry cortex_a76_sync_exception_aarch32
- vector_entry cortex_a76_irq_aarch32
- #if WORKAROUND_CVE_2022_23960
- apply_cve_2022_23960_bhb_wa CORTEX_A76_BHB_LOOP_COUNT
- #endif /* WORKAROUND_CVE_2022_23960 */
- #if DYNAMIC_WORKAROUND_CVE_2018_3639
- apply_cve_2018_3639_wa _is_sync_exception=0 _esr_el3_val=ESR_EL3_A32_SMC0
- #endif /* DYNAMIC_WORKAROUND_CVE_2018_3639*/
- b irq_aarch32
- end_vector_entry cortex_a76_irq_aarch32
- vector_entry cortex_a76_fiq_aarch32
- #if WORKAROUND_CVE_2022_23960
- apply_cve_2022_23960_bhb_wa CORTEX_A76_BHB_LOOP_COUNT
- #endif /* WORKAROUND_CVE_2022_23960 */
- #if DYNAMIC_WORKAROUND_CVE_2018_3639
- apply_cve_2018_3639_wa _is_sync_exception=0 _esr_el3_val=ESR_EL3_A32_SMC0
- #endif /* DYNAMIC_WORKAROUND_CVE_2018_3639*/
- b fiq_aarch32
- end_vector_entry cortex_a76_fiq_aarch32
- vector_entry cortex_a76_serror_aarch32
- #if WORKAROUND_CVE_2022_23960
- apply_cve_2022_23960_bhb_wa CORTEX_A76_BHB_LOOP_COUNT
- #endif /* WORKAROUND_CVE_2022_23960 */
- #if DYNAMIC_WORKAROUND_CVE_2018_3639
- apply_cve_2018_3639_wa _is_sync_exception=0 _esr_el3_val=ESR_EL3_A32_SMC0
- #endif /* DYNAMIC_WORKAROUND_CVE_2018_3639*/
- b serror_aarch32
- end_vector_entry cortex_a76_serror_aarch32
- #endif /* DYNAMIC_WORKAROUND_CVE_2018_3639 || WORKAROUND_CVE_2022_23960 */
- #if DYNAMIC_WORKAROUND_CVE_2018_3639
- /*
- * -----------------------------------------------------------------
- * This function applies the mitigation for CVE-2018-3639
- * specifically for sync exceptions. It implements a fast path
- * where `SMCCC_ARCH_WORKAROUND_2` SMC calls from a lower EL
- * running in AArch64 will go through the fast and return early.
- *
- * In the fast path x0-x3 registers do not need to be restored as the
- * calling context will have saved them.
- *
- * Caller must pass value of esr_el3 to compare via x2.
- * Save and restore these registers outside of this function from the
- * context before jumping to the main runtime vector table entry.
- *
- * Shall clobber: x0-x3, x30
- * -----------------------------------------------------------------
- */
- func apply_cve_2018_3639_sync_wa
- /*
- * Ensure SMC is coming from A64/A32 state on #0
- * with W0 = SMCCC_ARCH_WORKAROUND_2
- *
- * This sequence evaluates as:
- * (W0==SMCCC_ARCH_WORKAROUND_2) ? (ESR_EL3==SMC#0) : (NE)
- * allowing use of a single branch operation
- * X2 populated outside this function with the SMC FID.
- */
- orr w3, wzr, #SMCCC_ARCH_WORKAROUND_2
- cmp x0, x3
- mrs x3, esr_el3
- ccmp w2, w3, #0, eq
- /*
- * Static predictor will predict a fall-through, optimizing
- * the `SMCCC_ARCH_WORKAROUND_2` fast path.
- */
- bne 1f
- /*
- * The sequence below implements the `SMCCC_ARCH_WORKAROUND_2`
- * fast path.
- */
- cmp x1, xzr /* enable/disable check */
- /*
- * When the calling context wants mitigation disabled,
- * we program the mitigation disable function in the
- * CPU context, which gets invoked on subsequent exits from
- * EL3 via the `el3_exit` function. Otherwise NULL is
- * programmed in the CPU context, which results in caller's
- * inheriting the EL3 mitigation state (enabled) on subsequent
- * `el3_exit`.
- */
- mov x0, xzr
- adr x1, cortex_a76_disable_wa_cve_2018_3639
- csel x1, x1, x0, eq
- str x1, [sp, #CTX_CVE_2018_3639_OFFSET + CTX_CVE_2018_3639_DISABLE]
- mrs x2, CORTEX_A76_CPUACTLR2_EL1
- orr x1, x2, #CORTEX_A76_CPUACTLR2_EL1_DISABLE_LOAD_PASS_STORE
- bic x3, x2, #CORTEX_A76_CPUACTLR2_EL1_DISABLE_LOAD_PASS_STORE
- csel x3, x3, x1, eq
- msr CORTEX_A76_CPUACTLR2_EL1, x3
- ldp x29, x30, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X29]
- /*
- * `SMCCC_ARCH_WORKAROUND_2`fast path return to lower EL.
- */
- exception_return /* exception_return contains ISB */
- 1:
- ret
- endfunc apply_cve_2018_3639_sync_wa
- #endif /* DYNAMIC_WORKAROUND_CVE_2018_3639 */
- workaround_reset_start cortex_a76, ERRATUM(1073348), ERRATA_A76_1073348
- sysreg_bit_set CORTEX_A76_CPUACTLR_EL1 ,CORTEX_A76_CPUACTLR_EL1_DISABLE_STATIC_PREDICTION
- workaround_reset_end cortex_a76, ERRATUM(1073348)
- check_erratum_ls cortex_a76, ERRATUM(1073348), CPU_REV(1, 0)
- workaround_reset_start cortex_a76, ERRATUM(1130799), ERRATA_A76_1130799
- sysreg_bit_set CORTEX_A76_CPUACTLR2_EL1, CORTEX_A76_CPUACTLR2_EL1_BIT_59
- msr CORTEX_A76_CPUACTLR2_EL1, x1
- workaround_reset_end cortex_a76, ERRATUM(1130799)
- check_erratum_ls cortex_a76, ERRATUM(1130799), CPU_REV(2, 0)
- workaround_reset_start cortex_a76, ERRATUM(1220197), ERRATA_A76_1220197
- sysreg_bit_set CORTEX_A76_CPUECTLR_EL1, CORTEX_A76_CPUECTLR_EL1_WS_THR_L2
- workaround_reset_end cortex_a76, ERRATUM(1220197)
- check_erratum_ls cortex_a76, ERRATUM(1220197), CPU_REV(2, 0)
- workaround_reset_start cortex_a76, ERRATUM(1257314), ERRATA_A76_1257314
- sysreg_bit_set CORTEX_A76_CPUACTLR3_EL1, CORTEX_A76_CPUACTLR3_EL1_BIT_10
- workaround_reset_end cortex_a76, ERRATUM(1257314)
- check_erratum_ls cortex_a76, ERRATUM(1257314), CPU_REV(3, 0)
- workaround_reset_start cortex_a76, ERRATUM(1262606), ERRATA_A76_1262606
- sysreg_bit_set CORTEX_A76_CPUACTLR_EL1, CORTEX_A76_CPUACTLR_EL1_BIT_13
- workaround_reset_end cortex_a76, ERRATUM(1262606)
- check_erratum_ls cortex_a76, ERRATUM(1262606), CPU_REV(3, 0)
- workaround_reset_start cortex_a76, ERRATUM(1262888), ERRATA_A76_1262888
- sysreg_bit_set CORTEX_A76_CPUECTLR_EL1, CORTEX_A76_CPUECTLR_EL1_BIT_51
- workaround_reset_end cortex_a76, ERRATUM(1262888)
- check_erratum_ls cortex_a76, ERRATUM(1262888), CPU_REV(3, 0)
- workaround_reset_start cortex_a76, ERRATUM(1275112), ERRATA_A76_1275112
- sysreg_bit_set CORTEX_A76_CPUACTLR_EL1, CORTEX_A76_CPUACTLR_EL1_BIT_13
- workaround_reset_end cortex_a76, ERRATUM(1275112)
- check_erratum_ls cortex_a76, ERRATUM(1275112), CPU_REV(3, 0)
- check_erratum_custom_start cortex_a76, ERRATUM(1286807)
- #if ERRATA_A76_1286807
- mov x0, #ERRATA_APPLIES
- ret
- #else
- mov x1, #0x30
- b cpu_rev_var_ls
- #endif
- check_erratum_custom_end cortex_a76, ERRATUM(1286807)
- workaround_reset_start cortex_a76, ERRATUM(1791580), ERRATA_A76_1791580
- sysreg_bit_set CORTEX_A76_CPUACTLR2_EL1, CORTEX_A76_CPUACTLR2_EL1_BIT_2
- workaround_reset_end cortex_a76, ERRATUM(1791580)
- check_erratum_ls cortex_a76, ERRATUM(1791580), CPU_REV(4, 0)
- workaround_reset_start cortex_a76, ERRATUM(1868343), ERRATA_A76_1868343
- sysreg_bit_set CORTEX_A76_CPUACTLR_EL1, CORTEX_A76_CPUACTLR_EL1_BIT_13
- workaround_reset_end cortex_a76, ERRATUM(1868343)
- check_erratum_ls cortex_a76, ERRATUM(1868343), CPU_REV(4, 0)
- workaround_reset_start cortex_a76, ERRATUM(1946160), ERRATA_A76_1946160
- mov x0, #3
- msr S3_6_C15_C8_0, x0
- ldr x0, =0x10E3900002
- msr S3_6_C15_C8_2, x0
- ldr x0, =0x10FFF00083
- msr S3_6_C15_C8_3, x0
- ldr x0, =0x2001003FF
- msr S3_6_C15_C8_1, x0
- mov x0, #4
- msr S3_6_C15_C8_0, x0
- ldr x0, =0x10E3800082
- msr S3_6_C15_C8_2, x0
- ldr x0, =0x10FFF00083
- msr S3_6_C15_C8_3, x0
- ldr x0, =0x2001003FF
- msr S3_6_C15_C8_1, x0
- mov x0, #5
- msr S3_6_C15_C8_0, x0
- ldr x0, =0x10E3800200
- msr S3_6_C15_C8_2, x0
- ldr x0, =0x10FFF003E0
- msr S3_6_C15_C8_3, x0
- ldr x0, =0x2001003FF
- msr S3_6_C15_C8_1, x0
- workaround_reset_end cortex_a76, ERRATUM(1946160)
- check_erratum_range cortex_a76, ERRATUM(1946160), CPU_REV(3, 0), CPU_REV(4, 1)
- workaround_runtime_start cortex_a76, ERRATUM(2743102), ERRATA_A76_2743102
- /* dsb before isb of power down sequence */
- dsb sy
- workaround_runtime_end cortex_a76, ERRATUM(2743102)
- check_erratum_ls cortex_a76, ERRATUM(2743102), CPU_REV(4, 1)
- check_erratum_chosen cortex_a76, CVE(2018, 3639), WORKAROUND_CVE_2018_3639
- func cortex_a76_disable_wa_cve_2018_3639
- sysreg_bit_clear CORTEX_A76_CPUACTLR2_EL1, CORTEX_A76_CPUACTLR2_EL1_DISABLE_LOAD_PASS_STORE
- isb
- ret
- endfunc cortex_a76_disable_wa_cve_2018_3639
- /* --------------------------------------------------------------
- * Errata Workaround for Cortex A76 Errata #1165522.
- * This applies only to revisions <= r3p0 of Cortex A76.
- * Due to the nature of the errata it is applied unconditionally
- * when built in, report it as applicable in this case
- * --------------------------------------------------------------
- */
- check_erratum_custom_start cortex_a76, ERRATUM(1165522)
- #if ERRATA_A76_1165522
- mov x0, #ERRATA_APPLIES
- ret
- #else
- mov x1, #0x30
- b cpu_rev_var_ls
- #endif
- check_erratum_custom_end cortex_a76, ERRATUM(1165522)
- check_erratum_chosen cortex_a76, CVE(2022, 23960), WORKAROUND_CVE_2022_23960
- /* erratum has no workaround in the cpu. Generic code must take care */
- add_erratum_entry cortex_a76, CVE(2022, 23960), WORKAROUND_CVE_2022_23960, NO_APPLY_AT_RESET
- /* ERRATA_DSU_798953 :
- * The errata is defined in dsu_helpers.S but applies to cortex_a76
- * as well. Henceforth creating symbolic names to the already existing errata
- * workaround functions to get them registered under the Errata Framework.
- */
- .equ check_erratum_cortex_a76_798953, check_errata_dsu_798953
- .equ erratum_cortex_a76_798953_wa, errata_dsu_798953_wa
- add_erratum_entry cortex_a76, ERRATUM(798953), ERRATA_DSU_798953, APPLY_AT_RESET
- /* ERRATA_DSU_936184 :
- * The errata is defined in dsu_helpers.S but applies to cortex_a76
- * as well. Henceforth creating symbolic names to the already existing errata
- * workaround functions to get them registered under the Errata Framework.
- */
- .equ check_erratum_cortex_a76_936184, check_errata_dsu_936184
- .equ erratum_cortex_a76_936184_wa, errata_dsu_936184_wa
- add_erratum_entry cortex_a76, ERRATUM(936184), ERRATA_DSU_936184, APPLY_AT_RESET
- cpu_reset_func_start cortex_a76
- #if WORKAROUND_CVE_2018_3639
- /* If the PE implements SSBS, we don't need the dynamic workaround */
- mrs x0, id_aa64pfr1_el1
- lsr x0, x0, #ID_AA64PFR1_EL1_SSBS_SHIFT
- and x0, x0, #ID_AA64PFR1_EL1_SSBS_MASK
- #if !DYNAMIC_WORKAROUND_CVE_2018_3639 && ENABLE_ASSERTIONS
- cmp x0, 0
- ASM_ASSERT(ne)
- #endif
- #if DYNAMIC_WORKAROUND_CVE_2018_3639
- cbnz x0, 1f
- sysreg_bit_set CORTEX_A76_CPUACTLR2_EL1, CORTEX_A76_CPUACTLR2_EL1_DISABLE_LOAD_PASS_STORE
- isb
- #ifdef IMAGE_BL31
- /*
- * The Cortex-A76 generic vectors are overwritten to use the vectors
- * defined above. This is required in order to apply mitigation
- * against CVE-2018-3639 on exception entry from lower ELs.
- * If the below vector table is used, skip overriding it again for
- * CVE_2022_23960 as both use the same vbar.
- */
- override_vector_table cortex_a76_wa_cve_vbar
- isb
- b 2f
- #endif /* IMAGE_BL31 */
- 1:
- #endif /* DYNAMIC_WORKAROUND_CVE_2018_3639 */
- #endif /* WORKAROUND_CVE_2018_3639 */
- #if IMAGE_BL31 && WORKAROUND_CVE_2022_23960
- /*
- * The Cortex-A76 generic vectors are overridden to apply errata
- * mitigation on exception entry from lower ELs. This will be bypassed
- * if DYNAMIC_WORKAROUND_CVE_2018_3639 has overridden the vectors.
- */
- override_vector_table cortex_a76_wa_cve_vbar
- isb
- #endif /* IMAGE_BL31 && WORKAROUND_CVE_2022_23960 */
- 2:
- cpu_reset_func_end cortex_a76
- /* ---------------------------------------------
- * HW will do the cache maintenance while powering down
- * ---------------------------------------------
- */
- func cortex_a76_core_pwr_dwn
- /* ---------------------------------------------
- * Enable CPU power down bit in power control register
- * ---------------------------------------------
- */
- sysreg_bit_set CORTEX_A76_CPUPWRCTLR_EL1, CORTEX_A76_CORE_PWRDN_EN_MASK
- apply_erratum cortex_a76, ERRATUM(2743102), ERRATA_A76_2743102
- isb
- ret
- endfunc cortex_a76_core_pwr_dwn
- /* ---------------------------------------------
- * This function provides cortex_a76 specific
- * register information for crash reporting.
- * It needs to return with x6 pointing to
- * a list of register names in ascii and
- * x8 - x15 having values of registers to be
- * reported.
- * ---------------------------------------------
- */
- .section .rodata.cortex_a76_regs, "aS"
- cortex_a76_regs: /* The ascii list of register names to be reported */
- .asciz "cpuectlr_el1", ""
- func cortex_a76_cpu_reg_dump
- adr x6, cortex_a76_regs
- mrs x8, CORTEX_A76_CPUECTLR_EL1
- ret
- endfunc cortex_a76_cpu_reg_dump
- declare_cpu_ops_wa cortex_a76, CORTEX_A76_MIDR, \
- cortex_a76_reset_func, \
- CPU_NO_EXTRA1_FUNC, \
- cortex_a76_disable_wa_cve_2018_3639, \
- CPU_NO_EXTRA3_FUNC, \
- cortex_a76_core_pwr_dwn
|