/* * Copyright (c) 2019-2022, ARM Limited and Contributors. All rights reserved. * * SPDX-License-Identifier: BSD-3-Clause */ #include #include #include #include #include #include #include #include #include #include #include #include #define CCGR(x) (0x4000 + (x) * 16) enum pu_domain_id { HSIOMIX, PCIE, OTG1, OTG2, GPUMIX, VPUMIX, VPU_G1, VPU_G2, VPU_H1, DISPMIX, MIPI, /* below two domain only for ATF internal use */ GPU2D, GPU3D, MAX_DOMAINS, }; /* PU domain */ static struct imx_pwr_domain pu_domains[] = { IMX_MIX_DOMAIN(HSIOMIX, false), IMX_PD_DOMAIN(PCIE, false), IMX_PD_DOMAIN(OTG1, true), IMX_PD_DOMAIN(OTG2, true), IMX_MIX_DOMAIN(GPUMIX, false), IMX_MIX_DOMAIN(VPUMIX, false), IMX_PD_DOMAIN(VPU_G1, false), IMX_PD_DOMAIN(VPU_G2, false), IMX_PD_DOMAIN(VPU_H1, false), IMX_MIX_DOMAIN(DISPMIX, false), IMX_PD_DOMAIN(MIPI, false), /* below two domain only for ATF internal use */ IMX_MIX_DOMAIN(GPU2D, false), IMX_MIX_DOMAIN(GPU3D, false), }; static unsigned int pu_domain_status; #define GPU_RCR 0x40 #define VPU_RCR 0x44 #define VPU_CTL_BASE 0x38330000 #define BLK_SFT_RSTN_CSR 0x0 #define H1_SFT_RSTN BIT(2) #define G1_SFT_RSTN BIT(1) #define G2_SFT_RSTN BIT(0) #define DISP_CTL_BASE 0x32e28000 void vpu_sft_reset_assert(uint32_t domain_id) { uint32_t val; val = mmio_read_32(VPU_CTL_BASE + BLK_SFT_RSTN_CSR); switch (domain_id) { case VPU_G1: val &= ~G1_SFT_RSTN; mmio_write_32(VPU_CTL_BASE + BLK_SFT_RSTN_CSR, val); break; case VPU_G2: val &= ~G2_SFT_RSTN; mmio_write_32(VPU_CTL_BASE + BLK_SFT_RSTN_CSR, val); break; case VPU_H1: val &= ~H1_SFT_RSTN; mmio_write_32(VPU_CTL_BASE + BLK_SFT_RSTN_CSR, val); break; default: break; } } void vpu_sft_reset_deassert(uint32_t domain_id) { uint32_t val; val = mmio_read_32(VPU_CTL_BASE + BLK_SFT_RSTN_CSR); switch (domain_id) { case VPU_G1: val |= G1_SFT_RSTN; mmio_write_32(VPU_CTL_BASE + BLK_SFT_RSTN_CSR, val); break; case VPU_G2: val |= G2_SFT_RSTN; mmio_write_32(VPU_CTL_BASE + BLK_SFT_RSTN_CSR, val); break; case VPU_H1: val |= H1_SFT_RSTN; mmio_write_32(VPU_CTL_BASE + BLK_SFT_RSTN_CSR, val); break; default: break; } } void imx_gpc_pm_domain_enable(uint32_t domain_id, bool on) { if (domain_id >= MAX_DOMAINS) { return; } struct imx_pwr_domain *pwr_domain = &pu_domains[domain_id]; if (on) { pu_domain_status |= (1 << domain_id); if (domain_id == VPU_G1 || domain_id == VPU_G2 || domain_id == VPU_H1) { vpu_sft_reset_assert(domain_id); } /* HSIOMIX has no PU bit, so skip for it */ if (domain_id != HSIOMIX) { /* clear the PGC bit */ mmio_clrbits_32(IMX_GPC_BASE + pwr_domain->pgc_offset, 0x1); /* power up the domain */ mmio_setbits_32(IMX_GPC_BASE + PU_PGC_UP_TRG, pwr_domain->pwr_req); /* wait for power request done */ while (mmio_read_32(IMX_GPC_BASE + PU_PGC_UP_TRG) & pwr_domain->pwr_req) { ; } } if (domain_id == VPU_G1 || domain_id == VPU_G2 || domain_id == VPU_H1) { vpu_sft_reset_deassert(domain_id); /* dealy for a while to make sure reset done */ udelay(100); } if (domain_id == GPUMIX) { /* assert reset */ mmio_write_32(IMX_SRC_BASE + GPU_RCR, 0x1); /* power up GPU2D */ mmio_clrbits_32(IMX_GPC_BASE + GPU2D_PGC, 0x1); mmio_setbits_32(IMX_GPC_BASE + PU_PGC_UP_TRG, GPU2D_PWR_REQ); /* wait for power request done */ while (mmio_read_32(IMX_GPC_BASE + PU_PGC_UP_TRG) & GPU2D_PWR_REQ) { ; } udelay(1); /* power up GPU3D */ mmio_clrbits_32(IMX_GPC_BASE + GPU3D_PGC, 0x1); mmio_setbits_32(IMX_GPC_BASE + PU_PGC_UP_TRG, GPU3D_PWR_REQ); /* wait for power request done */ while (mmio_read_32(IMX_GPC_BASE + PU_PGC_UP_TRG) & GPU3D_PWR_REQ) { ; } udelay(10); /* release the gpumix reset */ mmio_write_32(IMX_SRC_BASE + GPU_RCR, 0x0); udelay(10); } /* vpu sft clock enable */ if (domain_id == VPUMIX) { mmio_write_32(IMX_SRC_BASE + VPU_RCR, 0x1); udelay(5); mmio_write_32(IMX_SRC_BASE + VPU_RCR, 0x0); udelay(5); /* enable all clock */ mmio_write_32(VPU_CTL_BASE + 0x4, 0x7); } if (domain_id == DISPMIX) { /* special setting for DISPMIX */ mmio_write_32(DISP_CTL_BASE + 0x4, 0x1fff); mmio_write_32(DISP_CTL_BASE, 0x7f); mmio_write_32(DISP_CTL_BASE + 0x8, 0x30000); } /* handle the ADB400 sync */ if (pwr_domain->need_sync) { /* clear adb power down request */ mmio_setbits_32(IMX_GPC_BASE + GPC_PU_PWRHSK, pwr_domain->adb400_sync); /* wait for adb power request ack */ while (!(mmio_read_32(IMX_GPC_BASE + GPC_PU_PWRHSK) & pwr_domain->adb400_ack)) { ; } } if (domain_id == GPUMIX) { /* power up GPU2D ADB */ mmio_setbits_32(IMX_GPC_BASE + GPC_PU_PWRHSK, GPU2D_ADB400_SYNC); /* wait for adb power request ack */ while (!(mmio_read_32(IMX_GPC_BASE + GPC_PU_PWRHSK) & GPU2D_ADB400_ACK)) { ; } /* power up GPU3D ADB */ mmio_setbits_32(IMX_GPC_BASE + GPC_PU_PWRHSK, GPU3D_ADB400_SYNC); /* wait for adb power request ack */ while (!(mmio_read_32(IMX_GPC_BASE + GPC_PU_PWRHSK) & GPU3D_ADB400_ACK)) { ; } } } else { pu_domain_status &= ~(1 << domain_id); if (domain_id == OTG1 || domain_id == OTG2) { return; } /* GPU2D & GPU3D ADB power down */ if (domain_id == GPUMIX) { mmio_clrbits_32(IMX_GPC_BASE + GPC_PU_PWRHSK, GPU2D_ADB400_SYNC); /* wait for adb power request ack */ while ((mmio_read_32(IMX_GPC_BASE + GPC_PU_PWRHSK) & GPU2D_ADB400_ACK)) { ; } mmio_clrbits_32(IMX_GPC_BASE + GPC_PU_PWRHSK, GPU3D_ADB400_SYNC); /* wait for adb power request ack */ while ((mmio_read_32(IMX_GPC_BASE + GPC_PU_PWRHSK) & GPU3D_ADB400_ACK)) { ; } } /* handle the ADB400 sync */ if (pwr_domain->need_sync) { /* set adb power down request */ mmio_clrbits_32(IMX_GPC_BASE + GPC_PU_PWRHSK, pwr_domain->adb400_sync); /* wait for adb power request ack */ while ((mmio_read_32(IMX_GPC_BASE + GPC_PU_PWRHSK) & pwr_domain->adb400_ack)) { ; } } if (domain_id == GPUMIX) { /* power down GPU2D */ mmio_setbits_32(IMX_GPC_BASE + GPU2D_PGC, 0x1); mmio_setbits_32(IMX_GPC_BASE + PU_PGC_DN_TRG, GPU2D_PWR_REQ); /* wait for power request done */ while (mmio_read_32(IMX_GPC_BASE + PU_PGC_DN_TRG) & GPU2D_PWR_REQ) { ; } /* power down GPU3D */ mmio_setbits_32(IMX_GPC_BASE + GPU3D_PGC, 0x1); mmio_setbits_32(IMX_GPC_BASE + PU_PGC_DN_TRG, GPU3D_PWR_REQ); /* wait for power request done */ while (mmio_read_32(IMX_GPC_BASE + PU_PGC_DN_TRG) & GPU3D_PWR_REQ) { ; } } /* HSIOMIX has no PU bit, so skip for it */ if (domain_id != HSIOMIX) { /* set the PGC bit */ mmio_setbits_32(IMX_GPC_BASE + pwr_domain->pgc_offset, 0x1); /* power down the domain */ mmio_setbits_32(IMX_GPC_BASE + PU_PGC_DN_TRG, pwr_domain->pwr_req); /* wait for power request done */ while (mmio_read_32(IMX_GPC_BASE + PU_PGC_DN_TRG) & pwr_domain->pwr_req) { ; } } } } void imx_gpc_init(void) { unsigned int val; int i; /* mask all the wakeup irq by default */ for (i = 0; i < 4; i++) { mmio_write_32(IMX_GPC_BASE + IMR1_CORE0_A53 + i * 4, ~0x0); mmio_write_32(IMX_GPC_BASE + IMR1_CORE1_A53 + i * 4, ~0x0); mmio_write_32(IMX_GPC_BASE + IMR1_CORE2_A53 + i * 4, ~0x0); mmio_write_32(IMX_GPC_BASE + IMR1_CORE3_A53 + i * 4, ~0x0); mmio_write_32(IMX_GPC_BASE + IMR1_CORE0_M4 + i * 4, ~0x0); } val = mmio_read_32(IMX_GPC_BASE + LPCR_A53_BSC); /* use GIC wake_request to wakeup C0~C3 from LPM */ val |= 0x30c00000; /* clear the MASTER0 LPM handshake */ val &= ~(1 << 6); mmio_write_32(IMX_GPC_BASE + LPCR_A53_BSC, val); /* clear MASTER1 & MASTER2 mapping in CPU0(A53) */ mmio_clrbits_32(IMX_GPC_BASE + MST_CPU_MAPPING, (MASTER1_MAPPING | MASTER2_MAPPING)); /* set all mix/PU in A53 domain */ mmio_write_32(IMX_GPC_BASE + PGC_CPU_0_1_MAPPING, 0xffff); /* * Set the CORE & SCU power up timing: * SW = 0x1, SW2ISO = 0x1; * the CPU CORE and SCU power up timing counter * is drived by 32K OSC, each domain's power up * latency is (SW + SW2ISO) / 32768 */ mmio_write_32(IMX_GPC_BASE + COREx_PGC_PCR(0) + 0x4, 0x81); mmio_write_32(IMX_GPC_BASE + COREx_PGC_PCR(1) + 0x4, 0x81); mmio_write_32(IMX_GPC_BASE + COREx_PGC_PCR(2) + 0x4, 0x81); mmio_write_32(IMX_GPC_BASE + COREx_PGC_PCR(3) + 0x4, 0x81); mmio_write_32(IMX_GPC_BASE + PLAT_PGC_PCR + 0x4, 0x81); mmio_write_32(IMX_GPC_BASE + PGC_SCU_TIMING, (0x59 << 10) | 0x5B | (0x2 << 20)); /* set DUMMY PDN/PUP ACK by default for A53 domain */ mmio_write_32(IMX_GPC_BASE + PGC_ACK_SEL_A53, A53_DUMMY_PUP_ACK | A53_DUMMY_PDN_ACK); /* clear DSM by default */ val = mmio_read_32(IMX_GPC_BASE + SLPCR); val &= ~SLPCR_EN_DSM; /* enable the fast wakeup wait mode */ val |= SLPCR_A53_FASTWUP_WAIT_MODE; /* clear the RBC */ val &= ~(0x3f << SLPCR_RBC_COUNT_SHIFT); /* set the STBY_COUNT to 0x5, (128 * 30)us */ val &= ~(0x7 << SLPCR_STBY_COUNT_SHFT); val |= (0x5 << SLPCR_STBY_COUNT_SHFT); mmio_write_32(IMX_GPC_BASE + SLPCR, val); /* * USB PHY power up needs to make sure RESET bit in SRC is clear, * otherwise, the PU power up bit in GPC will NOT self-cleared. * only need to do it once. */ mmio_clrbits_32(IMX_SRC_BASE + SRC_OTG1PHY_SCR, 0x1); mmio_clrbits_32(IMX_SRC_BASE + SRC_OTG2PHY_SCR, 0x1); }