/* * Copyright (C) 2014 Freescale Semiconductor, Inc. * * The code contained herein is licensed under the GNU General Public * License. You may obtain a copy of the GNU General Public License * Version 2 or later at the following locations: * * http://www.opensource.org/licenses/gpl-license.html * http://www.gnu.org/copyleft/gpl.html */ #include #include #include #include "hardware.h" /* * ==================== low level suspend ==================== * * Better to follow below rules to use ARM registers: * r0: pm_info structure address; * r1 ~ r4: for saving pm_info members; * r5 ~ r10: free registers; * r11: io base address. * * suspend ocram space layout: * ======================== high address ====================== * . * . * . * ^ * ^ * ^ * imx6_suspend code * PM_INFO structure(imx6_cpu_pm_info) * ======================== low address ======================= */ /* * Below offsets are based on struct imx6_cpu_pm_info * which defined in arch/arm/mach-imx/pm-imx6q.c, this * structure contains necessary pm info for low level * suspend related code. */ #define PM_INFO_PBASE_OFFSET 0x0 #define PM_INFO_RESUME_ADDR_OFFSET 0x4 #define PM_INFO_DDR_TYPE_OFFSET 0x8 #define PM_INFO_PM_INFO_SIZE_OFFSET 0xC #define PM_INFO_MX6Q_MMDC_P_OFFSET 0x10 #define PM_INFO_MX6Q_MMDC_V_OFFSET 0x14 #define PM_INFO_MX6Q_SRC_P_OFFSET 0x18 #define PM_INFO_MX6Q_SRC_V_OFFSET 0x1C #define PM_INFO_MX6Q_IOMUXC_P_OFFSET 0x20 #define PM_INFO_MX6Q_IOMUXC_V_OFFSET 0x24 #define PM_INFO_MX6Q_CCM_P_OFFSET 0x28 #define PM_INFO_MX6Q_CCM_V_OFFSET 0x2C #define PM_INFO_MX6Q_GPC_P_OFFSET 0x30 #define PM_INFO_MX6Q_GPC_V_OFFSET 0x34 #define PM_INFO_MX6Q_L2_P_OFFSET 0x38 #define PM_INFO_MX6Q_L2_V_OFFSET 0x3C #define PM_INFO_MX6Q_ANATOP_P_OFFSET 0x40 #define PM_INFO_MX6Q_ANATOP_V_OFFSET 0x44 #define PM_INFO_MX6Q_TTBR1_V_OFFSET 0x48 #define PM_INFO_MMDC_IO_NUM_OFFSET 0x4c #define PM_INFO_MMDC_IO_VAL_OFFSET 0x50 /* below offsets depends on MX6_MAX_MMDC_IO_NUM(33) definition */ #define PM_INFO_MMDC_NUM_OFFSET 0x158 #define PM_INFO_MMDC_VAL_OFFSET 0x15c #define MX6Q_SRC_GPR1 0x20 #define MX6Q_SRC_GPR2 0x24 #define MX6Q_MMDC_MAPSR 0x404 #define MX6Q_MMDC_MPDGCTRL0 0x83c #define MX6Q_GPC_IMR1 0x08 #define MX6Q_GPC_IMR2 0x0c #define MX6Q_GPC_IMR3 0x10 #define MX6Q_GPC_IMR4 0x14 #define MX6Q_CCM_CCR 0x0 #define MX6Q_ANATOP_CORE 0x140 #define MX6_MAX_MMDC_IO_NUM 33 #define MX6_MAX_MMDC_NUM 34 .align 3 .macro sync_l2_cache /* sync L2 cache to drain L2's buffers to DRAM. */ #ifdef CONFIG_CACHE_L2X0 ldr r11, [r0, #PM_INFO_MX6Q_L2_V_OFFSET] mov r6, #0x0 str r6, [r11, #L2X0_CACHE_SYNC] 1: ldr r6, [r11, #L2X0_CACHE_SYNC] ands r6, r6, #0x1 bne 1b #endif .endm /* r11 must be MMDC base address */ .macro reset_read_fifo /* reset read FIFO, RST_RD_FIFO */ ldr r7, =MX6Q_MMDC_MPDGCTRL0 ldr r6, [r11, r7] orr r6, r6, #(1 << 31) str r6, [r11, r7] 2: ldr r6, [r11, r7] ands r6, r6, #(1 << 31) bne 2b /* reset FIFO a second time */ ldr r6, [r11, r7] orr r6, r6, #(1 << 31) str r6, [r11, r7] 3: ldr r6, [r11, r7] ands r6, r6, #(1 << 31) bne 3b .endm /* r11 must be MMDC base address */ .macro mmdc_out_and_auto_self_refresh /* let DDR out of self-refresh */ ldr r7, [r11, #MX6Q_MMDC_MAPSR] bic r7, r7, #(1 << 21) str r7, [r11, #MX6Q_MMDC_MAPSR] 4: ldr r7, [r11, #MX6Q_MMDC_MAPSR] ands r7, r7, #(1 << 25) bne 4b /* enable DDR auto power saving */ ldr r7, [r11, #MX6Q_MMDC_MAPSR] bic r7, r7, #0x1 str r7, [r11, #MX6Q_MMDC_MAPSR] .endm /* r10 must be iomuxc base address */ .macro resume_iomuxc_gpr add r10, r10, #0x4000 /* IOMUXC GPR DRAM_RESET_BYPASS */ ldr r4, [r10, #0x8] bic r4, r4, #(0x1 << 27) str r4, [r10, #0x8] /* IOMUXC GPR DRAM_CKE_BYPASS */ ldr r4, [r10, #0x8] bic r4, r4, #(0x1 << 31) str r4, [r10, #0x8] .endm .macro resume_io /* restore MMDC IO */ cmp r5, #0x0 ldreq r10, [r0, #PM_INFO_MX6Q_IOMUXC_V_OFFSET] ldrne r10, [r0, #PM_INFO_MX6Q_IOMUXC_P_OFFSET] ldr r6, [r0, #PM_INFO_MMDC_IO_NUM_OFFSET] ldr r7, =PM_INFO_MMDC_IO_VAL_OFFSET add r7, r7, r0 5: ldr r8, [r7], #0x4 ldr r9, [r7], #0x4 str r9, [r10, r8] subs r6, r6, #0x1 bne 5b cmp r5, #0x0 /* Here only MMDC0 is set */ ldreq r11, [r0, #PM_INFO_MX6Q_MMDC_V_OFFSET] ldrne r11, [r0, #PM_INFO_MX6Q_MMDC_P_OFFSET] cmp r3, #IMX_DDR_TYPE_LPDDR2 bne 6f reset_read_fifo 6: mmdc_out_and_auto_self_refresh .endm .macro resume_mmdc_io cmp r5, #0x0 ldreq r10, [r0, #PM_INFO_MX6Q_IOMUXC_V_OFFSET] ldrne r10, [r0, #PM_INFO_MX6Q_IOMUXC_P_OFFSET] ldreq r11, [r0, #PM_INFO_MX6Q_MMDC_V_OFFSET] ldrne r11, [r0, #PM_INFO_MX6Q_MMDC_P_OFFSET] /* resume mmdc iomuxc settings */ ldr r6, [r0, #PM_INFO_MMDC_IO_NUM_OFFSET] ldr r7, =PM_INFO_MMDC_IO_VAL_OFFSET add r7, r7, r0 7: ldr r8, [r7], #0x4 ldr r9, [r7], #0x4 str r9, [r10, r8] subs r6, r6, #0x1 bne 7b /* check whether we need to restore MMDC */ cmp r5, #0x0 beq 8f /* check whether last suspend is with M/F mix off */ ldr r9, [r0, #PM_INFO_MX6Q_GPC_P_OFFSET] ldr r6, [r9, #0x220] cmp r6, #0x0 bne 9f 8: resume_iomuxc_gpr b 13f 9: /* restore MMDC settings */ ldr r6, [r0, #PM_INFO_MMDC_NUM_OFFSET] ldr r7, =PM_INFO_MMDC_VAL_OFFSET add r7, r7, r0 10: ldr r8, [r7], #0x4 ldr r9, [r7], #0x4 str r9, [r11, r8] subs r6, r6, #0x1 bne 10b /* let DDR enter self-refresh */ ldr r7, [r11, #MX6Q_MMDC_MAPSR] orr r7, r7, #(1 << 20) str r7, [r11, #MX6Q_MMDC_MAPSR] 11: ldr r7, [r11, #MX6Q_MMDC_MAPSR] ands r7, r7, #(1 << 24) beq 11b resume_iomuxc_gpr reset_read_fifo /* let DDR out of self-refresh */ ldr r7, [r11, #MX6Q_MMDC_MAPSR] bic r7, r7, #(1 << 20) str r7, [r11, #MX6Q_MMDC_MAPSR] 12: ldr r7, [r11, #MX6Q_MMDC_MAPSR] ands r7, r7, #(1 << 24) bne 12b /* kick off MMDC */ ldr r4, =0x0 str r4, [r11, #0x1c] 13: mmdc_out_and_auto_self_refresh .endm .macro store_ttbr1 /* Store TTBR1 to pm_info->ttbr1 */ mrc p15, 0, r7, c2, c0, 1 str r7, [r0, #PM_INFO_MX6Q_TTBR1_V_OFFSET] /* Disable Branch Prediction, Z bit in SCTLR. */ mrc p15, 0, r6, c1, c0, 0 bic r6, r6, #0x800 mcr p15, 0, r6, c1, c0, 0 /* Flush the BTAC. */ ldr r6, =0x0 mcr p15, 0, r6, c7, c1, 6 ldr r6, =iram_tlb_phys_addr ldr r6, [r6] dsb isb /* Store the IRAM table in TTBR1 */ mcr p15, 0, r6, c2, c0, 1 /* Read TTBCR and set PD0=1, N = 1 */ mrc p15, 0, r6, c2, c0, 2 orr r6, r6, #0x11 mcr p15, 0, r6, c2, c0, 2 dsb isb /* flush the TLB */ ldr r6, =0x0 mcr p15, 0, r6, c8, c3, 0 /* Disable L1 data cache. */ mrc p15, 0, r6, c1, c0, 0 bic r6, r6, #0x4 mcr p15, 0, r6, c1, c0, 0 dsb isb #ifdef CONFIG_CACHE_L2X0 ldr r8, [r0, #PM_INFO_MX6Q_L2_V_OFFSET] mov r6, #0x0 str r6, [r8, #0x100] dsb isb #endif .endm .macro restore_ttbr1 #ifdef CONFIG_CACHE_L2X0 /* Enable L2. */ ldr r8, [r0, #PM_INFO_MX6Q_L2_V_OFFSET] ldr r7, =0x1 str r7, [r8, #0x100] #endif /* Enable L1 data cache. */ mrc p15, 0, r6, c1, c0, 0 orr r6, r6, #0x4 mcr p15, 0, r6, c1, c0, 0 dsb isb /* Restore TTBCR */ /* Read TTBCR and set PD0=0, N = 0 */ mrc p15, 0, r6, c2, c0, 2 bic r6, r6, #0x11 mcr p15, 0, r6, c2, c0, 2 dsb isb /* flush the TLB */ ldr r6, =0x0 mcr p15, 0, r6, c8, c3, 0 /* Enable Branch Prediction, Z bit in SCTLR. */ mrc p15, 0, r6, c1, c0, 0 orr r6, r6, #0x800 mcr p15, 0, r6, c1, c0, 0 /* Flush the Branch Target Address Cache (BTAC) */ ldr r6, =0x0 mcr p15, 0, r6, c7, c1, 6 /* Restore TTBR1, get the origin ttbr1 from pm info */ ldr r7, [r0, #PM_INFO_MX6Q_TTBR1_V_OFFSET] mcr p15, 0, r7, c2, c0, 1 .endm ENTRY(imx6_suspend) push {r4-r12} /* * The value of r0 is mapped the same in origin table and IRAM table, * thus no need to care r0 here. */ ldr r1, [r0, #PM_INFO_PBASE_OFFSET] ldr r2, [r0, #PM_INFO_RESUME_ADDR_OFFSET] ldr r3, [r0, #PM_INFO_DDR_TYPE_OFFSET] ldr r4, [r0, #PM_INFO_PM_INFO_SIZE_OFFSET] /* * counting the resume address in iram * to set it in SRC register. */ ldr r6, =imx6_suspend ldr r7, =resume sub r7, r7, r6 add r8, r1, r4 add r9, r8, r7 /* * make sure TLB contain the addr we want, * as we will access them after MMDC IO floated. */ ldr r11, [r0, #PM_INFO_MX6Q_CCM_V_OFFSET] ldr r6, [r11, #0x0] ldr r11, [r0, #PM_INFO_MX6Q_GPC_V_OFFSET] ldr r6, [r11, #0x0] ldr r11, [r0, #PM_INFO_MX6Q_SRC_V_OFFSET] /* store physical resume addr and pm_info address. */ str r9, [r11, #MX6Q_SRC_GPR1] str r1, [r11, #MX6Q_SRC_GPR2] /* need to sync L2 cache before DSM. */ sync_l2_cache store_ttbr1 ldr r11, [r0, #PM_INFO_MX6Q_MMDC_V_OFFSET] /* * put DDR explicitly into self-refresh and * disable automatic power savings. */ ldr r7, [r11, #MX6Q_MMDC_MAPSR] orr r7, r7, #0x1 str r7, [r11, #MX6Q_MMDC_MAPSR] /* make the DDR explicitly enter self-refresh. */ ldr r7, [r11, #MX6Q_MMDC_MAPSR] orr r7, r7, #(1 << 21) str r7, [r11, #MX6Q_MMDC_MAPSR] poll_dvfs_set: ldr r7, [r11, #MX6Q_MMDC_MAPSR] ands r7, r7, #(1 << 25) beq poll_dvfs_set /* use r11 to store the IO address */ ldr r11, [r0, #PM_INFO_MX6Q_IOMUXC_V_OFFSET] ldr r6, =0x0 ldr r7, [r0, #PM_INFO_MMDC_IO_NUM_OFFSET] ldr r8, =PM_INFO_MMDC_IO_VAL_OFFSET add r8, r8, r0 /* LPDDR2's last 3 IOs need special setting */ cmp r3, #IMX_DDR_TYPE_LPDDR2 subeq r7, r7, #0x3 set_mmdc_io_lpm: ldr r9, [r8], #0x8 str r6, [r11, r9] subs r7, r7, #0x1 bne set_mmdc_io_lpm cmp r3, #IMX_DDR_TYPE_LPDDR2 bne set_mmdc_io_lpm_done ldr r6, =0x1000 ldr r9, [r8], #0x8 str r6, [r11, r9] ldr r9, [r8], #0x8 str r6, [r11, r9] ldr r6, =0x80000 ldr r9, [r8] str r6, [r11, r9] set_mmdc_io_lpm_done: /* check whether it supports Mega/Fast off */ ldr r6, [r0, #PM_INFO_MMDC_NUM_OFFSET] cmp r6, #0x0 beq set_mmdc_lpm_done /* IOMUXC GPR DRAM_RESET */ add r11, r11, #0x4000 ldr r6, [r11, #0x8] orr r6, r6, #(0x1 << 28) str r6, [r11, #0x8] /* IOMUXC GPR DRAM_RESET_BYPASS */ ldr r6, [r11, #0x8] orr r6, r6, #(0x1 << 27) str r6, [r11, #0x8] /* IOMUXC GPR DRAM_CKE_BYPASS */ ldr r6, [r11, #0x8] orr r6, r6, #(0x1 << 31) str r6, [r11, #0x8] set_mmdc_lpm_done: /* * mask all GPC interrupts before * enabling the RBC counters to * avoid the counter starting too * early if an interupt is already * pending. */ ldr r11, [r0, #PM_INFO_MX6Q_GPC_V_OFFSET] ldr r6, [r11, #MX6Q_GPC_IMR1] ldr r7, [r11, #MX6Q_GPC_IMR2] ldr r8, [r11, #MX6Q_GPC_IMR3] ldr r9, [r11, #MX6Q_GPC_IMR4] ldr r10, =0xffffffff str r10, [r11, #MX6Q_GPC_IMR1] str r10, [r11, #MX6Q_GPC_IMR2] str r10, [r11, #MX6Q_GPC_IMR3] str r10, [r11, #MX6Q_GPC_IMR4] /* * enable the RBC bypass counter here * to hold off the interrupts. RBC counter * = 32 (1ms), Minimum RBC delay should be * 400us for the analog LDOs to power down. */ ldr r11, [r0, #PM_INFO_MX6Q_CCM_V_OFFSET] ldr r10, [r11, #MX6Q_CCM_CCR] bic r10, r10, #(0x3f << 21) orr r10, r10, #(0x20 << 21) str r10, [r11, #MX6Q_CCM_CCR] /* enable the counter. */ ldr r10, [r11, #MX6Q_CCM_CCR] orr r10, r10, #(0x1 << 27) str r10, [r11, #MX6Q_CCM_CCR] /* unmask all the GPC interrupts. */ ldr r11, [r0, #PM_INFO_MX6Q_GPC_V_OFFSET] str r6, [r11, #MX6Q_GPC_IMR1] str r7, [r11, #MX6Q_GPC_IMR2] str r8, [r11, #MX6Q_GPC_IMR3] str r9, [r11, #MX6Q_GPC_IMR4] /* * now delay for a short while (3usec) * ARM is at 1GHz at this point * so a short loop should be enough. * this delay is required to ensure that * the RBC counter can start counting in * case an interrupt is already pending * or in case an interrupt arrives just * as ARM is about to assert DSM_request. */ ldr r6, =2000 rbc_loop: subs r6, r6, #0x1 bne rbc_loop /* * ERR005852 Analog: Transition from Deep Sleep Mode to * LDO Bypass Mode may cause the slow response of the * VDDARM_CAP output. * * Software workaround: * if internal ldo(VDDARM) bypassed, switch to analog bypass * mode (0x1E), prio to entering DSM, and then, revert to the * normal bypass mode, when exiting from DSM. */ ldr r11, [r0, #PM_INFO_MX6Q_ANATOP_V_OFFSET] ldr r10, [r11, #MX6Q_ANATOP_CORE] and r10, r10, #0x1f cmp r10, #0x1f bne ldo_check_done1 ldo_analog_bypass: ldr r10, [r11, #MX6Q_ANATOP_CORE] bic r10, r10, #0x1f orr r10, r10, #0x1e str r10, [r11, #MX6Q_ANATOP_CORE] ldo_check_done1: /* Zzz, enter stop mode */ wfi nop nop nop nop /* * run to here means there is pending * wakeup source, system should auto * resume, we need to restore MMDC IO first */ /* restore it with 0x1f if use ldo bypass mode.*/ ldr r10, [r11, #MX6Q_ANATOP_CORE] and r10, r10, #0x1f cmp r10, #0x1e bne ldo_check_done2 ldo_bypass_restore: ldr r10, [r11, #MX6Q_ANATOP_CORE] orr r10, r10, #0x1f str r10, [r11, #MX6Q_ANATOP_CORE] ldo_check_done2: mov r5, #0x0 /* check whether it supports Mega/Fast off */ ldr r6, [r0, #PM_INFO_MMDC_NUM_OFFSET] cmp r6, #0x0 beq only_resume_io resume_mmdc_io b resume_mmdc_done only_resume_io: resume_io resume_mmdc_done: restore_ttbr1 pop {r4-r12} /* return to suspend finish */ mov pc, lr resume: /* invalidate L1 I-cache first */ mov r6, #0x0 mcr p15, 0, r6, c7, c5, 0 mcr p15, 0, r6, c7, c5, 6 /* enable the Icache and branch prediction */ mov r6, #0x1800 mcr p15, 0, r6, c1, c0, 0 isb /* restore it with 0x1f if use ldo bypass mode.*/ ldr r11, [r0, #PM_INFO_MX6Q_ANATOP_P_OFFSET] ldr r7, [r11, #MX6Q_ANATOP_CORE] and r7, r7, #0x1f cmp r7, #0x1e bne ldo_check_done3 ldr r7, [r11, #MX6Q_ANATOP_CORE] orr r7, r7, #0x1f str r7, [r11, #MX6Q_ANATOP_CORE] ldo_check_done3: /* get physical resume address from pm_info. */ ldr lr, [r0, #PM_INFO_RESUME_ADDR_OFFSET] /* clear core0's entry and parameter */ ldr r11, [r0, #PM_INFO_MX6Q_SRC_P_OFFSET] mov r7, #0x0 str r7, [r11, #MX6Q_SRC_GPR1] str r7, [r11, #MX6Q_SRC_GPR2] mov r5, #0x1 /* check whether it supports Mega/Fast off */ ldr r6, [r0, #PM_INFO_MMDC_NUM_OFFSET] cmp r6, #0x0 beq dsm_only_resume_io resume_mmdc_io b dsm_resume_mmdc_done dsm_only_resume_io: ldr r3, [r0, #PM_INFO_DDR_TYPE_OFFSET] resume_io dsm_resume_mmdc_done: mov pc, lr ENDPROC(imx6_suspend) /* * The following code must assume it is running from physical address * where absolute virtual addresses to the data section have to be * turned into relative ones. */ #ifdef CONFIG_CACHE_L2X0 .macro pl310_resume adr r0, l2x0_saved_regs_offset ldr r2, [r0] add r2, r2, r0 ldr r0, [r2, #L2X0_R_PHY_BASE] @ get physical base of l2x0 ldr r1, [r2, #L2X0_R_AUX_CTRL] @ get aux_ctrl value str r1, [r0, #L2X0_AUX_CTRL] @ restore aux_ctrl mov r1, #0x1 str r1, [r0, #L2X0_CTRL] @ re-enable L2 .endm l2x0_saved_regs_offset: .word l2x0_saved_regs - . #else .macro pl310_resume .endm #endif ENTRY(v7_cpu_resume) bl v7_invalidate_l1 pl310_resume b cpu_resume ENDPROC(v7_cpu_resume)