mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-28 11:18:45 +07:00
ccf4975dca
<generated/ti-pm-asm-offsets.h> is only generated and included by arch/arm/mach-omap2/, so it does not need to reside in the globally visible include/generated/. I renamed it to arch/arm/mach-omap2/pm-asm-offsets.h since the prefix 'ti-' is just redundant in mach-omap2/. My main motivation of this change is to avoid the race condition for the parallel build (-j) when CONFIG_IKHEADERS is enabled. When it is enabled, all the headers under include/ are archived into kernel/kheaders_data.tar.xz and exposed in the sysfs. In the parallel build, we have no idea in which order files are built. - If ti-pm-asm-offsets.h is built before kheaders_data.tar.xz, the header will be included in the archive. Probably nobody will use it, but it is harmless except that it will increase the archive size needlessly. - If kheaders_data.tar.xz is built before ti-pm-asm-offsets.h, the header will not be included in the archive. However, in the next build, the archive will be re-generated to include the newly-found ti-pm-asm-offsets.h. This is not nice from the build system point of view. - If ti-pm-asm-offsets.h and kheaders_data.tar.xz are built at the same time, the corrupted header might be included in the archive, which does not look nice either. This commit fixes the race. Signed-off-by: Masahiro Yamada <yamada.masahiro@socionext.com> Tested-by: Keerthy <j-keerthy@ti.com> Signed-off-by: Tony Lindgren <tony@atomide.com>
263 lines
5.5 KiB
ArmAsm
263 lines
5.5 KiB
ArmAsm
/* SPDX-License-Identifier: GPL-2.0 */
|
|
/*
|
|
* Low level suspend code for AM33XX SoCs
|
|
*
|
|
* Copyright (C) 2012-2018 Texas Instruments Incorporated - http://www.ti.com/
|
|
* Dave Gerlach, Vaibhav Bedia
|
|
*/
|
|
|
|
#include <linux/linkage.h>
|
|
#include <linux/platform_data/pm33xx.h>
|
|
#include <linux/ti-emif-sram.h>
|
|
#include <asm/assembler.h>
|
|
#include <asm/memory.h>
|
|
|
|
#include "iomap.h"
|
|
#include "cm33xx.h"
|
|
#include "pm-asm-offsets.h"
|
|
|
|
#define AM33XX_CM_CLKCTRL_MODULESTATE_DISABLED 0x00030000
|
|
#define AM33XX_CM_CLKCTRL_MODULEMODE_DISABLE 0x0003
|
|
#define AM33XX_CM_CLKCTRL_MODULEMODE_ENABLE 0x0002
|
|
|
|
/* replicated define because linux/bitops.h cannot be included in assembly */
|
|
#define BIT(nr) (1 << (nr))
|
|
|
|
.arm
|
|
.arch armv7-a
|
|
.align 3
|
|
|
|
ENTRY(am33xx_do_wfi)
|
|
stmfd sp!, {r4 - r11, lr} @ save registers on stack
|
|
|
|
/* Save wfi_flags arg to data space */
|
|
mov r4, r0
|
|
adr r3, am33xx_pm_ro_sram_data
|
|
ldr r2, [r3, #AMX3_PM_RO_SRAM_DATA_VIRT_OFFSET]
|
|
str r4, [r2, #AMX3_PM_WFI_FLAGS_OFFSET]
|
|
|
|
/* Only flush cache is we know we are losing MPU context */
|
|
tst r4, #WFI_FLAG_FLUSH_CACHE
|
|
beq cache_skip_flush
|
|
|
|
/*
|
|
* Flush all data from the L1 and L2 data cache before disabling
|
|
* SCTLR.C bit.
|
|
*/
|
|
ldr r1, kernel_flush
|
|
blx r1
|
|
|
|
/*
|
|
* Clear the SCTLR.C bit to prevent further data cache
|
|
* allocation. Clearing SCTLR.C would make all the data accesses
|
|
* strongly ordered and would not hit the cache.
|
|
*/
|
|
mrc p15, 0, r0, c1, c0, 0
|
|
bic r0, r0, #(1 << 2) @ Disable the C bit
|
|
mcr p15, 0, r0, c1, c0, 0
|
|
isb
|
|
|
|
/*
|
|
* Invalidate L1 and L2 data cache.
|
|
*/
|
|
ldr r1, kernel_flush
|
|
blx r1
|
|
|
|
adr r3, am33xx_pm_ro_sram_data
|
|
ldr r2, [r3, #AMX3_PM_RO_SRAM_DATA_VIRT_OFFSET]
|
|
ldr r4, [r2, #AMX3_PM_WFI_FLAGS_OFFSET]
|
|
|
|
cache_skip_flush:
|
|
/* Check if we want self refresh */
|
|
tst r4, #WFI_FLAG_SELF_REFRESH
|
|
beq emif_skip_enter_sr
|
|
|
|
adr r9, am33xx_emif_sram_table
|
|
|
|
ldr r3, [r9, #EMIF_PM_ENTER_SR_OFFSET]
|
|
blx r3
|
|
|
|
emif_skip_enter_sr:
|
|
/* Only necessary if PER is losing context */
|
|
tst r4, #WFI_FLAG_SAVE_EMIF
|
|
beq emif_skip_save
|
|
|
|
ldr r3, [r9, #EMIF_PM_SAVE_CONTEXT_OFFSET]
|
|
blx r3
|
|
|
|
emif_skip_save:
|
|
/* Only can disable EMIF if we have entered self refresh */
|
|
tst r4, #WFI_FLAG_SELF_REFRESH
|
|
beq emif_skip_disable
|
|
|
|
/* Disable EMIF */
|
|
ldr r1, virt_emif_clkctrl
|
|
ldr r2, [r1]
|
|
bic r2, r2, #AM33XX_CM_CLKCTRL_MODULEMODE_DISABLE
|
|
str r2, [r1]
|
|
|
|
ldr r1, virt_emif_clkctrl
|
|
wait_emif_disable:
|
|
ldr r2, [r1]
|
|
mov r3, #AM33XX_CM_CLKCTRL_MODULESTATE_DISABLED
|
|
cmp r2, r3
|
|
bne wait_emif_disable
|
|
|
|
emif_skip_disable:
|
|
tst r4, #WFI_FLAG_WAKE_M3
|
|
beq wkup_m3_skip
|
|
|
|
/*
|
|
* For the MPU WFI to be registered as an interrupt
|
|
* to WKUP_M3, MPU_CLKCTRL.MODULEMODE needs to be set
|
|
* to DISABLED
|
|
*/
|
|
ldr r1, virt_mpu_clkctrl
|
|
ldr r2, [r1]
|
|
bic r2, r2, #AM33XX_CM_CLKCTRL_MODULEMODE_DISABLE
|
|
str r2, [r1]
|
|
|
|
wkup_m3_skip:
|
|
/*
|
|
* Execute an ISB instruction to ensure that all of the
|
|
* CP15 register changes have been committed.
|
|
*/
|
|
isb
|
|
|
|
/*
|
|
* Execute a barrier instruction to ensure that all cache,
|
|
* TLB and branch predictor maintenance operations issued
|
|
* have completed.
|
|
*/
|
|
dsb
|
|
dmb
|
|
|
|
/*
|
|
* Execute a WFI instruction and wait until the
|
|
* STANDBYWFI output is asserted to indicate that the
|
|
* CPU is in idle and low power state. CPU can specualatively
|
|
* prefetch the instructions so add NOPs after WFI. Thirteen
|
|
* NOPs as per Cortex-A8 pipeline.
|
|
*/
|
|
wfi
|
|
|
|
nop
|
|
nop
|
|
nop
|
|
nop
|
|
nop
|
|
nop
|
|
nop
|
|
nop
|
|
nop
|
|
nop
|
|
nop
|
|
nop
|
|
nop
|
|
|
|
/* We come here in case of an abort due to a late interrupt */
|
|
|
|
/* Set MPU_CLKCTRL.MODULEMODE back to ENABLE */
|
|
ldr r1, virt_mpu_clkctrl
|
|
mov r2, #AM33XX_CM_CLKCTRL_MODULEMODE_ENABLE
|
|
str r2, [r1]
|
|
|
|
/* Re-enable EMIF */
|
|
ldr r1, virt_emif_clkctrl
|
|
mov r2, #AM33XX_CM_CLKCTRL_MODULEMODE_ENABLE
|
|
str r2, [r1]
|
|
wait_emif_enable:
|
|
ldr r3, [r1]
|
|
cmp r2, r3
|
|
bne wait_emif_enable
|
|
|
|
/* Only necessary if PER is losing context */
|
|
tst r4, #WFI_FLAG_SELF_REFRESH
|
|
beq emif_skip_exit_sr_abt
|
|
|
|
adr r9, am33xx_emif_sram_table
|
|
ldr r1, [r9, #EMIF_PM_ABORT_SR_OFFSET]
|
|
blx r1
|
|
|
|
emif_skip_exit_sr_abt:
|
|
tst r4, #WFI_FLAG_FLUSH_CACHE
|
|
beq cache_skip_restore
|
|
|
|
/*
|
|
* Set SCTLR.C bit to allow data cache allocation
|
|
*/
|
|
mrc p15, 0, r0, c1, c0, 0
|
|
orr r0, r0, #(1 << 2) @ Enable the C bit
|
|
mcr p15, 0, r0, c1, c0, 0
|
|
isb
|
|
|
|
cache_skip_restore:
|
|
/* Let the suspend code know about the abort */
|
|
mov r0, #1
|
|
ldmfd sp!, {r4 - r11, pc} @ restore regs and return
|
|
ENDPROC(am33xx_do_wfi)
|
|
|
|
.align
|
|
ENTRY(am33xx_resume_offset)
|
|
.word . - am33xx_do_wfi
|
|
|
|
ENTRY(am33xx_resume_from_deep_sleep)
|
|
/* Re-enable EMIF */
|
|
ldr r0, phys_emif_clkctrl
|
|
mov r1, #AM33XX_CM_CLKCTRL_MODULEMODE_ENABLE
|
|
str r1, [r0]
|
|
wait_emif_enable1:
|
|
ldr r2, [r0]
|
|
cmp r1, r2
|
|
bne wait_emif_enable1
|
|
|
|
adr r9, am33xx_emif_sram_table
|
|
|
|
ldr r1, [r9, #EMIF_PM_RESTORE_CONTEXT_OFFSET]
|
|
blx r1
|
|
|
|
ldr r1, [r9, #EMIF_PM_EXIT_SR_OFFSET]
|
|
blx r1
|
|
|
|
resume_to_ddr:
|
|
/* We are back. Branch to the common CPU resume routine */
|
|
mov r0, #0
|
|
ldr pc, resume_addr
|
|
ENDPROC(am33xx_resume_from_deep_sleep)
|
|
|
|
/*
|
|
* Local variables
|
|
*/
|
|
.align
|
|
kernel_flush:
|
|
.word v7_flush_dcache_all
|
|
virt_mpu_clkctrl:
|
|
.word AM33XX_CM_MPU_MPU_CLKCTRL
|
|
virt_emif_clkctrl:
|
|
.word AM33XX_CM_PER_EMIF_CLKCTRL
|
|
phys_emif_clkctrl:
|
|
.word (AM33XX_CM_BASE + AM33XX_CM_PER_MOD + \
|
|
AM33XX_CM_PER_EMIF_CLKCTRL_OFFSET)
|
|
|
|
.align 3
|
|
/* DDR related defines */
|
|
am33xx_emif_sram_table:
|
|
.space EMIF_PM_FUNCTIONS_SIZE
|
|
|
|
ENTRY(am33xx_pm_sram)
|
|
.word am33xx_do_wfi
|
|
.word am33xx_do_wfi_sz
|
|
.word am33xx_resume_offset
|
|
.word am33xx_emif_sram_table
|
|
.word am33xx_pm_ro_sram_data
|
|
|
|
resume_addr:
|
|
.word cpu_resume - PAGE_OFFSET + 0x80000000
|
|
|
|
.align 3
|
|
ENTRY(am33xx_pm_ro_sram_data)
|
|
.space AMX3_PM_RO_SRAM_DATA_SIZE
|
|
|
|
ENTRY(am33xx_do_wfi_sz)
|
|
.word . - am33xx_do_wfi
|