300 lines
6.4 KiB
C
300 lines
6.4 KiB
C
/*
|
|
* Blackfin power management
|
|
*
|
|
* Copyright 2006-2009 Analog Devices Inc.
|
|
*
|
|
* Licensed under the GPL-2
|
|
* based on arm/mach-omap/pm.c
|
|
* Copyright 2001, Cliff Brake <cbrake@accelent.com> and others
|
|
*/
|
|
|
|
#include <linux/suspend.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/proc_fs.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/io.h>
|
|
#include <linux/irq.h>
|
|
|
|
#include <asm/cplb.h>
|
|
#include <asm/gpio.h>
|
|
#include <asm/dma.h>
|
|
#include <asm/dpmc.h>
|
|
#include <asm/pm.h>
|
|
|
|
#ifdef CONFIG_BF60x
|
|
struct bfin_cpu_pm_fns *bfin_cpu_pm;
|
|
#endif
|
|
|
|
void bfin_pm_suspend_standby_enter(void)
|
|
{
|
|
#if !BFIN_GPIO_PINT
|
|
bfin_pm_standby_setup();
|
|
#endif
|
|
|
|
#ifdef CONFIG_BF60x
|
|
bfin_cpu_pm->enter(PM_SUSPEND_STANDBY);
|
|
#else
|
|
# ifdef CONFIG_PM_BFIN_SLEEP_DEEPER
|
|
sleep_deeper(bfin_sic_iwr[0], bfin_sic_iwr[1], bfin_sic_iwr[2]);
|
|
# else
|
|
sleep_mode(bfin_sic_iwr[0], bfin_sic_iwr[1], bfin_sic_iwr[2]);
|
|
# endif
|
|
#endif
|
|
|
|
#if !BFIN_GPIO_PINT
|
|
bfin_pm_standby_restore();
|
|
#endif
|
|
|
|
#ifndef CONFIG_BF60x
|
|
#ifdef SIC_IWR0
|
|
bfin_write_SIC_IWR0(IWR_DISABLE_ALL);
|
|
# ifdef SIC_IWR1
|
|
/* BF52x system reset does not properly reset SIC_IWR1 which
|
|
* will screw up the bootrom as it relies on MDMA0/1 waking it
|
|
* up from IDLE instructions. See this report for more info:
|
|
* http://blackfin.uclinux.org/gf/tracker/4323
|
|
*/
|
|
if (ANOMALY_05000435)
|
|
bfin_write_SIC_IWR1(IWR_ENABLE(10) | IWR_ENABLE(11));
|
|
else
|
|
bfin_write_SIC_IWR1(IWR_DISABLE_ALL);
|
|
# endif
|
|
# ifdef SIC_IWR2
|
|
bfin_write_SIC_IWR2(IWR_DISABLE_ALL);
|
|
# endif
|
|
#else
|
|
bfin_write_SIC_IWR(IWR_DISABLE_ALL);
|
|
#endif
|
|
|
|
#endif
|
|
}
|
|
|
|
int bf53x_suspend_l1_mem(unsigned char *memptr)
|
|
{
|
|
dma_memcpy_nocache(memptr, (const void *) L1_CODE_START,
|
|
L1_CODE_LENGTH);
|
|
dma_memcpy_nocache(memptr + L1_CODE_LENGTH,
|
|
(const void *) L1_DATA_A_START, L1_DATA_A_LENGTH);
|
|
dma_memcpy_nocache(memptr + L1_CODE_LENGTH + L1_DATA_A_LENGTH,
|
|
(const void *) L1_DATA_B_START, L1_DATA_B_LENGTH);
|
|
memcpy(memptr + L1_CODE_LENGTH + L1_DATA_A_LENGTH +
|
|
L1_DATA_B_LENGTH, (const void *) L1_SCRATCH_START,
|
|
L1_SCRATCH_LENGTH);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int bf53x_resume_l1_mem(unsigned char *memptr)
|
|
{
|
|
dma_memcpy_nocache((void *) L1_CODE_START, memptr, L1_CODE_LENGTH);
|
|
dma_memcpy_nocache((void *) L1_DATA_A_START, memptr + L1_CODE_LENGTH,
|
|
L1_DATA_A_LENGTH);
|
|
dma_memcpy_nocache((void *) L1_DATA_B_START, memptr + L1_CODE_LENGTH +
|
|
L1_DATA_A_LENGTH, L1_DATA_B_LENGTH);
|
|
memcpy((void *) L1_SCRATCH_START, memptr + L1_CODE_LENGTH +
|
|
L1_DATA_A_LENGTH + L1_DATA_B_LENGTH, L1_SCRATCH_LENGTH);
|
|
|
|
return 0;
|
|
}
|
|
|
|
#if defined(CONFIG_BFIN_EXTMEM_WRITEBACK) || defined(CONFIG_BFIN_L2_WRITEBACK)
|
|
# ifdef CONFIG_BF60x
|
|
__attribute__((l1_text))
|
|
# endif
|
|
static void flushinv_all_dcache(void)
|
|
{
|
|
register u32 way, bank, subbank, set;
|
|
register u32 status, addr;
|
|
u32 dmem_ctl = bfin_read_DMEM_CONTROL();
|
|
|
|
for (bank = 0; bank < 2; ++bank) {
|
|
if (!(dmem_ctl & (1 << (DMC1_P - bank))))
|
|
continue;
|
|
|
|
for (way = 0; way < 2; ++way)
|
|
for (subbank = 0; subbank < 4; ++subbank)
|
|
for (set = 0; set < 64; ++set) {
|
|
|
|
bfin_write_DTEST_COMMAND(
|
|
way << 26 |
|
|
bank << 23 |
|
|
subbank << 16 |
|
|
set << 5
|
|
);
|
|
CSYNC();
|
|
status = bfin_read_DTEST_DATA0();
|
|
|
|
/* only worry about valid/dirty entries */
|
|
if ((status & 0x3) != 0x3)
|
|
continue;
|
|
|
|
|
|
/* construct the address using the tag */
|
|
addr = (status & 0xFFFFC800) | (subbank << 12) | (set << 5);
|
|
|
|
/* flush it */
|
|
__asm__ __volatile__("FLUSHINV[%0];" : : "a"(addr));
|
|
}
|
|
}
|
|
}
|
|
#endif
|
|
|
|
int bfin_pm_suspend_mem_enter(void)
|
|
{
|
|
int ret;
|
|
#ifndef CONFIG_BF60x
|
|
int wakeup;
|
|
#endif
|
|
|
|
unsigned char *memptr = kmalloc(L1_CODE_LENGTH + L1_DATA_A_LENGTH
|
|
+ L1_DATA_B_LENGTH + L1_SCRATCH_LENGTH,
|
|
GFP_ATOMIC);
|
|
|
|
if (memptr == NULL) {
|
|
panic("bf53x_suspend_l1_mem malloc failed");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
#ifndef CONFIG_BF60x
|
|
wakeup = bfin_read_VR_CTL() & ~FREQ;
|
|
wakeup |= SCKELOW;
|
|
|
|
#ifdef CONFIG_PM_BFIN_WAKE_PH6
|
|
wakeup |= PHYWE;
|
|
#endif
|
|
#ifdef CONFIG_PM_BFIN_WAKE_GP
|
|
wakeup |= GPWE;
|
|
#endif
|
|
#endif
|
|
|
|
ret = blackfin_dma_suspend();
|
|
|
|
if (ret) {
|
|
kfree(memptr);
|
|
return ret;
|
|
}
|
|
|
|
#ifdef CONFIG_GPIO_ADI
|
|
bfin_gpio_pm_hibernate_suspend();
|
|
#endif
|
|
|
|
#if defined(CONFIG_BFIN_EXTMEM_WRITEBACK) || defined(CONFIG_BFIN_L2_WRITEBACK)
|
|
flushinv_all_dcache();
|
|
#endif
|
|
_disable_dcplb();
|
|
_disable_icplb();
|
|
bf53x_suspend_l1_mem(memptr);
|
|
|
|
#ifndef CONFIG_BF60x
|
|
do_hibernate(wakeup | vr_wakeup); /* See you later! */
|
|
#else
|
|
bfin_cpu_pm->enter(PM_SUSPEND_MEM);
|
|
#endif
|
|
|
|
bf53x_resume_l1_mem(memptr);
|
|
|
|
_enable_icplb();
|
|
_enable_dcplb();
|
|
|
|
#ifdef CONFIG_GPIO_ADI
|
|
bfin_gpio_pm_hibernate_restore();
|
|
#endif
|
|
blackfin_dma_resume();
|
|
|
|
kfree(memptr);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* bfin_pm_valid - Tell the PM core that we only support the standby sleep
|
|
* state
|
|
* @state: suspend state we're checking.
|
|
*
|
|
*/
|
|
static int bfin_pm_valid(suspend_state_t state)
|
|
{
|
|
return (state == PM_SUSPEND_STANDBY
|
|
#if !(defined(BF533_FAMILY) || defined(CONFIG_BF561))
|
|
/*
|
|
* On BF533/2/1:
|
|
* If we enter Hibernate the SCKE Pin is driven Low,
|
|
* so that the SDRAM enters Self Refresh Mode.
|
|
* However when the reset sequence that follows hibernate
|
|
* state is executed, SCKE is driven High, taking the
|
|
* SDRAM out of Self Refresh.
|
|
*
|
|
* If you reconfigure and access the SDRAM "very quickly",
|
|
* you are likely to avoid errors, otherwise the SDRAM
|
|
* start losing its contents.
|
|
* An external HW workaround is possible using logic gates.
|
|
*/
|
|
|| state == PM_SUSPEND_MEM
|
|
#endif
|
|
);
|
|
}
|
|
|
|
/*
|
|
* bfin_pm_enter - Actually enter a sleep state.
|
|
* @state: State we're entering.
|
|
*
|
|
*/
|
|
static int bfin_pm_enter(suspend_state_t state)
|
|
{
|
|
switch (state) {
|
|
case PM_SUSPEND_STANDBY:
|
|
bfin_pm_suspend_standby_enter();
|
|
break;
|
|
case PM_SUSPEND_MEM:
|
|
bfin_pm_suspend_mem_enter();
|
|
break;
|
|
default:
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
#ifdef CONFIG_BFIN_PM_WAKEUP_TIME_BENCH
|
|
void bfin_pm_end(void)
|
|
{
|
|
u32 cycle, cycle2;
|
|
u64 usec64;
|
|
u32 usec;
|
|
|
|
__asm__ __volatile__ (
|
|
"1: %0 = CYCLES2\n"
|
|
"%1 = CYCLES\n"
|
|
"%2 = CYCLES2\n"
|
|
"CC = %2 == %0\n"
|
|
"if ! CC jump 1b\n"
|
|
: "=d,a" (cycle2), "=d,a" (cycle), "=d,a" (usec) : : "CC"
|
|
);
|
|
|
|
usec64 = ((u64)cycle2 << 32) + cycle;
|
|
do_div(usec64, get_cclk() / USEC_PER_SEC);
|
|
usec = usec64;
|
|
if (usec == 0)
|
|
usec = 1;
|
|
|
|
pr_info("PM: resume of kernel completes after %ld msec %03ld usec\n",
|
|
usec / USEC_PER_MSEC, usec % USEC_PER_MSEC);
|
|
}
|
|
#endif
|
|
|
|
static const struct platform_suspend_ops bfin_pm_ops = {
|
|
.enter = bfin_pm_enter,
|
|
.valid = bfin_pm_valid,
|
|
#ifdef CONFIG_BFIN_PM_WAKEUP_TIME_BENCH
|
|
.end = bfin_pm_end,
|
|
#endif
|
|
};
|
|
|
|
static int __init bfin_pm_init(void)
|
|
{
|
|
suspend_set_ops(&bfin_pm_ops);
|
|
return 0;
|
|
}
|
|
|
|
__initcall(bfin_pm_init);
|