1185 lines
29 KiB
C
1185 lines
29 KiB
C
// SPDX-License-Identifier: GPL-2.0+
|
|
/*
|
|
* Copyright (C) 2018 Oleksij Rempel <linux@rempel-privat.de>
|
|
*
|
|
* Driver for Alcor Micro AU6601 and AU6621 controllers
|
|
*/
|
|
|
|
/* Note: this driver was created without any documentation. Based
|
|
* on sniffing, testing and in some cases mimic of original driver.
|
|
* As soon as some one with documentation or more experience in SD/MMC, or
|
|
* reverse engineering then me, please review this driver and question every
|
|
* thing what I did. 2018 Oleksij Rempel <linux@rempel-privat.de>
|
|
*/
|
|
|
|
#include <linux/delay.h>
|
|
#include <linux/pci.h>
|
|
#include <linux/module.h>
|
|
#include <linux/io.h>
|
|
#include <linux/pm.h>
|
|
#include <linux/irq.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/platform_device.h>
|
|
|
|
#include <linux/mmc/host.h>
|
|
#include <linux/mmc/mmc.h>
|
|
|
|
#include <linux/alcor_pci.h>
|
|
|
|
enum alcor_cookie {
|
|
COOKIE_UNMAPPED,
|
|
COOKIE_PRE_MAPPED,
|
|
COOKIE_MAPPED,
|
|
};
|
|
|
|
struct alcor_pll_conf {
|
|
unsigned int clk_src_freq;
|
|
unsigned int clk_src_reg;
|
|
unsigned int min_div;
|
|
unsigned int max_div;
|
|
};
|
|
|
|
struct alcor_sdmmc_host {
|
|
struct device *dev;
|
|
struct alcor_pci_priv *alcor_pci;
|
|
|
|
struct mmc_request *mrq;
|
|
struct mmc_command *cmd;
|
|
struct mmc_data *data;
|
|
unsigned int dma_on:1;
|
|
|
|
struct mutex cmd_mutex;
|
|
|
|
struct delayed_work timeout_work;
|
|
|
|
struct sg_mapping_iter sg_miter; /* SG state for PIO */
|
|
struct scatterlist *sg;
|
|
unsigned int blocks; /* remaining PIO blocks */
|
|
int sg_count;
|
|
|
|
u32 irq_status_sd;
|
|
unsigned char cur_power_mode;
|
|
};
|
|
|
|
static const struct alcor_pll_conf alcor_pll_cfg[] = {
|
|
/* MHZ, CLK src, max div, min div */
|
|
{ 31250000, AU6601_CLK_31_25_MHZ, 1, 511},
|
|
{ 48000000, AU6601_CLK_48_MHZ, 1, 511},
|
|
{125000000, AU6601_CLK_125_MHZ, 1, 511},
|
|
{384000000, AU6601_CLK_384_MHZ, 1, 511},
|
|
};
|
|
|
|
static inline void alcor_rmw8(struct alcor_sdmmc_host *host, unsigned int addr,
|
|
u8 clear, u8 set)
|
|
{
|
|
struct alcor_pci_priv *priv = host->alcor_pci;
|
|
u32 var;
|
|
|
|
var = alcor_read8(priv, addr);
|
|
var &= ~clear;
|
|
var |= set;
|
|
alcor_write8(priv, var, addr);
|
|
}
|
|
|
|
/* As soon as irqs are masked, some status updates may be missed.
|
|
* Use this with care.
|
|
*/
|
|
static inline void alcor_mask_sd_irqs(struct alcor_sdmmc_host *host)
|
|
{
|
|
struct alcor_pci_priv *priv = host->alcor_pci;
|
|
|
|
alcor_write32(priv, 0, AU6601_REG_INT_ENABLE);
|
|
}
|
|
|
|
static inline void alcor_unmask_sd_irqs(struct alcor_sdmmc_host *host)
|
|
{
|
|
struct alcor_pci_priv *priv = host->alcor_pci;
|
|
|
|
alcor_write32(priv, AU6601_INT_CMD_MASK | AU6601_INT_DATA_MASK |
|
|
AU6601_INT_CARD_INSERT | AU6601_INT_CARD_REMOVE |
|
|
AU6601_INT_OVER_CURRENT_ERR,
|
|
AU6601_REG_INT_ENABLE);
|
|
}
|
|
|
|
static void alcor_reset(struct alcor_sdmmc_host *host, u8 val)
|
|
{
|
|
struct alcor_pci_priv *priv = host->alcor_pci;
|
|
int i;
|
|
|
|
alcor_write8(priv, val | AU6601_BUF_CTRL_RESET,
|
|
AU6601_REG_SW_RESET);
|
|
for (i = 0; i < 100; i++) {
|
|
if (!(alcor_read8(priv, AU6601_REG_SW_RESET) & val))
|
|
return;
|
|
udelay(50);
|
|
}
|
|
dev_err(host->dev, "%s: timeout\n", __func__);
|
|
}
|
|
|
|
/*
|
|
* Perform DMA I/O of a single page.
|
|
*/
|
|
static void alcor_data_set_dma(struct alcor_sdmmc_host *host)
|
|
{
|
|
struct alcor_pci_priv *priv = host->alcor_pci;
|
|
u32 addr;
|
|
|
|
if (!host->sg_count)
|
|
return;
|
|
|
|
if (!host->sg) {
|
|
dev_err(host->dev, "have blocks, but no SG\n");
|
|
return;
|
|
}
|
|
|
|
if (!sg_dma_len(host->sg)) {
|
|
dev_err(host->dev, "DMA SG len == 0\n");
|
|
return;
|
|
}
|
|
|
|
|
|
addr = (u32)sg_dma_address(host->sg);
|
|
|
|
alcor_write32(priv, addr, AU6601_REG_SDMA_ADDR);
|
|
host->sg = sg_next(host->sg);
|
|
host->sg_count--;
|
|
}
|
|
|
|
static void alcor_trigger_data_transfer(struct alcor_sdmmc_host *host)
|
|
{
|
|
struct alcor_pci_priv *priv = host->alcor_pci;
|
|
struct mmc_data *data = host->data;
|
|
u8 ctrl = 0;
|
|
|
|
if (data->flags & MMC_DATA_WRITE)
|
|
ctrl |= AU6601_DATA_WRITE;
|
|
|
|
if (data->host_cookie == COOKIE_MAPPED) {
|
|
/*
|
|
* For DMA transfers, this function is called just once,
|
|
* at the start of the operation. The hardware can only
|
|
* perform DMA I/O on a single page at a time, so here
|
|
* we kick off the transfer with the first page, and expect
|
|
* subsequent pages to be transferred upon IRQ events
|
|
* indicating that the single-page DMA was completed.
|
|
*/
|
|
alcor_data_set_dma(host);
|
|
ctrl |= AU6601_DATA_DMA_MODE;
|
|
host->dma_on = 1;
|
|
alcor_write32(priv, data->sg_count * 0x1000,
|
|
AU6601_REG_BLOCK_SIZE);
|
|
} else {
|
|
/*
|
|
* For PIO transfers, we break down each operation
|
|
* into several sector-sized transfers. When one sector has
|
|
* complete, the IRQ handler will call this function again
|
|
* to kick off the transfer of the next sector.
|
|
*/
|
|
alcor_write32(priv, data->blksz, AU6601_REG_BLOCK_SIZE);
|
|
}
|
|
|
|
alcor_write8(priv, ctrl | AU6601_DATA_START_XFER,
|
|
AU6601_DATA_XFER_CTRL);
|
|
}
|
|
|
|
static void alcor_trf_block_pio(struct alcor_sdmmc_host *host, bool read)
|
|
{
|
|
struct alcor_pci_priv *priv = host->alcor_pci;
|
|
size_t blksize, len;
|
|
u8 *buf;
|
|
|
|
if (!host->blocks)
|
|
return;
|
|
|
|
if (host->dma_on) {
|
|
dev_err(host->dev, "configured DMA but got PIO request.\n");
|
|
return;
|
|
}
|
|
|
|
if (!!(host->data->flags & MMC_DATA_READ) != read) {
|
|
dev_err(host->dev, "got unexpected direction %i != %i\n",
|
|
!!(host->data->flags & MMC_DATA_READ), read);
|
|
}
|
|
|
|
if (!sg_miter_next(&host->sg_miter))
|
|
return;
|
|
|
|
blksize = host->data->blksz;
|
|
len = min(host->sg_miter.length, blksize);
|
|
|
|
dev_dbg(host->dev, "PIO, %s block size: 0x%zx\n",
|
|
read ? "read" : "write", blksize);
|
|
|
|
host->sg_miter.consumed = len;
|
|
host->blocks--;
|
|
|
|
buf = host->sg_miter.addr;
|
|
|
|
if (read)
|
|
ioread32_rep(priv->iobase + AU6601_REG_BUFFER, buf, len >> 2);
|
|
else
|
|
iowrite32_rep(priv->iobase + AU6601_REG_BUFFER, buf, len >> 2);
|
|
|
|
sg_miter_stop(&host->sg_miter);
|
|
}
|
|
|
|
static void alcor_prepare_sg_miter(struct alcor_sdmmc_host *host)
|
|
{
|
|
unsigned int flags = SG_MITER_ATOMIC;
|
|
struct mmc_data *data = host->data;
|
|
|
|
if (data->flags & MMC_DATA_READ)
|
|
flags |= SG_MITER_TO_SG;
|
|
else
|
|
flags |= SG_MITER_FROM_SG;
|
|
sg_miter_start(&host->sg_miter, data->sg, data->sg_len, flags);
|
|
}
|
|
|
|
static void alcor_prepare_data(struct alcor_sdmmc_host *host,
|
|
struct mmc_command *cmd)
|
|
{
|
|
struct alcor_pci_priv *priv = host->alcor_pci;
|
|
struct mmc_data *data = cmd->data;
|
|
|
|
if (!data)
|
|
return;
|
|
|
|
|
|
host->data = data;
|
|
host->data->bytes_xfered = 0;
|
|
host->blocks = data->blocks;
|
|
host->sg = data->sg;
|
|
host->sg_count = data->sg_count;
|
|
dev_dbg(host->dev, "prepare DATA: sg %i, blocks: %i\n",
|
|
host->sg_count, host->blocks);
|
|
|
|
if (data->host_cookie != COOKIE_MAPPED)
|
|
alcor_prepare_sg_miter(host);
|
|
|
|
alcor_write8(priv, 0, AU6601_DATA_XFER_CTRL);
|
|
}
|
|
|
|
static void alcor_send_cmd(struct alcor_sdmmc_host *host,
|
|
struct mmc_command *cmd, bool set_timeout)
|
|
{
|
|
struct alcor_pci_priv *priv = host->alcor_pci;
|
|
unsigned long timeout = 0;
|
|
u8 ctrl = 0;
|
|
|
|
host->cmd = cmd;
|
|
alcor_prepare_data(host, cmd);
|
|
|
|
dev_dbg(host->dev, "send CMD. opcode: 0x%02x, arg; 0x%08x\n",
|
|
cmd->opcode, cmd->arg);
|
|
alcor_write8(priv, cmd->opcode | 0x40, AU6601_REG_CMD_OPCODE);
|
|
alcor_write32be(priv, cmd->arg, AU6601_REG_CMD_ARG);
|
|
|
|
switch (mmc_resp_type(cmd)) {
|
|
case MMC_RSP_NONE:
|
|
ctrl = AU6601_CMD_NO_RESP;
|
|
break;
|
|
case MMC_RSP_R1:
|
|
ctrl = AU6601_CMD_6_BYTE_CRC;
|
|
break;
|
|
case MMC_RSP_R1B:
|
|
ctrl = AU6601_CMD_6_BYTE_CRC | AU6601_CMD_STOP_WAIT_RDY;
|
|
break;
|
|
case MMC_RSP_R2:
|
|
ctrl = AU6601_CMD_17_BYTE_CRC;
|
|
break;
|
|
case MMC_RSP_R3:
|
|
ctrl = AU6601_CMD_6_BYTE_WO_CRC;
|
|
break;
|
|
default:
|
|
dev_err(host->dev, "%s: cmd->flag (0x%02x) is not valid\n",
|
|
mmc_hostname(mmc_from_priv(host)), mmc_resp_type(cmd));
|
|
break;
|
|
}
|
|
|
|
if (set_timeout) {
|
|
if (!cmd->data && cmd->busy_timeout)
|
|
timeout = cmd->busy_timeout;
|
|
else
|
|
timeout = 10000;
|
|
|
|
schedule_delayed_work(&host->timeout_work,
|
|
msecs_to_jiffies(timeout));
|
|
}
|
|
|
|
dev_dbg(host->dev, "xfer ctrl: 0x%02x; timeout: %lu\n", ctrl, timeout);
|
|
alcor_write8(priv, ctrl | AU6601_CMD_START_XFER,
|
|
AU6601_CMD_XFER_CTRL);
|
|
}
|
|
|
|
static void alcor_request_complete(struct alcor_sdmmc_host *host,
|
|
bool cancel_timeout)
|
|
{
|
|
struct mmc_request *mrq;
|
|
|
|
/*
|
|
* If this work gets rescheduled while running, it will
|
|
* be run again afterwards but without any active request.
|
|
*/
|
|
if (!host->mrq)
|
|
return;
|
|
|
|
if (cancel_timeout)
|
|
cancel_delayed_work(&host->timeout_work);
|
|
|
|
mrq = host->mrq;
|
|
|
|
host->mrq = NULL;
|
|
host->cmd = NULL;
|
|
host->data = NULL;
|
|
host->dma_on = 0;
|
|
|
|
mmc_request_done(mmc_from_priv(host), mrq);
|
|
}
|
|
|
|
static void alcor_finish_data(struct alcor_sdmmc_host *host)
|
|
{
|
|
struct mmc_data *data;
|
|
|
|
data = host->data;
|
|
host->data = NULL;
|
|
host->dma_on = 0;
|
|
|
|
/*
|
|
* The specification states that the block count register must
|
|
* be updated, but it does not specify at what point in the
|
|
* data flow. That makes the register entirely useless to read
|
|
* back so we have to assume that nothing made it to the card
|
|
* in the event of an error.
|
|
*/
|
|
if (data->error)
|
|
data->bytes_xfered = 0;
|
|
else
|
|
data->bytes_xfered = data->blksz * data->blocks;
|
|
|
|
/*
|
|
* Need to send CMD12 if -
|
|
* a) open-ended multiblock transfer (no CMD23)
|
|
* b) error in multiblock transfer
|
|
*/
|
|
if (data->stop &&
|
|
(data->error ||
|
|
!host->mrq->sbc)) {
|
|
|
|
/*
|
|
* The controller needs a reset of internal state machines
|
|
* upon error conditions.
|
|
*/
|
|
if (data->error)
|
|
alcor_reset(host, AU6601_RESET_CMD | AU6601_RESET_DATA);
|
|
|
|
alcor_unmask_sd_irqs(host);
|
|
alcor_send_cmd(host, data->stop, false);
|
|
return;
|
|
}
|
|
|
|
alcor_request_complete(host, 1);
|
|
}
|
|
|
|
static void alcor_err_irq(struct alcor_sdmmc_host *host, u32 intmask)
|
|
{
|
|
dev_dbg(host->dev, "ERR IRQ %x\n", intmask);
|
|
|
|
if (host->cmd) {
|
|
if (intmask & AU6601_INT_CMD_TIMEOUT_ERR)
|
|
host->cmd->error = -ETIMEDOUT;
|
|
else
|
|
host->cmd->error = -EILSEQ;
|
|
}
|
|
|
|
if (host->data) {
|
|
if (intmask & AU6601_INT_DATA_TIMEOUT_ERR)
|
|
host->data->error = -ETIMEDOUT;
|
|
else
|
|
host->data->error = -EILSEQ;
|
|
|
|
host->data->bytes_xfered = 0;
|
|
}
|
|
|
|
alcor_reset(host, AU6601_RESET_CMD | AU6601_RESET_DATA);
|
|
alcor_request_complete(host, 1);
|
|
}
|
|
|
|
static int alcor_cmd_irq_done(struct alcor_sdmmc_host *host, u32 intmask)
|
|
{
|
|
struct alcor_pci_priv *priv = host->alcor_pci;
|
|
|
|
intmask &= AU6601_INT_CMD_END;
|
|
|
|
if (!intmask)
|
|
return true;
|
|
|
|
/* got CMD_END but no CMD is in progress, wake thread an process the
|
|
* error
|
|
*/
|
|
if (!host->cmd)
|
|
return false;
|
|
|
|
if (host->cmd->flags & MMC_RSP_PRESENT) {
|
|
struct mmc_command *cmd = host->cmd;
|
|
|
|
cmd->resp[0] = alcor_read32be(priv, AU6601_REG_CMD_RSP0);
|
|
dev_dbg(host->dev, "RSP0: 0x%04x\n", cmd->resp[0]);
|
|
if (host->cmd->flags & MMC_RSP_136) {
|
|
cmd->resp[1] =
|
|
alcor_read32be(priv, AU6601_REG_CMD_RSP1);
|
|
cmd->resp[2] =
|
|
alcor_read32be(priv, AU6601_REG_CMD_RSP2);
|
|
cmd->resp[3] =
|
|
alcor_read32be(priv, AU6601_REG_CMD_RSP3);
|
|
dev_dbg(host->dev, "RSP1,2,3: 0x%04x 0x%04x 0x%04x\n",
|
|
cmd->resp[1], cmd->resp[2], cmd->resp[3]);
|
|
}
|
|
|
|
}
|
|
|
|
host->cmd->error = 0;
|
|
|
|
/* Processed actual command. */
|
|
if (!host->data)
|
|
return false;
|
|
|
|
alcor_trigger_data_transfer(host);
|
|
host->cmd = NULL;
|
|
return true;
|
|
}
|
|
|
|
static void alcor_cmd_irq_thread(struct alcor_sdmmc_host *host, u32 intmask)
|
|
{
|
|
intmask &= AU6601_INT_CMD_END;
|
|
|
|
if (!intmask)
|
|
return;
|
|
|
|
if (!host->cmd && intmask & AU6601_INT_CMD_END) {
|
|
dev_dbg(host->dev, "Got command interrupt 0x%08x even though no command operation was in progress.\n",
|
|
intmask);
|
|
}
|
|
|
|
/* Processed actual command. */
|
|
if (!host->data)
|
|
alcor_request_complete(host, 1);
|
|
else
|
|
alcor_trigger_data_transfer(host);
|
|
host->cmd = NULL;
|
|
}
|
|
|
|
static int alcor_data_irq_done(struct alcor_sdmmc_host *host, u32 intmask)
|
|
{
|
|
u32 tmp;
|
|
|
|
intmask &= AU6601_INT_DATA_MASK;
|
|
|
|
/* nothing here to do */
|
|
if (!intmask)
|
|
return 1;
|
|
|
|
/* we was too fast and got DATA_END after it was processed?
|
|
* lets ignore it for now.
|
|
*/
|
|
if (!host->data && intmask == AU6601_INT_DATA_END)
|
|
return 1;
|
|
|
|
/* looks like an error, so lets handle it. */
|
|
if (!host->data)
|
|
return 0;
|
|
|
|
tmp = intmask & (AU6601_INT_READ_BUF_RDY | AU6601_INT_WRITE_BUF_RDY
|
|
| AU6601_INT_DMA_END);
|
|
switch (tmp) {
|
|
case 0:
|
|
break;
|
|
case AU6601_INT_READ_BUF_RDY:
|
|
alcor_trf_block_pio(host, true);
|
|
return 1;
|
|
case AU6601_INT_WRITE_BUF_RDY:
|
|
alcor_trf_block_pio(host, false);
|
|
return 1;
|
|
case AU6601_INT_DMA_END:
|
|
if (!host->sg_count)
|
|
break;
|
|
|
|
alcor_data_set_dma(host);
|
|
break;
|
|
default:
|
|
dev_err(host->dev, "Got READ_BUF_RDY and WRITE_BUF_RDY at same time\n");
|
|
break;
|
|
}
|
|
|
|
if (intmask & AU6601_INT_DATA_END) {
|
|
if (!host->dma_on && host->blocks) {
|
|
alcor_trigger_data_transfer(host);
|
|
return 1;
|
|
} else {
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
static void alcor_data_irq_thread(struct alcor_sdmmc_host *host, u32 intmask)
|
|
{
|
|
intmask &= AU6601_INT_DATA_MASK;
|
|
|
|
if (!intmask)
|
|
return;
|
|
|
|
if (!host->data) {
|
|
dev_dbg(host->dev, "Got data interrupt 0x%08x even though no data operation was in progress.\n",
|
|
intmask);
|
|
alcor_reset(host, AU6601_RESET_DATA);
|
|
return;
|
|
}
|
|
|
|
if (alcor_data_irq_done(host, intmask))
|
|
return;
|
|
|
|
if ((intmask & AU6601_INT_DATA_END) || !host->blocks ||
|
|
(host->dma_on && !host->sg_count))
|
|
alcor_finish_data(host);
|
|
}
|
|
|
|
static void alcor_cd_irq(struct alcor_sdmmc_host *host, u32 intmask)
|
|
{
|
|
dev_dbg(host->dev, "card %s\n",
|
|
intmask & AU6601_INT_CARD_REMOVE ? "removed" : "inserted");
|
|
|
|
if (host->mrq) {
|
|
dev_dbg(host->dev, "cancel all pending tasks.\n");
|
|
|
|
if (host->data)
|
|
host->data->error = -ENOMEDIUM;
|
|
|
|
if (host->cmd)
|
|
host->cmd->error = -ENOMEDIUM;
|
|
else
|
|
host->mrq->cmd->error = -ENOMEDIUM;
|
|
|
|
alcor_request_complete(host, 1);
|
|
}
|
|
|
|
mmc_detect_change(mmc_from_priv(host), msecs_to_jiffies(1));
|
|
}
|
|
|
|
static irqreturn_t alcor_irq_thread(int irq, void *d)
|
|
{
|
|
struct alcor_sdmmc_host *host = d;
|
|
irqreturn_t ret = IRQ_HANDLED;
|
|
u32 intmask, tmp;
|
|
|
|
mutex_lock(&host->cmd_mutex);
|
|
|
|
intmask = host->irq_status_sd;
|
|
|
|
/* some thing bad */
|
|
if (unlikely(!intmask || AU6601_INT_ALL_MASK == intmask)) {
|
|
dev_dbg(host->dev, "unexpected IRQ: 0x%04x\n", intmask);
|
|
ret = IRQ_NONE;
|
|
goto exit;
|
|
}
|
|
|
|
tmp = intmask & (AU6601_INT_CMD_MASK | AU6601_INT_DATA_MASK);
|
|
if (tmp) {
|
|
if (tmp & AU6601_INT_ERROR_MASK)
|
|
alcor_err_irq(host, tmp);
|
|
else {
|
|
alcor_cmd_irq_thread(host, tmp);
|
|
alcor_data_irq_thread(host, tmp);
|
|
}
|
|
intmask &= ~(AU6601_INT_CMD_MASK | AU6601_INT_DATA_MASK);
|
|
}
|
|
|
|
if (intmask & (AU6601_INT_CARD_INSERT | AU6601_INT_CARD_REMOVE)) {
|
|
alcor_cd_irq(host, intmask);
|
|
intmask &= ~(AU6601_INT_CARD_INSERT | AU6601_INT_CARD_REMOVE);
|
|
}
|
|
|
|
if (intmask & AU6601_INT_OVER_CURRENT_ERR) {
|
|
dev_warn(host->dev,
|
|
"warning: over current detected!\n");
|
|
intmask &= ~AU6601_INT_OVER_CURRENT_ERR;
|
|
}
|
|
|
|
if (intmask)
|
|
dev_dbg(host->dev, "got not handled IRQ: 0x%04x\n", intmask);
|
|
|
|
exit:
|
|
mutex_unlock(&host->cmd_mutex);
|
|
alcor_unmask_sd_irqs(host);
|
|
return ret;
|
|
}
|
|
|
|
|
|
static irqreturn_t alcor_irq(int irq, void *d)
|
|
{
|
|
struct alcor_sdmmc_host *host = d;
|
|
struct alcor_pci_priv *priv = host->alcor_pci;
|
|
u32 status, tmp;
|
|
irqreturn_t ret;
|
|
int cmd_done, data_done;
|
|
|
|
status = alcor_read32(priv, AU6601_REG_INT_STATUS);
|
|
if (!status)
|
|
return IRQ_NONE;
|
|
|
|
alcor_write32(priv, status, AU6601_REG_INT_STATUS);
|
|
|
|
tmp = status & (AU6601_INT_READ_BUF_RDY | AU6601_INT_WRITE_BUF_RDY
|
|
| AU6601_INT_DATA_END | AU6601_INT_DMA_END
|
|
| AU6601_INT_CMD_END);
|
|
if (tmp == status) {
|
|
cmd_done = alcor_cmd_irq_done(host, tmp);
|
|
data_done = alcor_data_irq_done(host, tmp);
|
|
/* use fast path for simple tasks */
|
|
if (cmd_done && data_done) {
|
|
ret = IRQ_HANDLED;
|
|
goto alcor_irq_done;
|
|
}
|
|
}
|
|
|
|
host->irq_status_sd = status;
|
|
ret = IRQ_WAKE_THREAD;
|
|
alcor_mask_sd_irqs(host);
|
|
alcor_irq_done:
|
|
return ret;
|
|
}
|
|
|
|
static void alcor_set_clock(struct alcor_sdmmc_host *host, unsigned int clock)
|
|
{
|
|
struct alcor_pci_priv *priv = host->alcor_pci;
|
|
int i, diff = 0x7fffffff, tmp_clock = 0;
|
|
u16 clk_src = 0;
|
|
u8 clk_div = 0;
|
|
|
|
if (clock == 0) {
|
|
alcor_write16(priv, 0, AU6601_CLK_SELECT);
|
|
return;
|
|
}
|
|
|
|
for (i = 0; i < ARRAY_SIZE(alcor_pll_cfg); i++) {
|
|
unsigned int tmp_div, tmp_diff;
|
|
const struct alcor_pll_conf *cfg = &alcor_pll_cfg[i];
|
|
|
|
tmp_div = DIV_ROUND_UP(cfg->clk_src_freq, clock);
|
|
if (cfg->min_div > tmp_div || tmp_div > cfg->max_div)
|
|
continue;
|
|
|
|
tmp_clock = DIV_ROUND_UP(cfg->clk_src_freq, tmp_div);
|
|
tmp_diff = abs(clock - tmp_clock);
|
|
|
|
if (tmp_diff >= 0 && tmp_diff < diff) {
|
|
diff = tmp_diff;
|
|
clk_src = cfg->clk_src_reg;
|
|
clk_div = tmp_div;
|
|
}
|
|
}
|
|
|
|
clk_src |= ((clk_div - 1) << 8);
|
|
clk_src |= AU6601_CLK_ENABLE;
|
|
|
|
dev_dbg(host->dev, "set freq %d cal freq %d, use div %d, mod %x\n",
|
|
clock, tmp_clock, clk_div, clk_src);
|
|
|
|
alcor_write16(priv, clk_src, AU6601_CLK_SELECT);
|
|
|
|
}
|
|
|
|
static void alcor_set_timing(struct mmc_host *mmc, struct mmc_ios *ios)
|
|
{
|
|
struct alcor_sdmmc_host *host = mmc_priv(mmc);
|
|
|
|
if (ios->timing == MMC_TIMING_LEGACY) {
|
|
alcor_rmw8(host, AU6601_CLK_DELAY,
|
|
AU6601_CLK_POSITIVE_EDGE_ALL, 0);
|
|
} else {
|
|
alcor_rmw8(host, AU6601_CLK_DELAY,
|
|
0, AU6601_CLK_POSITIVE_EDGE_ALL);
|
|
}
|
|
}
|
|
|
|
static void alcor_set_bus_width(struct mmc_host *mmc, struct mmc_ios *ios)
|
|
{
|
|
struct alcor_sdmmc_host *host = mmc_priv(mmc);
|
|
struct alcor_pci_priv *priv = host->alcor_pci;
|
|
|
|
if (ios->bus_width == MMC_BUS_WIDTH_1) {
|
|
alcor_write8(priv, 0, AU6601_REG_BUS_CTRL);
|
|
} else if (ios->bus_width == MMC_BUS_WIDTH_4) {
|
|
alcor_write8(priv, AU6601_BUS_WIDTH_4BIT,
|
|
AU6601_REG_BUS_CTRL);
|
|
} else
|
|
dev_err(host->dev, "Unknown BUS mode\n");
|
|
|
|
}
|
|
|
|
static int alcor_card_busy(struct mmc_host *mmc)
|
|
{
|
|
struct alcor_sdmmc_host *host = mmc_priv(mmc);
|
|
struct alcor_pci_priv *priv = host->alcor_pci;
|
|
u8 status;
|
|
|
|
/* Check whether dat[0:3] low */
|
|
status = alcor_read8(priv, AU6601_DATA_PIN_STATE);
|
|
|
|
return !(status & AU6601_BUS_STAT_DAT_MASK);
|
|
}
|
|
|
|
static int alcor_get_cd(struct mmc_host *mmc)
|
|
{
|
|
struct alcor_sdmmc_host *host = mmc_priv(mmc);
|
|
struct alcor_pci_priv *priv = host->alcor_pci;
|
|
u8 detect;
|
|
|
|
detect = alcor_read8(priv, AU6601_DETECT_STATUS)
|
|
& AU6601_DETECT_STATUS_M;
|
|
/* check if card is present then send command and data */
|
|
return (detect == AU6601_SD_DETECTED);
|
|
}
|
|
|
|
static int alcor_get_ro(struct mmc_host *mmc)
|
|
{
|
|
struct alcor_sdmmc_host *host = mmc_priv(mmc);
|
|
struct alcor_pci_priv *priv = host->alcor_pci;
|
|
u8 status;
|
|
|
|
/* get write protect pin status */
|
|
status = alcor_read8(priv, AU6601_INTERFACE_MODE_CTRL);
|
|
|
|
return !!(status & AU6601_SD_CARD_WP);
|
|
}
|
|
|
|
static void alcor_request(struct mmc_host *mmc, struct mmc_request *mrq)
|
|
{
|
|
struct alcor_sdmmc_host *host = mmc_priv(mmc);
|
|
|
|
mutex_lock(&host->cmd_mutex);
|
|
|
|
host->mrq = mrq;
|
|
|
|
/* check if card is present then send command and data */
|
|
if (alcor_get_cd(mmc))
|
|
alcor_send_cmd(host, mrq->cmd, true);
|
|
else {
|
|
mrq->cmd->error = -ENOMEDIUM;
|
|
alcor_request_complete(host, 1);
|
|
}
|
|
|
|
mutex_unlock(&host->cmd_mutex);
|
|
}
|
|
|
|
static void alcor_pre_req(struct mmc_host *mmc,
|
|
struct mmc_request *mrq)
|
|
{
|
|
struct alcor_sdmmc_host *host = mmc_priv(mmc);
|
|
struct mmc_data *data = mrq->data;
|
|
struct mmc_command *cmd = mrq->cmd;
|
|
struct scatterlist *sg;
|
|
unsigned int i, sg_len;
|
|
|
|
if (!data || !cmd)
|
|
return;
|
|
|
|
data->host_cookie = COOKIE_UNMAPPED;
|
|
|
|
/* FIXME: looks like the DMA engine works only with CMD18 */
|
|
if (cmd->opcode != MMC_READ_MULTIPLE_BLOCK
|
|
&& cmd->opcode != MMC_WRITE_MULTIPLE_BLOCK)
|
|
return;
|
|
/*
|
|
* We don't do DMA on "complex" transfers, i.e. with
|
|
* non-word-aligned buffers or lengths. A future improvement
|
|
* could be made to use temporary DMA bounce-buffers when these
|
|
* requirements are not met.
|
|
*
|
|
* Also, we don't bother with all the DMA setup overhead for
|
|
* short transfers.
|
|
*/
|
|
if (data->blocks * data->blksz < AU6601_MAX_DMA_BLOCK_SIZE)
|
|
return;
|
|
|
|
if (data->blksz & 3)
|
|
return;
|
|
|
|
for_each_sg(data->sg, sg, data->sg_len, i) {
|
|
if (sg->length != AU6601_MAX_DMA_BLOCK_SIZE)
|
|
return;
|
|
if (sg->offset != 0)
|
|
return;
|
|
}
|
|
|
|
/* This data might be unmapped at this time */
|
|
|
|
sg_len = dma_map_sg(host->dev, data->sg, data->sg_len,
|
|
mmc_get_dma_dir(data));
|
|
if (sg_len)
|
|
data->host_cookie = COOKIE_MAPPED;
|
|
|
|
data->sg_count = sg_len;
|
|
}
|
|
|
|
static void alcor_post_req(struct mmc_host *mmc,
|
|
struct mmc_request *mrq,
|
|
int err)
|
|
{
|
|
struct alcor_sdmmc_host *host = mmc_priv(mmc);
|
|
struct mmc_data *data = mrq->data;
|
|
|
|
if (!data)
|
|
return;
|
|
|
|
if (data->host_cookie == COOKIE_MAPPED) {
|
|
dma_unmap_sg(host->dev,
|
|
data->sg,
|
|
data->sg_len,
|
|
mmc_get_dma_dir(data));
|
|
}
|
|
|
|
data->host_cookie = COOKIE_UNMAPPED;
|
|
}
|
|
|
|
static void alcor_set_power_mode(struct mmc_host *mmc, struct mmc_ios *ios)
|
|
{
|
|
struct alcor_sdmmc_host *host = mmc_priv(mmc);
|
|
struct alcor_pci_priv *priv = host->alcor_pci;
|
|
|
|
switch (ios->power_mode) {
|
|
case MMC_POWER_OFF:
|
|
alcor_set_clock(host, ios->clock);
|
|
/* set all pins to input */
|
|
alcor_write8(priv, 0, AU6601_OUTPUT_ENABLE);
|
|
/* turn of VDD */
|
|
alcor_write8(priv, 0, AU6601_POWER_CONTROL);
|
|
break;
|
|
case MMC_POWER_UP:
|
|
break;
|
|
case MMC_POWER_ON:
|
|
/* This is most trickiest part. The order and timings of
|
|
* instructions seems to play important role. Any changes may
|
|
* confuse internal state engine if this HW.
|
|
* FIXME: If we will ever get access to documentation, then this
|
|
* part should be reviewed again.
|
|
*/
|
|
|
|
/* enable SD card mode */
|
|
alcor_write8(priv, AU6601_SD_CARD,
|
|
AU6601_ACTIVE_CTRL);
|
|
/* set signal voltage to 3.3V */
|
|
alcor_write8(priv, 0, AU6601_OPT);
|
|
/* no documentation about clk delay, for now just try to mimic
|
|
* original driver.
|
|
*/
|
|
alcor_write8(priv, 0x20, AU6601_CLK_DELAY);
|
|
/* set BUS width to 1 bit */
|
|
alcor_write8(priv, 0, AU6601_REG_BUS_CTRL);
|
|
/* set CLK first time */
|
|
alcor_set_clock(host, ios->clock);
|
|
/* power on VDD */
|
|
alcor_write8(priv, AU6601_SD_CARD,
|
|
AU6601_POWER_CONTROL);
|
|
/* wait until the CLK will get stable */
|
|
mdelay(20);
|
|
/* set CLK again, mimic original driver. */
|
|
alcor_set_clock(host, ios->clock);
|
|
|
|
/* enable output */
|
|
alcor_write8(priv, AU6601_SD_CARD,
|
|
AU6601_OUTPUT_ENABLE);
|
|
/* The clk will not work on au6621. We need to trigger data
|
|
* transfer.
|
|
*/
|
|
alcor_write8(priv, AU6601_DATA_WRITE,
|
|
AU6601_DATA_XFER_CTRL);
|
|
/* configure timeout. Not clear what exactly it means. */
|
|
alcor_write8(priv, 0x7d, AU6601_TIME_OUT_CTRL);
|
|
mdelay(100);
|
|
break;
|
|
default:
|
|
dev_err(host->dev, "Unknown power parameter\n");
|
|
}
|
|
}
|
|
|
|
static void alcor_set_ios(struct mmc_host *mmc, struct mmc_ios *ios)
|
|
{
|
|
struct alcor_sdmmc_host *host = mmc_priv(mmc);
|
|
|
|
mutex_lock(&host->cmd_mutex);
|
|
|
|
dev_dbg(host->dev, "set ios. bus width: %x, power mode: %x\n",
|
|
ios->bus_width, ios->power_mode);
|
|
|
|
if (ios->power_mode != host->cur_power_mode) {
|
|
alcor_set_power_mode(mmc, ios);
|
|
host->cur_power_mode = ios->power_mode;
|
|
} else {
|
|
alcor_set_timing(mmc, ios);
|
|
alcor_set_bus_width(mmc, ios);
|
|
alcor_set_clock(host, ios->clock);
|
|
}
|
|
|
|
mutex_unlock(&host->cmd_mutex);
|
|
}
|
|
|
|
static int alcor_signal_voltage_switch(struct mmc_host *mmc,
|
|
struct mmc_ios *ios)
|
|
{
|
|
struct alcor_sdmmc_host *host = mmc_priv(mmc);
|
|
|
|
mutex_lock(&host->cmd_mutex);
|
|
|
|
switch (ios->signal_voltage) {
|
|
case MMC_SIGNAL_VOLTAGE_330:
|
|
alcor_rmw8(host, AU6601_OPT, AU6601_OPT_SD_18V, 0);
|
|
break;
|
|
case MMC_SIGNAL_VOLTAGE_180:
|
|
alcor_rmw8(host, AU6601_OPT, 0, AU6601_OPT_SD_18V);
|
|
break;
|
|
default:
|
|
/* No signal voltage switch required */
|
|
break;
|
|
}
|
|
|
|
mutex_unlock(&host->cmd_mutex);
|
|
return 0;
|
|
}
|
|
|
|
static const struct mmc_host_ops alcor_sdc_ops = {
|
|
.card_busy = alcor_card_busy,
|
|
.get_cd = alcor_get_cd,
|
|
.get_ro = alcor_get_ro,
|
|
.post_req = alcor_post_req,
|
|
.pre_req = alcor_pre_req,
|
|
.request = alcor_request,
|
|
.set_ios = alcor_set_ios,
|
|
.start_signal_voltage_switch = alcor_signal_voltage_switch,
|
|
};
|
|
|
|
static void alcor_timeout_timer(struct work_struct *work)
|
|
{
|
|
struct delayed_work *d = to_delayed_work(work);
|
|
struct alcor_sdmmc_host *host = container_of(d, struct alcor_sdmmc_host,
|
|
timeout_work);
|
|
mutex_lock(&host->cmd_mutex);
|
|
|
|
dev_dbg(host->dev, "triggered timeout\n");
|
|
if (host->mrq) {
|
|
dev_err(host->dev, "Timeout waiting for hardware interrupt.\n");
|
|
|
|
if (host->data) {
|
|
host->data->error = -ETIMEDOUT;
|
|
} else {
|
|
if (host->cmd)
|
|
host->cmd->error = -ETIMEDOUT;
|
|
else
|
|
host->mrq->cmd->error = -ETIMEDOUT;
|
|
}
|
|
|
|
alcor_reset(host, AU6601_RESET_CMD | AU6601_RESET_DATA);
|
|
alcor_request_complete(host, 0);
|
|
}
|
|
|
|
mutex_unlock(&host->cmd_mutex);
|
|
}
|
|
|
|
static void alcor_hw_init(struct alcor_sdmmc_host *host)
|
|
{
|
|
struct alcor_pci_priv *priv = host->alcor_pci;
|
|
struct alcor_dev_cfg *cfg = priv->cfg;
|
|
|
|
/* FIXME: This part is a mimics HW init of original driver.
|
|
* If we will ever get access to documentation, then this part
|
|
* should be reviewed again.
|
|
*/
|
|
|
|
/* reset command state engine */
|
|
alcor_reset(host, AU6601_RESET_CMD);
|
|
|
|
alcor_write8(priv, 0, AU6601_DMA_BOUNDARY);
|
|
/* enable sd card mode */
|
|
alcor_write8(priv, AU6601_SD_CARD, AU6601_ACTIVE_CTRL);
|
|
|
|
/* set BUS width to 1 bit */
|
|
alcor_write8(priv, 0, AU6601_REG_BUS_CTRL);
|
|
|
|
/* reset data state engine */
|
|
alcor_reset(host, AU6601_RESET_DATA);
|
|
/* Not sure if a voodoo with AU6601_DMA_BOUNDARY is really needed */
|
|
alcor_write8(priv, 0, AU6601_DMA_BOUNDARY);
|
|
|
|
alcor_write8(priv, 0, AU6601_INTERFACE_MODE_CTRL);
|
|
/* not clear what we are doing here. */
|
|
alcor_write8(priv, 0x44, AU6601_PAD_DRIVE0);
|
|
alcor_write8(priv, 0x44, AU6601_PAD_DRIVE1);
|
|
alcor_write8(priv, 0x00, AU6601_PAD_DRIVE2);
|
|
|
|
/* for 6601 - dma_boundary; for 6621 - dma_page_cnt
|
|
* exact meaning of this register is not clear.
|
|
*/
|
|
alcor_write8(priv, cfg->dma, AU6601_DMA_BOUNDARY);
|
|
|
|
/* make sure all pins are set to input and VDD is off */
|
|
alcor_write8(priv, 0, AU6601_OUTPUT_ENABLE);
|
|
alcor_write8(priv, 0, AU6601_POWER_CONTROL);
|
|
|
|
alcor_write8(priv, AU6601_DETECT_EN, AU6601_DETECT_STATUS);
|
|
/* now we should be safe to enable IRQs */
|
|
alcor_unmask_sd_irqs(host);
|
|
}
|
|
|
|
static void alcor_hw_uninit(struct alcor_sdmmc_host *host)
|
|
{
|
|
struct alcor_pci_priv *priv = host->alcor_pci;
|
|
|
|
alcor_mask_sd_irqs(host);
|
|
alcor_reset(host, AU6601_RESET_CMD | AU6601_RESET_DATA);
|
|
|
|
alcor_write8(priv, 0, AU6601_DETECT_STATUS);
|
|
|
|
alcor_write8(priv, 0, AU6601_OUTPUT_ENABLE);
|
|
alcor_write8(priv, 0, AU6601_POWER_CONTROL);
|
|
|
|
alcor_write8(priv, 0, AU6601_OPT);
|
|
}
|
|
|
|
static void alcor_init_mmc(struct alcor_sdmmc_host *host)
|
|
{
|
|
struct mmc_host *mmc = mmc_from_priv(host);
|
|
|
|
mmc->f_min = AU6601_MIN_CLOCK;
|
|
mmc->f_max = AU6601_MAX_CLOCK;
|
|
mmc->ocr_avail = MMC_VDD_33_34;
|
|
mmc->caps = MMC_CAP_4_BIT_DATA | MMC_CAP_SD_HIGHSPEED
|
|
| MMC_CAP_UHS_SDR12 | MMC_CAP_UHS_SDR25 | MMC_CAP_UHS_SDR50
|
|
| MMC_CAP_UHS_SDR104 | MMC_CAP_UHS_DDR50;
|
|
mmc->caps2 = MMC_CAP2_NO_SDIO;
|
|
mmc->ops = &alcor_sdc_ops;
|
|
|
|
/* The hardware does DMA data transfer of 4096 bytes to/from a single
|
|
* buffer address. Scatterlists are not supported at the hardware
|
|
* level, however we can work with them at the driver level,
|
|
* provided that each segment is exactly 4096 bytes in size.
|
|
* Upon DMA completion of a single segment (signalled via IRQ), we
|
|
* immediately proceed to transfer the next segment from the
|
|
* scatterlist.
|
|
*
|
|
* The overall request is limited to 240 sectors, matching the
|
|
* original vendor driver.
|
|
*/
|
|
mmc->max_segs = AU6601_MAX_DMA_SEGMENTS;
|
|
mmc->max_seg_size = AU6601_MAX_DMA_BLOCK_SIZE;
|
|
mmc->max_blk_count = 240;
|
|
mmc->max_req_size = mmc->max_blk_count * mmc->max_blk_size;
|
|
dma_set_max_seg_size(host->dev, mmc->max_seg_size);
|
|
}
|
|
|
|
static int alcor_pci_sdmmc_drv_probe(struct platform_device *pdev)
|
|
{
|
|
struct alcor_pci_priv *priv = pdev->dev.platform_data;
|
|
struct mmc_host *mmc;
|
|
struct alcor_sdmmc_host *host;
|
|
int ret;
|
|
|
|
mmc = mmc_alloc_host(sizeof(*host), &pdev->dev);
|
|
if (!mmc) {
|
|
dev_err(&pdev->dev, "Can't allocate MMC\n");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
host = mmc_priv(mmc);
|
|
host->dev = &pdev->dev;
|
|
host->cur_power_mode = MMC_POWER_UNDEFINED;
|
|
host->alcor_pci = priv;
|
|
|
|
/* make sure irqs are disabled */
|
|
alcor_write32(priv, 0, AU6601_REG_INT_ENABLE);
|
|
alcor_write32(priv, 0, AU6601_MS_INT_ENABLE);
|
|
|
|
ret = devm_request_threaded_irq(&pdev->dev, priv->irq,
|
|
alcor_irq, alcor_irq_thread, IRQF_SHARED,
|
|
DRV_NAME_ALCOR_PCI_SDMMC, host);
|
|
|
|
if (ret) {
|
|
dev_err(&pdev->dev, "Failed to get irq for data line\n");
|
|
return ret;
|
|
}
|
|
|
|
mutex_init(&host->cmd_mutex);
|
|
INIT_DELAYED_WORK(&host->timeout_work, alcor_timeout_timer);
|
|
|
|
alcor_init_mmc(host);
|
|
alcor_hw_init(host);
|
|
|
|
dev_set_drvdata(&pdev->dev, host);
|
|
mmc_add_host(mmc);
|
|
return 0;
|
|
}
|
|
|
|
static int alcor_pci_sdmmc_drv_remove(struct platform_device *pdev)
|
|
{
|
|
struct alcor_sdmmc_host *host = dev_get_drvdata(&pdev->dev);
|
|
struct mmc_host *mmc = mmc_from_priv(host);
|
|
|
|
if (cancel_delayed_work_sync(&host->timeout_work))
|
|
alcor_request_complete(host, 0);
|
|
|
|
alcor_hw_uninit(host);
|
|
mmc_remove_host(mmc);
|
|
mmc_free_host(mmc);
|
|
|
|
return 0;
|
|
}
|
|
|
|
#ifdef CONFIG_PM_SLEEP
|
|
static int alcor_pci_sdmmc_suspend(struct device *dev)
|
|
{
|
|
struct alcor_sdmmc_host *host = dev_get_drvdata(dev);
|
|
|
|
if (cancel_delayed_work_sync(&host->timeout_work))
|
|
alcor_request_complete(host, 0);
|
|
|
|
alcor_hw_uninit(host);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int alcor_pci_sdmmc_resume(struct device *dev)
|
|
{
|
|
struct alcor_sdmmc_host *host = dev_get_drvdata(dev);
|
|
|
|
alcor_hw_init(host);
|
|
|
|
return 0;
|
|
}
|
|
#endif /* CONFIG_PM_SLEEP */
|
|
|
|
static SIMPLE_DEV_PM_OPS(alcor_mmc_pm_ops, alcor_pci_sdmmc_suspend,
|
|
alcor_pci_sdmmc_resume);
|
|
|
|
static const struct platform_device_id alcor_pci_sdmmc_ids[] = {
|
|
{
|
|
.name = DRV_NAME_ALCOR_PCI_SDMMC,
|
|
}, {
|
|
/* sentinel */
|
|
}
|
|
};
|
|
MODULE_DEVICE_TABLE(platform, alcor_pci_sdmmc_ids);
|
|
|
|
static struct platform_driver alcor_pci_sdmmc_driver = {
|
|
.probe = alcor_pci_sdmmc_drv_probe,
|
|
.remove = alcor_pci_sdmmc_drv_remove,
|
|
.id_table = alcor_pci_sdmmc_ids,
|
|
.driver = {
|
|
.name = DRV_NAME_ALCOR_PCI_SDMMC,
|
|
.pm = &alcor_mmc_pm_ops
|
|
},
|
|
};
|
|
module_platform_driver(alcor_pci_sdmmc_driver);
|
|
|
|
MODULE_AUTHOR("Oleksij Rempel <linux@rempel-privat.de>");
|
|
MODULE_DESCRIPTION("PCI driver for Alcor Micro AU6601 Secure Digital Host Controller Interface");
|
|
MODULE_LICENSE("GPL");
|