1566 lines
42 KiB
C
1566 lines
42 KiB
C
// SPDX-License-Identifier: GPL-2.0+
|
|
/*
|
|
* ***************************************************************************
|
|
* Marvell Armada-3700 Serial Driver
|
|
* Author: Wilson Ding <dingwei@marvell.com>
|
|
* Copyright (C) 2015 Marvell International Ltd.
|
|
* ***************************************************************************
|
|
*/
|
|
|
|
#include <linux/clk.h>
|
|
#include <linux/clk-provider.h>
|
|
#include <linux/console.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/device.h>
|
|
#include <linux/init.h>
|
|
#include <linux/io.h>
|
|
#include <linux/iopoll.h>
|
|
#include <linux/math64.h>
|
|
#include <linux/of.h>
|
|
#include <linux/of_address.h>
|
|
#include <linux/of_device.h>
|
|
#include <linux/of_irq.h>
|
|
#include <linux/of_platform.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/serial.h>
|
|
#include <linux/serial_core.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/tty.h>
|
|
#include <linux/tty_flip.h>
|
|
|
|
/* Register Map */
|
|
#define UART_STD_RBR 0x00
|
|
#define UART_EXT_RBR 0x18
|
|
|
|
#define UART_STD_TSH 0x04
|
|
#define UART_EXT_TSH 0x1C
|
|
|
|
#define UART_STD_CTRL1 0x08
|
|
#define UART_EXT_CTRL1 0x04
|
|
#define CTRL_SOFT_RST BIT(31)
|
|
#define CTRL_TXFIFO_RST BIT(15)
|
|
#define CTRL_RXFIFO_RST BIT(14)
|
|
#define CTRL_SND_BRK_SEQ BIT(11)
|
|
#define CTRL_BRK_DET_INT BIT(3)
|
|
#define CTRL_FRM_ERR_INT BIT(2)
|
|
#define CTRL_PAR_ERR_INT BIT(1)
|
|
#define CTRL_OVR_ERR_INT BIT(0)
|
|
#define CTRL_BRK_INT (CTRL_BRK_DET_INT | CTRL_FRM_ERR_INT | \
|
|
CTRL_PAR_ERR_INT | CTRL_OVR_ERR_INT)
|
|
|
|
#define UART_STD_CTRL2 UART_STD_CTRL1
|
|
#define UART_EXT_CTRL2 0x20
|
|
#define CTRL_STD_TX_RDY_INT BIT(5)
|
|
#define CTRL_EXT_TX_RDY_INT BIT(6)
|
|
#define CTRL_STD_RX_RDY_INT BIT(4)
|
|
#define CTRL_EXT_RX_RDY_INT BIT(5)
|
|
|
|
#define UART_STAT 0x0C
|
|
#define STAT_TX_FIFO_EMP BIT(13)
|
|
#define STAT_TX_FIFO_FUL BIT(11)
|
|
#define STAT_TX_EMP BIT(6)
|
|
#define STAT_STD_TX_RDY BIT(5)
|
|
#define STAT_EXT_TX_RDY BIT(15)
|
|
#define STAT_STD_RX_RDY BIT(4)
|
|
#define STAT_EXT_RX_RDY BIT(14)
|
|
#define STAT_BRK_DET BIT(3)
|
|
#define STAT_FRM_ERR BIT(2)
|
|
#define STAT_PAR_ERR BIT(1)
|
|
#define STAT_OVR_ERR BIT(0)
|
|
#define STAT_BRK_ERR (STAT_BRK_DET | STAT_FRM_ERR \
|
|
| STAT_PAR_ERR | STAT_OVR_ERR)
|
|
|
|
/*
|
|
* Marvell Armada 3700 Functional Specifications describes that bit 21 of UART
|
|
* Clock Control register controls UART1 and bit 20 controls UART2. But in
|
|
* reality bit 21 controls UART2 and bit 20 controls UART1. This seems to be an
|
|
* error in Marvell's documentation. Hence following CLK_DIS macros are swapped.
|
|
*/
|
|
|
|
#define UART_BRDV 0x10
|
|
/* These bits are located in UART1 address space and control UART2 */
|
|
#define UART2_CLK_DIS BIT(21)
|
|
/* These bits are located in UART1 address space and control UART1 */
|
|
#define UART1_CLK_DIS BIT(20)
|
|
/* These bits are located in UART1 address space and control both UARTs */
|
|
#define CLK_NO_XTAL BIT(19)
|
|
#define CLK_TBG_DIV1_SHIFT 15
|
|
#define CLK_TBG_DIV1_MASK 0x7
|
|
#define CLK_TBG_DIV1_MAX 6
|
|
#define CLK_TBG_DIV2_SHIFT 12
|
|
#define CLK_TBG_DIV2_MASK 0x7
|
|
#define CLK_TBG_DIV2_MAX 6
|
|
#define CLK_TBG_SEL_SHIFT 10
|
|
#define CLK_TBG_SEL_MASK 0x3
|
|
/* These bits are located in both UARTs address space */
|
|
#define BRDV_BAUD_MASK 0x3FF
|
|
#define BRDV_BAUD_MAX BRDV_BAUD_MASK
|
|
|
|
#define UART_OSAMP 0x14
|
|
#define OSAMP_DEFAULT_DIVISOR 16
|
|
#define OSAMP_DIVISORS_MASK 0x3F3F3F3F
|
|
#define OSAMP_MAX_DIVISOR 63
|
|
|
|
#define MVEBU_NR_UARTS 2
|
|
|
|
#define MVEBU_UART_TYPE "mvebu-uart"
|
|
#define DRIVER_NAME "mvebu_serial"
|
|
|
|
enum {
|
|
/* Either there is only one summed IRQ... */
|
|
UART_IRQ_SUM = 0,
|
|
/* ...or there are two separate IRQ for RX and TX */
|
|
UART_RX_IRQ = 0,
|
|
UART_TX_IRQ,
|
|
UART_IRQ_COUNT
|
|
};
|
|
|
|
/* Diverging register offsets */
|
|
struct uart_regs_layout {
|
|
unsigned int rbr;
|
|
unsigned int tsh;
|
|
unsigned int ctrl;
|
|
unsigned int intr;
|
|
};
|
|
|
|
/* Diverging flags */
|
|
struct uart_flags {
|
|
unsigned int ctrl_tx_rdy_int;
|
|
unsigned int ctrl_rx_rdy_int;
|
|
unsigned int stat_tx_rdy;
|
|
unsigned int stat_rx_rdy;
|
|
};
|
|
|
|
/* Driver data, a structure for each UART port */
|
|
struct mvebu_uart_driver_data {
|
|
bool is_ext;
|
|
struct uart_regs_layout regs;
|
|
struct uart_flags flags;
|
|
};
|
|
|
|
/* Saved registers during suspend */
|
|
struct mvebu_uart_pm_regs {
|
|
unsigned int rbr;
|
|
unsigned int tsh;
|
|
unsigned int ctrl;
|
|
unsigned int intr;
|
|
unsigned int stat;
|
|
unsigned int brdv;
|
|
unsigned int osamp;
|
|
};
|
|
|
|
/* MVEBU UART driver structure */
|
|
struct mvebu_uart {
|
|
struct uart_port *port;
|
|
struct clk *clk;
|
|
int irq[UART_IRQ_COUNT];
|
|
struct mvebu_uart_driver_data *data;
|
|
#if defined(CONFIG_PM)
|
|
struct mvebu_uart_pm_regs pm_regs;
|
|
#endif /* CONFIG_PM */
|
|
};
|
|
|
|
static struct mvebu_uart *to_mvuart(struct uart_port *port)
|
|
{
|
|
return (struct mvebu_uart *)port->private_data;
|
|
}
|
|
|
|
#define IS_EXTENDED(port) (to_mvuart(port)->data->is_ext)
|
|
|
|
#define UART_RBR(port) (to_mvuart(port)->data->regs.rbr)
|
|
#define UART_TSH(port) (to_mvuart(port)->data->regs.tsh)
|
|
#define UART_CTRL(port) (to_mvuart(port)->data->regs.ctrl)
|
|
#define UART_INTR(port) (to_mvuart(port)->data->regs.intr)
|
|
|
|
#define CTRL_TX_RDY_INT(port) (to_mvuart(port)->data->flags.ctrl_tx_rdy_int)
|
|
#define CTRL_RX_RDY_INT(port) (to_mvuart(port)->data->flags.ctrl_rx_rdy_int)
|
|
#define STAT_TX_RDY(port) (to_mvuart(port)->data->flags.stat_tx_rdy)
|
|
#define STAT_RX_RDY(port) (to_mvuart(port)->data->flags.stat_rx_rdy)
|
|
|
|
static struct uart_port mvebu_uart_ports[MVEBU_NR_UARTS];
|
|
|
|
static DEFINE_SPINLOCK(mvebu_uart_lock);
|
|
|
|
/* Core UART Driver Operations */
|
|
static unsigned int mvebu_uart_tx_empty(struct uart_port *port)
|
|
{
|
|
unsigned long flags;
|
|
unsigned int st;
|
|
|
|
spin_lock_irqsave(&port->lock, flags);
|
|
st = readl(port->membase + UART_STAT);
|
|
spin_unlock_irqrestore(&port->lock, flags);
|
|
|
|
return (st & STAT_TX_EMP) ? TIOCSER_TEMT : 0;
|
|
}
|
|
|
|
static unsigned int mvebu_uart_get_mctrl(struct uart_port *port)
|
|
{
|
|
return TIOCM_CTS | TIOCM_DSR | TIOCM_CAR;
|
|
}
|
|
|
|
static void mvebu_uart_set_mctrl(struct uart_port *port,
|
|
unsigned int mctrl)
|
|
{
|
|
/*
|
|
* Even if we do not support configuring the modem control lines, this
|
|
* function must be proided to the serial core
|
|
*/
|
|
}
|
|
|
|
static void mvebu_uart_stop_tx(struct uart_port *port)
|
|
{
|
|
unsigned int ctl = readl(port->membase + UART_INTR(port));
|
|
|
|
ctl &= ~CTRL_TX_RDY_INT(port);
|
|
writel(ctl, port->membase + UART_INTR(port));
|
|
}
|
|
|
|
static void mvebu_uart_start_tx(struct uart_port *port)
|
|
{
|
|
unsigned int ctl;
|
|
struct circ_buf *xmit = &port->state->xmit;
|
|
|
|
if (IS_EXTENDED(port) && !uart_circ_empty(xmit)) {
|
|
writel(xmit->buf[xmit->tail], port->membase + UART_TSH(port));
|
|
xmit->tail = (xmit->tail + 1) & (UART_XMIT_SIZE - 1);
|
|
port->icount.tx++;
|
|
}
|
|
|
|
ctl = readl(port->membase + UART_INTR(port));
|
|
ctl |= CTRL_TX_RDY_INT(port);
|
|
writel(ctl, port->membase + UART_INTR(port));
|
|
}
|
|
|
|
static void mvebu_uart_stop_rx(struct uart_port *port)
|
|
{
|
|
unsigned int ctl;
|
|
|
|
ctl = readl(port->membase + UART_CTRL(port));
|
|
ctl &= ~CTRL_BRK_INT;
|
|
writel(ctl, port->membase + UART_CTRL(port));
|
|
|
|
ctl = readl(port->membase + UART_INTR(port));
|
|
ctl &= ~CTRL_RX_RDY_INT(port);
|
|
writel(ctl, port->membase + UART_INTR(port));
|
|
}
|
|
|
|
static void mvebu_uart_break_ctl(struct uart_port *port, int brk)
|
|
{
|
|
unsigned int ctl;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&port->lock, flags);
|
|
ctl = readl(port->membase + UART_CTRL(port));
|
|
if (brk == -1)
|
|
ctl |= CTRL_SND_BRK_SEQ;
|
|
else
|
|
ctl &= ~CTRL_SND_BRK_SEQ;
|
|
writel(ctl, port->membase + UART_CTRL(port));
|
|
spin_unlock_irqrestore(&port->lock, flags);
|
|
}
|
|
|
|
static void mvebu_uart_rx_chars(struct uart_port *port, unsigned int status)
|
|
{
|
|
struct tty_port *tport = &port->state->port;
|
|
unsigned char ch = 0;
|
|
char flag = 0;
|
|
int ret;
|
|
|
|
do {
|
|
if (status & STAT_RX_RDY(port)) {
|
|
ch = readl(port->membase + UART_RBR(port));
|
|
ch &= 0xff;
|
|
flag = TTY_NORMAL;
|
|
port->icount.rx++;
|
|
|
|
if (status & STAT_PAR_ERR)
|
|
port->icount.parity++;
|
|
}
|
|
|
|
/*
|
|
* For UART2, error bits are not cleared on buffer read.
|
|
* This causes interrupt loop and system hang.
|
|
*/
|
|
if (IS_EXTENDED(port) && (status & STAT_BRK_ERR)) {
|
|
ret = readl(port->membase + UART_STAT);
|
|
ret |= STAT_BRK_ERR;
|
|
writel(ret, port->membase + UART_STAT);
|
|
}
|
|
|
|
if (status & STAT_BRK_DET) {
|
|
port->icount.brk++;
|
|
status &= ~(STAT_FRM_ERR | STAT_PAR_ERR);
|
|
if (uart_handle_break(port))
|
|
goto ignore_char;
|
|
}
|
|
|
|
if (status & STAT_OVR_ERR)
|
|
port->icount.overrun++;
|
|
|
|
if (status & STAT_FRM_ERR)
|
|
port->icount.frame++;
|
|
|
|
if (uart_handle_sysrq_char(port, ch))
|
|
goto ignore_char;
|
|
|
|
if (status & port->ignore_status_mask & STAT_PAR_ERR)
|
|
status &= ~STAT_RX_RDY(port);
|
|
|
|
status &= port->read_status_mask;
|
|
|
|
if (status & STAT_PAR_ERR)
|
|
flag = TTY_PARITY;
|
|
|
|
status &= ~port->ignore_status_mask;
|
|
|
|
if (status & STAT_RX_RDY(port))
|
|
tty_insert_flip_char(tport, ch, flag);
|
|
|
|
if (status & STAT_BRK_DET)
|
|
tty_insert_flip_char(tport, 0, TTY_BREAK);
|
|
|
|
if (status & STAT_FRM_ERR)
|
|
tty_insert_flip_char(tport, 0, TTY_FRAME);
|
|
|
|
if (status & STAT_OVR_ERR)
|
|
tty_insert_flip_char(tport, 0, TTY_OVERRUN);
|
|
|
|
ignore_char:
|
|
status = readl(port->membase + UART_STAT);
|
|
} while (status & (STAT_RX_RDY(port) | STAT_BRK_DET));
|
|
|
|
tty_flip_buffer_push(tport);
|
|
}
|
|
|
|
static void mvebu_uart_tx_chars(struct uart_port *port, unsigned int status)
|
|
{
|
|
struct circ_buf *xmit = &port->state->xmit;
|
|
unsigned int count;
|
|
unsigned int st;
|
|
|
|
if (port->x_char) {
|
|
writel(port->x_char, port->membase + UART_TSH(port));
|
|
port->icount.tx++;
|
|
port->x_char = 0;
|
|
return;
|
|
}
|
|
|
|
if (uart_circ_empty(xmit) || uart_tx_stopped(port)) {
|
|
mvebu_uart_stop_tx(port);
|
|
return;
|
|
}
|
|
|
|
for (count = 0; count < port->fifosize; count++) {
|
|
writel(xmit->buf[xmit->tail], port->membase + UART_TSH(port));
|
|
xmit->tail = (xmit->tail + 1) & (UART_XMIT_SIZE - 1);
|
|
port->icount.tx++;
|
|
|
|
if (uart_circ_empty(xmit))
|
|
break;
|
|
|
|
st = readl(port->membase + UART_STAT);
|
|
if (st & STAT_TX_FIFO_FUL)
|
|
break;
|
|
}
|
|
|
|
if (uart_circ_chars_pending(xmit) < WAKEUP_CHARS)
|
|
uart_write_wakeup(port);
|
|
|
|
if (uart_circ_empty(xmit))
|
|
mvebu_uart_stop_tx(port);
|
|
}
|
|
|
|
static irqreturn_t mvebu_uart_isr(int irq, void *dev_id)
|
|
{
|
|
struct uart_port *port = (struct uart_port *)dev_id;
|
|
unsigned int st = readl(port->membase + UART_STAT);
|
|
|
|
if (st & (STAT_RX_RDY(port) | STAT_OVR_ERR | STAT_FRM_ERR |
|
|
STAT_BRK_DET))
|
|
mvebu_uart_rx_chars(port, st);
|
|
|
|
if (st & STAT_TX_RDY(port))
|
|
mvebu_uart_tx_chars(port, st);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static irqreturn_t mvebu_uart_rx_isr(int irq, void *dev_id)
|
|
{
|
|
struct uart_port *port = (struct uart_port *)dev_id;
|
|
unsigned int st = readl(port->membase + UART_STAT);
|
|
|
|
if (st & (STAT_RX_RDY(port) | STAT_OVR_ERR | STAT_FRM_ERR |
|
|
STAT_BRK_DET))
|
|
mvebu_uart_rx_chars(port, st);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static irqreturn_t mvebu_uart_tx_isr(int irq, void *dev_id)
|
|
{
|
|
struct uart_port *port = (struct uart_port *)dev_id;
|
|
unsigned int st = readl(port->membase + UART_STAT);
|
|
|
|
if (st & STAT_TX_RDY(port))
|
|
mvebu_uart_tx_chars(port, st);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static int mvebu_uart_startup(struct uart_port *port)
|
|
{
|
|
struct mvebu_uart *mvuart = to_mvuart(port);
|
|
unsigned int ctl;
|
|
int ret;
|
|
|
|
writel(CTRL_TXFIFO_RST | CTRL_RXFIFO_RST,
|
|
port->membase + UART_CTRL(port));
|
|
udelay(1);
|
|
|
|
/* Clear the error bits of state register before IRQ request */
|
|
ret = readl(port->membase + UART_STAT);
|
|
ret |= STAT_BRK_ERR;
|
|
writel(ret, port->membase + UART_STAT);
|
|
|
|
writel(CTRL_BRK_INT, port->membase + UART_CTRL(port));
|
|
|
|
ctl = readl(port->membase + UART_INTR(port));
|
|
ctl |= CTRL_RX_RDY_INT(port);
|
|
writel(ctl, port->membase + UART_INTR(port));
|
|
|
|
if (!mvuart->irq[UART_TX_IRQ]) {
|
|
/* Old bindings with just one interrupt (UART0 only) */
|
|
ret = devm_request_irq(port->dev, mvuart->irq[UART_IRQ_SUM],
|
|
mvebu_uart_isr, port->irqflags,
|
|
dev_name(port->dev), port);
|
|
if (ret) {
|
|
dev_err(port->dev, "unable to request IRQ %d\n",
|
|
mvuart->irq[UART_IRQ_SUM]);
|
|
return ret;
|
|
}
|
|
} else {
|
|
/* New bindings with an IRQ for RX and TX (both UART) */
|
|
ret = devm_request_irq(port->dev, mvuart->irq[UART_RX_IRQ],
|
|
mvebu_uart_rx_isr, port->irqflags,
|
|
dev_name(port->dev), port);
|
|
if (ret) {
|
|
dev_err(port->dev, "unable to request IRQ %d\n",
|
|
mvuart->irq[UART_RX_IRQ]);
|
|
return ret;
|
|
}
|
|
|
|
ret = devm_request_irq(port->dev, mvuart->irq[UART_TX_IRQ],
|
|
mvebu_uart_tx_isr, port->irqflags,
|
|
dev_name(port->dev),
|
|
port);
|
|
if (ret) {
|
|
dev_err(port->dev, "unable to request IRQ %d\n",
|
|
mvuart->irq[UART_TX_IRQ]);
|
|
devm_free_irq(port->dev, mvuart->irq[UART_RX_IRQ],
|
|
port);
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void mvebu_uart_shutdown(struct uart_port *port)
|
|
{
|
|
struct mvebu_uart *mvuart = to_mvuart(port);
|
|
|
|
writel(0, port->membase + UART_INTR(port));
|
|
|
|
if (!mvuart->irq[UART_TX_IRQ]) {
|
|
devm_free_irq(port->dev, mvuart->irq[UART_IRQ_SUM], port);
|
|
} else {
|
|
devm_free_irq(port->dev, mvuart->irq[UART_RX_IRQ], port);
|
|
devm_free_irq(port->dev, mvuart->irq[UART_TX_IRQ], port);
|
|
}
|
|
}
|
|
|
|
static unsigned int mvebu_uart_baud_rate_set(struct uart_port *port, unsigned int baud)
|
|
{
|
|
unsigned int d_divisor, m_divisor;
|
|
unsigned long flags;
|
|
u32 brdv, osamp;
|
|
|
|
if (!port->uartclk)
|
|
return 0;
|
|
|
|
/*
|
|
* The baudrate is derived from the UART clock thanks to divisors:
|
|
* > d1 * d2 ("TBG divisors"): can divide only TBG clock from 1 to 6
|
|
* > D ("baud generator"): can divide the clock from 1 to 1023
|
|
* > M ("fractional divisor"): allows a better accuracy (from 1 to 63)
|
|
*
|
|
* Exact formulas for calculating baudrate:
|
|
*
|
|
* with default x16 scheme:
|
|
* baudrate = xtal / (d * 16)
|
|
* baudrate = tbg / (d1 * d2 * d * 16)
|
|
*
|
|
* with fractional divisor:
|
|
* baudrate = 10 * xtal / (d * (3 * (m1 + m2) + 2 * (m3 + m4)))
|
|
* baudrate = 10 * tbg / (d1*d2 * d * (3 * (m1 + m2) + 2 * (m3 + m4)))
|
|
*
|
|
* Oversampling value:
|
|
* osamp = (m1 << 0) | (m2 << 8) | (m3 << 16) | (m4 << 24);
|
|
*
|
|
* Where m1 controls number of clock cycles per bit for bits 1,2,3;
|
|
* m2 for bits 4,5,6; m3 for bits 7,8 and m4 for bits 9,10.
|
|
*
|
|
* To simplify baudrate setup set all the M prescalers to the same
|
|
* value. For baudrates 9600 Bd and higher, it is enough to use the
|
|
* default (x16) divisor or fractional divisor with M = 63, so there
|
|
* is no need to use real fractional support (where the M prescalers
|
|
* are not equal).
|
|
*
|
|
* When all the M prescalers are zeroed then default (x16) divisor is
|
|
* used. Default x16 scheme is more stable than M (fractional divisor),
|
|
* so use M only when D divisor is not enough to derive baudrate.
|
|
*
|
|
* Member port->uartclk is either xtal clock rate or TBG clock rate
|
|
* divided by (d1 * d2). So d1 and d2 are already set by the UART clock
|
|
* driver (and UART driver itself cannot change them). Moreover they are
|
|
* shared between both UARTs.
|
|
*/
|
|
|
|
m_divisor = OSAMP_DEFAULT_DIVISOR;
|
|
d_divisor = DIV_ROUND_CLOSEST(port->uartclk, baud * m_divisor);
|
|
|
|
if (d_divisor > BRDV_BAUD_MAX) {
|
|
/*
|
|
* Experiments show that small M divisors are unstable.
|
|
* Use maximal possible M = 63 and calculate D divisor.
|
|
*/
|
|
m_divisor = OSAMP_MAX_DIVISOR;
|
|
d_divisor = DIV_ROUND_CLOSEST(port->uartclk, baud * m_divisor);
|
|
}
|
|
|
|
if (d_divisor < 1)
|
|
d_divisor = 1;
|
|
else if (d_divisor > BRDV_BAUD_MAX)
|
|
d_divisor = BRDV_BAUD_MAX;
|
|
|
|
spin_lock_irqsave(&mvebu_uart_lock, flags);
|
|
brdv = readl(port->membase + UART_BRDV);
|
|
brdv &= ~BRDV_BAUD_MASK;
|
|
brdv |= d_divisor;
|
|
writel(brdv, port->membase + UART_BRDV);
|
|
spin_unlock_irqrestore(&mvebu_uart_lock, flags);
|
|
|
|
osamp = readl(port->membase + UART_OSAMP);
|
|
osamp &= ~OSAMP_DIVISORS_MASK;
|
|
if (m_divisor != OSAMP_DEFAULT_DIVISOR)
|
|
osamp |= (m_divisor << 0) | (m_divisor << 8) |
|
|
(m_divisor << 16) | (m_divisor << 24);
|
|
writel(osamp, port->membase + UART_OSAMP);
|
|
|
|
return DIV_ROUND_CLOSEST(port->uartclk, d_divisor * m_divisor);
|
|
}
|
|
|
|
static void mvebu_uart_set_termios(struct uart_port *port,
|
|
struct ktermios *termios,
|
|
struct ktermios *old)
|
|
{
|
|
unsigned long flags;
|
|
unsigned int baud, min_baud, max_baud;
|
|
|
|
spin_lock_irqsave(&port->lock, flags);
|
|
|
|
port->read_status_mask = STAT_RX_RDY(port) | STAT_OVR_ERR |
|
|
STAT_TX_RDY(port) | STAT_TX_FIFO_FUL;
|
|
|
|
if (termios->c_iflag & INPCK)
|
|
port->read_status_mask |= STAT_FRM_ERR | STAT_PAR_ERR;
|
|
|
|
port->ignore_status_mask = 0;
|
|
if (termios->c_iflag & IGNPAR)
|
|
port->ignore_status_mask |=
|
|
STAT_FRM_ERR | STAT_PAR_ERR | STAT_OVR_ERR;
|
|
|
|
if ((termios->c_cflag & CREAD) == 0)
|
|
port->ignore_status_mask |= STAT_RX_RDY(port) | STAT_BRK_ERR;
|
|
|
|
/*
|
|
* Maximal divisor is 1023 and maximal fractional divisor is 63. And
|
|
* experiments show that baudrates above 1/80 of parent clock rate are
|
|
* not stable. So disallow baudrates above 1/80 of the parent clock
|
|
* rate. If port->uartclk is not available, then
|
|
* mvebu_uart_baud_rate_set() fails, so values min_baud and max_baud
|
|
* in this case do not matter.
|
|
*/
|
|
min_baud = DIV_ROUND_UP(port->uartclk, BRDV_BAUD_MAX *
|
|
OSAMP_MAX_DIVISOR);
|
|
max_baud = port->uartclk / 80;
|
|
|
|
baud = uart_get_baud_rate(port, termios, old, min_baud, max_baud);
|
|
baud = mvebu_uart_baud_rate_set(port, baud);
|
|
|
|
/* In case baudrate cannot be changed, report previous old value */
|
|
if (baud == 0 && old)
|
|
baud = tty_termios_baud_rate(old);
|
|
|
|
/* Only the following flag changes are supported */
|
|
if (old) {
|
|
termios->c_iflag &= INPCK | IGNPAR;
|
|
termios->c_iflag |= old->c_iflag & ~(INPCK | IGNPAR);
|
|
termios->c_cflag &= CREAD | CBAUD;
|
|
termios->c_cflag |= old->c_cflag & ~(CREAD | CBAUD);
|
|
termios->c_cflag |= CS8;
|
|
}
|
|
|
|
if (baud != 0) {
|
|
tty_termios_encode_baud_rate(termios, baud, baud);
|
|
uart_update_timeout(port, termios->c_cflag, baud);
|
|
}
|
|
|
|
spin_unlock_irqrestore(&port->lock, flags);
|
|
}
|
|
|
|
static const char *mvebu_uart_type(struct uart_port *port)
|
|
{
|
|
return MVEBU_UART_TYPE;
|
|
}
|
|
|
|
static void mvebu_uart_release_port(struct uart_port *port)
|
|
{
|
|
/* Nothing to do here */
|
|
}
|
|
|
|
static int mvebu_uart_request_port(struct uart_port *port)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
#ifdef CONFIG_CONSOLE_POLL
|
|
static int mvebu_uart_get_poll_char(struct uart_port *port)
|
|
{
|
|
unsigned int st = readl(port->membase + UART_STAT);
|
|
|
|
if (!(st & STAT_RX_RDY(port)))
|
|
return NO_POLL_CHAR;
|
|
|
|
return readl(port->membase + UART_RBR(port));
|
|
}
|
|
|
|
static void mvebu_uart_put_poll_char(struct uart_port *port, unsigned char c)
|
|
{
|
|
unsigned int st;
|
|
|
|
for (;;) {
|
|
st = readl(port->membase + UART_STAT);
|
|
|
|
if (!(st & STAT_TX_FIFO_FUL))
|
|
break;
|
|
|
|
udelay(1);
|
|
}
|
|
|
|
writel(c, port->membase + UART_TSH(port));
|
|
}
|
|
#endif
|
|
|
|
static const struct uart_ops mvebu_uart_ops = {
|
|
.tx_empty = mvebu_uart_tx_empty,
|
|
.set_mctrl = mvebu_uart_set_mctrl,
|
|
.get_mctrl = mvebu_uart_get_mctrl,
|
|
.stop_tx = mvebu_uart_stop_tx,
|
|
.start_tx = mvebu_uart_start_tx,
|
|
.stop_rx = mvebu_uart_stop_rx,
|
|
.break_ctl = mvebu_uart_break_ctl,
|
|
.startup = mvebu_uart_startup,
|
|
.shutdown = mvebu_uart_shutdown,
|
|
.set_termios = mvebu_uart_set_termios,
|
|
.type = mvebu_uart_type,
|
|
.release_port = mvebu_uart_release_port,
|
|
.request_port = mvebu_uart_request_port,
|
|
#ifdef CONFIG_CONSOLE_POLL
|
|
.poll_get_char = mvebu_uart_get_poll_char,
|
|
.poll_put_char = mvebu_uart_put_poll_char,
|
|
#endif
|
|
};
|
|
|
|
/* Console Driver Operations */
|
|
|
|
#ifdef CONFIG_SERIAL_MVEBU_CONSOLE
|
|
/* Early Console */
|
|
static void mvebu_uart_putc(struct uart_port *port, unsigned char c)
|
|
{
|
|
unsigned int st;
|
|
|
|
for (;;) {
|
|
st = readl(port->membase + UART_STAT);
|
|
if (!(st & STAT_TX_FIFO_FUL))
|
|
break;
|
|
}
|
|
|
|
/* At early stage, DT is not parsed yet, only use UART0 */
|
|
writel(c, port->membase + UART_STD_TSH);
|
|
|
|
for (;;) {
|
|
st = readl(port->membase + UART_STAT);
|
|
if (st & STAT_TX_FIFO_EMP)
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void mvebu_uart_putc_early_write(struct console *con,
|
|
const char *s,
|
|
unsigned int n)
|
|
{
|
|
struct earlycon_device *dev = con->data;
|
|
|
|
uart_console_write(&dev->port, s, n, mvebu_uart_putc);
|
|
}
|
|
|
|
static int __init
|
|
mvebu_uart_early_console_setup(struct earlycon_device *device,
|
|
const char *opt)
|
|
{
|
|
if (!device->port.membase)
|
|
return -ENODEV;
|
|
|
|
device->con->write = mvebu_uart_putc_early_write;
|
|
|
|
return 0;
|
|
}
|
|
|
|
EARLYCON_DECLARE(ar3700_uart, mvebu_uart_early_console_setup);
|
|
OF_EARLYCON_DECLARE(ar3700_uart, "marvell,armada-3700-uart",
|
|
mvebu_uart_early_console_setup);
|
|
|
|
static void wait_for_xmitr(struct uart_port *port)
|
|
{
|
|
u32 val;
|
|
|
|
readl_poll_timeout_atomic(port->membase + UART_STAT, val,
|
|
(val & STAT_TX_RDY(port)), 1, 10000);
|
|
}
|
|
|
|
static void wait_for_xmite(struct uart_port *port)
|
|
{
|
|
u32 val;
|
|
|
|
readl_poll_timeout_atomic(port->membase + UART_STAT, val,
|
|
(val & STAT_TX_EMP), 1, 10000);
|
|
}
|
|
|
|
static void mvebu_uart_console_putchar(struct uart_port *port, unsigned char ch)
|
|
{
|
|
wait_for_xmitr(port);
|
|
writel(ch, port->membase + UART_TSH(port));
|
|
}
|
|
|
|
static void mvebu_uart_console_write(struct console *co, const char *s,
|
|
unsigned int count)
|
|
{
|
|
struct uart_port *port = &mvebu_uart_ports[co->index];
|
|
unsigned long flags;
|
|
unsigned int ier, intr, ctl;
|
|
int locked = 1;
|
|
|
|
if (oops_in_progress)
|
|
locked = spin_trylock_irqsave(&port->lock, flags);
|
|
else
|
|
spin_lock_irqsave(&port->lock, flags);
|
|
|
|
ier = readl(port->membase + UART_CTRL(port)) & CTRL_BRK_INT;
|
|
intr = readl(port->membase + UART_INTR(port)) &
|
|
(CTRL_RX_RDY_INT(port) | CTRL_TX_RDY_INT(port));
|
|
writel(0, port->membase + UART_CTRL(port));
|
|
writel(0, port->membase + UART_INTR(port));
|
|
|
|
uart_console_write(port, s, count, mvebu_uart_console_putchar);
|
|
|
|
wait_for_xmite(port);
|
|
|
|
if (ier)
|
|
writel(ier, port->membase + UART_CTRL(port));
|
|
|
|
if (intr) {
|
|
ctl = intr | readl(port->membase + UART_INTR(port));
|
|
writel(ctl, port->membase + UART_INTR(port));
|
|
}
|
|
|
|
if (locked)
|
|
spin_unlock_irqrestore(&port->lock, flags);
|
|
}
|
|
|
|
static int mvebu_uart_console_setup(struct console *co, char *options)
|
|
{
|
|
struct uart_port *port;
|
|
int baud = 9600;
|
|
int bits = 8;
|
|
int parity = 'n';
|
|
int flow = 'n';
|
|
|
|
if (co->index < 0 || co->index >= MVEBU_NR_UARTS)
|
|
return -EINVAL;
|
|
|
|
port = &mvebu_uart_ports[co->index];
|
|
|
|
if (!port->mapbase || !port->membase) {
|
|
pr_debug("console on ttyMV%i not present\n", co->index);
|
|
return -ENODEV;
|
|
}
|
|
|
|
if (options)
|
|
uart_parse_options(options, &baud, &parity, &bits, &flow);
|
|
|
|
return uart_set_options(port, co, baud, parity, bits, flow);
|
|
}
|
|
|
|
static struct uart_driver mvebu_uart_driver;
|
|
|
|
static struct console mvebu_uart_console = {
|
|
.name = "ttyMV",
|
|
.write = mvebu_uart_console_write,
|
|
.device = uart_console_device,
|
|
.setup = mvebu_uart_console_setup,
|
|
.flags = CON_PRINTBUFFER,
|
|
.index = -1,
|
|
.data = &mvebu_uart_driver,
|
|
};
|
|
|
|
static int __init mvebu_uart_console_init(void)
|
|
{
|
|
register_console(&mvebu_uart_console);
|
|
return 0;
|
|
}
|
|
|
|
console_initcall(mvebu_uart_console_init);
|
|
|
|
|
|
#endif /* CONFIG_SERIAL_MVEBU_CONSOLE */
|
|
|
|
static struct uart_driver mvebu_uart_driver = {
|
|
.owner = THIS_MODULE,
|
|
.driver_name = DRIVER_NAME,
|
|
.dev_name = "ttyMV",
|
|
.nr = MVEBU_NR_UARTS,
|
|
#ifdef CONFIG_SERIAL_MVEBU_CONSOLE
|
|
.cons = &mvebu_uart_console,
|
|
#endif
|
|
};
|
|
|
|
#if defined(CONFIG_PM)
|
|
static int mvebu_uart_suspend(struct device *dev)
|
|
{
|
|
struct mvebu_uart *mvuart = dev_get_drvdata(dev);
|
|
struct uart_port *port = mvuart->port;
|
|
unsigned long flags;
|
|
|
|
uart_suspend_port(&mvebu_uart_driver, port);
|
|
|
|
mvuart->pm_regs.rbr = readl(port->membase + UART_RBR(port));
|
|
mvuart->pm_regs.tsh = readl(port->membase + UART_TSH(port));
|
|
mvuart->pm_regs.ctrl = readl(port->membase + UART_CTRL(port));
|
|
mvuart->pm_regs.intr = readl(port->membase + UART_INTR(port));
|
|
mvuart->pm_regs.stat = readl(port->membase + UART_STAT);
|
|
spin_lock_irqsave(&mvebu_uart_lock, flags);
|
|
mvuart->pm_regs.brdv = readl(port->membase + UART_BRDV);
|
|
spin_unlock_irqrestore(&mvebu_uart_lock, flags);
|
|
mvuart->pm_regs.osamp = readl(port->membase + UART_OSAMP);
|
|
|
|
device_set_wakeup_enable(dev, true);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int mvebu_uart_resume(struct device *dev)
|
|
{
|
|
struct mvebu_uart *mvuart = dev_get_drvdata(dev);
|
|
struct uart_port *port = mvuart->port;
|
|
unsigned long flags;
|
|
|
|
writel(mvuart->pm_regs.rbr, port->membase + UART_RBR(port));
|
|
writel(mvuart->pm_regs.tsh, port->membase + UART_TSH(port));
|
|
writel(mvuart->pm_regs.ctrl, port->membase + UART_CTRL(port));
|
|
writel(mvuart->pm_regs.intr, port->membase + UART_INTR(port));
|
|
writel(mvuart->pm_regs.stat, port->membase + UART_STAT);
|
|
spin_lock_irqsave(&mvebu_uart_lock, flags);
|
|
writel(mvuart->pm_regs.brdv, port->membase + UART_BRDV);
|
|
spin_unlock_irqrestore(&mvebu_uart_lock, flags);
|
|
writel(mvuart->pm_regs.osamp, port->membase + UART_OSAMP);
|
|
|
|
uart_resume_port(&mvebu_uart_driver, port);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct dev_pm_ops mvebu_uart_pm_ops = {
|
|
.suspend = mvebu_uart_suspend,
|
|
.resume = mvebu_uart_resume,
|
|
};
|
|
#endif /* CONFIG_PM */
|
|
|
|
static const struct of_device_id mvebu_uart_of_match[];
|
|
|
|
/* Counter to keep track of each UART port id when not using CONFIG_OF */
|
|
static int uart_num_counter;
|
|
|
|
static int mvebu_uart_probe(struct platform_device *pdev)
|
|
{
|
|
struct resource *reg = platform_get_resource(pdev, IORESOURCE_MEM, 0);
|
|
const struct of_device_id *match = of_match_device(mvebu_uart_of_match,
|
|
&pdev->dev);
|
|
struct uart_port *port;
|
|
struct mvebu_uart *mvuart;
|
|
int id, irq;
|
|
|
|
if (!reg) {
|
|
dev_err(&pdev->dev, "no registers defined\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* Assume that all UART ports have a DT alias or none has */
|
|
id = of_alias_get_id(pdev->dev.of_node, "serial");
|
|
if (!pdev->dev.of_node || id < 0)
|
|
pdev->id = uart_num_counter++;
|
|
else
|
|
pdev->id = id;
|
|
|
|
if (pdev->id >= MVEBU_NR_UARTS) {
|
|
dev_err(&pdev->dev, "cannot have more than %d UART ports\n",
|
|
MVEBU_NR_UARTS);
|
|
return -EINVAL;
|
|
}
|
|
|
|
port = &mvebu_uart_ports[pdev->id];
|
|
|
|
spin_lock_init(&port->lock);
|
|
|
|
port->dev = &pdev->dev;
|
|
port->type = PORT_MVEBU;
|
|
port->ops = &mvebu_uart_ops;
|
|
port->regshift = 0;
|
|
|
|
port->fifosize = 32;
|
|
port->iotype = UPIO_MEM32;
|
|
port->flags = UPF_FIXED_PORT;
|
|
port->line = pdev->id;
|
|
|
|
/*
|
|
* IRQ number is not stored in this structure because we may have two of
|
|
* them per port (RX and TX). Instead, use the driver UART structure
|
|
* array so called ->irq[].
|
|
*/
|
|
port->irq = 0;
|
|
port->irqflags = 0;
|
|
port->mapbase = reg->start;
|
|
|
|
port->membase = devm_ioremap_resource(&pdev->dev, reg);
|
|
if (IS_ERR(port->membase))
|
|
return PTR_ERR(port->membase);
|
|
|
|
mvuart = devm_kzalloc(&pdev->dev, sizeof(struct mvebu_uart),
|
|
GFP_KERNEL);
|
|
if (!mvuart)
|
|
return -ENOMEM;
|
|
|
|
/* Get controller data depending on the compatible string */
|
|
mvuart->data = (struct mvebu_uart_driver_data *)match->data;
|
|
mvuart->port = port;
|
|
|
|
port->private_data = mvuart;
|
|
platform_set_drvdata(pdev, mvuart);
|
|
|
|
/* Get fixed clock frequency */
|
|
mvuart->clk = devm_clk_get(&pdev->dev, NULL);
|
|
if (IS_ERR(mvuart->clk)) {
|
|
if (PTR_ERR(mvuart->clk) == -EPROBE_DEFER)
|
|
return PTR_ERR(mvuart->clk);
|
|
|
|
if (IS_EXTENDED(port)) {
|
|
dev_err(&pdev->dev, "unable to get UART clock\n");
|
|
return PTR_ERR(mvuart->clk);
|
|
}
|
|
} else {
|
|
if (!clk_prepare_enable(mvuart->clk))
|
|
port->uartclk = clk_get_rate(mvuart->clk);
|
|
}
|
|
|
|
/* Manage interrupts */
|
|
if (platform_irq_count(pdev) == 1) {
|
|
/* Old bindings: no name on the single unamed UART0 IRQ */
|
|
irq = platform_get_irq(pdev, 0);
|
|
if (irq < 0)
|
|
return irq;
|
|
|
|
mvuart->irq[UART_IRQ_SUM] = irq;
|
|
} else {
|
|
/*
|
|
* New bindings: named interrupts (RX, TX) for both UARTS,
|
|
* only make use of uart-rx and uart-tx interrupts, do not use
|
|
* uart-sum of UART0 port.
|
|
*/
|
|
irq = platform_get_irq_byname(pdev, "uart-rx");
|
|
if (irq < 0)
|
|
return irq;
|
|
|
|
mvuart->irq[UART_RX_IRQ] = irq;
|
|
|
|
irq = platform_get_irq_byname(pdev, "uart-tx");
|
|
if (irq < 0)
|
|
return irq;
|
|
|
|
mvuart->irq[UART_TX_IRQ] = irq;
|
|
}
|
|
|
|
/* UART Soft Reset*/
|
|
writel(CTRL_SOFT_RST, port->membase + UART_CTRL(port));
|
|
udelay(1);
|
|
writel(0, port->membase + UART_CTRL(port));
|
|
|
|
return uart_add_one_port(&mvebu_uart_driver, port);
|
|
}
|
|
|
|
static struct mvebu_uart_driver_data uart_std_driver_data = {
|
|
.is_ext = false,
|
|
.regs.rbr = UART_STD_RBR,
|
|
.regs.tsh = UART_STD_TSH,
|
|
.regs.ctrl = UART_STD_CTRL1,
|
|
.regs.intr = UART_STD_CTRL2,
|
|
.flags.ctrl_tx_rdy_int = CTRL_STD_TX_RDY_INT,
|
|
.flags.ctrl_rx_rdy_int = CTRL_STD_RX_RDY_INT,
|
|
.flags.stat_tx_rdy = STAT_STD_TX_RDY,
|
|
.flags.stat_rx_rdy = STAT_STD_RX_RDY,
|
|
};
|
|
|
|
static struct mvebu_uart_driver_data uart_ext_driver_data = {
|
|
.is_ext = true,
|
|
.regs.rbr = UART_EXT_RBR,
|
|
.regs.tsh = UART_EXT_TSH,
|
|
.regs.ctrl = UART_EXT_CTRL1,
|
|
.regs.intr = UART_EXT_CTRL2,
|
|
.flags.ctrl_tx_rdy_int = CTRL_EXT_TX_RDY_INT,
|
|
.flags.ctrl_rx_rdy_int = CTRL_EXT_RX_RDY_INT,
|
|
.flags.stat_tx_rdy = STAT_EXT_TX_RDY,
|
|
.flags.stat_rx_rdy = STAT_EXT_RX_RDY,
|
|
};
|
|
|
|
/* Match table for of_platform binding */
|
|
static const struct of_device_id mvebu_uart_of_match[] = {
|
|
{
|
|
.compatible = "marvell,armada-3700-uart",
|
|
.data = (void *)&uart_std_driver_data,
|
|
},
|
|
{
|
|
.compatible = "marvell,armada-3700-uart-ext",
|
|
.data = (void *)&uart_ext_driver_data,
|
|
},
|
|
{}
|
|
};
|
|
|
|
static struct platform_driver mvebu_uart_platform_driver = {
|
|
.probe = mvebu_uart_probe,
|
|
.driver = {
|
|
.name = "mvebu-uart",
|
|
.of_match_table = of_match_ptr(mvebu_uart_of_match),
|
|
.suppress_bind_attrs = true,
|
|
#if defined(CONFIG_PM)
|
|
.pm = &mvebu_uart_pm_ops,
|
|
#endif /* CONFIG_PM */
|
|
},
|
|
};
|
|
|
|
/* This code is based on clk-fixed-factor.c driver and modified. */
|
|
|
|
struct mvebu_uart_clock {
|
|
struct clk_hw clk_hw;
|
|
int clock_idx;
|
|
u32 pm_context_reg1;
|
|
u32 pm_context_reg2;
|
|
};
|
|
|
|
struct mvebu_uart_clock_base {
|
|
struct mvebu_uart_clock clocks[2];
|
|
unsigned int parent_rates[5];
|
|
int parent_idx;
|
|
unsigned int div;
|
|
void __iomem *reg1;
|
|
void __iomem *reg2;
|
|
bool configured;
|
|
};
|
|
|
|
#define PARENT_CLOCK_XTAL 4
|
|
|
|
#define to_uart_clock(hw) container_of(hw, struct mvebu_uart_clock, clk_hw)
|
|
#define to_uart_clock_base(uart_clock) container_of(uart_clock, \
|
|
struct mvebu_uart_clock_base, clocks[uart_clock->clock_idx])
|
|
|
|
static int mvebu_uart_clock_prepare(struct clk_hw *hw)
|
|
{
|
|
struct mvebu_uart_clock *uart_clock = to_uart_clock(hw);
|
|
struct mvebu_uart_clock_base *uart_clock_base =
|
|
to_uart_clock_base(uart_clock);
|
|
unsigned int prev_clock_idx, prev_clock_rate, prev_d1d2;
|
|
unsigned int parent_clock_idx, parent_clock_rate;
|
|
unsigned long flags;
|
|
unsigned int d1, d2;
|
|
u64 divisor;
|
|
u32 val;
|
|
|
|
/*
|
|
* This function just reconfigures UART Clock Control register (located
|
|
* in UART1 address space which controls both UART1 and UART2) to
|
|
* selected UART base clock and recalculates current UART1/UART2
|
|
* divisors in their address spaces, so that final baudrate will not be
|
|
* changed by switching UART parent clock. This is required for
|
|
* otherwise kernel's boot log stops working - we need to ensure that
|
|
* UART baudrate does not change during this setup. It is a one time
|
|
* operation, it will execute only once and set `configured` to true,
|
|
* and be skipped on subsequent calls. Because this UART Clock Control
|
|
* register (UART_BRDV) is shared between UART1 baudrate function,
|
|
* UART1 clock selector and UART2 clock selector, every access to
|
|
* UART_BRDV (reg1) needs to be protected by a lock.
|
|
*/
|
|
|
|
spin_lock_irqsave(&mvebu_uart_lock, flags);
|
|
|
|
if (uart_clock_base->configured) {
|
|
spin_unlock_irqrestore(&mvebu_uart_lock, flags);
|
|
return 0;
|
|
}
|
|
|
|
parent_clock_idx = uart_clock_base->parent_idx;
|
|
parent_clock_rate = uart_clock_base->parent_rates[parent_clock_idx];
|
|
|
|
val = readl(uart_clock_base->reg1);
|
|
|
|
if (uart_clock_base->div > CLK_TBG_DIV1_MAX) {
|
|
d1 = CLK_TBG_DIV1_MAX;
|
|
d2 = uart_clock_base->div / CLK_TBG_DIV1_MAX;
|
|
} else {
|
|
d1 = uart_clock_base->div;
|
|
d2 = 1;
|
|
}
|
|
|
|
if (val & CLK_NO_XTAL) {
|
|
prev_clock_idx = (val >> CLK_TBG_SEL_SHIFT) & CLK_TBG_SEL_MASK;
|
|
prev_d1d2 = ((val >> CLK_TBG_DIV1_SHIFT) & CLK_TBG_DIV1_MASK) *
|
|
((val >> CLK_TBG_DIV2_SHIFT) & CLK_TBG_DIV2_MASK);
|
|
} else {
|
|
prev_clock_idx = PARENT_CLOCK_XTAL;
|
|
prev_d1d2 = 1;
|
|
}
|
|
|
|
/* Note that uart_clock_base->parent_rates[i] may not be available */
|
|
prev_clock_rate = uart_clock_base->parent_rates[prev_clock_idx];
|
|
|
|
/* Recalculate UART1 divisor so UART1 baudrate does not change */
|
|
if (prev_clock_rate) {
|
|
divisor = DIV_U64_ROUND_CLOSEST((u64)(val & BRDV_BAUD_MASK) *
|
|
parent_clock_rate * prev_d1d2,
|
|
prev_clock_rate * d1 * d2);
|
|
if (divisor < 1)
|
|
divisor = 1;
|
|
else if (divisor > BRDV_BAUD_MAX)
|
|
divisor = BRDV_BAUD_MAX;
|
|
val = (val & ~BRDV_BAUD_MASK) | divisor;
|
|
}
|
|
|
|
if (parent_clock_idx != PARENT_CLOCK_XTAL) {
|
|
/* Do not use XTAL, select TBG clock and TBG d1 * d2 divisors */
|
|
val |= CLK_NO_XTAL;
|
|
val &= ~(CLK_TBG_DIV1_MASK << CLK_TBG_DIV1_SHIFT);
|
|
val |= d1 << CLK_TBG_DIV1_SHIFT;
|
|
val &= ~(CLK_TBG_DIV2_MASK << CLK_TBG_DIV2_SHIFT);
|
|
val |= d2 << CLK_TBG_DIV2_SHIFT;
|
|
val &= ~(CLK_TBG_SEL_MASK << CLK_TBG_SEL_SHIFT);
|
|
val |= parent_clock_idx << CLK_TBG_SEL_SHIFT;
|
|
} else {
|
|
/* Use XTAL, TBG bits are then ignored */
|
|
val &= ~CLK_NO_XTAL;
|
|
}
|
|
|
|
writel(val, uart_clock_base->reg1);
|
|
|
|
/* Recalculate UART2 divisor so UART2 baudrate does not change */
|
|
if (prev_clock_rate) {
|
|
val = readl(uart_clock_base->reg2);
|
|
divisor = DIV_U64_ROUND_CLOSEST((u64)(val & BRDV_BAUD_MASK) *
|
|
parent_clock_rate * prev_d1d2,
|
|
prev_clock_rate * d1 * d2);
|
|
if (divisor < 1)
|
|
divisor = 1;
|
|
else if (divisor > BRDV_BAUD_MAX)
|
|
divisor = BRDV_BAUD_MAX;
|
|
val = (val & ~BRDV_BAUD_MASK) | divisor;
|
|
writel(val, uart_clock_base->reg2);
|
|
}
|
|
|
|
uart_clock_base->configured = true;
|
|
|
|
spin_unlock_irqrestore(&mvebu_uart_lock, flags);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int mvebu_uart_clock_enable(struct clk_hw *hw)
|
|
{
|
|
struct mvebu_uart_clock *uart_clock = to_uart_clock(hw);
|
|
struct mvebu_uart_clock_base *uart_clock_base =
|
|
to_uart_clock_base(uart_clock);
|
|
unsigned long flags;
|
|
u32 val;
|
|
|
|
spin_lock_irqsave(&mvebu_uart_lock, flags);
|
|
|
|
val = readl(uart_clock_base->reg1);
|
|
|
|
if (uart_clock->clock_idx == 0)
|
|
val &= ~UART1_CLK_DIS;
|
|
else
|
|
val &= ~UART2_CLK_DIS;
|
|
|
|
writel(val, uart_clock_base->reg1);
|
|
|
|
spin_unlock_irqrestore(&mvebu_uart_lock, flags);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void mvebu_uart_clock_disable(struct clk_hw *hw)
|
|
{
|
|
struct mvebu_uart_clock *uart_clock = to_uart_clock(hw);
|
|
struct mvebu_uart_clock_base *uart_clock_base =
|
|
to_uart_clock_base(uart_clock);
|
|
unsigned long flags;
|
|
u32 val;
|
|
|
|
spin_lock_irqsave(&mvebu_uart_lock, flags);
|
|
|
|
val = readl(uart_clock_base->reg1);
|
|
|
|
if (uart_clock->clock_idx == 0)
|
|
val |= UART1_CLK_DIS;
|
|
else
|
|
val |= UART2_CLK_DIS;
|
|
|
|
writel(val, uart_clock_base->reg1);
|
|
|
|
spin_unlock_irqrestore(&mvebu_uart_lock, flags);
|
|
}
|
|
|
|
static int mvebu_uart_clock_is_enabled(struct clk_hw *hw)
|
|
{
|
|
struct mvebu_uart_clock *uart_clock = to_uart_clock(hw);
|
|
struct mvebu_uart_clock_base *uart_clock_base =
|
|
to_uart_clock_base(uart_clock);
|
|
u32 val;
|
|
|
|
val = readl(uart_clock_base->reg1);
|
|
|
|
if (uart_clock->clock_idx == 0)
|
|
return !(val & UART1_CLK_DIS);
|
|
else
|
|
return !(val & UART2_CLK_DIS);
|
|
}
|
|
|
|
static int mvebu_uart_clock_save_context(struct clk_hw *hw)
|
|
{
|
|
struct mvebu_uart_clock *uart_clock = to_uart_clock(hw);
|
|
struct mvebu_uart_clock_base *uart_clock_base =
|
|
to_uart_clock_base(uart_clock);
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&mvebu_uart_lock, flags);
|
|
uart_clock->pm_context_reg1 = readl(uart_clock_base->reg1);
|
|
uart_clock->pm_context_reg2 = readl(uart_clock_base->reg2);
|
|
spin_unlock_irqrestore(&mvebu_uart_lock, flags);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void mvebu_uart_clock_restore_context(struct clk_hw *hw)
|
|
{
|
|
struct mvebu_uart_clock *uart_clock = to_uart_clock(hw);
|
|
struct mvebu_uart_clock_base *uart_clock_base =
|
|
to_uart_clock_base(uart_clock);
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&mvebu_uart_lock, flags);
|
|
writel(uart_clock->pm_context_reg1, uart_clock_base->reg1);
|
|
writel(uart_clock->pm_context_reg2, uart_clock_base->reg2);
|
|
spin_unlock_irqrestore(&mvebu_uart_lock, flags);
|
|
}
|
|
|
|
static unsigned long mvebu_uart_clock_recalc_rate(struct clk_hw *hw,
|
|
unsigned long parent_rate)
|
|
{
|
|
struct mvebu_uart_clock *uart_clock = to_uart_clock(hw);
|
|
struct mvebu_uart_clock_base *uart_clock_base =
|
|
to_uart_clock_base(uart_clock);
|
|
|
|
return parent_rate / uart_clock_base->div;
|
|
}
|
|
|
|
static long mvebu_uart_clock_round_rate(struct clk_hw *hw, unsigned long rate,
|
|
unsigned long *parent_rate)
|
|
{
|
|
struct mvebu_uart_clock *uart_clock = to_uart_clock(hw);
|
|
struct mvebu_uart_clock_base *uart_clock_base =
|
|
to_uart_clock_base(uart_clock);
|
|
|
|
return *parent_rate / uart_clock_base->div;
|
|
}
|
|
|
|
static int mvebu_uart_clock_set_rate(struct clk_hw *hw, unsigned long rate,
|
|
unsigned long parent_rate)
|
|
{
|
|
/*
|
|
* We must report success but we can do so unconditionally because
|
|
* mvebu_uart_clock_round_rate returns values that ensure this call is a
|
|
* nop.
|
|
*/
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct clk_ops mvebu_uart_clock_ops = {
|
|
.prepare = mvebu_uart_clock_prepare,
|
|
.enable = mvebu_uart_clock_enable,
|
|
.disable = mvebu_uart_clock_disable,
|
|
.is_enabled = mvebu_uart_clock_is_enabled,
|
|
.save_context = mvebu_uart_clock_save_context,
|
|
.restore_context = mvebu_uart_clock_restore_context,
|
|
.round_rate = mvebu_uart_clock_round_rate,
|
|
.set_rate = mvebu_uart_clock_set_rate,
|
|
.recalc_rate = mvebu_uart_clock_recalc_rate,
|
|
};
|
|
|
|
static int mvebu_uart_clock_register(struct device *dev,
|
|
struct mvebu_uart_clock *uart_clock,
|
|
const char *name,
|
|
const char *parent_name)
|
|
{
|
|
struct clk_init_data init = { };
|
|
|
|
uart_clock->clk_hw.init = &init;
|
|
|
|
init.name = name;
|
|
init.ops = &mvebu_uart_clock_ops;
|
|
init.flags = 0;
|
|
init.num_parents = 1;
|
|
init.parent_names = &parent_name;
|
|
|
|
return devm_clk_hw_register(dev, &uart_clock->clk_hw);
|
|
}
|
|
|
|
static int mvebu_uart_clock_probe(struct platform_device *pdev)
|
|
{
|
|
static const char *const uart_clk_names[] = { "uart_1", "uart_2" };
|
|
static const char *const parent_clk_names[] = { "TBG-A-P", "TBG-B-P",
|
|
"TBG-A-S", "TBG-B-S",
|
|
"xtal" };
|
|
struct clk *parent_clks[ARRAY_SIZE(parent_clk_names)];
|
|
struct mvebu_uart_clock_base *uart_clock_base;
|
|
struct clk_hw_onecell_data *hw_clk_data;
|
|
struct device *dev = &pdev->dev;
|
|
int i, parent_clk_idx, ret;
|
|
unsigned long div, rate;
|
|
struct resource *res;
|
|
unsigned int d1, d2;
|
|
|
|
BUILD_BUG_ON(ARRAY_SIZE(uart_clk_names) !=
|
|
ARRAY_SIZE(uart_clock_base->clocks));
|
|
BUILD_BUG_ON(ARRAY_SIZE(parent_clk_names) !=
|
|
ARRAY_SIZE(uart_clock_base->parent_rates));
|
|
|
|
uart_clock_base = devm_kzalloc(dev,
|
|
sizeof(*uart_clock_base),
|
|
GFP_KERNEL);
|
|
if (!uart_clock_base)
|
|
return -ENOMEM;
|
|
|
|
res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
|
|
if (!res) {
|
|
dev_err(dev, "Couldn't get first register\n");
|
|
return -ENOENT;
|
|
}
|
|
|
|
/*
|
|
* UART Clock Control register (reg1 / UART_BRDV) is in the address
|
|
* space of UART1 (standard UART variant), controls parent clock and
|
|
* dividers for both UART1 and UART2 and is supplied via DT as the first
|
|
* resource. Therefore use ioremap() rather than ioremap_resource() to
|
|
* avoid conflicts with UART1 driver. Access to UART_BRDV is protected
|
|
* by a lock shared between clock and UART driver.
|
|
*/
|
|
uart_clock_base->reg1 = devm_ioremap(dev, res->start,
|
|
resource_size(res));
|
|
if (!uart_clock_base->reg1)
|
|
return -ENOMEM;
|
|
|
|
res = platform_get_resource(pdev, IORESOURCE_MEM, 1);
|
|
if (!res) {
|
|
dev_err(dev, "Couldn't get second register\n");
|
|
return -ENOENT;
|
|
}
|
|
|
|
/*
|
|
* UART 2 Baud Rate Divisor register (reg2 / UART_BRDV) is in address
|
|
* space of UART2 (extended UART variant), controls only one UART2
|
|
* specific divider and is supplied via DT as second resource.
|
|
* Therefore use ioremap() rather than ioremap_resource() to avoid
|
|
* conflicts with UART2 driver. Access to UART_BRDV is protected by a
|
|
* by lock shared between clock and UART driver.
|
|
*/
|
|
uart_clock_base->reg2 = devm_ioremap(dev, res->start,
|
|
resource_size(res));
|
|
if (!uart_clock_base->reg2)
|
|
return -ENOMEM;
|
|
|
|
hw_clk_data = devm_kzalloc(dev,
|
|
struct_size(hw_clk_data, hws,
|
|
ARRAY_SIZE(uart_clk_names)),
|
|
GFP_KERNEL);
|
|
if (!hw_clk_data)
|
|
return -ENOMEM;
|
|
|
|
hw_clk_data->num = ARRAY_SIZE(uart_clk_names);
|
|
for (i = 0; i < ARRAY_SIZE(uart_clk_names); i++) {
|
|
hw_clk_data->hws[i] = &uart_clock_base->clocks[i].clk_hw;
|
|
uart_clock_base->clocks[i].clock_idx = i;
|
|
}
|
|
|
|
parent_clk_idx = -1;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(parent_clk_names); i++) {
|
|
parent_clks[i] = devm_clk_get(dev, parent_clk_names[i]);
|
|
if (IS_ERR(parent_clks[i])) {
|
|
if (PTR_ERR(parent_clks[i]) == -EPROBE_DEFER)
|
|
return -EPROBE_DEFER;
|
|
dev_warn(dev, "Couldn't get the parent clock %s: %ld\n",
|
|
parent_clk_names[i], PTR_ERR(parent_clks[i]));
|
|
continue;
|
|
}
|
|
|
|
ret = clk_prepare_enable(parent_clks[i]);
|
|
if (ret) {
|
|
dev_warn(dev, "Couldn't enable parent clock %s: %d\n",
|
|
parent_clk_names[i], ret);
|
|
continue;
|
|
}
|
|
rate = clk_get_rate(parent_clks[i]);
|
|
uart_clock_base->parent_rates[i] = rate;
|
|
|
|
if (i != PARENT_CLOCK_XTAL) {
|
|
/*
|
|
* Calculate the smallest TBG d1 and d2 divisors that
|
|
* still can provide 9600 baudrate.
|
|
*/
|
|
d1 = DIV_ROUND_UP(rate, 9600 * OSAMP_MAX_DIVISOR *
|
|
BRDV_BAUD_MAX);
|
|
if (d1 < 1)
|
|
d1 = 1;
|
|
else if (d1 > CLK_TBG_DIV1_MAX)
|
|
d1 = CLK_TBG_DIV1_MAX;
|
|
|
|
d2 = DIV_ROUND_UP(rate, 9600 * OSAMP_MAX_DIVISOR *
|
|
BRDV_BAUD_MAX * d1);
|
|
if (d2 < 1)
|
|
d2 = 1;
|
|
else if (d2 > CLK_TBG_DIV2_MAX)
|
|
d2 = CLK_TBG_DIV2_MAX;
|
|
} else {
|
|
/*
|
|
* When UART clock uses XTAL clock as a source then it
|
|
* is not possible to use d1 and d2 divisors.
|
|
*/
|
|
d1 = d2 = 1;
|
|
}
|
|
|
|
/* Skip clock source which cannot provide 9600 baudrate */
|
|
if (rate > 9600 * OSAMP_MAX_DIVISOR * BRDV_BAUD_MAX * d1 * d2)
|
|
continue;
|
|
|
|
/*
|
|
* Choose TBG clock source with the smallest divisors. Use XTAL
|
|
* clock source only in case TBG is not available as XTAL cannot
|
|
* be used for baudrates higher than 230400.
|
|
*/
|
|
if (parent_clk_idx == -1 ||
|
|
(i != PARENT_CLOCK_XTAL && div > d1 * d2)) {
|
|
parent_clk_idx = i;
|
|
div = d1 * d2;
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < ARRAY_SIZE(parent_clk_names); i++) {
|
|
if (i == parent_clk_idx || IS_ERR(parent_clks[i]))
|
|
continue;
|
|
clk_disable_unprepare(parent_clks[i]);
|
|
devm_clk_put(dev, parent_clks[i]);
|
|
}
|
|
|
|
if (parent_clk_idx == -1) {
|
|
dev_err(dev, "No usable parent clock\n");
|
|
return -ENOENT;
|
|
}
|
|
|
|
uart_clock_base->parent_idx = parent_clk_idx;
|
|
uart_clock_base->div = div;
|
|
|
|
dev_notice(dev, "Using parent clock %s as base UART clock\n",
|
|
__clk_get_name(parent_clks[parent_clk_idx]));
|
|
|
|
for (i = 0; i < ARRAY_SIZE(uart_clk_names); i++) {
|
|
ret = mvebu_uart_clock_register(dev,
|
|
&uart_clock_base->clocks[i],
|
|
uart_clk_names[i],
|
|
__clk_get_name(parent_clks[parent_clk_idx]));
|
|
if (ret) {
|
|
dev_err(dev, "Can't register UART clock %d: %d\n",
|
|
i, ret);
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
return devm_of_clk_add_hw_provider(dev, of_clk_hw_onecell_get,
|
|
hw_clk_data);
|
|
}
|
|
|
|
static const struct of_device_id mvebu_uart_clock_of_match[] = {
|
|
{ .compatible = "marvell,armada-3700-uart-clock", },
|
|
{ }
|
|
};
|
|
|
|
static struct platform_driver mvebu_uart_clock_platform_driver = {
|
|
.probe = mvebu_uart_clock_probe,
|
|
.driver = {
|
|
.name = "mvebu-uart-clock",
|
|
.of_match_table = mvebu_uart_clock_of_match,
|
|
},
|
|
};
|
|
|
|
static int __init mvebu_uart_init(void)
|
|
{
|
|
int ret;
|
|
|
|
ret = uart_register_driver(&mvebu_uart_driver);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = platform_driver_register(&mvebu_uart_clock_platform_driver);
|
|
if (ret) {
|
|
uart_unregister_driver(&mvebu_uart_driver);
|
|
return ret;
|
|
}
|
|
|
|
ret = platform_driver_register(&mvebu_uart_platform_driver);
|
|
if (ret) {
|
|
platform_driver_unregister(&mvebu_uart_clock_platform_driver);
|
|
uart_unregister_driver(&mvebu_uart_driver);
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
arch_initcall(mvebu_uart_init);
|