1286 lines
32 KiB
C
1286 lines
32 KiB
C
/*
|
|
* Freescale lpuart serial port driver
|
|
*
|
|
* Copyright 2012-2013 Freescale Semiconductor, Inc.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
* (at your option) any later version.
|
|
*/
|
|
|
|
#if defined(CONFIG_SERIAL_FSL_LPUART_CONSOLE) && defined(CONFIG_MAGIC_SYSRQ)
|
|
#define SUPPORT_SYSRQ
|
|
#endif
|
|
|
|
#include <linux/clk.h>
|
|
#include <linux/console.h>
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/dmaengine.h>
|
|
#include <linux/dmapool.h>
|
|
#include <linux/io.h>
|
|
#include <linux/irq.h>
|
|
#include <linux/module.h>
|
|
#include <linux/of.h>
|
|
#include <linux/of_device.h>
|
|
#include <linux/of_dma.h>
|
|
#include <linux/serial_core.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/tty_flip.h>
|
|
|
|
/* All registers are 8-bit width */
|
|
#define UARTBDH 0x00
|
|
#define UARTBDL 0x01
|
|
#define UARTCR1 0x02
|
|
#define UARTCR2 0x03
|
|
#define UARTSR1 0x04
|
|
#define UARTCR3 0x06
|
|
#define UARTDR 0x07
|
|
#define UARTCR4 0x0a
|
|
#define UARTCR5 0x0b
|
|
#define UARTMODEM 0x0d
|
|
#define UARTPFIFO 0x10
|
|
#define UARTCFIFO 0x11
|
|
#define UARTSFIFO 0x12
|
|
#define UARTTWFIFO 0x13
|
|
#define UARTTCFIFO 0x14
|
|
#define UARTRWFIFO 0x15
|
|
|
|
#define UARTBDH_LBKDIE 0x80
|
|
#define UARTBDH_RXEDGIE 0x40
|
|
#define UARTBDH_SBR_MASK 0x1f
|
|
|
|
#define UARTCR1_LOOPS 0x80
|
|
#define UARTCR1_RSRC 0x20
|
|
#define UARTCR1_M 0x10
|
|
#define UARTCR1_WAKE 0x08
|
|
#define UARTCR1_ILT 0x04
|
|
#define UARTCR1_PE 0x02
|
|
#define UARTCR1_PT 0x01
|
|
|
|
#define UARTCR2_TIE 0x80
|
|
#define UARTCR2_TCIE 0x40
|
|
#define UARTCR2_RIE 0x20
|
|
#define UARTCR2_ILIE 0x10
|
|
#define UARTCR2_TE 0x08
|
|
#define UARTCR2_RE 0x04
|
|
#define UARTCR2_RWU 0x02
|
|
#define UARTCR2_SBK 0x01
|
|
|
|
#define UARTSR1_TDRE 0x80
|
|
#define UARTSR1_TC 0x40
|
|
#define UARTSR1_RDRF 0x20
|
|
#define UARTSR1_IDLE 0x10
|
|
#define UARTSR1_OR 0x08
|
|
#define UARTSR1_NF 0x04
|
|
#define UARTSR1_FE 0x02
|
|
#define UARTSR1_PE 0x01
|
|
|
|
#define UARTCR3_R8 0x80
|
|
#define UARTCR3_T8 0x40
|
|
#define UARTCR3_TXDIR 0x20
|
|
#define UARTCR3_TXINV 0x10
|
|
#define UARTCR3_ORIE 0x08
|
|
#define UARTCR3_NEIE 0x04
|
|
#define UARTCR3_FEIE 0x02
|
|
#define UARTCR3_PEIE 0x01
|
|
|
|
#define UARTCR4_MAEN1 0x80
|
|
#define UARTCR4_MAEN2 0x40
|
|
#define UARTCR4_M10 0x20
|
|
#define UARTCR4_BRFA_MASK 0x1f
|
|
#define UARTCR4_BRFA_OFF 0
|
|
|
|
#define UARTCR5_TDMAS 0x80
|
|
#define UARTCR5_RDMAS 0x20
|
|
|
|
#define UARTMODEM_RXRTSE 0x08
|
|
#define UARTMODEM_TXRTSPOL 0x04
|
|
#define UARTMODEM_TXRTSE 0x02
|
|
#define UARTMODEM_TXCTSE 0x01
|
|
|
|
#define UARTPFIFO_TXFE 0x80
|
|
#define UARTPFIFO_FIFOSIZE_MASK 0x7
|
|
#define UARTPFIFO_TXSIZE_OFF 4
|
|
#define UARTPFIFO_RXFE 0x08
|
|
#define UARTPFIFO_RXSIZE_OFF 0
|
|
|
|
#define UARTCFIFO_TXFLUSH 0x80
|
|
#define UARTCFIFO_RXFLUSH 0x40
|
|
#define UARTCFIFO_RXOFE 0x04
|
|
#define UARTCFIFO_TXOFE 0x02
|
|
#define UARTCFIFO_RXUFE 0x01
|
|
|
|
#define UARTSFIFO_TXEMPT 0x80
|
|
#define UARTSFIFO_RXEMPT 0x40
|
|
#define UARTSFIFO_RXOF 0x04
|
|
#define UARTSFIFO_TXOF 0x02
|
|
#define UARTSFIFO_RXUF 0x01
|
|
|
|
#define DMA_MAXBURST 16
|
|
#define DMA_MAXBURST_MASK (DMA_MAXBURST - 1)
|
|
#define FSL_UART_RX_DMA_BUFFER_SIZE 64
|
|
|
|
#define DRIVER_NAME "fsl-lpuart"
|
|
#define DEV_NAME "ttyLP"
|
|
#define UART_NR 6
|
|
|
|
struct lpuart_port {
|
|
struct uart_port port;
|
|
struct clk *clk;
|
|
unsigned int txfifo_size;
|
|
unsigned int rxfifo_size;
|
|
|
|
bool lpuart_dma_use;
|
|
struct dma_chan *dma_tx_chan;
|
|
struct dma_chan *dma_rx_chan;
|
|
struct dma_async_tx_descriptor *dma_tx_desc;
|
|
struct dma_async_tx_descriptor *dma_rx_desc;
|
|
dma_addr_t dma_tx_buf_bus;
|
|
dma_addr_t dma_rx_buf_bus;
|
|
dma_cookie_t dma_tx_cookie;
|
|
dma_cookie_t dma_rx_cookie;
|
|
unsigned char *dma_tx_buf_virt;
|
|
unsigned char *dma_rx_buf_virt;
|
|
unsigned int dma_tx_bytes;
|
|
unsigned int dma_rx_bytes;
|
|
int dma_tx_in_progress;
|
|
int dma_rx_in_progress;
|
|
unsigned int dma_rx_timeout;
|
|
struct timer_list lpuart_timer;
|
|
};
|
|
|
|
static struct of_device_id lpuart_dt_ids[] = {
|
|
{
|
|
.compatible = "fsl,vf610-lpuart",
|
|
},
|
|
{ /* sentinel */ }
|
|
};
|
|
MODULE_DEVICE_TABLE(of, lpuart_dt_ids);
|
|
|
|
/* Forward declare this for the dma callbacks*/
|
|
static void lpuart_dma_tx_complete(void *arg);
|
|
static void lpuart_dma_rx_complete(void *arg);
|
|
|
|
static void lpuart_stop_tx(struct uart_port *port)
|
|
{
|
|
unsigned char temp;
|
|
|
|
temp = readb(port->membase + UARTCR2);
|
|
temp &= ~(UARTCR2_TIE | UARTCR2_TCIE);
|
|
writeb(temp, port->membase + UARTCR2);
|
|
}
|
|
|
|
static void lpuart_stop_rx(struct uart_port *port)
|
|
{
|
|
unsigned char temp;
|
|
|
|
temp = readb(port->membase + UARTCR2);
|
|
writeb(temp & ~UARTCR2_RE, port->membase + UARTCR2);
|
|
}
|
|
|
|
static void lpuart_enable_ms(struct uart_port *port)
|
|
{
|
|
}
|
|
|
|
static void lpuart_copy_rx_to_tty(struct lpuart_port *sport,
|
|
struct tty_port *tty, int count)
|
|
{
|
|
int copied;
|
|
|
|
sport->port.icount.rx += count;
|
|
|
|
if (!tty) {
|
|
dev_err(sport->port.dev, "No tty port\n");
|
|
return;
|
|
}
|
|
|
|
dma_sync_single_for_cpu(sport->port.dev, sport->dma_rx_buf_bus,
|
|
FSL_UART_RX_DMA_BUFFER_SIZE, DMA_FROM_DEVICE);
|
|
copied = tty_insert_flip_string(tty,
|
|
((unsigned char *)(sport->dma_rx_buf_virt)), count);
|
|
|
|
if (copied != count) {
|
|
WARN_ON(1);
|
|
dev_err(sport->port.dev, "RxData copy to tty layer failed\n");
|
|
}
|
|
|
|
dma_sync_single_for_device(sport->port.dev, sport->dma_rx_buf_bus,
|
|
FSL_UART_RX_DMA_BUFFER_SIZE, DMA_TO_DEVICE);
|
|
}
|
|
|
|
static void lpuart_pio_tx(struct lpuart_port *sport)
|
|
{
|
|
struct circ_buf *xmit = &sport->port.state->xmit;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&sport->port.lock, flags);
|
|
|
|
while (!uart_circ_empty(xmit) &&
|
|
readb(sport->port.membase + UARTTCFIFO) < sport->txfifo_size) {
|
|
writeb(xmit->buf[xmit->tail], sport->port.membase + UARTDR);
|
|
xmit->tail = (xmit->tail + 1) & (UART_XMIT_SIZE - 1);
|
|
sport->port.icount.tx++;
|
|
}
|
|
|
|
if (uart_circ_chars_pending(xmit) < WAKEUP_CHARS)
|
|
uart_write_wakeup(&sport->port);
|
|
|
|
if (uart_circ_empty(xmit))
|
|
writeb(readb(sport->port.membase + UARTCR5) | UARTCR5_TDMAS,
|
|
sport->port.membase + UARTCR5);
|
|
|
|
spin_unlock_irqrestore(&sport->port.lock, flags);
|
|
}
|
|
|
|
static int lpuart_dma_tx(struct lpuart_port *sport, unsigned long count)
|
|
{
|
|
struct circ_buf *xmit = &sport->port.state->xmit;
|
|
dma_addr_t tx_bus_addr;
|
|
|
|
dma_sync_single_for_device(sport->port.dev, sport->dma_tx_buf_bus,
|
|
UART_XMIT_SIZE, DMA_TO_DEVICE);
|
|
sport->dma_tx_bytes = count & ~(DMA_MAXBURST_MASK);
|
|
tx_bus_addr = sport->dma_tx_buf_bus + xmit->tail;
|
|
sport->dma_tx_desc = dmaengine_prep_slave_single(sport->dma_tx_chan,
|
|
tx_bus_addr, sport->dma_tx_bytes,
|
|
DMA_MEM_TO_DEV, DMA_PREP_INTERRUPT);
|
|
|
|
if (!sport->dma_tx_desc) {
|
|
dev_err(sport->port.dev, "Not able to get desc for tx\n");
|
|
return -EIO;
|
|
}
|
|
|
|
sport->dma_tx_desc->callback = lpuart_dma_tx_complete;
|
|
sport->dma_tx_desc->callback_param = sport;
|
|
sport->dma_tx_in_progress = 1;
|
|
sport->dma_tx_cookie = dmaengine_submit(sport->dma_tx_desc);
|
|
dma_async_issue_pending(sport->dma_tx_chan);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void lpuart_prepare_tx(struct lpuart_port *sport)
|
|
{
|
|
struct circ_buf *xmit = &sport->port.state->xmit;
|
|
unsigned long count = CIRC_CNT_TO_END(xmit->head,
|
|
xmit->tail, UART_XMIT_SIZE);
|
|
|
|
if (!count)
|
|
return;
|
|
|
|
if (count < DMA_MAXBURST)
|
|
writeb(readb(sport->port.membase + UARTCR5) & ~UARTCR5_TDMAS,
|
|
sport->port.membase + UARTCR5);
|
|
else {
|
|
writeb(readb(sport->port.membase + UARTCR5) | UARTCR5_TDMAS,
|
|
sport->port.membase + UARTCR5);
|
|
lpuart_dma_tx(sport, count);
|
|
}
|
|
}
|
|
|
|
static void lpuart_dma_tx_complete(void *arg)
|
|
{
|
|
struct lpuart_port *sport = arg;
|
|
struct circ_buf *xmit = &sport->port.state->xmit;
|
|
unsigned long flags;
|
|
|
|
async_tx_ack(sport->dma_tx_desc);
|
|
|
|
spin_lock_irqsave(&sport->port.lock, flags);
|
|
|
|
xmit->tail = (xmit->tail + sport->dma_tx_bytes) & (UART_XMIT_SIZE - 1);
|
|
sport->dma_tx_in_progress = 0;
|
|
|
|
if (uart_circ_chars_pending(xmit) < WAKEUP_CHARS)
|
|
uart_write_wakeup(&sport->port);
|
|
|
|
lpuart_prepare_tx(sport);
|
|
|
|
spin_unlock_irqrestore(&sport->port.lock, flags);
|
|
}
|
|
|
|
static int lpuart_dma_rx(struct lpuart_port *sport)
|
|
{
|
|
dma_sync_single_for_device(sport->port.dev, sport->dma_rx_buf_bus,
|
|
FSL_UART_RX_DMA_BUFFER_SIZE, DMA_TO_DEVICE);
|
|
sport->dma_rx_desc = dmaengine_prep_slave_single(sport->dma_rx_chan,
|
|
sport->dma_rx_buf_bus, FSL_UART_RX_DMA_BUFFER_SIZE,
|
|
DMA_DEV_TO_MEM, DMA_PREP_INTERRUPT);
|
|
|
|
if (!sport->dma_rx_desc) {
|
|
dev_err(sport->port.dev, "Not able to get desc for rx\n");
|
|
return -EIO;
|
|
}
|
|
|
|
sport->dma_rx_desc->callback = lpuart_dma_rx_complete;
|
|
sport->dma_rx_desc->callback_param = sport;
|
|
sport->dma_rx_in_progress = 1;
|
|
sport->dma_rx_cookie = dmaengine_submit(sport->dma_rx_desc);
|
|
dma_async_issue_pending(sport->dma_rx_chan);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void lpuart_dma_rx_complete(void *arg)
|
|
{
|
|
struct lpuart_port *sport = arg;
|
|
struct tty_port *port = &sport->port.state->port;
|
|
unsigned long flags;
|
|
|
|
async_tx_ack(sport->dma_rx_desc);
|
|
|
|
spin_lock_irqsave(&sport->port.lock, flags);
|
|
|
|
sport->dma_rx_in_progress = 0;
|
|
lpuart_copy_rx_to_tty(sport, port, FSL_UART_RX_DMA_BUFFER_SIZE);
|
|
tty_flip_buffer_push(port);
|
|
lpuart_dma_rx(sport);
|
|
|
|
spin_unlock_irqrestore(&sport->port.lock, flags);
|
|
}
|
|
|
|
static void lpuart_timer_func(unsigned long data)
|
|
{
|
|
struct lpuart_port *sport = (struct lpuart_port *)data;
|
|
struct tty_port *port = &sport->port.state->port;
|
|
struct dma_tx_state state;
|
|
unsigned long flags;
|
|
unsigned char temp;
|
|
int count;
|
|
|
|
del_timer(&sport->lpuart_timer);
|
|
dmaengine_pause(sport->dma_rx_chan);
|
|
dmaengine_tx_status(sport->dma_rx_chan, sport->dma_rx_cookie, &state);
|
|
dmaengine_terminate_all(sport->dma_rx_chan);
|
|
count = FSL_UART_RX_DMA_BUFFER_SIZE - state.residue;
|
|
async_tx_ack(sport->dma_rx_desc);
|
|
|
|
spin_lock_irqsave(&sport->port.lock, flags);
|
|
|
|
sport->dma_rx_in_progress = 0;
|
|
lpuart_copy_rx_to_tty(sport, port, count);
|
|
tty_flip_buffer_push(port);
|
|
temp = readb(sport->port.membase + UARTCR5);
|
|
writeb(temp & ~UARTCR5_RDMAS, sport->port.membase + UARTCR5);
|
|
|
|
spin_unlock_irqrestore(&sport->port.lock, flags);
|
|
}
|
|
|
|
static inline void lpuart_prepare_rx(struct lpuart_port *sport)
|
|
{
|
|
unsigned long flags;
|
|
unsigned char temp;
|
|
|
|
spin_lock_irqsave(&sport->port.lock, flags);
|
|
|
|
init_timer(&sport->lpuart_timer);
|
|
sport->lpuart_timer.function = lpuart_timer_func;
|
|
sport->lpuart_timer.data = (unsigned long)sport;
|
|
sport->lpuart_timer.expires = jiffies + sport->dma_rx_timeout;
|
|
add_timer(&sport->lpuart_timer);
|
|
|
|
lpuart_dma_rx(sport);
|
|
temp = readb(sport->port.membase + UARTCR5);
|
|
writeb(temp | UARTCR5_RDMAS, sport->port.membase + UARTCR5);
|
|
|
|
spin_unlock_irqrestore(&sport->port.lock, flags);
|
|
}
|
|
|
|
static inline void lpuart_transmit_buffer(struct lpuart_port *sport)
|
|
{
|
|
struct circ_buf *xmit = &sport->port.state->xmit;
|
|
|
|
while (!uart_circ_empty(xmit) &&
|
|
(readb(sport->port.membase + UARTTCFIFO) < sport->txfifo_size)) {
|
|
writeb(xmit->buf[xmit->tail], sport->port.membase + UARTDR);
|
|
xmit->tail = (xmit->tail + 1) & (UART_XMIT_SIZE - 1);
|
|
sport->port.icount.tx++;
|
|
}
|
|
|
|
if (uart_circ_chars_pending(xmit) < WAKEUP_CHARS)
|
|
uart_write_wakeup(&sport->port);
|
|
|
|
if (uart_circ_empty(xmit))
|
|
lpuart_stop_tx(&sport->port);
|
|
}
|
|
|
|
static void lpuart_start_tx(struct uart_port *port)
|
|
{
|
|
struct lpuart_port *sport = container_of(port,
|
|
struct lpuart_port, port);
|
|
struct circ_buf *xmit = &sport->port.state->xmit;
|
|
unsigned char temp;
|
|
|
|
temp = readb(port->membase + UARTCR2);
|
|
writeb(temp | UARTCR2_TIE, port->membase + UARTCR2);
|
|
|
|
if (sport->lpuart_dma_use) {
|
|
if (!uart_circ_empty(xmit) && !sport->dma_tx_in_progress)
|
|
lpuart_prepare_tx(sport);
|
|
} else {
|
|
if (readb(port->membase + UARTSR1) & UARTSR1_TDRE)
|
|
lpuart_transmit_buffer(sport);
|
|
}
|
|
}
|
|
|
|
static irqreturn_t lpuart_txint(int irq, void *dev_id)
|
|
{
|
|
struct lpuart_port *sport = dev_id;
|
|
struct circ_buf *xmit = &sport->port.state->xmit;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&sport->port.lock, flags);
|
|
if (sport->port.x_char) {
|
|
writeb(sport->port.x_char, sport->port.membase + UARTDR);
|
|
goto out;
|
|
}
|
|
|
|
if (uart_circ_empty(xmit) || uart_tx_stopped(&sport->port)) {
|
|
lpuart_stop_tx(&sport->port);
|
|
goto out;
|
|
}
|
|
|
|
lpuart_transmit_buffer(sport);
|
|
|
|
if (uart_circ_chars_pending(xmit) < WAKEUP_CHARS)
|
|
uart_write_wakeup(&sport->port);
|
|
|
|
out:
|
|
spin_unlock_irqrestore(&sport->port.lock, flags);
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static irqreturn_t lpuart_rxint(int irq, void *dev_id)
|
|
{
|
|
struct lpuart_port *sport = dev_id;
|
|
unsigned int flg, ignored = 0;
|
|
struct tty_port *port = &sport->port.state->port;
|
|
unsigned long flags;
|
|
unsigned char rx, sr;
|
|
|
|
spin_lock_irqsave(&sport->port.lock, flags);
|
|
|
|
while (!(readb(sport->port.membase + UARTSFIFO) & UARTSFIFO_RXEMPT)) {
|
|
flg = TTY_NORMAL;
|
|
sport->port.icount.rx++;
|
|
/*
|
|
* to clear the FE, OR, NF, FE, PE flags,
|
|
* read SR1 then read DR
|
|
*/
|
|
sr = readb(sport->port.membase + UARTSR1);
|
|
rx = readb(sport->port.membase + UARTDR);
|
|
|
|
if (uart_handle_sysrq_char(&sport->port, (unsigned char)rx))
|
|
continue;
|
|
|
|
if (sr & (UARTSR1_PE | UARTSR1_OR | UARTSR1_FE)) {
|
|
if (sr & UARTSR1_PE)
|
|
sport->port.icount.parity++;
|
|
else if (sr & UARTSR1_FE)
|
|
sport->port.icount.frame++;
|
|
|
|
if (sr & UARTSR1_OR)
|
|
sport->port.icount.overrun++;
|
|
|
|
if (sr & sport->port.ignore_status_mask) {
|
|
if (++ignored > 100)
|
|
goto out;
|
|
continue;
|
|
}
|
|
|
|
sr &= sport->port.read_status_mask;
|
|
|
|
if (sr & UARTSR1_PE)
|
|
flg = TTY_PARITY;
|
|
else if (sr & UARTSR1_FE)
|
|
flg = TTY_FRAME;
|
|
|
|
if (sr & UARTSR1_OR)
|
|
flg = TTY_OVERRUN;
|
|
|
|
#ifdef SUPPORT_SYSRQ
|
|
sport->port.sysrq = 0;
|
|
#endif
|
|
}
|
|
|
|
tty_insert_flip_char(port, rx, flg);
|
|
}
|
|
|
|
out:
|
|
spin_unlock_irqrestore(&sport->port.lock, flags);
|
|
|
|
tty_flip_buffer_push(port);
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static irqreturn_t lpuart_int(int irq, void *dev_id)
|
|
{
|
|
struct lpuart_port *sport = dev_id;
|
|
unsigned char sts;
|
|
|
|
sts = readb(sport->port.membase + UARTSR1);
|
|
|
|
if (sts & UARTSR1_RDRF) {
|
|
if (sport->lpuart_dma_use)
|
|
lpuart_prepare_rx(sport);
|
|
else
|
|
lpuart_rxint(irq, dev_id);
|
|
}
|
|
if (sts & UARTSR1_TDRE &&
|
|
!(readb(sport->port.membase + UARTCR5) & UARTCR5_TDMAS)) {
|
|
if (sport->lpuart_dma_use)
|
|
lpuart_pio_tx(sport);
|
|
else
|
|
lpuart_txint(irq, dev_id);
|
|
}
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
/* return TIOCSER_TEMT when transmitter is not busy */
|
|
static unsigned int lpuart_tx_empty(struct uart_port *port)
|
|
{
|
|
return (readb(port->membase + UARTSR1) & UARTSR1_TC) ?
|
|
TIOCSER_TEMT : 0;
|
|
}
|
|
|
|
static unsigned int lpuart_get_mctrl(struct uart_port *port)
|
|
{
|
|
unsigned int temp = 0;
|
|
unsigned char reg;
|
|
|
|
reg = readb(port->membase + UARTMODEM);
|
|
if (reg & UARTMODEM_TXCTSE)
|
|
temp |= TIOCM_CTS;
|
|
|
|
if (reg & UARTMODEM_RXRTSE)
|
|
temp |= TIOCM_RTS;
|
|
|
|
return temp;
|
|
}
|
|
|
|
static void lpuart_set_mctrl(struct uart_port *port, unsigned int mctrl)
|
|
{
|
|
unsigned char temp;
|
|
|
|
temp = readb(port->membase + UARTMODEM) &
|
|
~(UARTMODEM_RXRTSE | UARTMODEM_TXCTSE);
|
|
|
|
if (mctrl & TIOCM_RTS)
|
|
temp |= UARTMODEM_RXRTSE;
|
|
|
|
if (mctrl & TIOCM_CTS)
|
|
temp |= UARTMODEM_TXCTSE;
|
|
|
|
writeb(temp, port->membase + UARTMODEM);
|
|
}
|
|
|
|
static void lpuart_break_ctl(struct uart_port *port, int break_state)
|
|
{
|
|
unsigned char temp;
|
|
|
|
temp = readb(port->membase + UARTCR2) & ~UARTCR2_SBK;
|
|
|
|
if (break_state != 0)
|
|
temp |= UARTCR2_SBK;
|
|
|
|
writeb(temp, port->membase + UARTCR2);
|
|
}
|
|
|
|
static void lpuart_setup_watermark(struct lpuart_port *sport)
|
|
{
|
|
unsigned char val, cr2;
|
|
unsigned char cr2_saved;
|
|
|
|
cr2 = readb(sport->port.membase + UARTCR2);
|
|
cr2_saved = cr2;
|
|
cr2 &= ~(UARTCR2_TIE | UARTCR2_TCIE | UARTCR2_TE |
|
|
UARTCR2_RIE | UARTCR2_RE);
|
|
writeb(cr2, sport->port.membase + UARTCR2);
|
|
|
|
/* determine FIFO size and enable FIFO mode */
|
|
val = readb(sport->port.membase + UARTPFIFO);
|
|
|
|
sport->txfifo_size = 0x1 << (((val >> UARTPFIFO_TXSIZE_OFF) &
|
|
UARTPFIFO_FIFOSIZE_MASK) + 1);
|
|
|
|
sport->rxfifo_size = 0x1 << (((val >> UARTPFIFO_RXSIZE_OFF) &
|
|
UARTPFIFO_FIFOSIZE_MASK) + 1);
|
|
|
|
writeb(val | UARTPFIFO_TXFE | UARTPFIFO_RXFE,
|
|
sport->port.membase + UARTPFIFO);
|
|
|
|
/* flush Tx and Rx FIFO */
|
|
writeb(UARTCFIFO_TXFLUSH | UARTCFIFO_RXFLUSH,
|
|
sport->port.membase + UARTCFIFO);
|
|
|
|
writeb(0, sport->port.membase + UARTTWFIFO);
|
|
writeb(1, sport->port.membase + UARTRWFIFO);
|
|
|
|
/* Restore cr2 */
|
|
writeb(cr2_saved, sport->port.membase + UARTCR2);
|
|
}
|
|
|
|
static int lpuart_dma_tx_request(struct uart_port *port)
|
|
{
|
|
struct lpuart_port *sport = container_of(port,
|
|
struct lpuart_port, port);
|
|
struct dma_chan *tx_chan;
|
|
struct dma_slave_config dma_tx_sconfig;
|
|
dma_addr_t dma_bus;
|
|
unsigned char *dma_buf;
|
|
int ret;
|
|
|
|
tx_chan = dma_request_slave_channel(sport->port.dev, "tx");
|
|
|
|
if (!tx_chan) {
|
|
dev_err(sport->port.dev, "Dma tx channel request failed!\n");
|
|
return -ENODEV;
|
|
}
|
|
|
|
dma_bus = dma_map_single(tx_chan->device->dev,
|
|
sport->port.state->xmit.buf,
|
|
UART_XMIT_SIZE, DMA_TO_DEVICE);
|
|
|
|
if (dma_mapping_error(tx_chan->device->dev, dma_bus)) {
|
|
dev_err(sport->port.dev, "dma_map_single tx failed\n");
|
|
dma_release_channel(tx_chan);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
dma_buf = sport->port.state->xmit.buf;
|
|
dma_tx_sconfig.dst_addr = sport->port.mapbase + UARTDR;
|
|
dma_tx_sconfig.dst_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE;
|
|
dma_tx_sconfig.dst_maxburst = DMA_MAXBURST;
|
|
dma_tx_sconfig.direction = DMA_MEM_TO_DEV;
|
|
ret = dmaengine_slave_config(tx_chan, &dma_tx_sconfig);
|
|
|
|
if (ret < 0) {
|
|
dev_err(sport->port.dev,
|
|
"Dma slave config failed, err = %d\n", ret);
|
|
dma_release_channel(tx_chan);
|
|
return ret;
|
|
}
|
|
|
|
sport->dma_tx_chan = tx_chan;
|
|
sport->dma_tx_buf_virt = dma_buf;
|
|
sport->dma_tx_buf_bus = dma_bus;
|
|
sport->dma_tx_in_progress = 0;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int lpuart_dma_rx_request(struct uart_port *port)
|
|
{
|
|
struct lpuart_port *sport = container_of(port,
|
|
struct lpuart_port, port);
|
|
struct dma_chan *rx_chan;
|
|
struct dma_slave_config dma_rx_sconfig;
|
|
dma_addr_t dma_bus;
|
|
unsigned char *dma_buf;
|
|
int ret;
|
|
|
|
rx_chan = dma_request_slave_channel(sport->port.dev, "rx");
|
|
|
|
if (!rx_chan) {
|
|
dev_err(sport->port.dev, "Dma rx channel request failed!\n");
|
|
return -ENODEV;
|
|
}
|
|
|
|
dma_buf = devm_kzalloc(sport->port.dev,
|
|
FSL_UART_RX_DMA_BUFFER_SIZE, GFP_KERNEL);
|
|
|
|
if (!dma_buf) {
|
|
dev_err(sport->port.dev, "Dma rx alloc failed\n");
|
|
dma_release_channel(rx_chan);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
dma_bus = dma_map_single(rx_chan->device->dev, dma_buf,
|
|
FSL_UART_RX_DMA_BUFFER_SIZE, DMA_FROM_DEVICE);
|
|
|
|
if (dma_mapping_error(rx_chan->device->dev, dma_bus)) {
|
|
dev_err(sport->port.dev, "dma_map_single rx failed\n");
|
|
dma_release_channel(rx_chan);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
dma_rx_sconfig.src_addr = sport->port.mapbase + UARTDR;
|
|
dma_rx_sconfig.src_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE;
|
|
dma_rx_sconfig.src_maxburst = 1;
|
|
dma_rx_sconfig.direction = DMA_DEV_TO_MEM;
|
|
ret = dmaengine_slave_config(rx_chan, &dma_rx_sconfig);
|
|
|
|
if (ret < 0) {
|
|
dev_err(sport->port.dev,
|
|
"Dma slave config failed, err = %d\n", ret);
|
|
dma_release_channel(rx_chan);
|
|
return ret;
|
|
}
|
|
|
|
sport->dma_rx_chan = rx_chan;
|
|
sport->dma_rx_buf_virt = dma_buf;
|
|
sport->dma_rx_buf_bus = dma_bus;
|
|
sport->dma_rx_in_progress = 0;
|
|
|
|
sport->dma_rx_timeout = (sport->port.timeout - HZ / 50) *
|
|
FSL_UART_RX_DMA_BUFFER_SIZE * 3 /
|
|
sport->rxfifo_size / 2;
|
|
|
|
if (sport->dma_rx_timeout < msecs_to_jiffies(20))
|
|
sport->dma_rx_timeout = msecs_to_jiffies(20);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void lpuart_dma_tx_free(struct uart_port *port)
|
|
{
|
|
struct lpuart_port *sport = container_of(port,
|
|
struct lpuart_port, port);
|
|
struct dma_chan *dma_chan;
|
|
|
|
dma_unmap_single(sport->port.dev, sport->dma_tx_buf_bus,
|
|
UART_XMIT_SIZE, DMA_TO_DEVICE);
|
|
dma_chan = sport->dma_tx_chan;
|
|
sport->dma_tx_chan = NULL;
|
|
sport->dma_tx_buf_bus = 0;
|
|
sport->dma_tx_buf_virt = NULL;
|
|
dma_release_channel(dma_chan);
|
|
}
|
|
|
|
static void lpuart_dma_rx_free(struct uart_port *port)
|
|
{
|
|
struct lpuart_port *sport = container_of(port,
|
|
struct lpuart_port, port);
|
|
struct dma_chan *dma_chan;
|
|
|
|
dma_unmap_single(sport->port.dev, sport->dma_rx_buf_bus,
|
|
FSL_UART_RX_DMA_BUFFER_SIZE, DMA_FROM_DEVICE);
|
|
|
|
dma_chan = sport->dma_rx_chan;
|
|
sport->dma_rx_chan = NULL;
|
|
sport->dma_rx_buf_bus = 0;
|
|
sport->dma_rx_buf_virt = NULL;
|
|
dma_release_channel(dma_chan);
|
|
}
|
|
|
|
static int lpuart_startup(struct uart_port *port)
|
|
{
|
|
struct lpuart_port *sport = container_of(port, struct lpuart_port, port);
|
|
int ret;
|
|
unsigned long flags;
|
|
unsigned char temp;
|
|
|
|
/*whether use dma support by dma request results*/
|
|
if (lpuart_dma_tx_request(port) || lpuart_dma_rx_request(port)) {
|
|
sport->lpuart_dma_use = false;
|
|
} else {
|
|
sport->lpuart_dma_use = true;
|
|
temp = readb(port->membase + UARTCR5);
|
|
writeb(temp | UARTCR5_TDMAS, port->membase + UARTCR5);
|
|
}
|
|
|
|
ret = devm_request_irq(port->dev, port->irq, lpuart_int, 0,
|
|
DRIVER_NAME, sport);
|
|
if (ret)
|
|
return ret;
|
|
|
|
spin_lock_irqsave(&sport->port.lock, flags);
|
|
|
|
lpuart_setup_watermark(sport);
|
|
|
|
temp = readb(sport->port.membase + UARTCR2);
|
|
temp |= (UARTCR2_RIE | UARTCR2_TIE | UARTCR2_RE | UARTCR2_TE);
|
|
writeb(temp, sport->port.membase + UARTCR2);
|
|
|
|
spin_unlock_irqrestore(&sport->port.lock, flags);
|
|
return 0;
|
|
}
|
|
|
|
static void lpuart_shutdown(struct uart_port *port)
|
|
{
|
|
struct lpuart_port *sport = container_of(port, struct lpuart_port, port);
|
|
unsigned char temp;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&port->lock, flags);
|
|
|
|
/* disable Rx/Tx and interrupts */
|
|
temp = readb(port->membase + UARTCR2);
|
|
temp &= ~(UARTCR2_TE | UARTCR2_RE |
|
|
UARTCR2_TIE | UARTCR2_TCIE | UARTCR2_RIE);
|
|
writeb(temp, port->membase + UARTCR2);
|
|
|
|
spin_unlock_irqrestore(&port->lock, flags);
|
|
|
|
devm_free_irq(port->dev, port->irq, sport);
|
|
|
|
if (sport->lpuart_dma_use) {
|
|
lpuart_dma_tx_free(port);
|
|
lpuart_dma_rx_free(port);
|
|
}
|
|
}
|
|
|
|
static void
|
|
lpuart_set_termios(struct uart_port *port, struct ktermios *termios,
|
|
struct ktermios *old)
|
|
{
|
|
struct lpuart_port *sport = container_of(port, struct lpuart_port, port);
|
|
unsigned long flags;
|
|
unsigned char cr1, old_cr1, old_cr2, cr4, bdh, modem;
|
|
unsigned int baud;
|
|
unsigned int old_csize = old ? old->c_cflag & CSIZE : CS8;
|
|
unsigned int sbr, brfa;
|
|
|
|
cr1 = old_cr1 = readb(sport->port.membase + UARTCR1);
|
|
old_cr2 = readb(sport->port.membase + UARTCR2);
|
|
cr4 = readb(sport->port.membase + UARTCR4);
|
|
bdh = readb(sport->port.membase + UARTBDH);
|
|
modem = readb(sport->port.membase + UARTMODEM);
|
|
/*
|
|
* only support CS8 and CS7, and for CS7 must enable PE.
|
|
* supported mode:
|
|
* - (7,e/o,1)
|
|
* - (8,n,1)
|
|
* - (8,m/s,1)
|
|
* - (8,e/o,1)
|
|
*/
|
|
while ((termios->c_cflag & CSIZE) != CS8 &&
|
|
(termios->c_cflag & CSIZE) != CS7) {
|
|
termios->c_cflag &= ~CSIZE;
|
|
termios->c_cflag |= old_csize;
|
|
old_csize = CS8;
|
|
}
|
|
|
|
if ((termios->c_cflag & CSIZE) == CS8 ||
|
|
(termios->c_cflag & CSIZE) == CS7)
|
|
cr1 = old_cr1 & ~UARTCR1_M;
|
|
|
|
if (termios->c_cflag & CMSPAR) {
|
|
if ((termios->c_cflag & CSIZE) != CS8) {
|
|
termios->c_cflag &= ~CSIZE;
|
|
termios->c_cflag |= CS8;
|
|
}
|
|
cr1 |= UARTCR1_M;
|
|
}
|
|
|
|
if (termios->c_cflag & CRTSCTS) {
|
|
modem |= (UARTMODEM_RXRTSE | UARTMODEM_TXCTSE);
|
|
} else {
|
|
termios->c_cflag &= ~CRTSCTS;
|
|
modem &= ~(UARTMODEM_RXRTSE | UARTMODEM_TXCTSE);
|
|
}
|
|
|
|
if (termios->c_cflag & CSTOPB)
|
|
termios->c_cflag &= ~CSTOPB;
|
|
|
|
/* parity must be enabled when CS7 to match 8-bits format */
|
|
if ((termios->c_cflag & CSIZE) == CS7)
|
|
termios->c_cflag |= PARENB;
|
|
|
|
if ((termios->c_cflag & PARENB)) {
|
|
if (termios->c_cflag & CMSPAR) {
|
|
cr1 &= ~UARTCR1_PE;
|
|
cr1 |= UARTCR1_M;
|
|
} else {
|
|
cr1 |= UARTCR1_PE;
|
|
if ((termios->c_cflag & CSIZE) == CS8)
|
|
cr1 |= UARTCR1_M;
|
|
if (termios->c_cflag & PARODD)
|
|
cr1 |= UARTCR1_PT;
|
|
else
|
|
cr1 &= ~UARTCR1_PT;
|
|
}
|
|
}
|
|
|
|
/* ask the core to calculate the divisor */
|
|
baud = uart_get_baud_rate(port, termios, old, 50, port->uartclk / 16);
|
|
|
|
spin_lock_irqsave(&sport->port.lock, flags);
|
|
|
|
sport->port.read_status_mask = 0;
|
|
if (termios->c_iflag & INPCK)
|
|
sport->port.read_status_mask |= (UARTSR1_FE | UARTSR1_PE);
|
|
if (termios->c_iflag & (BRKINT | PARMRK))
|
|
sport->port.read_status_mask |= UARTSR1_FE;
|
|
|
|
/* characters to ignore */
|
|
sport->port.ignore_status_mask = 0;
|
|
if (termios->c_iflag & IGNPAR)
|
|
sport->port.ignore_status_mask |= UARTSR1_PE;
|
|
if (termios->c_iflag & IGNBRK) {
|
|
sport->port.ignore_status_mask |= UARTSR1_FE;
|
|
/*
|
|
* if we're ignoring parity and break indicators,
|
|
* ignore overruns too (for real raw support).
|
|
*/
|
|
if (termios->c_iflag & IGNPAR)
|
|
sport->port.ignore_status_mask |= UARTSR1_OR;
|
|
}
|
|
|
|
/* update the per-port timeout */
|
|
uart_update_timeout(port, termios->c_cflag, baud);
|
|
|
|
/* wait transmit engin complete */
|
|
while (!(readb(sport->port.membase + UARTSR1) & UARTSR1_TC))
|
|
barrier();
|
|
|
|
/* disable transmit and receive */
|
|
writeb(old_cr2 & ~(UARTCR2_TE | UARTCR2_RE),
|
|
sport->port.membase + UARTCR2);
|
|
|
|
sbr = sport->port.uartclk / (16 * baud);
|
|
brfa = ((sport->port.uartclk - (16 * sbr * baud)) * 2) / baud;
|
|
bdh &= ~UARTBDH_SBR_MASK;
|
|
bdh |= (sbr >> 8) & 0x1F;
|
|
cr4 &= ~UARTCR4_BRFA_MASK;
|
|
brfa &= UARTCR4_BRFA_MASK;
|
|
writeb(cr4 | brfa, sport->port.membase + UARTCR4);
|
|
writeb(bdh, sport->port.membase + UARTBDH);
|
|
writeb(sbr & 0xFF, sport->port.membase + UARTBDL);
|
|
writeb(cr1, sport->port.membase + UARTCR1);
|
|
writeb(modem, sport->port.membase + UARTMODEM);
|
|
|
|
/* restore control register */
|
|
writeb(old_cr2, sport->port.membase + UARTCR2);
|
|
|
|
spin_unlock_irqrestore(&sport->port.lock, flags);
|
|
}
|
|
|
|
static const char *lpuart_type(struct uart_port *port)
|
|
{
|
|
return "FSL_LPUART";
|
|
}
|
|
|
|
static void lpuart_release_port(struct uart_port *port)
|
|
{
|
|
/* nothing to do */
|
|
}
|
|
|
|
static int lpuart_request_port(struct uart_port *port)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
/* configure/autoconfigure the port */
|
|
static void lpuart_config_port(struct uart_port *port, int flags)
|
|
{
|
|
if (flags & UART_CONFIG_TYPE)
|
|
port->type = PORT_LPUART;
|
|
}
|
|
|
|
static int lpuart_verify_port(struct uart_port *port, struct serial_struct *ser)
|
|
{
|
|
int ret = 0;
|
|
|
|
if (ser->type != PORT_UNKNOWN && ser->type != PORT_LPUART)
|
|
ret = -EINVAL;
|
|
if (port->irq != ser->irq)
|
|
ret = -EINVAL;
|
|
if (ser->io_type != UPIO_MEM)
|
|
ret = -EINVAL;
|
|
if (port->uartclk / 16 != ser->baud_base)
|
|
ret = -EINVAL;
|
|
if (port->iobase != ser->port)
|
|
ret = -EINVAL;
|
|
if (ser->hub6 != 0)
|
|
ret = -EINVAL;
|
|
return ret;
|
|
}
|
|
|
|
static struct uart_ops lpuart_pops = {
|
|
.tx_empty = lpuart_tx_empty,
|
|
.set_mctrl = lpuart_set_mctrl,
|
|
.get_mctrl = lpuart_get_mctrl,
|
|
.stop_tx = lpuart_stop_tx,
|
|
.start_tx = lpuart_start_tx,
|
|
.stop_rx = lpuart_stop_rx,
|
|
.enable_ms = lpuart_enable_ms,
|
|
.break_ctl = lpuart_break_ctl,
|
|
.startup = lpuart_startup,
|
|
.shutdown = lpuart_shutdown,
|
|
.set_termios = lpuart_set_termios,
|
|
.type = lpuart_type,
|
|
.request_port = lpuart_request_port,
|
|
.release_port = lpuart_release_port,
|
|
.config_port = lpuart_config_port,
|
|
.verify_port = lpuart_verify_port,
|
|
};
|
|
|
|
static struct lpuart_port *lpuart_ports[UART_NR];
|
|
|
|
#ifdef CONFIG_SERIAL_FSL_LPUART_CONSOLE
|
|
static void lpuart_console_putchar(struct uart_port *port, int ch)
|
|
{
|
|
while (!(readb(port->membase + UARTSR1) & UARTSR1_TDRE))
|
|
barrier();
|
|
|
|
writeb(ch, port->membase + UARTDR);
|
|
}
|
|
|
|
static void
|
|
lpuart_console_write(struct console *co, const char *s, unsigned int count)
|
|
{
|
|
struct lpuart_port *sport = lpuart_ports[co->index];
|
|
unsigned char old_cr2, cr2;
|
|
|
|
/* first save CR2 and then disable interrupts */
|
|
cr2 = old_cr2 = readb(sport->port.membase + UARTCR2);
|
|
cr2 |= (UARTCR2_TE | UARTCR2_RE);
|
|
cr2 &= ~(UARTCR2_TIE | UARTCR2_TCIE | UARTCR2_RIE);
|
|
writeb(cr2, sport->port.membase + UARTCR2);
|
|
|
|
uart_console_write(&sport->port, s, count, lpuart_console_putchar);
|
|
|
|
/* wait for transmitter finish complete and restore CR2 */
|
|
while (!(readb(sport->port.membase + UARTSR1) & UARTSR1_TC))
|
|
barrier();
|
|
|
|
writeb(old_cr2, sport->port.membase + UARTCR2);
|
|
}
|
|
|
|
/*
|
|
* if the port was already initialised (eg, by a boot loader),
|
|
* try to determine the current setup.
|
|
*/
|
|
static void __init
|
|
lpuart_console_get_options(struct lpuart_port *sport, int *baud,
|
|
int *parity, int *bits)
|
|
{
|
|
unsigned char cr, bdh, bdl, brfa;
|
|
unsigned int sbr, uartclk, baud_raw;
|
|
|
|
cr = readb(sport->port.membase + UARTCR2);
|
|
cr &= UARTCR2_TE | UARTCR2_RE;
|
|
if (!cr)
|
|
return;
|
|
|
|
/* ok, the port was enabled */
|
|
|
|
cr = readb(sport->port.membase + UARTCR1);
|
|
|
|
*parity = 'n';
|
|
if (cr & UARTCR1_PE) {
|
|
if (cr & UARTCR1_PT)
|
|
*parity = 'o';
|
|
else
|
|
*parity = 'e';
|
|
}
|
|
|
|
if (cr & UARTCR1_M)
|
|
*bits = 9;
|
|
else
|
|
*bits = 8;
|
|
|
|
bdh = readb(sport->port.membase + UARTBDH);
|
|
bdh &= UARTBDH_SBR_MASK;
|
|
bdl = readb(sport->port.membase + UARTBDL);
|
|
sbr = bdh;
|
|
sbr <<= 8;
|
|
sbr |= bdl;
|
|
brfa = readb(sport->port.membase + UARTCR4);
|
|
brfa &= UARTCR4_BRFA_MASK;
|
|
|
|
uartclk = clk_get_rate(sport->clk);
|
|
/*
|
|
* baud = mod_clk/(16*(sbr[13]+(brfa)/32)
|
|
*/
|
|
baud_raw = uartclk / (16 * (sbr + brfa / 32));
|
|
|
|
if (*baud != baud_raw)
|
|
printk(KERN_INFO "Serial: Console lpuart rounded baud rate"
|
|
"from %d to %d\n", baud_raw, *baud);
|
|
}
|
|
|
|
static int __init lpuart_console_setup(struct console *co, char *options)
|
|
{
|
|
struct lpuart_port *sport;
|
|
int baud = 115200;
|
|
int bits = 8;
|
|
int parity = 'n';
|
|
int flow = 'n';
|
|
|
|
/*
|
|
* check whether an invalid uart number has been specified, and
|
|
* if so, search for the first available port that does have
|
|
* console support.
|
|
*/
|
|
if (co->index == -1 || co->index >= ARRAY_SIZE(lpuart_ports))
|
|
co->index = 0;
|
|
|
|
sport = lpuart_ports[co->index];
|
|
if (sport == NULL)
|
|
return -ENODEV;
|
|
|
|
if (options)
|
|
uart_parse_options(options, &baud, &parity, &bits, &flow);
|
|
else
|
|
lpuart_console_get_options(sport, &baud, &parity, &bits);
|
|
|
|
lpuart_setup_watermark(sport);
|
|
|
|
return uart_set_options(&sport->port, co, baud, parity, bits, flow);
|
|
}
|
|
|
|
static struct uart_driver lpuart_reg;
|
|
static struct console lpuart_console = {
|
|
.name = DEV_NAME,
|
|
.write = lpuart_console_write,
|
|
.device = uart_console_device,
|
|
.setup = lpuart_console_setup,
|
|
.flags = CON_PRINTBUFFER,
|
|
.index = -1,
|
|
.data = &lpuart_reg,
|
|
};
|
|
|
|
#define LPUART_CONSOLE (&lpuart_console)
|
|
#else
|
|
#define LPUART_CONSOLE NULL
|
|
#endif
|
|
|
|
static struct uart_driver lpuart_reg = {
|
|
.owner = THIS_MODULE,
|
|
.driver_name = DRIVER_NAME,
|
|
.dev_name = DEV_NAME,
|
|
.nr = ARRAY_SIZE(lpuart_ports),
|
|
.cons = LPUART_CONSOLE,
|
|
};
|
|
|
|
static int lpuart_probe(struct platform_device *pdev)
|
|
{
|
|
struct device_node *np = pdev->dev.of_node;
|
|
struct lpuart_port *sport;
|
|
struct resource *res;
|
|
int ret;
|
|
|
|
sport = devm_kzalloc(&pdev->dev, sizeof(*sport), GFP_KERNEL);
|
|
if (!sport)
|
|
return -ENOMEM;
|
|
|
|
pdev->dev.coherent_dma_mask = 0;
|
|
|
|
ret = of_alias_get_id(np, "serial");
|
|
if (ret < 0) {
|
|
dev_err(&pdev->dev, "failed to get alias id, errno %d\n", ret);
|
|
return ret;
|
|
}
|
|
sport->port.line = ret;
|
|
|
|
res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
|
|
if (!res)
|
|
return -ENODEV;
|
|
|
|
sport->port.mapbase = res->start;
|
|
sport->port.membase = devm_ioremap_resource(&pdev->dev, res);
|
|
if (IS_ERR(sport->port.membase))
|
|
return PTR_ERR(sport->port.membase);
|
|
|
|
sport->port.dev = &pdev->dev;
|
|
sport->port.type = PORT_LPUART;
|
|
sport->port.iotype = UPIO_MEM;
|
|
sport->port.irq = platform_get_irq(pdev, 0);
|
|
sport->port.ops = &lpuart_pops;
|
|
sport->port.flags = UPF_BOOT_AUTOCONF;
|
|
|
|
sport->clk = devm_clk_get(&pdev->dev, "ipg");
|
|
if (IS_ERR(sport->clk)) {
|
|
ret = PTR_ERR(sport->clk);
|
|
dev_err(&pdev->dev, "failed to get uart clk: %d\n", ret);
|
|
return ret;
|
|
}
|
|
|
|
ret = clk_prepare_enable(sport->clk);
|
|
if (ret) {
|
|
dev_err(&pdev->dev, "failed to enable uart clk: %d\n", ret);
|
|
return ret;
|
|
}
|
|
|
|
sport->port.uartclk = clk_get_rate(sport->clk);
|
|
|
|
lpuart_ports[sport->port.line] = sport;
|
|
|
|
platform_set_drvdata(pdev, &sport->port);
|
|
|
|
ret = uart_add_one_port(&lpuart_reg, &sport->port);
|
|
if (ret) {
|
|
clk_disable_unprepare(sport->clk);
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int lpuart_remove(struct platform_device *pdev)
|
|
{
|
|
struct lpuart_port *sport = platform_get_drvdata(pdev);
|
|
|
|
uart_remove_one_port(&lpuart_reg, &sport->port);
|
|
|
|
clk_disable_unprepare(sport->clk);
|
|
|
|
return 0;
|
|
}
|
|
|
|
#ifdef CONFIG_PM_SLEEP
|
|
static int lpuart_suspend(struct device *dev)
|
|
{
|
|
struct lpuart_port *sport = dev_get_drvdata(dev);
|
|
|
|
uart_suspend_port(&lpuart_reg, &sport->port);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int lpuart_resume(struct device *dev)
|
|
{
|
|
struct lpuart_port *sport = dev_get_drvdata(dev);
|
|
|
|
uart_resume_port(&lpuart_reg, &sport->port);
|
|
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
static SIMPLE_DEV_PM_OPS(lpuart_pm_ops, lpuart_suspend, lpuart_resume);
|
|
|
|
static struct platform_driver lpuart_driver = {
|
|
.probe = lpuart_probe,
|
|
.remove = lpuart_remove,
|
|
.driver = {
|
|
.name = "fsl-lpuart",
|
|
.owner = THIS_MODULE,
|
|
.of_match_table = lpuart_dt_ids,
|
|
.pm = &lpuart_pm_ops,
|
|
},
|
|
};
|
|
|
|
static int __init lpuart_serial_init(void)
|
|
{
|
|
int ret;
|
|
|
|
pr_info("serial: Freescale lpuart driver\n");
|
|
|
|
ret = uart_register_driver(&lpuart_reg);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = platform_driver_register(&lpuart_driver);
|
|
if (ret)
|
|
uart_unregister_driver(&lpuart_reg);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void __exit lpuart_serial_exit(void)
|
|
{
|
|
platform_driver_unregister(&lpuart_driver);
|
|
uart_unregister_driver(&lpuart_reg);
|
|
}
|
|
|
|
module_init(lpuart_serial_init);
|
|
module_exit(lpuart_serial_exit);
|
|
|
|
MODULE_DESCRIPTION("Freescale lpuart serial port driver");
|
|
MODULE_LICENSE("GPL v2");
|