Merge git://git.kernel.org/pub/scm/linux/kernel/git/davem/net-2.6
* git://git.kernel.org/pub/scm/linux/kernel/git/davem/net-2.6:
smc91x: let smc91x work well under netpoll
pxaficp-ir: remove incorrect net_device_ops
NET: llc, zero sockaddr_llc struct
drivers/net: fixed drivers that support netpoll use ndo_start_xmit()
netpoll: warning for ndo_start_xmit returns with interrupts enabled
net: Fix Micrel KSZ8842 Kconfig description
netfilter: xt_quota: fix wrong return value (error case)
ipv6: Fix commit 63d9950b08
(ipv6: Make v4-mapped bindings consistent with IPv4)
E100: fix interaction with swiotlb on X86.
pkt_sched: Convert CBQ to tasklet_hrtimer.
pkt_sched: Convert qdisc_watchdog to tasklet_hrtimer
rtl8187: always set MSR_LINK_ENEDCA flag with RTL8187B
ibm_newemac: emac_close() needs to call netif_carrier_off()
net: fix ks8851 build errors
net: Rename MAC platform driver for w90p910 platform
yellowfin: Fix buffer underrun after dev_alloc_skb() failure
orinoco: correct key bounds check in orinoco_hw_get_tkip_iv
mac80211: fix todo lock
This commit is contained in:
commit
1cac6ec9b7
|
@ -1727,12 +1727,14 @@ config KS8842
|
||||||
tristate "Micrel KSZ8842"
|
tristate "Micrel KSZ8842"
|
||||||
depends on HAS_IOMEM
|
depends on HAS_IOMEM
|
||||||
help
|
help
|
||||||
This platform driver is for Micrel KSZ8842 chip.
|
This platform driver is for Micrel KSZ8842 / KS8842
|
||||||
|
2-port ethernet switch chip (managed, VLAN, QoS).
|
||||||
|
|
||||||
config KS8851
|
config KS8851
|
||||||
tristate "Micrel KS8851 SPI"
|
tristate "Micrel KS8851 SPI"
|
||||||
depends on SPI
|
depends on SPI
|
||||||
select MII
|
select MII
|
||||||
|
select CRC32
|
||||||
help
|
help
|
||||||
SPI driver for Micrel KS8851 SPI attached network chip.
|
SPI driver for Micrel KS8851 SPI attached network chip.
|
||||||
|
|
||||||
|
|
|
@ -1080,7 +1080,7 @@ static struct platform_driver w90p910_ether_driver = {
|
||||||
.probe = w90p910_ether_probe,
|
.probe = w90p910_ether_probe,
|
||||||
.remove = __devexit_p(w90p910_ether_remove),
|
.remove = __devexit_p(w90p910_ether_remove),
|
||||||
.driver = {
|
.driver = {
|
||||||
.name = "w90p910-emc",
|
.name = "nuc900-emc",
|
||||||
.owner = THIS_MODULE,
|
.owner = THIS_MODULE,
|
||||||
},
|
},
|
||||||
};
|
};
|
||||||
|
@ -1101,5 +1101,5 @@ module_exit(w90p910_ether_exit);
|
||||||
MODULE_AUTHOR("Wan ZongShun <mcuos.com@gmail.com>");
|
MODULE_AUTHOR("Wan ZongShun <mcuos.com@gmail.com>");
|
||||||
MODULE_DESCRIPTION("w90p910 MAC driver!");
|
MODULE_DESCRIPTION("w90p910 MAC driver!");
|
||||||
MODULE_LICENSE("GPL");
|
MODULE_LICENSE("GPL");
|
||||||
MODULE_ALIAS("platform:w90p910-emc");
|
MODULE_ALIAS("platform:nuc900-emc");
|
||||||
|
|
||||||
|
|
|
@ -1899,7 +1899,7 @@ static int e100_rx_indicate(struct nic *nic, struct rx *rx,
|
||||||
nic->ru_running = RU_SUSPENDED;
|
nic->ru_running = RU_SUSPENDED;
|
||||||
pci_dma_sync_single_for_device(nic->pdev, rx->dma_addr,
|
pci_dma_sync_single_for_device(nic->pdev, rx->dma_addr,
|
||||||
sizeof(struct rfd),
|
sizeof(struct rfd),
|
||||||
PCI_DMA_BIDIRECTIONAL);
|
PCI_DMA_FROMDEVICE);
|
||||||
return -ENODATA;
|
return -ENODATA;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -309,6 +309,7 @@ static int mpc52xx_fec_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||||
{
|
{
|
||||||
struct mpc52xx_fec_priv *priv = netdev_priv(dev);
|
struct mpc52xx_fec_priv *priv = netdev_priv(dev);
|
||||||
struct bcom_fec_bd *bd;
|
struct bcom_fec_bd *bd;
|
||||||
|
unsigned long flags;
|
||||||
|
|
||||||
if (bcom_queue_full(priv->tx_dmatsk)) {
|
if (bcom_queue_full(priv->tx_dmatsk)) {
|
||||||
if (net_ratelimit())
|
if (net_ratelimit())
|
||||||
|
@ -316,7 +317,7 @@ static int mpc52xx_fec_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||||
return NETDEV_TX_BUSY;
|
return NETDEV_TX_BUSY;
|
||||||
}
|
}
|
||||||
|
|
||||||
spin_lock_irq(&priv->lock);
|
spin_lock_irqsave(&priv->lock, flags);
|
||||||
dev->trans_start = jiffies;
|
dev->trans_start = jiffies;
|
||||||
|
|
||||||
bd = (struct bcom_fec_bd *)
|
bd = (struct bcom_fec_bd *)
|
||||||
|
@ -332,7 +333,7 @@ static int mpc52xx_fec_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||||
netif_stop_queue(dev);
|
netif_stop_queue(dev);
|
||||||
}
|
}
|
||||||
|
|
||||||
spin_unlock_irq(&priv->lock);
|
spin_unlock_irqrestore(&priv->lock, flags);
|
||||||
|
|
||||||
return NETDEV_TX_OK;
|
return NETDEV_TX_OK;
|
||||||
}
|
}
|
||||||
|
|
|
@ -1305,6 +1305,8 @@ static int emac_close(struct net_device *ndev)
|
||||||
|
|
||||||
free_irq(dev->emac_irq, dev);
|
free_irq(dev->emac_irq, dev);
|
||||||
|
|
||||||
|
netif_carrier_off(ndev);
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -803,9 +803,6 @@ static const struct net_device_ops pxa_irda_netdev_ops = {
|
||||||
.ndo_stop = pxa_irda_stop,
|
.ndo_stop = pxa_irda_stop,
|
||||||
.ndo_start_xmit = pxa_irda_hard_xmit,
|
.ndo_start_xmit = pxa_irda_hard_xmit,
|
||||||
.ndo_do_ioctl = pxa_irda_ioctl,
|
.ndo_do_ioctl = pxa_irda_ioctl,
|
||||||
.ndo_change_mtu = eth_change_mtu,
|
|
||||||
.ndo_validate_addr = eth_validate_addr,
|
|
||||||
.ndo_set_mac_address = eth_mac_addr,
|
|
||||||
};
|
};
|
||||||
|
|
||||||
static int pxa_irda_probe(struct platform_device *pdev)
|
static int pxa_irda_probe(struct platform_device *pdev)
|
||||||
|
@ -830,6 +827,7 @@ static int pxa_irda_probe(struct platform_device *pdev)
|
||||||
if (!dev)
|
if (!dev)
|
||||||
goto err_mem_3;
|
goto err_mem_3;
|
||||||
|
|
||||||
|
SET_NETDEV_DEV(dev, &pdev->dev);
|
||||||
si = netdev_priv(dev);
|
si = netdev_priv(dev);
|
||||||
si->dev = &pdev->dev;
|
si->dev = &pdev->dev;
|
||||||
si->pdata = pdev->dev.platform_data;
|
si->pdata = pdev->dev.platform_data;
|
||||||
|
|
|
@ -41,6 +41,7 @@ static int ixpdev_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||||
struct ixpdev_priv *ip = netdev_priv(dev);
|
struct ixpdev_priv *ip = netdev_priv(dev);
|
||||||
struct ixpdev_tx_desc *desc;
|
struct ixpdev_tx_desc *desc;
|
||||||
int entry;
|
int entry;
|
||||||
|
unsigned long flags;
|
||||||
|
|
||||||
if (unlikely(skb->len > PAGE_SIZE)) {
|
if (unlikely(skb->len > PAGE_SIZE)) {
|
||||||
/* @@@ Count drops. */
|
/* @@@ Count drops. */
|
||||||
|
@ -63,11 +64,11 @@ static int ixpdev_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||||
|
|
||||||
dev->trans_start = jiffies;
|
dev->trans_start = jiffies;
|
||||||
|
|
||||||
local_irq_disable();
|
local_irq_save(flags);
|
||||||
ip->tx_queue_entries++;
|
ip->tx_queue_entries++;
|
||||||
if (ip->tx_queue_entries == TX_BUF_COUNT_PER_CHAN)
|
if (ip->tx_queue_entries == TX_BUF_COUNT_PER_CHAN)
|
||||||
netif_stop_queue(dev);
|
netif_stop_queue(dev);
|
||||||
local_irq_enable();
|
local_irq_restore(flags);
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
|
@ -620,6 +620,7 @@ static int macb_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||||
dma_addr_t mapping;
|
dma_addr_t mapping;
|
||||||
unsigned int len, entry;
|
unsigned int len, entry;
|
||||||
u32 ctrl;
|
u32 ctrl;
|
||||||
|
unsigned long flags;
|
||||||
|
|
||||||
#ifdef DEBUG
|
#ifdef DEBUG
|
||||||
int i;
|
int i;
|
||||||
|
@ -635,12 +636,12 @@ static int macb_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
len = skb->len;
|
len = skb->len;
|
||||||
spin_lock_irq(&bp->lock);
|
spin_lock_irqsave(&bp->lock, flags);
|
||||||
|
|
||||||
/* This is a hard error, log it. */
|
/* This is a hard error, log it. */
|
||||||
if (TX_BUFFS_AVAIL(bp) < 1) {
|
if (TX_BUFFS_AVAIL(bp) < 1) {
|
||||||
netif_stop_queue(dev);
|
netif_stop_queue(dev);
|
||||||
spin_unlock_irq(&bp->lock);
|
spin_unlock_irqrestore(&bp->lock, flags);
|
||||||
dev_err(&bp->pdev->dev,
|
dev_err(&bp->pdev->dev,
|
||||||
"BUG! Tx Ring full when queue awake!\n");
|
"BUG! Tx Ring full when queue awake!\n");
|
||||||
dev_dbg(&bp->pdev->dev, "tx_head = %u, tx_tail = %u\n",
|
dev_dbg(&bp->pdev->dev, "tx_head = %u, tx_tail = %u\n",
|
||||||
|
@ -674,7 +675,7 @@ static int macb_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||||
if (TX_BUFFS_AVAIL(bp) < 1)
|
if (TX_BUFFS_AVAIL(bp) < 1)
|
||||||
netif_stop_queue(dev);
|
netif_stop_queue(dev);
|
||||||
|
|
||||||
spin_unlock_irq(&bp->lock);
|
spin_unlock_irqrestore(&bp->lock, flags);
|
||||||
|
|
||||||
dev->trans_start = jiffies;
|
dev->trans_start = jiffies;
|
||||||
|
|
||||||
|
|
|
@ -437,6 +437,7 @@ static inline void mlx4_en_xmit_poll(struct mlx4_en_priv *priv, int tx_ind)
|
||||||
{
|
{
|
||||||
struct mlx4_en_cq *cq = &priv->tx_cq[tx_ind];
|
struct mlx4_en_cq *cq = &priv->tx_cq[tx_ind];
|
||||||
struct mlx4_en_tx_ring *ring = &priv->tx_ring[tx_ind];
|
struct mlx4_en_tx_ring *ring = &priv->tx_ring[tx_ind];
|
||||||
|
unsigned long flags;
|
||||||
|
|
||||||
/* If we don't have a pending timer, set one up to catch our recent
|
/* If we don't have a pending timer, set one up to catch our recent
|
||||||
post in case the interface becomes idle */
|
post in case the interface becomes idle */
|
||||||
|
@ -445,9 +446,9 @@ static inline void mlx4_en_xmit_poll(struct mlx4_en_priv *priv, int tx_ind)
|
||||||
|
|
||||||
/* Poll the CQ every mlx4_en_TX_MODER_POLL packets */
|
/* Poll the CQ every mlx4_en_TX_MODER_POLL packets */
|
||||||
if ((++ring->poll_cnt & (MLX4_EN_TX_POLL_MODER - 1)) == 0)
|
if ((++ring->poll_cnt & (MLX4_EN_TX_POLL_MODER - 1)) == 0)
|
||||||
if (spin_trylock_irq(&ring->comp_lock)) {
|
if (spin_trylock_irqsave(&ring->comp_lock, flags)) {
|
||||||
mlx4_en_process_tx_cq(priv->dev, cq);
|
mlx4_en_process_tx_cq(priv->dev, cq);
|
||||||
spin_unlock_irq(&ring->comp_lock);
|
spin_unlock_irqrestore(&ring->comp_lock, flags);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -196,21 +196,23 @@ static void PRINT_PKT(u_char *buf, int length)
|
||||||
/* this enables an interrupt in the interrupt mask register */
|
/* this enables an interrupt in the interrupt mask register */
|
||||||
#define SMC_ENABLE_INT(lp, x) do { \
|
#define SMC_ENABLE_INT(lp, x) do { \
|
||||||
unsigned char mask; \
|
unsigned char mask; \
|
||||||
spin_lock_irq(&lp->lock); \
|
unsigned long smc_enable_flags; \
|
||||||
|
spin_lock_irqsave(&lp->lock, smc_enable_flags); \
|
||||||
mask = SMC_GET_INT_MASK(lp); \
|
mask = SMC_GET_INT_MASK(lp); \
|
||||||
mask |= (x); \
|
mask |= (x); \
|
||||||
SMC_SET_INT_MASK(lp, mask); \
|
SMC_SET_INT_MASK(lp, mask); \
|
||||||
spin_unlock_irq(&lp->lock); \
|
spin_unlock_irqrestore(&lp->lock, smc_enable_flags); \
|
||||||
} while (0)
|
} while (0)
|
||||||
|
|
||||||
/* this disables an interrupt from the interrupt mask register */
|
/* this disables an interrupt from the interrupt mask register */
|
||||||
#define SMC_DISABLE_INT(lp, x) do { \
|
#define SMC_DISABLE_INT(lp, x) do { \
|
||||||
unsigned char mask; \
|
unsigned char mask; \
|
||||||
spin_lock_irq(&lp->lock); \
|
unsigned long smc_disable_flags; \
|
||||||
|
spin_lock_irqsave(&lp->lock, smc_disable_flags); \
|
||||||
mask = SMC_GET_INT_MASK(lp); \
|
mask = SMC_GET_INT_MASK(lp); \
|
||||||
mask &= ~(x); \
|
mask &= ~(x); \
|
||||||
SMC_SET_INT_MASK(lp, mask); \
|
SMC_SET_INT_MASK(lp, mask); \
|
||||||
spin_unlock_irq(&lp->lock); \
|
spin_unlock_irqrestore(&lp->lock, smc_disable_flags); \
|
||||||
} while (0)
|
} while (0)
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -520,21 +522,21 @@ static inline void smc_rcv(struct net_device *dev)
|
||||||
* any other concurrent access and C would always interrupt B. But life
|
* any other concurrent access and C would always interrupt B. But life
|
||||||
* isn't that easy in a SMP world...
|
* isn't that easy in a SMP world...
|
||||||
*/
|
*/
|
||||||
#define smc_special_trylock(lock) \
|
#define smc_special_trylock(lock, flags) \
|
||||||
({ \
|
({ \
|
||||||
int __ret; \
|
int __ret; \
|
||||||
local_irq_disable(); \
|
local_irq_save(flags); \
|
||||||
__ret = spin_trylock(lock); \
|
__ret = spin_trylock(lock); \
|
||||||
if (!__ret) \
|
if (!__ret) \
|
||||||
local_irq_enable(); \
|
local_irq_restore(flags); \
|
||||||
__ret; \
|
__ret; \
|
||||||
})
|
})
|
||||||
#define smc_special_lock(lock) spin_lock_irq(lock)
|
#define smc_special_lock(lock, flags) spin_lock_irq(lock, flags)
|
||||||
#define smc_special_unlock(lock) spin_unlock_irq(lock)
|
#define smc_special_unlock(lock, flags) spin_unlock_irqrestore(lock, flags)
|
||||||
#else
|
#else
|
||||||
#define smc_special_trylock(lock) (1)
|
#define smc_special_trylock(lock, flags) (1)
|
||||||
#define smc_special_lock(lock) do { } while (0)
|
#define smc_special_lock(lock, flags) do { } while (0)
|
||||||
#define smc_special_unlock(lock) do { } while (0)
|
#define smc_special_unlock(lock, flags) do { } while (0)
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -548,10 +550,11 @@ static void smc_hardware_send_pkt(unsigned long data)
|
||||||
struct sk_buff *skb;
|
struct sk_buff *skb;
|
||||||
unsigned int packet_no, len;
|
unsigned int packet_no, len;
|
||||||
unsigned char *buf;
|
unsigned char *buf;
|
||||||
|
unsigned long flags;
|
||||||
|
|
||||||
DBG(3, "%s: %s\n", dev->name, __func__);
|
DBG(3, "%s: %s\n", dev->name, __func__);
|
||||||
|
|
||||||
if (!smc_special_trylock(&lp->lock)) {
|
if (!smc_special_trylock(&lp->lock, flags)) {
|
||||||
netif_stop_queue(dev);
|
netif_stop_queue(dev);
|
||||||
tasklet_schedule(&lp->tx_task);
|
tasklet_schedule(&lp->tx_task);
|
||||||
return;
|
return;
|
||||||
|
@ -559,7 +562,7 @@ static void smc_hardware_send_pkt(unsigned long data)
|
||||||
|
|
||||||
skb = lp->pending_tx_skb;
|
skb = lp->pending_tx_skb;
|
||||||
if (unlikely(!skb)) {
|
if (unlikely(!skb)) {
|
||||||
smc_special_unlock(&lp->lock);
|
smc_special_unlock(&lp->lock, flags);
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
lp->pending_tx_skb = NULL;
|
lp->pending_tx_skb = NULL;
|
||||||
|
@ -569,7 +572,7 @@ static void smc_hardware_send_pkt(unsigned long data)
|
||||||
printk("%s: Memory allocation failed.\n", dev->name);
|
printk("%s: Memory allocation failed.\n", dev->name);
|
||||||
dev->stats.tx_errors++;
|
dev->stats.tx_errors++;
|
||||||
dev->stats.tx_fifo_errors++;
|
dev->stats.tx_fifo_errors++;
|
||||||
smc_special_unlock(&lp->lock);
|
smc_special_unlock(&lp->lock, flags);
|
||||||
goto done;
|
goto done;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -608,7 +611,7 @@ static void smc_hardware_send_pkt(unsigned long data)
|
||||||
|
|
||||||
/* queue the packet for TX */
|
/* queue the packet for TX */
|
||||||
SMC_SET_MMU_CMD(lp, MC_ENQUEUE);
|
SMC_SET_MMU_CMD(lp, MC_ENQUEUE);
|
||||||
smc_special_unlock(&lp->lock);
|
smc_special_unlock(&lp->lock, flags);
|
||||||
|
|
||||||
dev->trans_start = jiffies;
|
dev->trans_start = jiffies;
|
||||||
dev->stats.tx_packets++;
|
dev->stats.tx_packets++;
|
||||||
|
@ -633,6 +636,7 @@ static int smc_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||||
struct smc_local *lp = netdev_priv(dev);
|
struct smc_local *lp = netdev_priv(dev);
|
||||||
void __iomem *ioaddr = lp->base;
|
void __iomem *ioaddr = lp->base;
|
||||||
unsigned int numPages, poll_count, status;
|
unsigned int numPages, poll_count, status;
|
||||||
|
unsigned long flags;
|
||||||
|
|
||||||
DBG(3, "%s: %s\n", dev->name, __func__);
|
DBG(3, "%s: %s\n", dev->name, __func__);
|
||||||
|
|
||||||
|
@ -658,7 +662,7 @@ static int smc_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
smc_special_lock(&lp->lock);
|
smc_special_lock(&lp->lock, flags);
|
||||||
|
|
||||||
/* now, try to allocate the memory */
|
/* now, try to allocate the memory */
|
||||||
SMC_SET_MMU_CMD(lp, MC_ALLOC | numPages);
|
SMC_SET_MMU_CMD(lp, MC_ALLOC | numPages);
|
||||||
|
@ -676,7 +680,7 @@ static int smc_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||||
}
|
}
|
||||||
} while (--poll_count);
|
} while (--poll_count);
|
||||||
|
|
||||||
smc_special_unlock(&lp->lock);
|
smc_special_unlock(&lp->lock, flags);
|
||||||
|
|
||||||
lp->pending_tx_skb = skb;
|
lp->pending_tx_skb = skb;
|
||||||
if (!poll_count) {
|
if (!poll_count) {
|
||||||
|
|
|
@ -70,7 +70,7 @@ int orinoco_hw_get_tkip_iv(struct orinoco_private *priv, int key, u8 *tsc)
|
||||||
int err = 0;
|
int err = 0;
|
||||||
u8 tsc_arr[4][IW_ENCODE_SEQ_MAX_SIZE];
|
u8 tsc_arr[4][IW_ENCODE_SEQ_MAX_SIZE];
|
||||||
|
|
||||||
if ((key < 0) || (key > 4))
|
if ((key < 0) || (key >= 4))
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
|
|
||||||
err = hermes_read_ltv(hw, USER_BAP, HERMES_RID_CURRENT_TKIP_IV,
|
err = hermes_read_ltv(hw, USER_BAP, HERMES_RID_CURRENT_TKIP_IV,
|
||||||
|
|
|
@ -869,6 +869,9 @@ static int rtl8187b_init_hw(struct ieee80211_hw *dev)
|
||||||
priv->aifsn[3] = 3; /* AIFSN[AC_BE] */
|
priv->aifsn[3] = 3; /* AIFSN[AC_BE] */
|
||||||
rtl818x_iowrite8(priv, &priv->map->ACM_CONTROL, 0);
|
rtl818x_iowrite8(priv, &priv->map->ACM_CONTROL, 0);
|
||||||
|
|
||||||
|
/* ENEDCA flag must always be set, transmit issues? */
|
||||||
|
rtl818x_iowrite8(priv, &priv->map->MSR, RTL818X_MSR_ENEDCA);
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1173,13 +1176,16 @@ static void rtl8187_bss_info_changed(struct ieee80211_hw *dev,
|
||||||
rtl818x_iowrite8(priv, &priv->map->BSSID[i],
|
rtl818x_iowrite8(priv, &priv->map->BSSID[i],
|
||||||
info->bssid[i]);
|
info->bssid[i]);
|
||||||
|
|
||||||
|
if (priv->is_rtl8187b)
|
||||||
|
reg = RTL818X_MSR_ENEDCA;
|
||||||
|
else
|
||||||
|
reg = 0;
|
||||||
|
|
||||||
if (is_valid_ether_addr(info->bssid)) {
|
if (is_valid_ether_addr(info->bssid)) {
|
||||||
reg = RTL818X_MSR_INFRA;
|
reg |= RTL818X_MSR_INFRA;
|
||||||
if (priv->is_rtl8187b)
|
|
||||||
reg |= RTL818X_MSR_ENEDCA;
|
|
||||||
rtl818x_iowrite8(priv, &priv->map->MSR, reg);
|
rtl818x_iowrite8(priv, &priv->map->MSR, reg);
|
||||||
} else {
|
} else {
|
||||||
reg = RTL818X_MSR_NO_LINK;
|
reg |= RTL818X_MSR_NO_LINK;
|
||||||
rtl818x_iowrite8(priv, &priv->map->MSR, reg);
|
rtl818x_iowrite8(priv, &priv->map->MSR, reg);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -346,7 +346,7 @@ static int netdev_ioctl(struct net_device *dev, struct ifreq *rq, int cmd);
|
||||||
static int yellowfin_open(struct net_device *dev);
|
static int yellowfin_open(struct net_device *dev);
|
||||||
static void yellowfin_timer(unsigned long data);
|
static void yellowfin_timer(unsigned long data);
|
||||||
static void yellowfin_tx_timeout(struct net_device *dev);
|
static void yellowfin_tx_timeout(struct net_device *dev);
|
||||||
static void yellowfin_init_ring(struct net_device *dev);
|
static int yellowfin_init_ring(struct net_device *dev);
|
||||||
static int yellowfin_start_xmit(struct sk_buff *skb, struct net_device *dev);
|
static int yellowfin_start_xmit(struct sk_buff *skb, struct net_device *dev);
|
||||||
static irqreturn_t yellowfin_interrupt(int irq, void *dev_instance);
|
static irqreturn_t yellowfin_interrupt(int irq, void *dev_instance);
|
||||||
static int yellowfin_rx(struct net_device *dev);
|
static int yellowfin_rx(struct net_device *dev);
|
||||||
|
@ -573,19 +573,24 @@ static int yellowfin_open(struct net_device *dev)
|
||||||
{
|
{
|
||||||
struct yellowfin_private *yp = netdev_priv(dev);
|
struct yellowfin_private *yp = netdev_priv(dev);
|
||||||
void __iomem *ioaddr = yp->base;
|
void __iomem *ioaddr = yp->base;
|
||||||
int i;
|
int i, ret;
|
||||||
|
|
||||||
/* Reset the chip. */
|
/* Reset the chip. */
|
||||||
iowrite32(0x80000000, ioaddr + DMACtrl);
|
iowrite32(0x80000000, ioaddr + DMACtrl);
|
||||||
|
|
||||||
i = request_irq(dev->irq, &yellowfin_interrupt, IRQF_SHARED, dev->name, dev);
|
ret = request_irq(dev->irq, &yellowfin_interrupt, IRQF_SHARED, dev->name, dev);
|
||||||
if (i) return i;
|
if (ret)
|
||||||
|
return ret;
|
||||||
|
|
||||||
if (yellowfin_debug > 1)
|
if (yellowfin_debug > 1)
|
||||||
printk(KERN_DEBUG "%s: yellowfin_open() irq %d.\n",
|
printk(KERN_DEBUG "%s: yellowfin_open() irq %d.\n",
|
||||||
dev->name, dev->irq);
|
dev->name, dev->irq);
|
||||||
|
|
||||||
yellowfin_init_ring(dev);
|
ret = yellowfin_init_ring(dev);
|
||||||
|
if (ret) {
|
||||||
|
free_irq(dev->irq, dev);
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
iowrite32(yp->rx_ring_dma, ioaddr + RxPtr);
|
iowrite32(yp->rx_ring_dma, ioaddr + RxPtr);
|
||||||
iowrite32(yp->tx_ring_dma, ioaddr + TxPtr);
|
iowrite32(yp->tx_ring_dma, ioaddr + TxPtr);
|
||||||
|
@ -725,10 +730,10 @@ static void yellowfin_tx_timeout(struct net_device *dev)
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Initialize the Rx and Tx rings, along with various 'dev' bits. */
|
/* Initialize the Rx and Tx rings, along with various 'dev' bits. */
|
||||||
static void yellowfin_init_ring(struct net_device *dev)
|
static int yellowfin_init_ring(struct net_device *dev)
|
||||||
{
|
{
|
||||||
struct yellowfin_private *yp = netdev_priv(dev);
|
struct yellowfin_private *yp = netdev_priv(dev);
|
||||||
int i;
|
int i, j;
|
||||||
|
|
||||||
yp->tx_full = 0;
|
yp->tx_full = 0;
|
||||||
yp->cur_rx = yp->cur_tx = 0;
|
yp->cur_rx = yp->cur_tx = 0;
|
||||||
|
@ -753,6 +758,11 @@ static void yellowfin_init_ring(struct net_device *dev)
|
||||||
yp->rx_ring[i].addr = cpu_to_le32(pci_map_single(yp->pci_dev,
|
yp->rx_ring[i].addr = cpu_to_le32(pci_map_single(yp->pci_dev,
|
||||||
skb->data, yp->rx_buf_sz, PCI_DMA_FROMDEVICE));
|
skb->data, yp->rx_buf_sz, PCI_DMA_FROMDEVICE));
|
||||||
}
|
}
|
||||||
|
if (i != RX_RING_SIZE) {
|
||||||
|
for (j = 0; j < i; j++)
|
||||||
|
dev_kfree_skb(yp->rx_skbuff[j]);
|
||||||
|
return -ENOMEM;
|
||||||
|
}
|
||||||
yp->rx_ring[i-1].dbdma_cmd = cpu_to_le32(CMD_STOP);
|
yp->rx_ring[i-1].dbdma_cmd = cpu_to_le32(CMD_STOP);
|
||||||
yp->dirty_rx = (unsigned int)(i - RX_RING_SIZE);
|
yp->dirty_rx = (unsigned int)(i - RX_RING_SIZE);
|
||||||
|
|
||||||
|
@ -769,8 +779,6 @@ static void yellowfin_init_ring(struct net_device *dev)
|
||||||
yp->tx_ring[--i].dbdma_cmd = cpu_to_le32(CMD_STOP | BRANCH_ALWAYS);
|
yp->tx_ring[--i].dbdma_cmd = cpu_to_le32(CMD_STOP | BRANCH_ALWAYS);
|
||||||
#else
|
#else
|
||||||
{
|
{
|
||||||
int j;
|
|
||||||
|
|
||||||
/* Tx ring needs a pair of descriptors, the second for the status. */
|
/* Tx ring needs a pair of descriptors, the second for the status. */
|
||||||
for (i = 0; i < TX_RING_SIZE; i++) {
|
for (i = 0; i < TX_RING_SIZE; i++) {
|
||||||
j = 2*i;
|
j = 2*i;
|
||||||
|
@ -805,7 +813,7 @@ static void yellowfin_init_ring(struct net_device *dev)
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
yp->tx_tail_desc = &yp->tx_status[0];
|
yp->tx_tail_desc = &yp->tx_status[0];
|
||||||
return;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int yellowfin_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
static int yellowfin_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||||
|
|
|
@ -61,8 +61,8 @@ psched_tdiff_bounded(psched_time_t tv1, psched_time_t tv2, psched_time_t bound)
|
||||||
}
|
}
|
||||||
|
|
||||||
struct qdisc_watchdog {
|
struct qdisc_watchdog {
|
||||||
struct hrtimer timer;
|
struct tasklet_hrtimer timer;
|
||||||
struct Qdisc *qdisc;
|
struct Qdisc *qdisc;
|
||||||
};
|
};
|
||||||
|
|
||||||
extern void qdisc_watchdog_init(struct qdisc_watchdog *wd, struct Qdisc *qdisc);
|
extern void qdisc_watchdog_init(struct qdisc_watchdog *wd, struct Qdisc *qdisc);
|
||||||
|
|
|
@ -319,6 +319,11 @@ static void netpoll_send_skb(struct netpoll *np, struct sk_buff *skb)
|
||||||
|
|
||||||
udelay(USEC_PER_POLL);
|
udelay(USEC_PER_POLL);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
WARN_ONCE(!irqs_disabled(),
|
||||||
|
"netpoll_send_skb(): %s enabled interrupts in poll (%pF)\n",
|
||||||
|
dev->name, ops->ndo_start_xmit);
|
||||||
|
|
||||||
local_irq_restore(flags);
|
local_irq_restore(flags);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -306,8 +306,10 @@ int inet6_bind(struct socket *sock, struct sockaddr *uaddr, int addr_len)
|
||||||
v4addr != htonl(INADDR_ANY) &&
|
v4addr != htonl(INADDR_ANY) &&
|
||||||
chk_addr_ret != RTN_LOCAL &&
|
chk_addr_ret != RTN_LOCAL &&
|
||||||
chk_addr_ret != RTN_MULTICAST &&
|
chk_addr_ret != RTN_MULTICAST &&
|
||||||
chk_addr_ret != RTN_BROADCAST)
|
chk_addr_ret != RTN_BROADCAST) {
|
||||||
|
err = -EADDRNOTAVAIL;
|
||||||
goto out;
|
goto out;
|
||||||
|
}
|
||||||
} else {
|
} else {
|
||||||
if (addr_type != IPV6_ADDR_ANY) {
|
if (addr_type != IPV6_ADDR_ANY) {
|
||||||
struct net_device *dev = NULL;
|
struct net_device *dev = NULL;
|
||||||
|
|
|
@ -914,6 +914,7 @@ static int llc_ui_getname(struct socket *sock, struct sockaddr *uaddr,
|
||||||
struct llc_sock *llc = llc_sk(sk);
|
struct llc_sock *llc = llc_sk(sk);
|
||||||
int rc = 0;
|
int rc = 0;
|
||||||
|
|
||||||
|
memset(&sllc, 0, sizeof(sllc));
|
||||||
lock_sock(sk);
|
lock_sock(sk);
|
||||||
if (sock_flag(sk, SOCK_ZAPPED))
|
if (sock_flag(sk, SOCK_ZAPPED))
|
||||||
goto out;
|
goto out;
|
||||||
|
|
|
@ -67,6 +67,8 @@ static DECLARE_WORK(todo_work, key_todo);
|
||||||
*
|
*
|
||||||
* @key: key to add to do item for
|
* @key: key to add to do item for
|
||||||
* @flag: todo flag(s)
|
* @flag: todo flag(s)
|
||||||
|
*
|
||||||
|
* Must be called with IRQs or softirqs disabled.
|
||||||
*/
|
*/
|
||||||
static void add_todo(struct ieee80211_key *key, u32 flag)
|
static void add_todo(struct ieee80211_key *key, u32 flag)
|
||||||
{
|
{
|
||||||
|
@ -140,9 +142,9 @@ static void ieee80211_key_enable_hw_accel(struct ieee80211_key *key)
|
||||||
ret = drv_set_key(key->local, SET_KEY, &sdata->vif, sta, &key->conf);
|
ret = drv_set_key(key->local, SET_KEY, &sdata->vif, sta, &key->conf);
|
||||||
|
|
||||||
if (!ret) {
|
if (!ret) {
|
||||||
spin_lock(&todo_lock);
|
spin_lock_bh(&todo_lock);
|
||||||
key->flags |= KEY_FLAG_UPLOADED_TO_HARDWARE;
|
key->flags |= KEY_FLAG_UPLOADED_TO_HARDWARE;
|
||||||
spin_unlock(&todo_lock);
|
spin_unlock_bh(&todo_lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (ret && ret != -ENOSPC && ret != -EOPNOTSUPP)
|
if (ret && ret != -ENOSPC && ret != -EOPNOTSUPP)
|
||||||
|
@ -164,12 +166,12 @@ static void ieee80211_key_disable_hw_accel(struct ieee80211_key *key)
|
||||||
if (!key || !key->local->ops->set_key)
|
if (!key || !key->local->ops->set_key)
|
||||||
return;
|
return;
|
||||||
|
|
||||||
spin_lock(&todo_lock);
|
spin_lock_bh(&todo_lock);
|
||||||
if (!(key->flags & KEY_FLAG_UPLOADED_TO_HARDWARE)) {
|
if (!(key->flags & KEY_FLAG_UPLOADED_TO_HARDWARE)) {
|
||||||
spin_unlock(&todo_lock);
|
spin_unlock_bh(&todo_lock);
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
spin_unlock(&todo_lock);
|
spin_unlock_bh(&todo_lock);
|
||||||
|
|
||||||
sta = get_sta_for_key(key);
|
sta = get_sta_for_key(key);
|
||||||
sdata = key->sdata;
|
sdata = key->sdata;
|
||||||
|
@ -188,9 +190,9 @@ static void ieee80211_key_disable_hw_accel(struct ieee80211_key *key)
|
||||||
wiphy_name(key->local->hw.wiphy),
|
wiphy_name(key->local->hw.wiphy),
|
||||||
key->conf.keyidx, sta ? sta->addr : bcast_addr, ret);
|
key->conf.keyidx, sta ? sta->addr : bcast_addr, ret);
|
||||||
|
|
||||||
spin_lock(&todo_lock);
|
spin_lock_bh(&todo_lock);
|
||||||
key->flags &= ~KEY_FLAG_UPLOADED_TO_HARDWARE;
|
key->flags &= ~KEY_FLAG_UPLOADED_TO_HARDWARE;
|
||||||
spin_unlock(&todo_lock);
|
spin_unlock_bh(&todo_lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void __ieee80211_set_default_key(struct ieee80211_sub_if_data *sdata,
|
static void __ieee80211_set_default_key(struct ieee80211_sub_if_data *sdata,
|
||||||
|
@ -437,14 +439,14 @@ void ieee80211_key_link(struct ieee80211_key *key,
|
||||||
|
|
||||||
__ieee80211_key_replace(sdata, sta, old_key, key);
|
__ieee80211_key_replace(sdata, sta, old_key, key);
|
||||||
|
|
||||||
spin_unlock_irqrestore(&sdata->local->key_lock, flags);
|
|
||||||
|
|
||||||
/* free old key later */
|
/* free old key later */
|
||||||
add_todo(old_key, KEY_FLAG_TODO_DELETE);
|
add_todo(old_key, KEY_FLAG_TODO_DELETE);
|
||||||
|
|
||||||
add_todo(key, KEY_FLAG_TODO_ADD_DEBUGFS);
|
add_todo(key, KEY_FLAG_TODO_ADD_DEBUGFS);
|
||||||
if (netif_running(sdata->dev))
|
if (netif_running(sdata->dev))
|
||||||
add_todo(key, KEY_FLAG_TODO_HWACCEL_ADD);
|
add_todo(key, KEY_FLAG_TODO_HWACCEL_ADD);
|
||||||
|
|
||||||
|
spin_unlock_irqrestore(&sdata->local->key_lock, flags);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void __ieee80211_key_free(struct ieee80211_key *key)
|
static void __ieee80211_key_free(struct ieee80211_key *key)
|
||||||
|
@ -547,7 +549,7 @@ static void __ieee80211_key_todo(void)
|
||||||
*/
|
*/
|
||||||
synchronize_rcu();
|
synchronize_rcu();
|
||||||
|
|
||||||
spin_lock(&todo_lock);
|
spin_lock_bh(&todo_lock);
|
||||||
while (!list_empty(&todo_list)) {
|
while (!list_empty(&todo_list)) {
|
||||||
key = list_first_entry(&todo_list, struct ieee80211_key, todo);
|
key = list_first_entry(&todo_list, struct ieee80211_key, todo);
|
||||||
list_del_init(&key->todo);
|
list_del_init(&key->todo);
|
||||||
|
@ -558,7 +560,7 @@ static void __ieee80211_key_todo(void)
|
||||||
KEY_FLAG_TODO_HWACCEL_REMOVE |
|
KEY_FLAG_TODO_HWACCEL_REMOVE |
|
||||||
KEY_FLAG_TODO_DELETE);
|
KEY_FLAG_TODO_DELETE);
|
||||||
key->flags &= ~todoflags;
|
key->flags &= ~todoflags;
|
||||||
spin_unlock(&todo_lock);
|
spin_unlock_bh(&todo_lock);
|
||||||
|
|
||||||
work_done = false;
|
work_done = false;
|
||||||
|
|
||||||
|
@ -591,9 +593,9 @@ static void __ieee80211_key_todo(void)
|
||||||
|
|
||||||
WARN_ON(!work_done);
|
WARN_ON(!work_done);
|
||||||
|
|
||||||
spin_lock(&todo_lock);
|
spin_lock_bh(&todo_lock);
|
||||||
}
|
}
|
||||||
spin_unlock(&todo_lock);
|
spin_unlock_bh(&todo_lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
void ieee80211_key_todo(void)
|
void ieee80211_key_todo(void)
|
||||||
|
|
|
@ -52,7 +52,7 @@ static bool quota_mt_check(const struct xt_mtchk_param *par)
|
||||||
|
|
||||||
q->master = kmalloc(sizeof(*q->master), GFP_KERNEL);
|
q->master = kmalloc(sizeof(*q->master), GFP_KERNEL);
|
||||||
if (q->master == NULL)
|
if (q->master == NULL)
|
||||||
return -ENOMEM;
|
return false;
|
||||||
|
|
||||||
q->master->quota = q->quota;
|
q->master->quota = q->quota;
|
||||||
return true;
|
return true;
|
||||||
|
|
|
@ -468,8 +468,8 @@ static enum hrtimer_restart qdisc_watchdog(struct hrtimer *timer)
|
||||||
|
|
||||||
void qdisc_watchdog_init(struct qdisc_watchdog *wd, struct Qdisc *qdisc)
|
void qdisc_watchdog_init(struct qdisc_watchdog *wd, struct Qdisc *qdisc)
|
||||||
{
|
{
|
||||||
hrtimer_init(&wd->timer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS);
|
tasklet_hrtimer_init(&wd->timer, qdisc_watchdog,
|
||||||
wd->timer.function = qdisc_watchdog;
|
CLOCK_MONOTONIC, HRTIMER_MODE_ABS);
|
||||||
wd->qdisc = qdisc;
|
wd->qdisc = qdisc;
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(qdisc_watchdog_init);
|
EXPORT_SYMBOL(qdisc_watchdog_init);
|
||||||
|
@ -485,13 +485,13 @@ void qdisc_watchdog_schedule(struct qdisc_watchdog *wd, psched_time_t expires)
|
||||||
wd->qdisc->flags |= TCQ_F_THROTTLED;
|
wd->qdisc->flags |= TCQ_F_THROTTLED;
|
||||||
time = ktime_set(0, 0);
|
time = ktime_set(0, 0);
|
||||||
time = ktime_add_ns(time, PSCHED_TICKS2NS(expires));
|
time = ktime_add_ns(time, PSCHED_TICKS2NS(expires));
|
||||||
hrtimer_start(&wd->timer, time, HRTIMER_MODE_ABS);
|
tasklet_hrtimer_start(&wd->timer, time, HRTIMER_MODE_ABS);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(qdisc_watchdog_schedule);
|
EXPORT_SYMBOL(qdisc_watchdog_schedule);
|
||||||
|
|
||||||
void qdisc_watchdog_cancel(struct qdisc_watchdog *wd)
|
void qdisc_watchdog_cancel(struct qdisc_watchdog *wd)
|
||||||
{
|
{
|
||||||
hrtimer_cancel(&wd->timer);
|
tasklet_hrtimer_cancel(&wd->timer);
|
||||||
wd->qdisc->flags &= ~TCQ_F_THROTTLED;
|
wd->qdisc->flags &= ~TCQ_F_THROTTLED;
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(qdisc_watchdog_cancel);
|
EXPORT_SYMBOL(qdisc_watchdog_cancel);
|
||||||
|
|
|
@ -163,7 +163,7 @@ struct cbq_sched_data
|
||||||
psched_time_t now_rt; /* Cached real time */
|
psched_time_t now_rt; /* Cached real time */
|
||||||
unsigned pmask;
|
unsigned pmask;
|
||||||
|
|
||||||
struct hrtimer delay_timer;
|
struct tasklet_hrtimer delay_timer;
|
||||||
struct qdisc_watchdog watchdog; /* Watchdog timer,
|
struct qdisc_watchdog watchdog; /* Watchdog timer,
|
||||||
started when CBQ has
|
started when CBQ has
|
||||||
backlog, but cannot
|
backlog, but cannot
|
||||||
|
@ -503,6 +503,8 @@ static void cbq_ovl_delay(struct cbq_class *cl)
|
||||||
cl->undertime = q->now + delay;
|
cl->undertime = q->now + delay;
|
||||||
|
|
||||||
if (delay > 0) {
|
if (delay > 0) {
|
||||||
|
struct hrtimer *ht;
|
||||||
|
|
||||||
sched += delay + cl->penalty;
|
sched += delay + cl->penalty;
|
||||||
cl->penalized = sched;
|
cl->penalized = sched;
|
||||||
cl->cpriority = TC_CBQ_MAXPRIO;
|
cl->cpriority = TC_CBQ_MAXPRIO;
|
||||||
|
@ -510,12 +512,12 @@ static void cbq_ovl_delay(struct cbq_class *cl)
|
||||||
|
|
||||||
expires = ktime_set(0, 0);
|
expires = ktime_set(0, 0);
|
||||||
expires = ktime_add_ns(expires, PSCHED_TICKS2NS(sched));
|
expires = ktime_add_ns(expires, PSCHED_TICKS2NS(sched));
|
||||||
if (hrtimer_try_to_cancel(&q->delay_timer) &&
|
ht = &q->delay_timer.timer;
|
||||||
ktime_to_ns(ktime_sub(
|
if (hrtimer_try_to_cancel(ht) &&
|
||||||
hrtimer_get_expires(&q->delay_timer),
|
ktime_to_ns(ktime_sub(hrtimer_get_expires(ht),
|
||||||
expires)) > 0)
|
expires)) > 0)
|
||||||
hrtimer_set_expires(&q->delay_timer, expires);
|
hrtimer_set_expires(ht, expires);
|
||||||
hrtimer_restart(&q->delay_timer);
|
hrtimer_restart(ht);
|
||||||
cl->delayed = 1;
|
cl->delayed = 1;
|
||||||
cl->xstats.overactions++;
|
cl->xstats.overactions++;
|
||||||
return;
|
return;
|
||||||
|
@ -621,7 +623,7 @@ static enum hrtimer_restart cbq_undelay(struct hrtimer *timer)
|
||||||
|
|
||||||
time = ktime_set(0, 0);
|
time = ktime_set(0, 0);
|
||||||
time = ktime_add_ns(time, PSCHED_TICKS2NS(now + delay));
|
time = ktime_add_ns(time, PSCHED_TICKS2NS(now + delay));
|
||||||
hrtimer_start(&q->delay_timer, time, HRTIMER_MODE_ABS);
|
tasklet_hrtimer_start(&q->delay_timer, time, HRTIMER_MODE_ABS);
|
||||||
}
|
}
|
||||||
|
|
||||||
sch->flags &= ~TCQ_F_THROTTLED;
|
sch->flags &= ~TCQ_F_THROTTLED;
|
||||||
|
@ -1214,7 +1216,7 @@ cbq_reset(struct Qdisc* sch)
|
||||||
q->tx_class = NULL;
|
q->tx_class = NULL;
|
||||||
q->tx_borrowed = NULL;
|
q->tx_borrowed = NULL;
|
||||||
qdisc_watchdog_cancel(&q->watchdog);
|
qdisc_watchdog_cancel(&q->watchdog);
|
||||||
hrtimer_cancel(&q->delay_timer);
|
tasklet_hrtimer_cancel(&q->delay_timer);
|
||||||
q->toplevel = TC_CBQ_MAXLEVEL;
|
q->toplevel = TC_CBQ_MAXLEVEL;
|
||||||
q->now = psched_get_time();
|
q->now = psched_get_time();
|
||||||
q->now_rt = q->now;
|
q->now_rt = q->now;
|
||||||
|
@ -1397,7 +1399,8 @@ static int cbq_init(struct Qdisc *sch, struct nlattr *opt)
|
||||||
q->link.minidle = -0x7FFFFFFF;
|
q->link.minidle = -0x7FFFFFFF;
|
||||||
|
|
||||||
qdisc_watchdog_init(&q->watchdog, sch);
|
qdisc_watchdog_init(&q->watchdog, sch);
|
||||||
hrtimer_init(&q->delay_timer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS);
|
tasklet_hrtimer_init(&q->delay_timer, cbq_undelay,
|
||||||
|
CLOCK_MONOTONIC, HRTIMER_MODE_ABS);
|
||||||
q->delay_timer.function = cbq_undelay;
|
q->delay_timer.function = cbq_undelay;
|
||||||
q->toplevel = TC_CBQ_MAXLEVEL;
|
q->toplevel = TC_CBQ_MAXLEVEL;
|
||||||
q->now = psched_get_time();
|
q->now = psched_get_time();
|
||||||
|
|
Loading…
Reference in New Issue