drivers:net: Convert dma_alloc_coherent(...__GFP_ZERO) to dma_zalloc_coherent
__GFP_ZERO is an uncommon flag and perhaps is better not used. static inline dma_zalloc_coherent exists so convert the uses of dma_alloc_coherent with __GFP_ZERO to the more common kernel style with zalloc. Remove memset from the static inline dma_zalloc_coherent and add just one use of __GFP_ZERO instead. Trivially reduces the size of the existing uses of dma_zalloc_coherent. Realign arguments as appropriate. Signed-off-by: Joe Perches <joe@perches.com> Acked-by: Neil Horman <nhorman@tuxdriver.com> Acked-by: Jesse Brandeburg <jesse.brandeburg@intel.com> Acked-by: Jeff Kirsher <jeffrey.t.kirsher@intel.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
afe4fd0624
commit
ede23fa816
|
@ -1464,18 +1464,18 @@ static int greth_of_probe(struct platform_device *ofdev)
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Allocate TX descriptor ring in coherent memory */
|
/* Allocate TX descriptor ring in coherent memory */
|
||||||
greth->tx_bd_base = dma_alloc_coherent(greth->dev, 1024,
|
greth->tx_bd_base = dma_zalloc_coherent(greth->dev, 1024,
|
||||||
&greth->tx_bd_base_phys,
|
&greth->tx_bd_base_phys,
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
GFP_KERNEL);
|
||||||
if (!greth->tx_bd_base) {
|
if (!greth->tx_bd_base) {
|
||||||
err = -ENOMEM;
|
err = -ENOMEM;
|
||||||
goto error3;
|
goto error3;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Allocate RX descriptor ring in coherent memory */
|
/* Allocate RX descriptor ring in coherent memory */
|
||||||
greth->rx_bd_base = dma_alloc_coherent(greth->dev, 1024,
|
greth->rx_bd_base = dma_zalloc_coherent(greth->dev, 1024,
|
||||||
&greth->rx_bd_base_phys,
|
&greth->rx_bd_base_phys,
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
GFP_KERNEL);
|
||||||
if (!greth->rx_bd_base) {
|
if (!greth->rx_bd_base) {
|
||||||
err = -ENOMEM;
|
err = -ENOMEM;
|
||||||
goto error4;
|
goto error4;
|
||||||
|
|
|
@ -948,8 +948,7 @@ static int bcm_enet_open(struct net_device *dev)
|
||||||
|
|
||||||
/* allocate rx dma ring */
|
/* allocate rx dma ring */
|
||||||
size = priv->rx_ring_size * sizeof(struct bcm_enet_desc);
|
size = priv->rx_ring_size * sizeof(struct bcm_enet_desc);
|
||||||
p = dma_alloc_coherent(kdev, size, &priv->rx_desc_dma,
|
p = dma_zalloc_coherent(kdev, size, &priv->rx_desc_dma, GFP_KERNEL);
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
|
||||||
if (!p) {
|
if (!p) {
|
||||||
ret = -ENOMEM;
|
ret = -ENOMEM;
|
||||||
goto out_freeirq_tx;
|
goto out_freeirq_tx;
|
||||||
|
@ -960,8 +959,7 @@ static int bcm_enet_open(struct net_device *dev)
|
||||||
|
|
||||||
/* allocate tx dma ring */
|
/* allocate tx dma ring */
|
||||||
size = priv->tx_ring_size * sizeof(struct bcm_enet_desc);
|
size = priv->tx_ring_size * sizeof(struct bcm_enet_desc);
|
||||||
p = dma_alloc_coherent(kdev, size, &priv->tx_desc_dma,
|
p = dma_zalloc_coherent(kdev, size, &priv->tx_desc_dma, GFP_KERNEL);
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
|
||||||
if (!p) {
|
if (!p) {
|
||||||
ret = -ENOMEM;
|
ret = -ENOMEM;
|
||||||
goto out_free_rx_ring;
|
goto out_free_rx_ring;
|
||||||
|
|
|
@ -853,9 +853,8 @@ bnx2_alloc_mem(struct bnx2 *bp)
|
||||||
bp->status_stats_size = status_blk_size +
|
bp->status_stats_size = status_blk_size +
|
||||||
sizeof(struct statistics_block);
|
sizeof(struct statistics_block);
|
||||||
|
|
||||||
status_blk = dma_alloc_coherent(&bp->pdev->dev, bp->status_stats_size,
|
status_blk = dma_zalloc_coherent(&bp->pdev->dev, bp->status_stats_size,
|
||||||
&bp->status_blk_mapping,
|
&bp->status_blk_mapping, GFP_KERNEL);
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
|
||||||
if (status_blk == NULL)
|
if (status_blk == NULL)
|
||||||
goto alloc_mem_err;
|
goto alloc_mem_err;
|
||||||
|
|
||||||
|
|
|
@ -2069,9 +2069,8 @@ static inline u32 reg_poll(struct bnx2x *bp, u32 reg, u32 expected, int ms,
|
||||||
void bnx2x_igu_clear_sb_gen(struct bnx2x *bp, u8 func, u8 idu_sb_id,
|
void bnx2x_igu_clear_sb_gen(struct bnx2x *bp, u8 func, u8 idu_sb_id,
|
||||||
bool is_pf);
|
bool is_pf);
|
||||||
|
|
||||||
#define BNX2X_ILT_ZALLOC(x, y, size) \
|
#define BNX2X_ILT_ZALLOC(x, y, size) \
|
||||||
x = dma_alloc_coherent(&bp->pdev->dev, size, y, \
|
x = dma_zalloc_coherent(&bp->pdev->dev, size, y, GFP_KERNEL)
|
||||||
GFP_KERNEL | __GFP_ZERO)
|
|
||||||
|
|
||||||
#define BNX2X_ILT_FREE(x, y, size) \
|
#define BNX2X_ILT_FREE(x, y, size) \
|
||||||
do { \
|
do { \
|
||||||
|
|
|
@ -51,8 +51,7 @@ extern int int_mode;
|
||||||
|
|
||||||
#define BNX2X_PCI_ALLOC(x, y, size) \
|
#define BNX2X_PCI_ALLOC(x, y, size) \
|
||||||
do { \
|
do { \
|
||||||
x = dma_alloc_coherent(&bp->pdev->dev, size, y, \
|
x = dma_zalloc_coherent(&bp->pdev->dev, size, y, GFP_KERNEL); \
|
||||||
GFP_KERNEL | __GFP_ZERO); \
|
|
||||||
if (x == NULL) \
|
if (x == NULL) \
|
||||||
goto alloc_mem_err; \
|
goto alloc_mem_err; \
|
||||||
DP(NETIF_MSG_HW, "BNX2X_PCI_ALLOC: Physical %Lx Virtual %p\n", \
|
DP(NETIF_MSG_HW, "BNX2X_PCI_ALLOC: Physical %Lx Virtual %p\n", \
|
||||||
|
|
|
@ -8591,10 +8591,10 @@ static int tg3_mem_rx_acquire(struct tg3 *tp)
|
||||||
if (!i && tg3_flag(tp, ENABLE_RSS))
|
if (!i && tg3_flag(tp, ENABLE_RSS))
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
tnapi->rx_rcb = dma_alloc_coherent(&tp->pdev->dev,
|
tnapi->rx_rcb = dma_zalloc_coherent(&tp->pdev->dev,
|
||||||
TG3_RX_RCB_RING_BYTES(tp),
|
TG3_RX_RCB_RING_BYTES(tp),
|
||||||
&tnapi->rx_rcb_mapping,
|
&tnapi->rx_rcb_mapping,
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
GFP_KERNEL);
|
||||||
if (!tnapi->rx_rcb)
|
if (!tnapi->rx_rcb)
|
||||||
goto err_out;
|
goto err_out;
|
||||||
}
|
}
|
||||||
|
@ -8643,10 +8643,9 @@ static int tg3_alloc_consistent(struct tg3 *tp)
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
tp->hw_stats = dma_alloc_coherent(&tp->pdev->dev,
|
tp->hw_stats = dma_zalloc_coherent(&tp->pdev->dev,
|
||||||
sizeof(struct tg3_hw_stats),
|
sizeof(struct tg3_hw_stats),
|
||||||
&tp->stats_mapping,
|
&tp->stats_mapping, GFP_KERNEL);
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
|
||||||
if (!tp->hw_stats)
|
if (!tp->hw_stats)
|
||||||
goto err_out;
|
goto err_out;
|
||||||
|
|
||||||
|
@ -8654,10 +8653,10 @@ static int tg3_alloc_consistent(struct tg3 *tp)
|
||||||
struct tg3_napi *tnapi = &tp->napi[i];
|
struct tg3_napi *tnapi = &tp->napi[i];
|
||||||
struct tg3_hw_status *sblk;
|
struct tg3_hw_status *sblk;
|
||||||
|
|
||||||
tnapi->hw_status = dma_alloc_coherent(&tp->pdev->dev,
|
tnapi->hw_status = dma_zalloc_coherent(&tp->pdev->dev,
|
||||||
TG3_HW_STATUS_SIZE,
|
TG3_HW_STATUS_SIZE,
|
||||||
&tnapi->status_mapping,
|
&tnapi->status_mapping,
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
GFP_KERNEL);
|
||||||
if (!tnapi->hw_status)
|
if (!tnapi->hw_status)
|
||||||
goto err_out;
|
goto err_out;
|
||||||
|
|
||||||
|
|
|
@ -145,8 +145,8 @@ static int be_queue_alloc(struct be_adapter *adapter, struct be_queue_info *q,
|
||||||
q->len = len;
|
q->len = len;
|
||||||
q->entry_size = entry_size;
|
q->entry_size = entry_size;
|
||||||
mem->size = len * entry_size;
|
mem->size = len * entry_size;
|
||||||
mem->va = dma_alloc_coherent(&adapter->pdev->dev, mem->size, &mem->dma,
|
mem->va = dma_zalloc_coherent(&adapter->pdev->dev, mem->size, &mem->dma,
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
GFP_KERNEL);
|
||||||
if (!mem->va)
|
if (!mem->va)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
return 0;
|
return 0;
|
||||||
|
@ -2642,8 +2642,8 @@ static int be_setup_wol(struct be_adapter *adapter, bool enable)
|
||||||
memset(mac, 0, ETH_ALEN);
|
memset(mac, 0, ETH_ALEN);
|
||||||
|
|
||||||
cmd.size = sizeof(struct be_cmd_req_acpi_wol_magic_config);
|
cmd.size = sizeof(struct be_cmd_req_acpi_wol_magic_config);
|
||||||
cmd.va = dma_alloc_coherent(&adapter->pdev->dev, cmd.size, &cmd.dma,
|
cmd.va = dma_zalloc_coherent(&adapter->pdev->dev, cmd.size, &cmd.dma,
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
GFP_KERNEL);
|
||||||
if (cmd.va == NULL)
|
if (cmd.va == NULL)
|
||||||
return -1;
|
return -1;
|
||||||
|
|
||||||
|
@ -3946,9 +3946,9 @@ static int be_ctrl_init(struct be_adapter *adapter)
|
||||||
memset(mbox_mem_align->va, 0, sizeof(struct be_mcc_mailbox));
|
memset(mbox_mem_align->va, 0, sizeof(struct be_mcc_mailbox));
|
||||||
|
|
||||||
rx_filter->size = sizeof(struct be_cmd_req_rx_filter);
|
rx_filter->size = sizeof(struct be_cmd_req_rx_filter);
|
||||||
rx_filter->va = dma_alloc_coherent(&adapter->pdev->dev, rx_filter->size,
|
rx_filter->va = dma_zalloc_coherent(&adapter->pdev->dev,
|
||||||
&rx_filter->dma,
|
rx_filter->size, &rx_filter->dma,
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
GFP_KERNEL);
|
||||||
if (rx_filter->va == NULL) {
|
if (rx_filter->va == NULL) {
|
||||||
status = -ENOMEM;
|
status = -ENOMEM;
|
||||||
goto free_mbox;
|
goto free_mbox;
|
||||||
|
@ -3994,8 +3994,8 @@ static int be_stats_init(struct be_adapter *adapter)
|
||||||
/* BE3 and Skyhawk */
|
/* BE3 and Skyhawk */
|
||||||
cmd->size = sizeof(struct be_cmd_req_get_stats_v1);
|
cmd->size = sizeof(struct be_cmd_req_get_stats_v1);
|
||||||
|
|
||||||
cmd->va = dma_alloc_coherent(&adapter->pdev->dev, cmd->size, &cmd->dma,
|
cmd->va = dma_zalloc_coherent(&adapter->pdev->dev, cmd->size, &cmd->dma,
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
GFP_KERNEL);
|
||||||
if (cmd->va == NULL)
|
if (cmd->va == NULL)
|
||||||
return -1;
|
return -1;
|
||||||
return 0;
|
return 0;
|
||||||
|
|
|
@ -778,10 +778,9 @@ static int ftgmac100_alloc_buffers(struct ftgmac100 *priv)
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
priv->descs = dma_alloc_coherent(priv->dev,
|
priv->descs = dma_zalloc_coherent(priv->dev,
|
||||||
sizeof(struct ftgmac100_descs),
|
sizeof(struct ftgmac100_descs),
|
||||||
&priv->descs_dma_addr,
|
&priv->descs_dma_addr, GFP_KERNEL);
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
|
||||||
if (!priv->descs)
|
if (!priv->descs)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
|
|
|
@ -732,10 +732,10 @@ static int ftmac100_alloc_buffers(struct ftmac100 *priv)
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
priv->descs = dma_alloc_coherent(priv->dev,
|
priv->descs = dma_zalloc_coherent(priv->dev,
|
||||||
sizeof(struct ftmac100_descs),
|
sizeof(struct ftmac100_descs),
|
||||||
&priv->descs_dma_addr,
|
&priv->descs_dma_addr,
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
GFP_KERNEL);
|
||||||
if (!priv->descs)
|
if (!priv->descs)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
|
|
|
@ -637,8 +637,8 @@ static int mal_probe(struct platform_device *ofdev)
|
||||||
bd_size = sizeof(struct mal_descriptor) *
|
bd_size = sizeof(struct mal_descriptor) *
|
||||||
(NUM_TX_BUFF * mal->num_tx_chans +
|
(NUM_TX_BUFF * mal->num_tx_chans +
|
||||||
NUM_RX_BUFF * mal->num_rx_chans);
|
NUM_RX_BUFF * mal->num_rx_chans);
|
||||||
mal->bd_virt = dma_alloc_coherent(&ofdev->dev, bd_size, &mal->bd_dma,
|
mal->bd_virt = dma_zalloc_coherent(&ofdev->dev, bd_size, &mal->bd_dma,
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
GFP_KERNEL);
|
||||||
if (mal->bd_virt == NULL) {
|
if (mal->bd_virt == NULL) {
|
||||||
err = -ENOMEM;
|
err = -ENOMEM;
|
||||||
goto fail_unmap;
|
goto fail_unmap;
|
||||||
|
|
|
@ -1019,8 +1019,8 @@ static int e1000_setup_desc_rings(struct e1000_adapter *adapter)
|
||||||
|
|
||||||
txdr->size = txdr->count * sizeof(struct e1000_tx_desc);
|
txdr->size = txdr->count * sizeof(struct e1000_tx_desc);
|
||||||
txdr->size = ALIGN(txdr->size, 4096);
|
txdr->size = ALIGN(txdr->size, 4096);
|
||||||
txdr->desc = dma_alloc_coherent(&pdev->dev, txdr->size, &txdr->dma,
|
txdr->desc = dma_zalloc_coherent(&pdev->dev, txdr->size, &txdr->dma,
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
GFP_KERNEL);
|
||||||
if (!txdr->desc) {
|
if (!txdr->desc) {
|
||||||
ret_val = 2;
|
ret_val = 2;
|
||||||
goto err_nomem;
|
goto err_nomem;
|
||||||
|
@ -1077,8 +1077,8 @@ static int e1000_setup_desc_rings(struct e1000_adapter *adapter)
|
||||||
}
|
}
|
||||||
|
|
||||||
rxdr->size = rxdr->count * sizeof(struct e1000_rx_desc);
|
rxdr->size = rxdr->count * sizeof(struct e1000_rx_desc);
|
||||||
rxdr->desc = dma_alloc_coherent(&pdev->dev, rxdr->size, &rxdr->dma,
|
rxdr->desc = dma_zalloc_coherent(&pdev->dev, rxdr->size, &rxdr->dma,
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
GFP_KERNEL);
|
||||||
if (!rxdr->desc) {
|
if (!rxdr->desc) {
|
||||||
ret_val = 6;
|
ret_val = 6;
|
||||||
goto err_nomem;
|
goto err_nomem;
|
||||||
|
|
|
@ -718,8 +718,8 @@ ixgb_setup_tx_resources(struct ixgb_adapter *adapter)
|
||||||
txdr->size = txdr->count * sizeof(struct ixgb_tx_desc);
|
txdr->size = txdr->count * sizeof(struct ixgb_tx_desc);
|
||||||
txdr->size = ALIGN(txdr->size, 4096);
|
txdr->size = ALIGN(txdr->size, 4096);
|
||||||
|
|
||||||
txdr->desc = dma_alloc_coherent(&pdev->dev, txdr->size, &txdr->dma,
|
txdr->desc = dma_zalloc_coherent(&pdev->dev, txdr->size, &txdr->dma,
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
GFP_KERNEL);
|
||||||
if (!txdr->desc) {
|
if (!txdr->desc) {
|
||||||
vfree(txdr->buffer_info);
|
vfree(txdr->buffer_info);
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
|
@ -583,10 +583,9 @@ static int init_hash_table(struct pxa168_eth_private *pep)
|
||||||
* table is full.
|
* table is full.
|
||||||
*/
|
*/
|
||||||
if (pep->htpr == NULL) {
|
if (pep->htpr == NULL) {
|
||||||
pep->htpr = dma_alloc_coherent(pep->dev->dev.parent,
|
pep->htpr = dma_zalloc_coherent(pep->dev->dev.parent,
|
||||||
HASH_ADDR_TABLE_SIZE,
|
HASH_ADDR_TABLE_SIZE,
|
||||||
&pep->htpr_dma,
|
&pep->htpr_dma, GFP_KERNEL);
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
|
||||||
if (pep->htpr == NULL)
|
if (pep->htpr == NULL)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
} else {
|
} else {
|
||||||
|
@ -1024,9 +1023,9 @@ static int rxq_init(struct net_device *dev)
|
||||||
pep->rx_desc_count = 0;
|
pep->rx_desc_count = 0;
|
||||||
size = pep->rx_ring_size * sizeof(struct rx_desc);
|
size = pep->rx_ring_size * sizeof(struct rx_desc);
|
||||||
pep->rx_desc_area_size = size;
|
pep->rx_desc_area_size = size;
|
||||||
pep->p_rx_desc_area = dma_alloc_coherent(pep->dev->dev.parent, size,
|
pep->p_rx_desc_area = dma_zalloc_coherent(pep->dev->dev.parent, size,
|
||||||
&pep->rx_desc_dma,
|
&pep->rx_desc_dma,
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
GFP_KERNEL);
|
||||||
if (!pep->p_rx_desc_area)
|
if (!pep->p_rx_desc_area)
|
||||||
goto out;
|
goto out;
|
||||||
|
|
||||||
|
@ -1085,9 +1084,9 @@ static int txq_init(struct net_device *dev)
|
||||||
pep->tx_desc_count = 0;
|
pep->tx_desc_count = 0;
|
||||||
size = pep->tx_ring_size * sizeof(struct tx_desc);
|
size = pep->tx_ring_size * sizeof(struct tx_desc);
|
||||||
pep->tx_desc_area_size = size;
|
pep->tx_desc_area_size = size;
|
||||||
pep->p_tx_desc_area = dma_alloc_coherent(pep->dev->dev.parent, size,
|
pep->p_tx_desc_area = dma_zalloc_coherent(pep->dev->dev.parent, size,
|
||||||
&pep->tx_desc_dma,
|
&pep->tx_desc_dma,
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
GFP_KERNEL);
|
||||||
if (!pep->p_tx_desc_area)
|
if (!pep->p_tx_desc_area)
|
||||||
goto out;
|
goto out;
|
||||||
/* Initialize the next_desc_ptr links in the Tx descriptors ring */
|
/* Initialize the next_desc_ptr links in the Tx descriptors ring */
|
||||||
|
|
|
@ -3783,9 +3783,9 @@ static int myri10ge_alloc_slices(struct myri10ge_priv *mgp)
|
||||||
for (i = 0; i < mgp->num_slices; i++) {
|
for (i = 0; i < mgp->num_slices; i++) {
|
||||||
ss = &mgp->ss[i];
|
ss = &mgp->ss[i];
|
||||||
bytes = mgp->max_intr_slots * sizeof(*ss->rx_done.entry);
|
bytes = mgp->max_intr_slots * sizeof(*ss->rx_done.entry);
|
||||||
ss->rx_done.entry = dma_alloc_coherent(&pdev->dev, bytes,
|
ss->rx_done.entry = dma_zalloc_coherent(&pdev->dev, bytes,
|
||||||
&ss->rx_done.bus,
|
&ss->rx_done.bus,
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
GFP_KERNEL);
|
||||||
if (ss->rx_done.entry == NULL)
|
if (ss->rx_done.entry == NULL)
|
||||||
goto abort;
|
goto abort;
|
||||||
bytes = sizeof(*ss->fw_stats);
|
bytes = sizeof(*ss->fw_stats);
|
||||||
|
|
|
@ -1491,9 +1491,9 @@ pch_gbe_alloc_rx_buffers_pool(struct pch_gbe_adapter *adapter,
|
||||||
bufsz = adapter->rx_buffer_len;
|
bufsz = adapter->rx_buffer_len;
|
||||||
|
|
||||||
size = rx_ring->count * bufsz + PCH_GBE_RESERVE_MEMORY;
|
size = rx_ring->count * bufsz + PCH_GBE_RESERVE_MEMORY;
|
||||||
rx_ring->rx_buff_pool = dma_alloc_coherent(&pdev->dev, size,
|
rx_ring->rx_buff_pool =
|
||||||
&rx_ring->rx_buff_pool_logic,
|
dma_zalloc_coherent(&pdev->dev, size,
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
&rx_ring->rx_buff_pool_logic, GFP_KERNEL);
|
||||||
if (!rx_ring->rx_buff_pool)
|
if (!rx_ring->rx_buff_pool)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
|
@ -1807,9 +1807,8 @@ int pch_gbe_setup_tx_resources(struct pch_gbe_adapter *adapter,
|
||||||
|
|
||||||
tx_ring->size = tx_ring->count * (int)sizeof(struct pch_gbe_tx_desc);
|
tx_ring->size = tx_ring->count * (int)sizeof(struct pch_gbe_tx_desc);
|
||||||
|
|
||||||
tx_ring->desc = dma_alloc_coherent(&pdev->dev, tx_ring->size,
|
tx_ring->desc = dma_zalloc_coherent(&pdev->dev, tx_ring->size,
|
||||||
&tx_ring->dma,
|
&tx_ring->dma, GFP_KERNEL);
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
|
||||||
if (!tx_ring->desc) {
|
if (!tx_ring->desc) {
|
||||||
vfree(tx_ring->buffer_info);
|
vfree(tx_ring->buffer_info);
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
@ -1852,9 +1851,8 @@ int pch_gbe_setup_rx_resources(struct pch_gbe_adapter *adapter,
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
rx_ring->size = rx_ring->count * (int)sizeof(struct pch_gbe_rx_desc);
|
rx_ring->size = rx_ring->count * (int)sizeof(struct pch_gbe_rx_desc);
|
||||||
rx_ring->desc = dma_alloc_coherent(&pdev->dev, rx_ring->size,
|
rx_ring->desc = dma_zalloc_coherent(&pdev->dev, rx_ring->size,
|
||||||
&rx_ring->dma,
|
&rx_ring->dma, GFP_KERNEL);
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
|
||||||
if (!rx_ring->desc) {
|
if (!rx_ring->desc) {
|
||||||
vfree(rx_ring->buffer_info);
|
vfree(rx_ring->buffer_info);
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
|
@ -440,10 +440,9 @@ static int pasemi_mac_setup_rx_resources(const struct net_device *dev)
|
||||||
if (pasemi_dma_alloc_ring(&ring->chan, RX_RING_SIZE))
|
if (pasemi_dma_alloc_ring(&ring->chan, RX_RING_SIZE))
|
||||||
goto out_ring_desc;
|
goto out_ring_desc;
|
||||||
|
|
||||||
ring->buffers = dma_alloc_coherent(&mac->dma_pdev->dev,
|
ring->buffers = dma_zalloc_coherent(&mac->dma_pdev->dev,
|
||||||
RX_RING_SIZE * sizeof(u64),
|
RX_RING_SIZE * sizeof(u64),
|
||||||
&ring->buf_dma,
|
&ring->buf_dma, GFP_KERNEL);
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
|
||||||
if (!ring->buffers)
|
if (!ring->buffers)
|
||||||
goto out_ring_desc;
|
goto out_ring_desc;
|
||||||
|
|
||||||
|
|
|
@ -448,14 +448,14 @@ int qlcnic_82xx_fw_cmd_create_tx_ctx(struct qlcnic_adapter *adapter,
|
||||||
*(tx_ring->hw_consumer) = 0;
|
*(tx_ring->hw_consumer) = 0;
|
||||||
|
|
||||||
rq_size = SIZEOF_HOSTRQ_TX(struct qlcnic_hostrq_tx_ctx);
|
rq_size = SIZEOF_HOSTRQ_TX(struct qlcnic_hostrq_tx_ctx);
|
||||||
rq_addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size,
|
rq_addr = dma_zalloc_coherent(&adapter->pdev->dev, rq_size,
|
||||||
&rq_phys_addr, GFP_KERNEL | __GFP_ZERO);
|
&rq_phys_addr, GFP_KERNEL);
|
||||||
if (!rq_addr)
|
if (!rq_addr)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
rsp_size = SIZEOF_CARDRSP_TX(struct qlcnic_cardrsp_tx_ctx);
|
rsp_size = SIZEOF_CARDRSP_TX(struct qlcnic_cardrsp_tx_ctx);
|
||||||
rsp_addr = dma_alloc_coherent(&adapter->pdev->dev, rsp_size,
|
rsp_addr = dma_zalloc_coherent(&adapter->pdev->dev, rsp_size,
|
||||||
&rsp_phys_addr, GFP_KERNEL | __GFP_ZERO);
|
&rsp_phys_addr, GFP_KERNEL);
|
||||||
if (!rsp_addr) {
|
if (!rsp_addr) {
|
||||||
err = -ENOMEM;
|
err = -ENOMEM;
|
||||||
goto out_free_rq;
|
goto out_free_rq;
|
||||||
|
@ -865,8 +865,8 @@ int qlcnic_82xx_get_nic_info(struct qlcnic_adapter *adapter,
|
||||||
struct qlcnic_cmd_args cmd;
|
struct qlcnic_cmd_args cmd;
|
||||||
size_t nic_size = sizeof(struct qlcnic_info_le);
|
size_t nic_size = sizeof(struct qlcnic_info_le);
|
||||||
|
|
||||||
nic_info_addr = dma_alloc_coherent(&adapter->pdev->dev, nic_size,
|
nic_info_addr = dma_zalloc_coherent(&adapter->pdev->dev, nic_size,
|
||||||
&nic_dma_t, GFP_KERNEL | __GFP_ZERO);
|
&nic_dma_t, GFP_KERNEL);
|
||||||
if (!nic_info_addr)
|
if (!nic_info_addr)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
|
@ -919,8 +919,8 @@ int qlcnic_82xx_set_nic_info(struct qlcnic_adapter *adapter,
|
||||||
if (adapter->ahw->op_mode != QLCNIC_MGMT_FUNC)
|
if (adapter->ahw->op_mode != QLCNIC_MGMT_FUNC)
|
||||||
return err;
|
return err;
|
||||||
|
|
||||||
nic_info_addr = dma_alloc_coherent(&adapter->pdev->dev, nic_size,
|
nic_info_addr = dma_zalloc_coherent(&adapter->pdev->dev, nic_size,
|
||||||
&nic_dma_t, GFP_KERNEL | __GFP_ZERO);
|
&nic_dma_t, GFP_KERNEL);
|
||||||
if (!nic_info_addr)
|
if (!nic_info_addr)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
|
@ -972,9 +972,8 @@ int qlcnic_82xx_get_pci_info(struct qlcnic_adapter *adapter,
|
||||||
size_t npar_size = sizeof(struct qlcnic_pci_info_le);
|
size_t npar_size = sizeof(struct qlcnic_pci_info_le);
|
||||||
size_t pci_size = npar_size * QLCNIC_MAX_PCI_FUNC;
|
size_t pci_size = npar_size * QLCNIC_MAX_PCI_FUNC;
|
||||||
|
|
||||||
pci_info_addr = dma_alloc_coherent(&adapter->pdev->dev, pci_size,
|
pci_info_addr = dma_zalloc_coherent(&adapter->pdev->dev, pci_size,
|
||||||
&pci_info_dma_t,
|
&pci_info_dma_t, GFP_KERNEL);
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
|
||||||
if (!pci_info_addr)
|
if (!pci_info_addr)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
|
@ -1074,8 +1073,8 @@ int qlcnic_get_port_stats(struct qlcnic_adapter *adapter, const u8 func,
|
||||||
return -EIO;
|
return -EIO;
|
||||||
}
|
}
|
||||||
|
|
||||||
stats_addr = dma_alloc_coherent(&adapter->pdev->dev, stats_size,
|
stats_addr = dma_zalloc_coherent(&adapter->pdev->dev, stats_size,
|
||||||
&stats_dma_t, GFP_KERNEL | __GFP_ZERO);
|
&stats_dma_t, GFP_KERNEL);
|
||||||
if (!stats_addr)
|
if (!stats_addr)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
|
@ -1130,8 +1129,8 @@ int qlcnic_get_mac_stats(struct qlcnic_adapter *adapter,
|
||||||
if (mac_stats == NULL)
|
if (mac_stats == NULL)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
stats_addr = dma_alloc_coherent(&adapter->pdev->dev, stats_size,
|
stats_addr = dma_zalloc_coherent(&adapter->pdev->dev, stats_size,
|
||||||
&stats_dma_t, GFP_KERNEL | __GFP_ZERO);
|
&stats_dma_t, GFP_KERNEL);
|
||||||
if (!stats_addr)
|
if (!stats_addr)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
|
|
|
@ -33,9 +33,8 @@
|
||||||
int efx_nic_alloc_buffer(struct efx_nic *efx, struct efx_buffer *buffer,
|
int efx_nic_alloc_buffer(struct efx_nic *efx, struct efx_buffer *buffer,
|
||||||
unsigned int len, gfp_t gfp_flags)
|
unsigned int len, gfp_t gfp_flags)
|
||||||
{
|
{
|
||||||
buffer->addr = dma_alloc_coherent(&efx->pci_dev->dev, len,
|
buffer->addr = dma_zalloc_coherent(&efx->pci_dev->dev, len,
|
||||||
&buffer->dma_addr,
|
&buffer->dma_addr, gfp_flags);
|
||||||
gfp_flags | __GFP_ZERO);
|
|
||||||
if (!buffer->addr)
|
if (!buffer->addr)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
buffer->len = len;
|
buffer->len = len;
|
||||||
|
|
|
@ -212,9 +212,8 @@ static void meth_check_link(struct net_device *dev)
|
||||||
static int meth_init_tx_ring(struct meth_private *priv)
|
static int meth_init_tx_ring(struct meth_private *priv)
|
||||||
{
|
{
|
||||||
/* Init TX ring */
|
/* Init TX ring */
|
||||||
priv->tx_ring = dma_alloc_coherent(NULL, TX_RING_BUFFER_SIZE,
|
priv->tx_ring = dma_zalloc_coherent(NULL, TX_RING_BUFFER_SIZE,
|
||||||
&priv->tx_ring_dma,
|
&priv->tx_ring_dma, GFP_ATOMIC);
|
||||||
GFP_ATOMIC | __GFP_ZERO);
|
|
||||||
if (!priv->tx_ring)
|
if (!priv->tx_ring)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
|
|
|
@ -1308,13 +1308,13 @@ static int tsi108_open(struct net_device *dev)
|
||||||
data->id, dev->irq, dev->name);
|
data->id, dev->irq, dev->name);
|
||||||
}
|
}
|
||||||
|
|
||||||
data->rxring = dma_alloc_coherent(NULL, rxring_size, &data->rxdma,
|
data->rxring = dma_zalloc_coherent(NULL, rxring_size, &data->rxdma,
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
GFP_KERNEL);
|
||||||
if (!data->rxring)
|
if (!data->rxring)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
data->txring = dma_alloc_coherent(NULL, txring_size, &data->txdma,
|
data->txring = dma_zalloc_coherent(NULL, txring_size, &data->txdma,
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
GFP_KERNEL);
|
||||||
if (!data->txring) {
|
if (!data->txring) {
|
||||||
pci_free_consistent(0, rxring_size, data->rxring, data->rxdma);
|
pci_free_consistent(0, rxring_size, data->rxring, data->rxdma);
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
|
@ -243,15 +243,15 @@ static int temac_dma_bd_init(struct net_device *ndev)
|
||||||
|
|
||||||
/* allocate the tx and rx ring buffer descriptors. */
|
/* allocate the tx and rx ring buffer descriptors. */
|
||||||
/* returns a virtual address and a physical address. */
|
/* returns a virtual address and a physical address. */
|
||||||
lp->tx_bd_v = dma_alloc_coherent(ndev->dev.parent,
|
lp->tx_bd_v = dma_zalloc_coherent(ndev->dev.parent,
|
||||||
sizeof(*lp->tx_bd_v) * TX_BD_NUM,
|
sizeof(*lp->tx_bd_v) * TX_BD_NUM,
|
||||||
&lp->tx_bd_p, GFP_KERNEL | __GFP_ZERO);
|
&lp->tx_bd_p, GFP_KERNEL);
|
||||||
if (!lp->tx_bd_v)
|
if (!lp->tx_bd_v)
|
||||||
goto out;
|
goto out;
|
||||||
|
|
||||||
lp->rx_bd_v = dma_alloc_coherent(ndev->dev.parent,
|
lp->rx_bd_v = dma_zalloc_coherent(ndev->dev.parent,
|
||||||
sizeof(*lp->rx_bd_v) * RX_BD_NUM,
|
sizeof(*lp->rx_bd_v) * RX_BD_NUM,
|
||||||
&lp->rx_bd_p, GFP_KERNEL | __GFP_ZERO);
|
&lp->rx_bd_p, GFP_KERNEL);
|
||||||
if (!lp->rx_bd_v)
|
if (!lp->rx_bd_v)
|
||||||
goto out;
|
goto out;
|
||||||
|
|
||||||
|
|
|
@ -201,17 +201,15 @@ static int axienet_dma_bd_init(struct net_device *ndev)
|
||||||
/*
|
/*
|
||||||
* Allocate the Tx and Rx buffer descriptors.
|
* Allocate the Tx and Rx buffer descriptors.
|
||||||
*/
|
*/
|
||||||
lp->tx_bd_v = dma_alloc_coherent(ndev->dev.parent,
|
lp->tx_bd_v = dma_zalloc_coherent(ndev->dev.parent,
|
||||||
sizeof(*lp->tx_bd_v) * TX_BD_NUM,
|
sizeof(*lp->tx_bd_v) * TX_BD_NUM,
|
||||||
&lp->tx_bd_p,
|
&lp->tx_bd_p, GFP_KERNEL);
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
|
||||||
if (!lp->tx_bd_v)
|
if (!lp->tx_bd_v)
|
||||||
goto out;
|
goto out;
|
||||||
|
|
||||||
lp->rx_bd_v = dma_alloc_coherent(ndev->dev.parent,
|
lp->rx_bd_v = dma_zalloc_coherent(ndev->dev.parent,
|
||||||
sizeof(*lp->rx_bd_v) * RX_BD_NUM,
|
sizeof(*lp->rx_bd_v) * RX_BD_NUM,
|
||||||
&lp->rx_bd_p,
|
&lp->rx_bd_p, GFP_KERNEL);
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
|
||||||
if (!lp->rx_bd_v)
|
if (!lp->rx_bd_v)
|
||||||
goto out;
|
goto out;
|
||||||
|
|
||||||
|
|
|
@ -1068,9 +1068,9 @@ static int dfx_driver_init(struct net_device *dev, const char *print_name,
|
||||||
#endif
|
#endif
|
||||||
sizeof(PI_CONSUMER_BLOCK) +
|
sizeof(PI_CONSUMER_BLOCK) +
|
||||||
(PI_ALIGN_K_DESC_BLK - 1);
|
(PI_ALIGN_K_DESC_BLK - 1);
|
||||||
bp->kmalloced = top_v = dma_alloc_coherent(bp->bus_dev, alloc_size,
|
bp->kmalloced = top_v = dma_zalloc_coherent(bp->bus_dev, alloc_size,
|
||||||
&bp->kmalloced_dma,
|
&bp->kmalloced_dma,
|
||||||
GFP_ATOMIC | __GFP_ZERO);
|
GFP_ATOMIC);
|
||||||
if (top_v == NULL)
|
if (top_v == NULL)
|
||||||
return DFX_K_FAILURE;
|
return DFX_K_FAILURE;
|
||||||
|
|
||||||
|
|
|
@ -351,16 +351,16 @@ static int ali_ircc_open(int i, chipio_t *info)
|
||||||
|
|
||||||
/* Allocate memory if needed */
|
/* Allocate memory if needed */
|
||||||
self->rx_buff.head =
|
self->rx_buff.head =
|
||||||
dma_alloc_coherent(NULL, self->rx_buff.truesize,
|
dma_zalloc_coherent(NULL, self->rx_buff.truesize,
|
||||||
&self->rx_buff_dma, GFP_KERNEL | __GFP_ZERO);
|
&self->rx_buff_dma, GFP_KERNEL);
|
||||||
if (self->rx_buff.head == NULL) {
|
if (self->rx_buff.head == NULL) {
|
||||||
err = -ENOMEM;
|
err = -ENOMEM;
|
||||||
goto err_out2;
|
goto err_out2;
|
||||||
}
|
}
|
||||||
|
|
||||||
self->tx_buff.head =
|
self->tx_buff.head =
|
||||||
dma_alloc_coherent(NULL, self->tx_buff.truesize,
|
dma_zalloc_coherent(NULL, self->tx_buff.truesize,
|
||||||
&self->tx_buff_dma, GFP_KERNEL | __GFP_ZERO);
|
&self->tx_buff_dma, GFP_KERNEL);
|
||||||
if (self->tx_buff.head == NULL) {
|
if (self->tx_buff.head == NULL) {
|
||||||
err = -ENOMEM;
|
err = -ENOMEM;
|
||||||
goto err_out3;
|
goto err_out3;
|
||||||
|
|
|
@ -430,8 +430,8 @@ static int __init nsc_ircc_open(chipio_t *info)
|
||||||
|
|
||||||
/* Allocate memory if needed */
|
/* Allocate memory if needed */
|
||||||
self->rx_buff.head =
|
self->rx_buff.head =
|
||||||
dma_alloc_coherent(NULL, self->rx_buff.truesize,
|
dma_zalloc_coherent(NULL, self->rx_buff.truesize,
|
||||||
&self->rx_buff_dma, GFP_KERNEL | __GFP_ZERO);
|
&self->rx_buff_dma, GFP_KERNEL);
|
||||||
if (self->rx_buff.head == NULL) {
|
if (self->rx_buff.head == NULL) {
|
||||||
err = -ENOMEM;
|
err = -ENOMEM;
|
||||||
goto out2;
|
goto out2;
|
||||||
|
@ -439,8 +439,8 @@ static int __init nsc_ircc_open(chipio_t *info)
|
||||||
}
|
}
|
||||||
|
|
||||||
self->tx_buff.head =
|
self->tx_buff.head =
|
||||||
dma_alloc_coherent(NULL, self->tx_buff.truesize,
|
dma_zalloc_coherent(NULL, self->tx_buff.truesize,
|
||||||
&self->tx_buff_dma, GFP_KERNEL | __GFP_ZERO);
|
&self->tx_buff_dma, GFP_KERNEL);
|
||||||
if (self->tx_buff.head == NULL) {
|
if (self->tx_buff.head == NULL) {
|
||||||
err = -ENOMEM;
|
err = -ENOMEM;
|
||||||
goto out3;
|
goto out3;
|
||||||
|
|
|
@ -562,14 +562,14 @@ static int smsc_ircc_open(unsigned int fir_base, unsigned int sir_base, u8 dma,
|
||||||
self->tx_buff.truesize = SMSC_IRCC2_TX_BUFF_TRUESIZE;
|
self->tx_buff.truesize = SMSC_IRCC2_TX_BUFF_TRUESIZE;
|
||||||
|
|
||||||
self->rx_buff.head =
|
self->rx_buff.head =
|
||||||
dma_alloc_coherent(NULL, self->rx_buff.truesize,
|
dma_zalloc_coherent(NULL, self->rx_buff.truesize,
|
||||||
&self->rx_buff_dma, GFP_KERNEL | __GFP_ZERO);
|
&self->rx_buff_dma, GFP_KERNEL);
|
||||||
if (self->rx_buff.head == NULL)
|
if (self->rx_buff.head == NULL)
|
||||||
goto err_out2;
|
goto err_out2;
|
||||||
|
|
||||||
self->tx_buff.head =
|
self->tx_buff.head =
|
||||||
dma_alloc_coherent(NULL, self->tx_buff.truesize,
|
dma_zalloc_coherent(NULL, self->tx_buff.truesize,
|
||||||
&self->tx_buff_dma, GFP_KERNEL | __GFP_ZERO);
|
&self->tx_buff_dma, GFP_KERNEL);
|
||||||
if (self->tx_buff.head == NULL)
|
if (self->tx_buff.head == NULL)
|
||||||
goto err_out3;
|
goto err_out3;
|
||||||
|
|
||||||
|
|
|
@ -361,16 +361,16 @@ static int via_ircc_open(struct pci_dev *pdev, chipio_t *info, unsigned int id)
|
||||||
|
|
||||||
/* Allocate memory if needed */
|
/* Allocate memory if needed */
|
||||||
self->rx_buff.head =
|
self->rx_buff.head =
|
||||||
dma_alloc_coherent(&pdev->dev, self->rx_buff.truesize,
|
dma_zalloc_coherent(&pdev->dev, self->rx_buff.truesize,
|
||||||
&self->rx_buff_dma, GFP_KERNEL | __GFP_ZERO);
|
&self->rx_buff_dma, GFP_KERNEL);
|
||||||
if (self->rx_buff.head == NULL) {
|
if (self->rx_buff.head == NULL) {
|
||||||
err = -ENOMEM;
|
err = -ENOMEM;
|
||||||
goto err_out2;
|
goto err_out2;
|
||||||
}
|
}
|
||||||
|
|
||||||
self->tx_buff.head =
|
self->tx_buff.head =
|
||||||
dma_alloc_coherent(&pdev->dev, self->tx_buff.truesize,
|
dma_zalloc_coherent(&pdev->dev, self->tx_buff.truesize,
|
||||||
&self->tx_buff_dma, GFP_KERNEL | __GFP_ZERO);
|
&self->tx_buff_dma, GFP_KERNEL);
|
||||||
if (self->tx_buff.head == NULL) {
|
if (self->tx_buff.head == NULL) {
|
||||||
err = -ENOMEM;
|
err = -ENOMEM;
|
||||||
goto err_out3;
|
goto err_out3;
|
||||||
|
|
|
@ -215,16 +215,16 @@ static int w83977af_open(int i, unsigned int iobase, unsigned int irq,
|
||||||
|
|
||||||
/* Allocate memory if needed */
|
/* Allocate memory if needed */
|
||||||
self->rx_buff.head =
|
self->rx_buff.head =
|
||||||
dma_alloc_coherent(NULL, self->rx_buff.truesize,
|
dma_zalloc_coherent(NULL, self->rx_buff.truesize,
|
||||||
&self->rx_buff_dma, GFP_KERNEL | __GFP_ZERO);
|
&self->rx_buff_dma, GFP_KERNEL);
|
||||||
if (self->rx_buff.head == NULL) {
|
if (self->rx_buff.head == NULL) {
|
||||||
err = -ENOMEM;
|
err = -ENOMEM;
|
||||||
goto err_out1;
|
goto err_out1;
|
||||||
}
|
}
|
||||||
|
|
||||||
self->tx_buff.head =
|
self->tx_buff.head =
|
||||||
dma_alloc_coherent(NULL, self->tx_buff.truesize,
|
dma_zalloc_coherent(NULL, self->tx_buff.truesize,
|
||||||
&self->tx_buff_dma, GFP_KERNEL | __GFP_ZERO);
|
&self->tx_buff_dma, GFP_KERNEL);
|
||||||
if (self->tx_buff.head == NULL) {
|
if (self->tx_buff.head == NULL) {
|
||||||
err = -ENOMEM;
|
err = -ENOMEM;
|
||||||
goto err_out2;
|
goto err_out2;
|
||||||
|
|
|
@ -431,9 +431,9 @@ static int alloc_ringmemory(struct b43_dmaring *ring)
|
||||||
u16 ring_mem_size = (ring->type == B43_DMA_64BIT) ?
|
u16 ring_mem_size = (ring->type == B43_DMA_64BIT) ?
|
||||||
B43_DMA64_RINGMEMSIZE : B43_DMA32_RINGMEMSIZE;
|
B43_DMA64_RINGMEMSIZE : B43_DMA32_RINGMEMSIZE;
|
||||||
|
|
||||||
ring->descbase = dma_alloc_coherent(ring->dev->dev->dma_dev,
|
ring->descbase = dma_zalloc_coherent(ring->dev->dev->dma_dev,
|
||||||
ring_mem_size, &(ring->dmabase),
|
ring_mem_size, &(ring->dmabase),
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
GFP_KERNEL);
|
||||||
if (!ring->descbase)
|
if (!ring->descbase)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
|
|
|
@ -331,10 +331,9 @@ void free_descriptor_buffer(struct b43legacy_dmaring *ring,
|
||||||
static int alloc_ringmemory(struct b43legacy_dmaring *ring)
|
static int alloc_ringmemory(struct b43legacy_dmaring *ring)
|
||||||
{
|
{
|
||||||
/* GFP flags must match the flags in free_ringmemory()! */
|
/* GFP flags must match the flags in free_ringmemory()! */
|
||||||
ring->descbase = dma_alloc_coherent(ring->dev->dev->dma_dev,
|
ring->descbase = dma_zalloc_coherent(ring->dev->dev->dma_dev,
|
||||||
B43legacy_DMA_RINGMEMSIZE,
|
B43legacy_DMA_RINGMEMSIZE,
|
||||||
&(ring->dmabase),
|
&(ring->dmabase), GFP_KERNEL);
|
||||||
GFP_KERNEL | __GFP_ZERO);
|
|
||||||
if (!ring->descbase)
|
if (!ring->descbase)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
|
|
|
@ -132,9 +132,8 @@ static inline int dma_set_seg_boundary(struct device *dev, unsigned long mask)
|
||||||
static inline void *dma_zalloc_coherent(struct device *dev, size_t size,
|
static inline void *dma_zalloc_coherent(struct device *dev, size_t size,
|
||||||
dma_addr_t *dma_handle, gfp_t flag)
|
dma_addr_t *dma_handle, gfp_t flag)
|
||||||
{
|
{
|
||||||
void *ret = dma_alloc_coherent(dev, size, dma_handle, flag);
|
void *ret = dma_alloc_coherent(dev, size, dma_handle,
|
||||||
if (ret)
|
flag | __GFP_ZERO);
|
||||||
memset(ret, 0, size);
|
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue