mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-28 11:18:45 +07:00
ipw2x00: switch from 'pci_' to 'dma_' API
The wrappers in include/linux/pci-dma-compat.h should go away. The patch has been generated with the coccinelle script below and has been hand modified to replace GFP_ with a correct flag. It has been compile tested. When memory is allocated in 'ipw2100_msg_allocate()' (ipw2100.c), GFP_KERNEL can be used because it is called from the probe function. The call chain is: ipw2100_pci_init_one (the probe function) --> ipw2100_queues_allocate --> ipw2100_msg_allocate Moreover, 'ipw2100_msg_allocate()' already uses GFP_KERNEL for some other memory allocations. When memory is allocated in 'status_queue_allocate()' (ipw2100.c), GFP_KERNEL can be used because it is called from the probe function. The call chain is: ipw2100_pci_init_one (the probe function) --> ipw2100_queues_allocate --> ipw2100_rx_allocate --> status_queue_allocate Moreover, 'ipw2100_rx_allocate()' already uses GFP_KERNEL for some other memory allocations. When memory is allocated in 'bd_queue_allocate()' (ipw2100.c), GFP_KERNEL can be used because it is called from the probe function. The call chain is: ipw2100_pci_init_one (the probe function) --> ipw2100_queues_allocate --> ipw2100_rx_allocate --> bd_queue_allocate Moreover, 'ipw2100_rx_allocate()' already uses GFP_KERNEL for some other memory allocations. When memory is allocated in 'ipw2100_tx_allocate()' (ipw2100.c), GFP_KERNEL can be used because it is called from the probe function. The call chain is: ipw2100_pci_init_one (the probe function) --> ipw2100_queues_allocate --> ipw2100_tx_allocate Moreover, 'ipw2100_tx_allocate()' already uses GFP_KERNEL for some other memory allocations. When memory is allocated in 'ipw_queue_tx_init()' (ipw2200.c), GFP_KERNEL can be used because it is called from a call chain that already uses GFP_KERNEL and no spin_lock is taken in the between. The call chain is: ipw_up --> ipw_load --> ipw_queue_reset --> ipw_queue_tx_init 'ipw_up()' already uses GFP_KERNEL for some other memory allocations. @@ @@ - PCI_DMA_BIDIRECTIONAL + DMA_BIDIRECTIONAL @@ @@ - PCI_DMA_TODEVICE + DMA_TO_DEVICE @@ @@ - PCI_DMA_FROMDEVICE + DMA_FROM_DEVICE @@ @@ - PCI_DMA_NONE + DMA_NONE @@ expression e1, e2, e3; @@ - pci_alloc_consistent(e1, e2, e3) + dma_alloc_coherent(&e1->dev, e2, e3, GFP_) @@ expression e1, e2, e3; @@ - pci_zalloc_consistent(e1, e2, e3) + dma_alloc_coherent(&e1->dev, e2, e3, GFP_) @@ expression e1, e2, e3, e4; @@ - pci_free_consistent(e1, e2, e3, e4) + dma_free_coherent(&e1->dev, e2, e3, e4) @@ expression e1, e2, e3, e4; @@ - pci_map_single(e1, e2, e3, e4) + dma_map_single(&e1->dev, e2, e3, e4) @@ expression e1, e2, e3, e4; @@ - pci_unmap_single(e1, e2, e3, e4) + dma_unmap_single(&e1->dev, e2, e3, e4) @@ expression e1, e2, e3, e4, e5; @@ - pci_map_page(e1, e2, e3, e4, e5) + dma_map_page(&e1->dev, e2, e3, e4, e5) @@ expression e1, e2, e3, e4; @@ - pci_unmap_page(e1, e2, e3, e4) + dma_unmap_page(&e1->dev, e2, e3, e4) @@ expression e1, e2, e3, e4; @@ - pci_map_sg(e1, e2, e3, e4) + dma_map_sg(&e1->dev, e2, e3, e4) @@ expression e1, e2, e3, e4; @@ - pci_unmap_sg(e1, e2, e3, e4) + dma_unmap_sg(&e1->dev, e2, e3, e4) @@ expression e1, e2, e3, e4; @@ - pci_dma_sync_single_for_cpu(e1, e2, e3, e4) + dma_sync_single_for_cpu(&e1->dev, e2, e3, e4) @@ expression e1, e2, e3, e4; @@ - pci_dma_sync_single_for_device(e1, e2, e3, e4) + dma_sync_single_for_device(&e1->dev, e2, e3, e4) @@ expression e1, e2, e3, e4; @@ - pci_dma_sync_sg_for_cpu(e1, e2, e3, e4) + dma_sync_sg_for_cpu(&e1->dev, e2, e3, e4) @@ expression e1, e2, e3, e4; @@ - pci_dma_sync_sg_for_device(e1, e2, e3, e4) + dma_sync_sg_for_device(&e1->dev, e2, e3, e4) @@ expression e1, e2; @@ - pci_dma_mapping_error(e1, e2) + dma_mapping_error(&e1->dev, e2) @@ expression e1, e2; @@ - pci_set_dma_mask(e1, e2) + dma_set_mask(&e1->dev, e2) @@ expression e1, e2; @@ - pci_set_consistent_dma_mask(e1, e2) + dma_set_coherent_mask(&e1->dev, e2) Signed-off-by: Christophe JAILLET <christophe.jaillet@wanadoo.fr> Signed-off-by: Kalle Valo <kvalo@codeaurora.org> Link: https://lore.kernel.org/r/20200722101716.26185-1-christophe.jaillet@wanadoo.fr
This commit is contained in:
parent
9130559cf8
commit
e52525c0c3
@ -2295,10 +2295,11 @@ static int ipw2100_alloc_skb(struct ipw2100_priv *priv,
|
||||
return -ENOMEM;
|
||||
|
||||
packet->rxp = (struct ipw2100_rx *)packet->skb->data;
|
||||
packet->dma_addr = pci_map_single(priv->pci_dev, packet->skb->data,
|
||||
packet->dma_addr = dma_map_single(&priv->pci_dev->dev,
|
||||
packet->skb->data,
|
||||
sizeof(struct ipw2100_rx),
|
||||
PCI_DMA_FROMDEVICE);
|
||||
if (pci_dma_mapping_error(priv->pci_dev, packet->dma_addr)) {
|
||||
DMA_FROM_DEVICE);
|
||||
if (dma_mapping_error(&priv->pci_dev->dev, packet->dma_addr)) {
|
||||
dev_kfree_skb(packet->skb);
|
||||
return -ENOMEM;
|
||||
}
|
||||
@ -2479,9 +2480,8 @@ static void isr_rx(struct ipw2100_priv *priv, int i,
|
||||
return;
|
||||
}
|
||||
|
||||
pci_unmap_single(priv->pci_dev,
|
||||
packet->dma_addr,
|
||||
sizeof(struct ipw2100_rx), PCI_DMA_FROMDEVICE);
|
||||
dma_unmap_single(&priv->pci_dev->dev, packet->dma_addr,
|
||||
sizeof(struct ipw2100_rx), DMA_FROM_DEVICE);
|
||||
|
||||
skb_put(packet->skb, status->frame_size);
|
||||
|
||||
@ -2563,8 +2563,8 @@ static void isr_rx_monitor(struct ipw2100_priv *priv, int i,
|
||||
return;
|
||||
}
|
||||
|
||||
pci_unmap_single(priv->pci_dev, packet->dma_addr,
|
||||
sizeof(struct ipw2100_rx), PCI_DMA_FROMDEVICE);
|
||||
dma_unmap_single(&priv->pci_dev->dev, packet->dma_addr,
|
||||
sizeof(struct ipw2100_rx), DMA_FROM_DEVICE);
|
||||
memmove(packet->skb->data + sizeof(struct ipw_rt_hdr),
|
||||
packet->skb->data, status->frame_size);
|
||||
|
||||
@ -2689,9 +2689,9 @@ static void __ipw2100_rx_process(struct ipw2100_priv *priv)
|
||||
|
||||
/* Sync the DMA for the RX buffer so CPU is sure to get
|
||||
* the correct values */
|
||||
pci_dma_sync_single_for_cpu(priv->pci_dev, packet->dma_addr,
|
||||
sizeof(struct ipw2100_rx),
|
||||
PCI_DMA_FROMDEVICE);
|
||||
dma_sync_single_for_cpu(&priv->pci_dev->dev, packet->dma_addr,
|
||||
sizeof(struct ipw2100_rx),
|
||||
DMA_FROM_DEVICE);
|
||||
|
||||
if (unlikely(ipw2100_corruption_check(priv, i))) {
|
||||
ipw2100_corruption_detected(priv, i);
|
||||
@ -2923,9 +2923,8 @@ static int __ipw2100_tx_process(struct ipw2100_priv *priv)
|
||||
(packet->index + 1 + i) % txq->entries,
|
||||
tbd->host_addr, tbd->buf_length);
|
||||
|
||||
pci_unmap_single(priv->pci_dev,
|
||||
tbd->host_addr,
|
||||
tbd->buf_length, PCI_DMA_TODEVICE);
|
||||
dma_unmap_single(&priv->pci_dev->dev, tbd->host_addr,
|
||||
tbd->buf_length, DMA_TO_DEVICE);
|
||||
}
|
||||
|
||||
libipw_txb_free(packet->info.d_struct.txb);
|
||||
@ -3165,15 +3164,13 @@ static void ipw2100_tx_send_data(struct ipw2100_priv *priv)
|
||||
tbd->buf_length = packet->info.d_struct.txb->
|
||||
fragments[i]->len - LIBIPW_3ADDR_LEN;
|
||||
|
||||
tbd->host_addr = pci_map_single(priv->pci_dev,
|
||||
tbd->host_addr = dma_map_single(&priv->pci_dev->dev,
|
||||
packet->info.d_struct.
|
||||
txb->fragments[i]->
|
||||
data +
|
||||
txb->fragments[i]->data +
|
||||
LIBIPW_3ADDR_LEN,
|
||||
tbd->buf_length,
|
||||
PCI_DMA_TODEVICE);
|
||||
if (pci_dma_mapping_error(priv->pci_dev,
|
||||
tbd->host_addr)) {
|
||||
DMA_TO_DEVICE);
|
||||
if (dma_mapping_error(&priv->pci_dev->dev, tbd->host_addr)) {
|
||||
IPW_DEBUG_TX("dma mapping error\n");
|
||||
break;
|
||||
}
|
||||
@ -3182,10 +3179,10 @@ static void ipw2100_tx_send_data(struct ipw2100_priv *priv)
|
||||
txq->next, tbd->host_addr,
|
||||
tbd->buf_length);
|
||||
|
||||
pci_dma_sync_single_for_device(priv->pci_dev,
|
||||
tbd->host_addr,
|
||||
tbd->buf_length,
|
||||
PCI_DMA_TODEVICE);
|
||||
dma_sync_single_for_device(&priv->pci_dev->dev,
|
||||
tbd->host_addr,
|
||||
tbd->buf_length,
|
||||
DMA_TO_DEVICE);
|
||||
|
||||
txq->next++;
|
||||
txq->next %= txq->entries;
|
||||
@ -3440,9 +3437,9 @@ static int ipw2100_msg_allocate(struct ipw2100_priv *priv)
|
||||
return -ENOMEM;
|
||||
|
||||
for (i = 0; i < IPW_COMMAND_POOL_SIZE; i++) {
|
||||
v = pci_zalloc_consistent(priv->pci_dev,
|
||||
sizeof(struct ipw2100_cmd_header),
|
||||
&p);
|
||||
v = dma_alloc_coherent(&priv->pci_dev->dev,
|
||||
sizeof(struct ipw2100_cmd_header), &p,
|
||||
GFP_KERNEL);
|
||||
if (!v) {
|
||||
printk(KERN_ERR DRV_NAME ": "
|
||||
"%s: PCI alloc failed for msg "
|
||||
@ -3461,11 +3458,10 @@ static int ipw2100_msg_allocate(struct ipw2100_priv *priv)
|
||||
return 0;
|
||||
|
||||
for (j = 0; j < i; j++) {
|
||||
pci_free_consistent(priv->pci_dev,
|
||||
sizeof(struct ipw2100_cmd_header),
|
||||
priv->msg_buffers[j].info.c_struct.cmd,
|
||||
priv->msg_buffers[j].info.c_struct.
|
||||
cmd_phys);
|
||||
dma_free_coherent(&priv->pci_dev->dev,
|
||||
sizeof(struct ipw2100_cmd_header),
|
||||
priv->msg_buffers[j].info.c_struct.cmd,
|
||||
priv->msg_buffers[j].info.c_struct.cmd_phys);
|
||||
}
|
||||
|
||||
kfree(priv->msg_buffers);
|
||||
@ -3496,11 +3492,10 @@ static void ipw2100_msg_free(struct ipw2100_priv *priv)
|
||||
return;
|
||||
|
||||
for (i = 0; i < IPW_COMMAND_POOL_SIZE; i++) {
|
||||
pci_free_consistent(priv->pci_dev,
|
||||
sizeof(struct ipw2100_cmd_header),
|
||||
priv->msg_buffers[i].info.c_struct.cmd,
|
||||
priv->msg_buffers[i].info.c_struct.
|
||||
cmd_phys);
|
||||
dma_free_coherent(&priv->pci_dev->dev,
|
||||
sizeof(struct ipw2100_cmd_header),
|
||||
priv->msg_buffers[i].info.c_struct.cmd,
|
||||
priv->msg_buffers[i].info.c_struct.cmd_phys);
|
||||
}
|
||||
|
||||
kfree(priv->msg_buffers);
|
||||
@ -4323,7 +4318,8 @@ static int status_queue_allocate(struct ipw2100_priv *priv, int entries)
|
||||
IPW_DEBUG_INFO("enter\n");
|
||||
|
||||
q->size = entries * sizeof(struct ipw2100_status);
|
||||
q->drv = pci_zalloc_consistent(priv->pci_dev, q->size, &q->nic);
|
||||
q->drv = dma_alloc_coherent(&priv->pci_dev->dev, q->size, &q->nic,
|
||||
GFP_KERNEL);
|
||||
if (!q->drv) {
|
||||
IPW_DEBUG_WARNING("Can not allocate status queue.\n");
|
||||
return -ENOMEM;
|
||||
@ -4339,9 +4335,10 @@ static void status_queue_free(struct ipw2100_priv *priv)
|
||||
IPW_DEBUG_INFO("enter\n");
|
||||
|
||||
if (priv->status_queue.drv) {
|
||||
pci_free_consistent(priv->pci_dev, priv->status_queue.size,
|
||||
priv->status_queue.drv,
|
||||
priv->status_queue.nic);
|
||||
dma_free_coherent(&priv->pci_dev->dev,
|
||||
priv->status_queue.size,
|
||||
priv->status_queue.drv,
|
||||
priv->status_queue.nic);
|
||||
priv->status_queue.drv = NULL;
|
||||
}
|
||||
|
||||
@ -4357,7 +4354,8 @@ static int bd_queue_allocate(struct ipw2100_priv *priv,
|
||||
|
||||
q->entries = entries;
|
||||
q->size = entries * sizeof(struct ipw2100_bd);
|
||||
q->drv = pci_zalloc_consistent(priv->pci_dev, q->size, &q->nic);
|
||||
q->drv = dma_alloc_coherent(&priv->pci_dev->dev, q->size, &q->nic,
|
||||
GFP_KERNEL);
|
||||
if (!q->drv) {
|
||||
IPW_DEBUG_INFO
|
||||
("can't allocate shared memory for buffer descriptors\n");
|
||||
@ -4377,7 +4375,8 @@ static void bd_queue_free(struct ipw2100_priv *priv, struct ipw2100_bd_queue *q)
|
||||
return;
|
||||
|
||||
if (q->drv) {
|
||||
pci_free_consistent(priv->pci_dev, q->size, q->drv, q->nic);
|
||||
dma_free_coherent(&priv->pci_dev->dev, q->size, q->drv,
|
||||
q->nic);
|
||||
q->drv = NULL;
|
||||
}
|
||||
|
||||
@ -4437,9 +4436,9 @@ static int ipw2100_tx_allocate(struct ipw2100_priv *priv)
|
||||
}
|
||||
|
||||
for (i = 0; i < TX_PENDED_QUEUE_LENGTH; i++) {
|
||||
v = pci_alloc_consistent(priv->pci_dev,
|
||||
sizeof(struct ipw2100_data_header),
|
||||
&p);
|
||||
v = dma_alloc_coherent(&priv->pci_dev->dev,
|
||||
sizeof(struct ipw2100_data_header), &p,
|
||||
GFP_KERNEL);
|
||||
if (!v) {
|
||||
printk(KERN_ERR DRV_NAME
|
||||
": %s: PCI alloc failed for tx " "buffers.\n",
|
||||
@ -4459,11 +4458,10 @@ static int ipw2100_tx_allocate(struct ipw2100_priv *priv)
|
||||
return 0;
|
||||
|
||||
for (j = 0; j < i; j++) {
|
||||
pci_free_consistent(priv->pci_dev,
|
||||
sizeof(struct ipw2100_data_header),
|
||||
priv->tx_buffers[j].info.d_struct.data,
|
||||
priv->tx_buffers[j].info.d_struct.
|
||||
data_phys);
|
||||
dma_free_coherent(&priv->pci_dev->dev,
|
||||
sizeof(struct ipw2100_data_header),
|
||||
priv->tx_buffers[j].info.d_struct.data,
|
||||
priv->tx_buffers[j].info.d_struct.data_phys);
|
||||
}
|
||||
|
||||
kfree(priv->tx_buffers);
|
||||
@ -4540,12 +4538,10 @@ static void ipw2100_tx_free(struct ipw2100_priv *priv)
|
||||
priv->tx_buffers[i].info.d_struct.txb = NULL;
|
||||
}
|
||||
if (priv->tx_buffers[i].info.d_struct.data)
|
||||
pci_free_consistent(priv->pci_dev,
|
||||
sizeof(struct ipw2100_data_header),
|
||||
priv->tx_buffers[i].info.d_struct.
|
||||
data,
|
||||
priv->tx_buffers[i].info.d_struct.
|
||||
data_phys);
|
||||
dma_free_coherent(&priv->pci_dev->dev,
|
||||
sizeof(struct ipw2100_data_header),
|
||||
priv->tx_buffers[i].info.d_struct.data,
|
||||
priv->tx_buffers[i].info.d_struct.data_phys);
|
||||
}
|
||||
|
||||
kfree(priv->tx_buffers);
|
||||
@ -4608,9 +4604,10 @@ static int ipw2100_rx_allocate(struct ipw2100_priv *priv)
|
||||
return 0;
|
||||
|
||||
for (j = 0; j < i; j++) {
|
||||
pci_unmap_single(priv->pci_dev, priv->rx_buffers[j].dma_addr,
|
||||
dma_unmap_single(&priv->pci_dev->dev,
|
||||
priv->rx_buffers[j].dma_addr,
|
||||
sizeof(struct ipw2100_rx_packet),
|
||||
PCI_DMA_FROMDEVICE);
|
||||
DMA_FROM_DEVICE);
|
||||
dev_kfree_skb(priv->rx_buffers[j].skb);
|
||||
}
|
||||
|
||||
@ -4662,10 +4659,10 @@ static void ipw2100_rx_free(struct ipw2100_priv *priv)
|
||||
|
||||
for (i = 0; i < RX_QUEUE_LENGTH; i++) {
|
||||
if (priv->rx_buffers[i].rxp) {
|
||||
pci_unmap_single(priv->pci_dev,
|
||||
dma_unmap_single(&priv->pci_dev->dev,
|
||||
priv->rx_buffers[i].dma_addr,
|
||||
sizeof(struct ipw2100_rx),
|
||||
PCI_DMA_FROMDEVICE);
|
||||
DMA_FROM_DEVICE);
|
||||
dev_kfree_skb(priv->rx_buffers[i].skb);
|
||||
}
|
||||
}
|
||||
@ -6196,7 +6193,7 @@ static int ipw2100_pci_init_one(struct pci_dev *pci_dev,
|
||||
pci_set_master(pci_dev);
|
||||
pci_set_drvdata(pci_dev, priv);
|
||||
|
||||
err = pci_set_dma_mask(pci_dev, DMA_BIT_MASK(32));
|
||||
err = dma_set_mask(&pci_dev->dev, DMA_BIT_MASK(32));
|
||||
if (err) {
|
||||
printk(KERN_WARNING DRV_NAME
|
||||
"Error calling pci_set_dma_mask.\n");
|
||||
|
@ -3442,8 +3442,9 @@ static void ipw_rx_queue_reset(struct ipw_priv *priv,
|
||||
/* In the reset function, these buffers may have been allocated
|
||||
* to an SKB, so we need to unmap and free potential storage */
|
||||
if (rxq->pool[i].skb != NULL) {
|
||||
pci_unmap_single(priv->pci_dev, rxq->pool[i].dma_addr,
|
||||
IPW_RX_BUF_SIZE, PCI_DMA_FROMDEVICE);
|
||||
dma_unmap_single(&priv->pci_dev->dev,
|
||||
rxq->pool[i].dma_addr,
|
||||
IPW_RX_BUF_SIZE, DMA_FROM_DEVICE);
|
||||
dev_kfree_skb(rxq->pool[i].skb);
|
||||
rxq->pool[i].skb = NULL;
|
||||
}
|
||||
@ -3774,7 +3775,8 @@ static int ipw_queue_tx_init(struct ipw_priv *priv,
|
||||
return -ENOMEM;
|
||||
|
||||
q->bd =
|
||||
pci_alloc_consistent(dev, sizeof(q->bd[0]) * count, &q->q.dma_addr);
|
||||
dma_alloc_coherent(&dev->dev, sizeof(q->bd[0]) * count,
|
||||
&q->q.dma_addr, GFP_KERNEL);
|
||||
if (!q->bd) {
|
||||
IPW_ERROR("pci_alloc_consistent(%zd) failed\n",
|
||||
sizeof(q->bd[0]) * count);
|
||||
@ -3816,9 +3818,10 @@ static void ipw_queue_tx_free_tfd(struct ipw_priv *priv,
|
||||
|
||||
/* unmap chunks if any */
|
||||
for (i = 0; i < le32_to_cpu(bd->u.data.num_chunks); i++) {
|
||||
pci_unmap_single(dev, le32_to_cpu(bd->u.data.chunk_ptr[i]),
|
||||
dma_unmap_single(&dev->dev,
|
||||
le32_to_cpu(bd->u.data.chunk_ptr[i]),
|
||||
le16_to_cpu(bd->u.data.chunk_len[i]),
|
||||
PCI_DMA_TODEVICE);
|
||||
DMA_TO_DEVICE);
|
||||
if (txq->txb[txq->q.last_used]) {
|
||||
libipw_txb_free(txq->txb[txq->q.last_used]);
|
||||
txq->txb[txq->q.last_used] = NULL;
|
||||
@ -3850,8 +3853,8 @@ static void ipw_queue_tx_free(struct ipw_priv *priv, struct clx2_tx_queue *txq)
|
||||
}
|
||||
|
||||
/* free buffers belonging to queue itself */
|
||||
pci_free_consistent(dev, sizeof(txq->bd[0]) * q->n_bd, txq->bd,
|
||||
q->dma_addr);
|
||||
dma_free_coherent(&dev->dev, sizeof(txq->bd[0]) * q->n_bd, txq->bd,
|
||||
q->dma_addr);
|
||||
kfree(txq->txb);
|
||||
|
||||
/* 0 fill whole structure */
|
||||
@ -5196,8 +5199,8 @@ static void ipw_rx_queue_replenish(void *data)
|
||||
list_del(element);
|
||||
|
||||
rxb->dma_addr =
|
||||
pci_map_single(priv->pci_dev, rxb->skb->data,
|
||||
IPW_RX_BUF_SIZE, PCI_DMA_FROMDEVICE);
|
||||
dma_map_single(&priv->pci_dev->dev, rxb->skb->data,
|
||||
IPW_RX_BUF_SIZE, DMA_FROM_DEVICE);
|
||||
|
||||
list_add_tail(&rxb->list, &rxq->rx_free);
|
||||
rxq->free_count++;
|
||||
@ -5230,8 +5233,9 @@ static void ipw_rx_queue_free(struct ipw_priv *priv, struct ipw_rx_queue *rxq)
|
||||
|
||||
for (i = 0; i < RX_QUEUE_SIZE + RX_FREE_BUFFERS; i++) {
|
||||
if (rxq->pool[i].skb != NULL) {
|
||||
pci_unmap_single(priv->pci_dev, rxq->pool[i].dma_addr,
|
||||
IPW_RX_BUF_SIZE, PCI_DMA_FROMDEVICE);
|
||||
dma_unmap_single(&priv->pci_dev->dev,
|
||||
rxq->pool[i].dma_addr,
|
||||
IPW_RX_BUF_SIZE, DMA_FROM_DEVICE);
|
||||
dev_kfree_skb(rxq->pool[i].skb);
|
||||
}
|
||||
}
|
||||
@ -8263,9 +8267,8 @@ static void ipw_rx(struct ipw_priv *priv)
|
||||
}
|
||||
priv->rxq->queue[i] = NULL;
|
||||
|
||||
pci_dma_sync_single_for_cpu(priv->pci_dev, rxb->dma_addr,
|
||||
IPW_RX_BUF_SIZE,
|
||||
PCI_DMA_FROMDEVICE);
|
||||
dma_sync_single_for_cpu(&priv->pci_dev->dev, rxb->dma_addr,
|
||||
IPW_RX_BUF_SIZE, DMA_FROM_DEVICE);
|
||||
|
||||
pkt = (struct ipw_rx_packet *)rxb->skb->data;
|
||||
IPW_DEBUG_RX("Packet: type=%02X seq=%02X bits=%02X\n",
|
||||
@ -8417,8 +8420,8 @@ static void ipw_rx(struct ipw_priv *priv)
|
||||
rxb->skb = NULL;
|
||||
}
|
||||
|
||||
pci_unmap_single(priv->pci_dev, rxb->dma_addr,
|
||||
IPW_RX_BUF_SIZE, PCI_DMA_FROMDEVICE);
|
||||
dma_unmap_single(&priv->pci_dev->dev, rxb->dma_addr,
|
||||
IPW_RX_BUF_SIZE, DMA_FROM_DEVICE);
|
||||
list_add_tail(&rxb->list, &priv->rxq->rx_used);
|
||||
|
||||
i = (i + 1) % RX_QUEUE_SIZE;
|
||||
@ -10217,11 +10220,10 @@ static int ipw_tx_skb(struct ipw_priv *priv, struct libipw_txb *txb,
|
||||
txb->fragments[i]->len - hdr_len);
|
||||
|
||||
tfd->u.data.chunk_ptr[i] =
|
||||
cpu_to_le32(pci_map_single
|
||||
(priv->pci_dev,
|
||||
txb->fragments[i]->data + hdr_len,
|
||||
txb->fragments[i]->len - hdr_len,
|
||||
PCI_DMA_TODEVICE));
|
||||
cpu_to_le32(dma_map_single(&priv->pci_dev->dev,
|
||||
txb->fragments[i]->data + hdr_len,
|
||||
txb->fragments[i]->len - hdr_len,
|
||||
DMA_TO_DEVICE));
|
||||
tfd->u.data.chunk_len[i] =
|
||||
cpu_to_le16(txb->fragments[i]->len - hdr_len);
|
||||
}
|
||||
@ -10251,10 +10253,10 @@ static int ipw_tx_skb(struct ipw_priv *priv, struct libipw_txb *txb,
|
||||
dev_kfree_skb_any(txb->fragments[i]);
|
||||
txb->fragments[i] = skb;
|
||||
tfd->u.data.chunk_ptr[i] =
|
||||
cpu_to_le32(pci_map_single
|
||||
(priv->pci_dev, skb->data,
|
||||
remaining_bytes,
|
||||
PCI_DMA_TODEVICE));
|
||||
cpu_to_le32(dma_map_single(&priv->pci_dev->dev,
|
||||
skb->data,
|
||||
remaining_bytes,
|
||||
DMA_TO_DEVICE));
|
||||
|
||||
le32_add_cpu(&tfd->u.data.num_chunks, 1);
|
||||
}
|
||||
@ -11620,9 +11622,9 @@ static int ipw_pci_probe(struct pci_dev *pdev,
|
||||
|
||||
pci_set_master(pdev);
|
||||
|
||||
err = pci_set_dma_mask(pdev, DMA_BIT_MASK(32));
|
||||
err = dma_set_mask(&pdev->dev, DMA_BIT_MASK(32));
|
||||
if (!err)
|
||||
err = pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(32));
|
||||
err = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32));
|
||||
if (err) {
|
||||
printk(KERN_WARNING DRV_NAME ": No suitable DMA available.\n");
|
||||
goto out_pci_disable_device;
|
||||
|
Loading…
Reference in New Issue
Block a user