mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-28 11:18:45 +07:00
mt76usb: remove mt76u_buf and use urb directly
Put urb pointer in mt76_queue_entry directly instead of mt76u_buf structure. Signed-off-by: Stanislaw Gruszka <sgruszka@redhat.com> Signed-off-by: Felix Fietkau <nbd@nbd.name>
This commit is contained in:
parent
279ade99ed
commit
d7d4ea9ac8
@ -89,10 +89,6 @@ struct mt76_tx_info {
|
||||
u32 info;
|
||||
};
|
||||
|
||||
struct mt76u_buf {
|
||||
struct urb *urb;
|
||||
};
|
||||
|
||||
struct mt76_queue_entry {
|
||||
union {
|
||||
void *buf;
|
||||
@ -100,7 +96,7 @@ struct mt76_queue_entry {
|
||||
};
|
||||
union {
|
||||
struct mt76_txwi_cache *txwi;
|
||||
struct mt76u_buf ubuf;
|
||||
struct urb *urb;
|
||||
};
|
||||
enum mt76_txq_id qid;
|
||||
bool schedule;
|
||||
|
@ -283,12 +283,11 @@ mt76u_set_endpoints(struct usb_interface *intf,
|
||||
}
|
||||
|
||||
static int
|
||||
mt76u_fill_rx_sg(struct mt76_dev *dev, struct mt76u_buf *buf, int nsgs,
|
||||
mt76u_fill_rx_sg(struct mt76_dev *dev, struct urb *urb, int nsgs,
|
||||
gfp_t gfp)
|
||||
{
|
||||
struct mt76_queue *q = &dev->q_rx[MT_RXQ_MAIN];
|
||||
int sglen = SKB_WITH_OVERHEAD(q->buf_size);
|
||||
struct urb *urb = buf->urb;
|
||||
|
||||
int i;
|
||||
|
||||
@ -323,44 +322,43 @@ mt76u_fill_rx_sg(struct mt76_dev *dev, struct mt76u_buf *buf, int nsgs,
|
||||
|
||||
static int
|
||||
mt76u_refill_rx(struct mt76_dev *dev, struct mt76_queue *q,
|
||||
struct mt76u_buf *buf, int nsgs, gfp_t gfp)
|
||||
struct urb *urb, int nsgs, gfp_t gfp)
|
||||
{
|
||||
if (dev->usb.sg_en) {
|
||||
return mt76u_fill_rx_sg(dev, buf, nsgs, gfp);
|
||||
return mt76u_fill_rx_sg(dev, urb, nsgs, gfp);
|
||||
} else {
|
||||
buf->urb->transfer_buffer_length =
|
||||
SKB_WITH_OVERHEAD(q->buf_size);
|
||||
buf->urb->transfer_buffer =
|
||||
page_frag_alloc(&q->rx_page, q->buf_size, gfp);
|
||||
return buf->urb->transfer_buffer ? 0 : -ENOMEM;
|
||||
urb->transfer_buffer_length = SKB_WITH_OVERHEAD(q->buf_size);
|
||||
urb->transfer_buffer = page_frag_alloc(&q->rx_page,
|
||||
q->buf_size, gfp);
|
||||
return urb->transfer_buffer ? 0 : -ENOMEM;
|
||||
}
|
||||
}
|
||||
|
||||
static int
|
||||
mt76u_buf_alloc(struct mt76_dev *dev, struct mt76u_buf *buf)
|
||||
mt76u_urb_alloc(struct mt76_dev *dev, struct mt76_queue_entry *e)
|
||||
{
|
||||
struct mt76_queue *q = &dev->q_rx[MT_RXQ_MAIN];
|
||||
struct urb *urb;
|
||||
|
||||
buf->urb = usb_alloc_urb(0, GFP_KERNEL);
|
||||
if (!buf->urb)
|
||||
urb = usb_alloc_urb(0, GFP_KERNEL);
|
||||
if (!urb)
|
||||
return -ENOMEM;
|
||||
e->urb = urb;
|
||||
|
||||
if (dev->usb.sg_en) {
|
||||
buf->urb->sg = devm_kcalloc(dev->dev, MT_SG_MAX_SIZE,
|
||||
sizeof(*buf->urb->sg),
|
||||
GFP_KERNEL);
|
||||
if (!buf->urb->sg)
|
||||
urb->sg = devm_kcalloc(dev->dev, MT_SG_MAX_SIZE,
|
||||
sizeof(*urb->sg), GFP_KERNEL);
|
||||
if (!urb->sg)
|
||||
return -ENOMEM;
|
||||
|
||||
sg_init_table(buf->urb->sg, MT_SG_MAX_SIZE);
|
||||
sg_init_table(urb->sg, MT_SG_MAX_SIZE);
|
||||
}
|
||||
|
||||
return mt76u_refill_rx(dev, q, buf, MT_SG_MAX_SIZE, GFP_KERNEL);
|
||||
return mt76u_refill_rx(dev, q, urb, MT_SG_MAX_SIZE, GFP_KERNEL);
|
||||
}
|
||||
|
||||
static void mt76u_buf_free(struct mt76u_buf *buf)
|
||||
static void mt76u_urb_free(struct urb *urb)
|
||||
{
|
||||
struct urb *urb = buf->urb;
|
||||
int i;
|
||||
|
||||
for (i = 0; i < urb->num_sgs; i++)
|
||||
@ -369,12 +367,12 @@ static void mt76u_buf_free(struct mt76u_buf *buf)
|
||||
if (urb->transfer_buffer)
|
||||
skb_free_frag(urb->transfer_buffer);
|
||||
|
||||
usb_free_urb(buf->urb);
|
||||
usb_free_urb(urb);
|
||||
}
|
||||
|
||||
static void
|
||||
mt76u_fill_bulk_urb(struct mt76_dev *dev, int dir, int index,
|
||||
struct mt76u_buf *buf, usb_complete_t complete_fn,
|
||||
struct urb *urb, usb_complete_t complete_fn,
|
||||
void *context)
|
||||
{
|
||||
struct usb_device *udev = to_usb_device(dev->dev);
|
||||
@ -385,27 +383,27 @@ mt76u_fill_bulk_urb(struct mt76_dev *dev, int dir, int index,
|
||||
else
|
||||
pipe = usb_sndbulkpipe(udev, dev->usb.out_ep[index]);
|
||||
|
||||
buf->urb->dev = udev;
|
||||
buf->urb->pipe = pipe;
|
||||
buf->urb->complete = complete_fn;
|
||||
buf->urb->context = context;
|
||||
urb->dev = udev;
|
||||
urb->pipe = pipe;
|
||||
urb->complete = complete_fn;
|
||||
urb->context = context;
|
||||
}
|
||||
|
||||
static inline struct mt76u_buf
|
||||
*mt76u_get_next_rx_entry(struct mt76_queue *q)
|
||||
static inline struct urb *
|
||||
mt76u_get_next_rx_entry(struct mt76_queue *q)
|
||||
{
|
||||
struct mt76u_buf *buf = NULL;
|
||||
struct urb *urb = NULL;
|
||||
unsigned long flags;
|
||||
|
||||
spin_lock_irqsave(&q->lock, flags);
|
||||
if (q->queued > 0) {
|
||||
buf = &q->entry[q->head].ubuf;
|
||||
urb = q->entry[q->head].urb;
|
||||
q->head = (q->head + 1) % q->ndesc;
|
||||
q->queued--;
|
||||
}
|
||||
spin_unlock_irqrestore(&q->lock, flags);
|
||||
|
||||
return buf;
|
||||
return urb;
|
||||
}
|
||||
|
||||
static int mt76u_get_rx_entry_len(u8 *data, u32 data_len)
|
||||
@ -424,10 +422,9 @@ static int mt76u_get_rx_entry_len(u8 *data, u32 data_len)
|
||||
}
|
||||
|
||||
static int
|
||||
mt76u_process_rx_entry(struct mt76_dev *dev, struct mt76u_buf *buf)
|
||||
mt76u_process_rx_entry(struct mt76_dev *dev, struct urb *urb)
|
||||
{
|
||||
struct mt76_queue *q = &dev->q_rx[MT_RXQ_MAIN];
|
||||
struct urb *urb = buf->urb;
|
||||
u8 *data = urb->num_sgs ? sg_virt(&urb->sg[0]) : urb->transfer_buffer;
|
||||
int data_len = urb->num_sgs ? urb->sg[0].length : urb->actual_length;
|
||||
int len, nsgs = 1;
|
||||
@ -488,7 +485,7 @@ static void mt76u_complete_rx(struct urb *urb)
|
||||
}
|
||||
|
||||
spin_lock_irqsave(&q->lock, flags);
|
||||
if (WARN_ONCE(q->entry[q->tail].ubuf.urb != urb, "rx urb mismatch"))
|
||||
if (WARN_ONCE(q->entry[q->tail].urb != urb, "rx urb mismatch"))
|
||||
goto out;
|
||||
|
||||
q->tail = (q->tail + 1) % q->ndesc;
|
||||
@ -499,37 +496,37 @@ static void mt76u_complete_rx(struct urb *urb)
|
||||
}
|
||||
|
||||
static int
|
||||
mt76u_submit_rx_buf(struct mt76_dev *dev, struct mt76u_buf *buf)
|
||||
mt76u_submit_rx_buf(struct mt76_dev *dev, struct urb *urb)
|
||||
{
|
||||
mt76u_fill_bulk_urb(dev, USB_DIR_IN, MT_EP_IN_PKT_RX, buf,
|
||||
mt76u_fill_bulk_urb(dev, USB_DIR_IN, MT_EP_IN_PKT_RX, urb,
|
||||
mt76u_complete_rx, dev);
|
||||
trace_submit_urb(dev, buf->urb);
|
||||
trace_submit_urb(dev, urb);
|
||||
|
||||
return usb_submit_urb(buf->urb, GFP_ATOMIC);
|
||||
return usb_submit_urb(urb, GFP_ATOMIC);
|
||||
}
|
||||
|
||||
static void mt76u_rx_tasklet(unsigned long data)
|
||||
{
|
||||
struct mt76_dev *dev = (struct mt76_dev *)data;
|
||||
struct mt76_queue *q = &dev->q_rx[MT_RXQ_MAIN];
|
||||
struct mt76u_buf *buf;
|
||||
struct urb *urb;
|
||||
int err, count;
|
||||
|
||||
rcu_read_lock();
|
||||
|
||||
while (true) {
|
||||
buf = mt76u_get_next_rx_entry(q);
|
||||
if (!buf)
|
||||
urb = mt76u_get_next_rx_entry(q);
|
||||
if (!urb)
|
||||
break;
|
||||
|
||||
count = mt76u_process_rx_entry(dev, buf);
|
||||
count = mt76u_process_rx_entry(dev, urb);
|
||||
if (count > 0) {
|
||||
err = mt76u_refill_rx(dev, q, buf, count,
|
||||
err = mt76u_refill_rx(dev, q, urb, count,
|
||||
GFP_ATOMIC);
|
||||
if (err < 0)
|
||||
break;
|
||||
}
|
||||
mt76u_submit_rx_buf(dev, buf);
|
||||
mt76u_submit_rx_buf(dev, urb);
|
||||
}
|
||||
mt76_rx_poll_complete(dev, MT_RXQ_MAIN, NULL);
|
||||
|
||||
@ -544,7 +541,7 @@ int mt76u_submit_rx_buffers(struct mt76_dev *dev)
|
||||
|
||||
spin_lock_irqsave(&q->lock, flags);
|
||||
for (i = 0; i < q->ndesc; i++) {
|
||||
err = mt76u_submit_rx_buf(dev, &q->entry[i].ubuf);
|
||||
err = mt76u_submit_rx_buf(dev, q->entry[i].urb);
|
||||
if (err < 0)
|
||||
break;
|
||||
}
|
||||
@ -576,7 +573,7 @@ static int mt76u_alloc_rx(struct mt76_dev *dev)
|
||||
q->buf_size = dev->usb.sg_en ? MT_RX_BUF_SIZE : PAGE_SIZE;
|
||||
q->ndesc = MT_NUM_RX_ENTRIES;
|
||||
for (i = 0; i < q->ndesc; i++) {
|
||||
err = mt76u_buf_alloc(dev, &q->entry[i].ubuf);
|
||||
err = mt76u_urb_alloc(dev, &q->entry[i]);
|
||||
if (err < 0)
|
||||
return err;
|
||||
}
|
||||
@ -591,7 +588,7 @@ static void mt76u_free_rx(struct mt76_dev *dev)
|
||||
int i;
|
||||
|
||||
for (i = 0; i < q->ndesc; i++)
|
||||
mt76u_buf_free(&q->entry[i].ubuf);
|
||||
mt76u_urb_free(q->entry[i].urb);
|
||||
|
||||
if (!q->rx_page.va)
|
||||
return;
|
||||
@ -607,7 +604,7 @@ static void mt76u_stop_rx(struct mt76_dev *dev)
|
||||
int i;
|
||||
|
||||
for (i = 0; i < q->ndesc; i++)
|
||||
usb_kill_urb(q->entry[i].ubuf.urb);
|
||||
usb_kill_urb(q->entry[i].urb);
|
||||
}
|
||||
|
||||
static void mt76u_tx_tasklet(unsigned long data)
|
||||
@ -724,7 +721,7 @@ mt76u_tx_queue_skb(struct mt76_dev *dev, enum mt76_txq_id qid,
|
||||
struct ieee80211_sta *sta)
|
||||
{
|
||||
struct mt76_queue *q = dev->q_tx[qid].q;
|
||||
struct mt76u_buf *buf;
|
||||
struct urb *urb;
|
||||
u16 idx = q->tail;
|
||||
int err;
|
||||
|
||||
@ -737,13 +734,13 @@ mt76u_tx_queue_skb(struct mt76_dev *dev, enum mt76_txq_id qid,
|
||||
return err;
|
||||
|
||||
q->entry[idx].done = false;
|
||||
buf = &q->entry[idx].ubuf;
|
||||
err = mt76u_tx_setup_buffers(dev, skb, buf->urb);
|
||||
urb = q->entry[idx].urb;
|
||||
err = mt76u_tx_setup_buffers(dev, skb, urb);
|
||||
if (err < 0)
|
||||
return err;
|
||||
|
||||
mt76u_fill_bulk_urb(dev, USB_DIR_OUT, q2ep(q->hw_idx),
|
||||
buf, mt76u_complete_tx, &q->entry[idx]);
|
||||
urb, mt76u_complete_tx, &q->entry[idx]);
|
||||
|
||||
q->tail = (q->tail + 1) % q->ndesc;
|
||||
q->entry[idx].skb = skb;
|
||||
@ -754,14 +751,14 @@ mt76u_tx_queue_skb(struct mt76_dev *dev, enum mt76_txq_id qid,
|
||||
|
||||
static void mt76u_tx_kick(struct mt76_dev *dev, struct mt76_queue *q)
|
||||
{
|
||||
struct mt76u_buf *buf;
|
||||
struct urb *urb;
|
||||
int err;
|
||||
|
||||
while (q->first != q->tail) {
|
||||
buf = &q->entry[q->first].ubuf;
|
||||
urb = q->entry[q->first].urb;
|
||||
|
||||
trace_submit_urb(dev, buf->urb);
|
||||
err = usb_submit_urb(buf->urb, GFP_ATOMIC);
|
||||
trace_submit_urb(dev, urb);
|
||||
err = usb_submit_urb(urb, GFP_ATOMIC);
|
||||
if (err < 0) {
|
||||
if (err == -ENODEV)
|
||||
set_bit(MT76_REMOVED, &dev->state);
|
||||
@ -776,7 +773,7 @@ static void mt76u_tx_kick(struct mt76_dev *dev, struct mt76_queue *q)
|
||||
|
||||
static int mt76u_alloc_tx(struct mt76_dev *dev)
|
||||
{
|
||||
struct mt76u_buf *buf;
|
||||
struct urb *urb;
|
||||
struct mt76_queue *q;
|
||||
int i, j;
|
||||
|
||||
@ -804,19 +801,18 @@ static int mt76u_alloc_tx(struct mt76_dev *dev)
|
||||
|
||||
q->ndesc = MT_NUM_TX_ENTRIES;
|
||||
for (j = 0; j < q->ndesc; j++) {
|
||||
buf = &q->entry[j].ubuf;
|
||||
|
||||
buf->urb = usb_alloc_urb(0, GFP_KERNEL);
|
||||
if (!buf->urb)
|
||||
urb = usb_alloc_urb(0, GFP_KERNEL);
|
||||
if (!urb)
|
||||
return -ENOMEM;
|
||||
q->entry[j].urb = urb;
|
||||
|
||||
if (!dev->usb.sg_en)
|
||||
continue;
|
||||
|
||||
buf->urb->sg = devm_kcalloc(dev->dev, MT_SG_MAX_SIZE,
|
||||
sizeof(struct scatterlist),
|
||||
GFP_KERNEL);
|
||||
if (!buf->urb->sg)
|
||||
urb->sg = devm_kcalloc(dev->dev, MT_SG_MAX_SIZE,
|
||||
sizeof(struct scatterlist),
|
||||
GFP_KERNEL);
|
||||
if (!urb->sg)
|
||||
return -ENOMEM;
|
||||
}
|
||||
}
|
||||
@ -831,7 +827,7 @@ static void mt76u_free_tx(struct mt76_dev *dev)
|
||||
for (i = 0; i < IEEE80211_NUM_ACS; i++) {
|
||||
q = dev->q_tx[i].q;
|
||||
for (j = 0; j < q->ndesc; j++)
|
||||
usb_free_urb(q->entry[j].ubuf.urb);
|
||||
usb_free_urb(q->entry[j].urb);
|
||||
}
|
||||
}
|
||||
|
||||
@ -843,7 +839,7 @@ static void mt76u_stop_tx(struct mt76_dev *dev)
|
||||
for (i = 0; i < IEEE80211_NUM_ACS; i++) {
|
||||
q = dev->q_tx[i].q;
|
||||
for (j = 0; j < q->ndesc; j++)
|
||||
usb_kill_urb(q->entry[j].ubuf.urb);
|
||||
usb_kill_urb(q->entry[j].urb);
|
||||
}
|
||||
}
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user