summaryrefslogtreecommitdiff
path: root/drivers/net/mhi_net.c
diff options
context:
space:
mode:
authorLoic Poulain <loic.poulain@linaro.org>2021-01-11 19:07:42 +0100
committerJakub Kicinski <kuba@kernel.org>2021-01-29 19:42:06 -0800
commit6e10785ee148655577885b65605836210a741bee (patch)
tree818ea9b8588452c9b09841250bb18a13388f3227 /drivers/net/mhi_net.c
parente6ec3ccd4eb2b3b9cf0a4127cf00fae602570a52 (diff)
net: mhi: Get rid of local rx queue count
Use the new mhi_get_free_desc_count helper to track queue usage instead of relying on the locally maintained rx_queued count. Signed-off-by: Loic Poulain <loic.poulain@linaro.org> Signed-off-by: Jakub Kicinski <kuba@kernel.org>
Diffstat (limited to 'drivers/net/mhi_net.c')
-rw-r--r--drivers/net/mhi_net.c13
1 files changed, 5 insertions, 8 deletions
diff --git a/drivers/net/mhi_net.c b/drivers/net/mhi_net.c
index abdd2baa312a..4f512531b7d0 100644
--- a/drivers/net/mhi_net.c
+++ b/drivers/net/mhi_net.c
@@ -25,7 +25,6 @@ struct mhi_net_stats {
u64_stats_t tx_bytes;
u64_stats_t tx_errors;
u64_stats_t tx_dropped;
- atomic_t rx_queued;
struct u64_stats_sync tx_syncp;
struct u64_stats_sync rx_syncp;
};
@@ -138,9 +137,9 @@ static void mhi_net_dl_callback(struct mhi_device *mhi_dev,
{
struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev);
struct sk_buff *skb = mhi_res->buf_addr;
- int remaining;
+ int free_desc_count;
- remaining = atomic_dec_return(&mhi_netdev->stats.rx_queued);
+ free_desc_count = mhi_get_free_desc_count(mhi_dev, DMA_FROM_DEVICE);
if (unlikely(mhi_res->transaction_status)) {
dev_kfree_skb_any(skb);
@@ -175,7 +174,7 @@ static void mhi_net_dl_callback(struct mhi_device *mhi_dev,
}
/* Refill if RX buffers queue becomes low */
- if (remaining <= mhi_netdev->rx_queue_sz / 2)
+ if (free_desc_count >= mhi_netdev->rx_queue_sz / 2)
schedule_delayed_work(&mhi_netdev->rx_refill, 0);
}
@@ -222,7 +221,7 @@ static void mhi_net_rx_refill_work(struct work_struct *work)
struct sk_buff *skb;
int err;
- while (atomic_read(&mhi_netdev->stats.rx_queued) < mhi_netdev->rx_queue_sz) {
+ while (!mhi_queue_is_full(mdev, DMA_FROM_DEVICE)) {
skb = netdev_alloc_skb(ndev, size);
if (unlikely(!skb))
break;
@@ -235,8 +234,6 @@ static void mhi_net_rx_refill_work(struct work_struct *work)
break;
}
- atomic_inc(&mhi_netdev->stats.rx_queued);
-
/* Do not hog the CPU if rx buffers are consumed faster than
* queued (unlikely).
*/
@@ -244,7 +241,7 @@ static void mhi_net_rx_refill_work(struct work_struct *work)
}
/* If we're still starved of rx buffers, reschedule later */
- if (unlikely(!atomic_read(&mhi_netdev->stats.rx_queued)))
+ if (mhi_get_free_desc_count(mdev, DMA_FROM_DEVICE) == mhi_netdev->rx_queue_sz)
schedule_delayed_work(&mhi_netdev->rx_refill, HZ / 2);
}