]> git.proxmox.com Git - mirror_ubuntu-artful-kernel.git/commitdiff
virtio-net: napi helper functions
authorWillem de Bruijn <willemb@google.com>
Mon, 24 Apr 2017 17:49:26 +0000 (13:49 -0400)
committerDavid S. Miller <davem@davemloft.net>
Tue, 25 Apr 2017 03:55:19 +0000 (23:55 -0400)
Prepare virtio-net for tx napi by converting existing napi code to
use helper functions. This also deduplicates some logic.

Signed-off-by: Willem de Bruijn <willemb@google.com>
Signed-off-by: Jason Wang <jasowang@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
drivers/net/virtio_net.c

index 666ada6130abfe8912c7e2f045d91e4e45e4558c..b9c1df29892caf488ebbb0e2f5597023411fa7bb 100644 (file)
@@ -239,6 +239,26 @@ static struct page *get_a_page(struct receive_queue *rq, gfp_t gfp_mask)
        return p;
 }
 
+static void virtqueue_napi_schedule(struct napi_struct *napi,
+                                   struct virtqueue *vq)
+{
+       if (napi_schedule_prep(napi)) {
+               virtqueue_disable_cb(vq);
+               __napi_schedule(napi);
+       }
+}
+
+static void virtqueue_napi_complete(struct napi_struct *napi,
+                                   struct virtqueue *vq, int processed)
+{
+       int opaque;
+
+       opaque = virtqueue_enable_cb_prepare(vq);
+       if (napi_complete_done(napi, processed) &&
+           unlikely(virtqueue_poll(vq, opaque)))
+               virtqueue_napi_schedule(napi, vq);
+}
+
 static void skb_xmit_done(struct virtqueue *vq)
 {
        struct virtnet_info *vi = vq->vdev->priv;
@@ -936,27 +956,20 @@ static void skb_recv_done(struct virtqueue *rvq)
        struct virtnet_info *vi = rvq->vdev->priv;
        struct receive_queue *rq = &vi->rq[vq2rxq(rvq)];
 
-       /* Schedule NAPI, Suppress further interrupts if successful. */
-       if (napi_schedule_prep(&rq->napi)) {
-               virtqueue_disable_cb(rvq);
-               __napi_schedule(&rq->napi);
-       }
+       virtqueue_napi_schedule(&rq->napi, rvq);
 }
 
-static void virtnet_napi_enable(struct receive_queue *rq)
+static void virtnet_napi_enable(struct virtqueue *vq, struct napi_struct *napi)
 {
-       napi_enable(&rq->napi);
+       napi_enable(napi);
 
        /* If all buffers were filled by other side before we napi_enabled, we
-        * won't get another interrupt, so process any outstanding packets
-        * now.  virtnet_poll wants re-enable the queue, so we disable here.
-        * We synchronize against interrupts via NAPI_STATE_SCHED */
-       if (napi_schedule_prep(&rq->napi)) {
-               virtqueue_disable_cb(rq->vq);
-               local_bh_disable();
-               __napi_schedule(&rq->napi);
-               local_bh_enable();
-       }
+        * won't get another interrupt, so process any outstanding packets now.
+        * Call local_bh_enable after to trigger softIRQ processing.
+        */
+       local_bh_disable();
+       virtqueue_napi_schedule(napi, vq);
+       local_bh_enable();
 }
 
 static void refill_work(struct work_struct *work)
@@ -971,7 +984,7 @@ static void refill_work(struct work_struct *work)
 
                napi_disable(&rq->napi);
                still_empty = !try_fill_recv(vi, rq, GFP_KERNEL);
-               virtnet_napi_enable(rq);
+               virtnet_napi_enable(rq->vq, &rq->napi);
 
                /* In theory, this can happen: if we don't get any buffers in
                 * we will *never* try to fill again.
@@ -1011,21 +1024,13 @@ static int virtnet_poll(struct napi_struct *napi, int budget)
 {
        struct receive_queue *rq =
                container_of(napi, struct receive_queue, napi);
-       unsigned int r, received;
+       unsigned int received;
 
        received = virtnet_receive(rq, budget);
 
        /* Out of packets? */
-       if (received < budget) {
-               r = virtqueue_enable_cb_prepare(rq->vq);
-               if (napi_complete_done(napi, received)) {
-                       if (unlikely(virtqueue_poll(rq->vq, r)) &&
-                           napi_schedule_prep(napi)) {
-                               virtqueue_disable_cb(rq->vq);
-                               __napi_schedule(napi);
-                       }
-               }
-       }
+       if (received < budget)
+               virtqueue_napi_complete(napi, rq->vq, received);
 
        return received;
 }
@@ -1040,7 +1045,7 @@ static int virtnet_open(struct net_device *dev)
                        /* Make sure we have some buffers: if oom use wq. */
                        if (!try_fill_recv(vi, &vi->rq[i], GFP_KERNEL))
                                schedule_delayed_work(&vi->refill, 0);
-               virtnet_napi_enable(&vi->rq[i]);
+               virtnet_napi_enable(vi->rq[i].vq, &vi->rq[i].napi);
        }
 
        return 0;
@@ -1747,7 +1752,7 @@ static int virtnet_restore_up(struct virtio_device *vdev)
                                schedule_delayed_work(&vi->refill, 0);
 
                for (i = 0; i < vi->max_queue_pairs; i++)
-                       virtnet_napi_enable(&vi->rq[i]);
+                       virtnet_napi_enable(vi->rq[i].vq, &vi->rq[i].napi);
        }
 
        netif_device_attach(vi->dev);