]> git.proxmox.com Git - mirror_ubuntu-jammy-kernel.git/commitdiff
soc: fsl: dpio: Replace QMAN array mode with ring mode enqueue
authorYouri Querry <youri.querry_1@nxp.com>
Thu, 12 Dec 2019 17:01:18 +0000 (17:01 +0000)
committerLi Yang <leoyang.li@nxp.com>
Thu, 20 Feb 2020 00:06:34 +0000 (18:06 -0600)
This change of algorithm will enable faster bulk enqueue.
This will greatly benefit XDP bulk enqueue.

Signed-off-by: Youri Querry <youri.querry_1@nxp.com>
Acked-by: Roy Pledge <roy.pledge@nxp.com>
Signed-off-by: Li Yang <leoyang.li@nxp.com>
drivers/soc/fsl/dpio/qbman-portal.c
drivers/soc/fsl/dpio/qbman-portal.h

index 0ffe018afb1791d8ebe01d5a0e4326c9160babdb..740ee0d1958206d0a30c1171524af8e857310ffe 100644 (file)
@@ -8,6 +8,7 @@
 #include <asm/cacheflush.h>
 #include <linux/io.h>
 #include <linux/slab.h>
+#include <linux/spinlock.h>
 #include <soc/fsl/dpaa2-global.h>
 
 #include "qbman-portal.h"
@@ -21,6 +22,7 @@
 
 /* CINH register offsets */
 #define QBMAN_CINH_SWP_EQCR_PI      0x800
+#define QBMAN_CINH_SWP_EQCR_CI     0x840
 #define QBMAN_CINH_SWP_EQAR    0x8c0
 #define QBMAN_CINH_SWP_CR_RT        0x900
 #define QBMAN_CINH_SWP_VDQCR_RT     0x940
@@ -44,6 +46,8 @@
 #define QBMAN_CENA_SWP_CR      0x600
 #define QBMAN_CENA_SWP_RR(vb)  (0x700 + ((u32)(vb) >> 1))
 #define QBMAN_CENA_SWP_VDQCR   0x780
+#define QBMAN_CENA_SWP_EQCR_CI 0x840
+#define QBMAN_CENA_SWP_EQCR_CI_MEMBACK 0x1840
 
 /* CENA register offsets in memory-backed mode */
 #define QBMAN_CENA_SWP_DQRR_MEM(n)  (0x800 + ((u32)(n) << 6))
 /* opaque token for static dequeues */
 #define QMAN_SDQCR_TOKEN    0xbb
 
+#define QBMAN_EQCR_DCA_IDXMASK          0x0f
+#define QBMAN_ENQUEUE_FLAG_DCA          (1ULL << 31)
+
+#define EQ_DESC_SIZE_WITHOUT_FD 29
+#define EQ_DESC_SIZE_FD_START 32
+
 enum qbman_sdqcr_dct {
        qbman_sdqcr_dct_null = 0,
        qbman_sdqcr_dct_prio_ics,
@@ -215,6 +225,15 @@ static inline u32 qbman_set_swp_cfg(u8 max_fill, u8 wn,    u8 est, u8 rpm, u8 dcm,
 
 #define QMAN_RT_MODE      0x00000100
 
+static inline u8 qm_cyc_diff(u8 ringsize, u8 first, u8 last)
+{
+       /* 'first' is included, 'last' is excluded */
+       if (first <= last)
+               return last - first;
+       else
+               return (2 * ringsize) - (first - last);
+}
+
 /**
  * qbman_swp_init() - Create a functional object representing the given
  *                    QBMan portal descriptor.
@@ -227,6 +246,10 @@ struct qbman_swp *qbman_swp_init(const struct qbman_swp_desc *d)
 {
        struct qbman_swp *p = kzalloc(sizeof(*p), GFP_KERNEL);
        u32 reg;
+       u32 mask_size;
+       u32 eqcr_pi;
+
+       spin_lock_init(&p->access_spinlock);
 
        if (!p)
                return NULL;
@@ -255,25 +278,38 @@ struct qbman_swp *qbman_swp_init(const struct qbman_swp_desc *d)
        p->addr_cena = d->cena_bar;
        p->addr_cinh = d->cinh_bar;
 
-       if ((p->desc->qman_version & QMAN_REV_MASK) >= QMAN_REV_5000)
-               memset(p->addr_cena, 0, 64 * 1024);
+       if ((p->desc->qman_version & QMAN_REV_MASK) < QMAN_REV_5000) {
 
-       reg = qbman_set_swp_cfg(p->dqrr.dqrr_size,
-                               1, /* Writes Non-cacheable */
-                               0, /* EQCR_CI stashing threshold */
-                               3, /* RPM: Valid bit mode, RCR in array mode */
-                               2, /* DCM: Discrete consumption ack mode */
-                               3, /* EPM: Valid bit mode, EQCR in array mode */
-                               1, /* mem stashing drop enable == TRUE */
-                               1, /* mem stashing priority == TRUE */
-                               1, /* mem stashing enable == TRUE */
-                               1, /* dequeue stashing priority == TRUE */
-                               0, /* dequeue stashing enable == FALSE */
-                               0); /* EQCR_CI stashing priority == FALSE */
-       if ((p->desc->qman_version & QMAN_REV_MASK) >= QMAN_REV_5000)
+               reg = qbman_set_swp_cfg(p->dqrr.dqrr_size,
+                       1, /* Writes Non-cacheable */
+                       0, /* EQCR_CI stashing threshold */
+                       3, /* RPM: RCR in array mode */
+                       2, /* DCM: Discrete consumption ack */
+                       2, /* EPM: EQCR in ring mode */
+                       1, /* mem stashing drop enable enable */
+                       1, /* mem stashing priority enable */
+                       1, /* mem stashing enable */
+                       1, /* dequeue stashing priority enable */
+                       0, /* dequeue stashing enable enable */
+                       0); /* EQCR_CI stashing priority enable */
+       } else {
+               memset(p->addr_cena, 0, 64 * 1024);
+               reg = qbman_set_swp_cfg(p->dqrr.dqrr_size,
+                       1, /* Writes Non-cacheable */
+                       1, /* EQCR_CI stashing threshold */
+                       3, /* RPM: RCR in array mode */
+                       2, /* DCM: Discrete consumption ack */
+                       0, /* EPM: EQCR in ring mode */
+                       1, /* mem stashing drop enable */
+                       1, /* mem stashing priority enable */
+                       1, /* mem stashing enable */
+                       1, /* dequeue stashing priority enable */
+                       0, /* dequeue stashing enable */
+                       0); /* EQCR_CI stashing priority enable */
                reg |= 1 << SWP_CFG_CPBS_SHIFT | /* memory-backed mode */
                       1 << SWP_CFG_VPM_SHIFT |  /* VDQCR read triggered mode */
                       1 << SWP_CFG_CPM_SHIFT;   /* CR read triggered mode */
+       }
 
        qbman_write_register(p, QBMAN_CINH_SWP_CFG, reg);
        reg = qbman_read_register(p, QBMAN_CINH_SWP_CFG);
@@ -295,7 +331,9 @@ struct qbman_swp *qbman_swp_init(const struct qbman_swp_desc *d)
         */
        qbman_write_register(p, QBMAN_CINH_SWP_SDQCR, 0);
 
+       p->eqcr.pi_ring_size = 8;
        if ((p->desc->qman_version & QMAN_REV_MASK) >= QMAN_REV_5000) {
+               p->eqcr.pi_ring_size = 32;
                qbman_swp_enqueue_ptr =
                        qbman_swp_enqueue_mem_back;
                qbman_swp_enqueue_multiple_ptr =
@@ -307,6 +345,15 @@ struct qbman_swp *qbman_swp_init(const struct qbman_swp_desc *d)
                qbman_swp_release_ptr = qbman_swp_release_mem_back;
        }
 
+       for (mask_size = p->eqcr.pi_ring_size; mask_size > 0; mask_size >>= 1)
+               p->eqcr.pi_ci_mask = (p->eqcr.pi_ci_mask << 1) + 1;
+       eqcr_pi = qbman_read_register(p, QBMAN_CINH_SWP_EQCR_PI);
+       p->eqcr.pi = eqcr_pi & p->eqcr.pi_ci_mask;
+       p->eqcr.pi_vb = eqcr_pi & QB_VALID_BIT;
+       p->eqcr.ci = qbman_read_register(p, QBMAN_CINH_SWP_EQCR_CI)
+                       & p->eqcr.pi_ci_mask;
+       p->eqcr.available = p->eqcr.pi_ring_size;
+
        return p;
 }
 
@@ -460,6 +507,7 @@ enum qb_enqueue_commands {
 #define QB_ENQUEUE_CMD_ORP_ENABLE_SHIFT      2
 #define QB_ENQUEUE_CMD_IRQ_ON_DISPATCH_SHIFT 3
 #define QB_ENQUEUE_CMD_TARGET_TYPE_SHIFT     4
+#define QB_ENQUEUE_CMD_DCA_EN_SHIFT          7
 
 /**
  * qbman_eq_desc_clear() - Clear the contents of a descriptor to
@@ -535,6 +583,7 @@ static inline void qbman_write_eqcr_am_rt_register(struct qbman_swp *p,
                                     QMAN_RT_MODE);
 }
 
+#define QB_RT_BIT ((u32)0x100)
 /**
  * qbman_swp_enqueue_direct() - Issue an enqueue command
  * @s:  the software portal used for enqueue
@@ -546,34 +595,19 @@ static inline void qbman_write_eqcr_am_rt_register(struct qbman_swp *p,
  *
  * Return 0 for successful enqueue, -EBUSY if the EQCR is not ready.
  */
-int qbman_swp_enqueue_direct(struct qbman_swp *s, const struct qbman_eq_desc *d,
-                     const struct dpaa2_fd *fd)
+static
+int qbman_swp_enqueue_direct(struct qbman_swp *s,
+                            const struct qbman_eq_desc *d,
+                            const struct dpaa2_fd *fd)
 {
-       struct qbman_eq_desc_with_fd *p;
-       u32 eqar = qbman_read_register(s, QBMAN_CINH_SWP_EQAR);
-
-       if (!EQAR_SUCCESS(eqar))
-               return -EBUSY;
+       int flags = 0;
+       int ret = qbman_swp_enqueue_multiple_direct(s, d, fd, &flags, 1);
 
-       p = qbman_get_cmd(s, QBMAN_CENA_SWP_EQCR(EQAR_IDX(eqar)));
-       /* This is mapped as DEVICE type memory, writes are
-        * with address alignment:
-        * desc.dca address alignment = 1
-        * desc.seqnum address alignment = 2
-        * desc.orpid address alignment = 4
-        * desc.tgtid address alignment = 8
-        */
-       p->desc.dca = d->dca;
-       p->desc.seqnum = d->seqnum;
-       p->desc.orpid = d->orpid;
-       memcpy(&p->desc.tgtid, &d->tgtid, 24);
-       memcpy(&p->fd, fd, sizeof(*fd));
-
-       /* Set the verb byte, have to substitute in the valid-bit */
-       dma_wmb();
-       p->desc.verb = d->verb | EQAR_VB(eqar);
-
-       return 0;
+       if (ret >= 0)
+               ret = 0;
+       else
+               ret = -EBUSY;
+       return  ret;
 }
 
 /**
@@ -587,35 +621,19 @@ int qbman_swp_enqueue_direct(struct qbman_swp *s, const struct qbman_eq_desc *d,
  *
  * Return 0 for successful enqueue, -EBUSY if the EQCR is not ready.
  */
+static
 int qbman_swp_enqueue_mem_back(struct qbman_swp *s,
                               const struct qbman_eq_desc *d,
                               const struct dpaa2_fd *fd)
 {
-       struct qbman_eq_desc_with_fd *p;
-       u32 eqar = qbman_read_register(s, QBMAN_CINH_SWP_EQAR);
-
-       if (!EQAR_SUCCESS(eqar))
-               return -EBUSY;
+       int flags = 0;
+       int ret = qbman_swp_enqueue_multiple_mem_back(s, d, fd, &flags, 1);
 
-       p = qbman_get_cmd(s, QBMAN_CENA_SWP_EQCR(EQAR_IDX(eqar)));
-       /* This is mapped as DEVICE type memory, writes are
-        * with address alignment:
-        * desc.dca address alignment = 1
-        * desc.seqnum address alignment = 2
-        * desc.orpid address alignment = 4
-        * desc.tgtid address alignment = 8
-        */
-       p->desc.dca = d->dca;
-       p->desc.seqnum = d->seqnum;
-       p->desc.orpid = d->orpid;
-       memcpy(&p->desc.tgtid, &d->tgtid, 24);
-       memcpy(&p->fd, fd, sizeof(*fd));
-
-       p->desc.verb = d->verb | EQAR_VB(eqar);
-       dma_wmb();
-       qbman_write_eqcr_am_rt_register(s, EQAR_IDX(eqar));
-
-       return 0;
+       if (ret >= 0)
+               ret = 0;
+       else
+               ret = -EBUSY;
+       return  ret;
 }
 
 /**
@@ -624,26 +642,82 @@ int qbman_swp_enqueue_mem_back(struct qbman_swp *s,
  * @s:  the software portal used for enqueue
  * @d:  the enqueue descriptor
  * @fd: table pointer of frame descriptor table to be enqueued
- * @flags: table pointer of flags, not used for the moment
+ * @flags: table pointer of QBMAN_ENQUEUE_FLAG_DCA flags, not used if NULL
  * @num_frames: number of fd to be enqueued
  *
  * Return the number of fd enqueued, or a negative error number.
  */
+static
 int qbman_swp_enqueue_multiple_direct(struct qbman_swp *s,
                                      const struct qbman_eq_desc *d,
                                      const struct dpaa2_fd *fd,
                                      uint32_t *flags,
                                      int num_frames)
 {
-       int count = 0;
+       uint32_t *p = NULL;
+       const uint32_t *cl = (uint32_t *)d;
+       uint32_t eqcr_ci, eqcr_pi, half_mask, full_mask;
+       int i, num_enqueued = 0;
+       uint64_t addr_cena;
+
+       spin_lock(&s->access_spinlock);
+       half_mask = (s->eqcr.pi_ci_mask>>1);
+       full_mask = s->eqcr.pi_ci_mask;
+
+       if (!s->eqcr.available) {
+               eqcr_ci = s->eqcr.ci;
+               p = s->addr_cena + QBMAN_CENA_SWP_EQCR_CI;
+               s->eqcr.ci = qbman_read_register(s, QBMAN_CINH_SWP_EQCR_CI);
+
+               s->eqcr.available = qm_cyc_diff(s->eqcr.pi_ring_size,
+                                       eqcr_ci, s->eqcr.ci);
+               if (!s->eqcr.available) {
+                       spin_unlock(&s->access_spinlock);
+                       return 0;
+               }
+       }
 
-       while (count < num_frames) {
-               if (qbman_swp_enqueue_direct(s, d, fd) != 0)
-                       break;
-               count++;
+       eqcr_pi = s->eqcr.pi;
+       num_enqueued = (s->eqcr.available < num_frames) ?
+                       s->eqcr.available : num_frames;
+       s->eqcr.available -= num_enqueued;
+       /* Fill in the EQCR ring */
+       for (i = 0; i < num_enqueued; i++) {
+               p = (s->addr_cena + QBMAN_CENA_SWP_EQCR(eqcr_pi & half_mask));
+               /* Skip copying the verb */
+               memcpy(&p[1], &cl[1], EQ_DESC_SIZE_WITHOUT_FD - 1);
+               memcpy(&p[EQ_DESC_SIZE_FD_START/sizeof(uint32_t)],
+                      &fd[i], sizeof(*fd));
+               eqcr_pi++;
        }
 
-       return count;
+       dma_wmb();
+
+       /* Set the verb byte, have to substitute in the valid-bit */
+       eqcr_pi = s->eqcr.pi;
+       for (i = 0; i < num_enqueued; i++) {
+               p = (s->addr_cena + QBMAN_CENA_SWP_EQCR(eqcr_pi & half_mask));
+               p[0] = cl[0] | s->eqcr.pi_vb;
+               if (flags && (flags[i] & QBMAN_ENQUEUE_FLAG_DCA)) {
+                       struct qbman_eq_desc *d = (struct qbman_eq_desc *)p;
+
+                       d->dca = (1 << QB_ENQUEUE_CMD_DCA_EN_SHIFT) |
+                               ((flags[i]) & QBMAN_EQCR_DCA_IDXMASK);
+               }
+               eqcr_pi++;
+               if (!(eqcr_pi & half_mask))
+                       s->eqcr.pi_vb ^= QB_VALID_BIT;
+       }
+
+       /* Flush all the cacheline without load/store in between */
+       eqcr_pi = s->eqcr.pi;
+       addr_cena = (size_t)s->addr_cena;
+       for (i = 0; i < num_enqueued; i++)
+               eqcr_pi++;
+       s->eqcr.pi = eqcr_pi & full_mask;
+       spin_unlock(&s->access_spinlock);
+
+       return num_enqueued;
 }
 
 /**
@@ -652,26 +726,80 @@ int qbman_swp_enqueue_multiple_direct(struct qbman_swp *s,
  * @s:  the software portal used for enqueue
  * @d:  the enqueue descriptor
  * @fd: table pointer of frame descriptor table to be enqueued
- * @flags: table pointer of flags, not used for the moment
+ * @flags: table pointer of QBMAN_ENQUEUE_FLAG_DCA flags, not used if NULL
  * @num_frames: number of fd to be enqueued
  *
  * Return the number of fd enqueued, or a negative error number.
  */
+static
 int qbman_swp_enqueue_multiple_mem_back(struct qbman_swp *s,
-                                     const struct qbman_eq_desc *d,
-                                     const struct dpaa2_fd *fd,
-                                     uint32_t *flags,
-                                     int num_frames)
+                                       const struct qbman_eq_desc *d,
+                                       const struct dpaa2_fd *fd,
+                                       uint32_t *flags,
+                                       int num_frames)
 {
-       int count = 0;
+       uint32_t *p = NULL;
+       const uint32_t *cl = (uint32_t *)(d);
+       uint32_t eqcr_ci, eqcr_pi, half_mask, full_mask;
+       int i, num_enqueued = 0;
+       unsigned long irq_flags;
+
+       spin_lock(&s->access_spinlock);
+       local_irq_save(irq_flags);
+
+       half_mask = (s->eqcr.pi_ci_mask>>1);
+       full_mask = s->eqcr.pi_ci_mask;
+       if (!s->eqcr.available) {
+               eqcr_ci = s->eqcr.ci;
+               p = s->addr_cena + QBMAN_CENA_SWP_EQCR_CI_MEMBACK;
+               s->eqcr.ci = __raw_readl(p) & full_mask;
+               s->eqcr.available = qm_cyc_diff(s->eqcr.pi_ring_size,
+                                       eqcr_ci, s->eqcr.ci);
+               if (!s->eqcr.available) {
+                       local_irq_restore(irq_flags);
+                       spin_unlock(&s->access_spinlock);
+                       return 0;
+               }
+       }
+
+       eqcr_pi = s->eqcr.pi;
+       num_enqueued = (s->eqcr.available < num_frames) ?
+                       s->eqcr.available : num_frames;
+       s->eqcr.available -= num_enqueued;
+       /* Fill in the EQCR ring */
+       for (i = 0; i < num_enqueued; i++) {
+               p = (s->addr_cena + QBMAN_CENA_SWP_EQCR(eqcr_pi & half_mask));
+               /* Skip copying the verb */
+               memcpy(&p[1], &cl[1], EQ_DESC_SIZE_WITHOUT_FD - 1);
+               memcpy(&p[EQ_DESC_SIZE_FD_START/sizeof(uint32_t)],
+                      &fd[i], sizeof(*fd));
+               eqcr_pi++;
+       }
 
-       while (count < num_frames) {
-               if (qbman_swp_enqueue_mem_back(s, d, fd) != 0)
-                       break;
-               count++;
+       /* Set the verb byte, have to substitute in the valid-bit */
+       eqcr_pi = s->eqcr.pi;
+       for (i = 0; i < num_enqueued; i++) {
+               p = (s->addr_cena + QBMAN_CENA_SWP_EQCR(eqcr_pi & half_mask));
+               p[0] = cl[0] | s->eqcr.pi_vb;
+               if (flags && (flags[i] & QBMAN_ENQUEUE_FLAG_DCA)) {
+                       struct qbman_eq_desc *d = (struct qbman_eq_desc *)p;
+
+                       d->dca = (1 << QB_ENQUEUE_CMD_DCA_EN_SHIFT) |
+                               ((flags[i]) & QBMAN_EQCR_DCA_IDXMASK);
+               }
+               eqcr_pi++;
+               if (!(eqcr_pi & half_mask))
+                       s->eqcr.pi_vb ^= QB_VALID_BIT;
        }
+       s->eqcr.pi = eqcr_pi & full_mask;
 
-       return count;
+       dma_wmb();
+       qbman_write_register(s, QBMAN_CINH_SWP_EQCR_PI,
+                               (QB_RT_BIT)|(s->eqcr.pi)|s->eqcr.pi_vb);
+       local_irq_restore(irq_flags);
+       spin_unlock(&s->access_spinlock);
+
+       return num_enqueued;
 }
 
 /**
@@ -684,20 +812,66 @@ int qbman_swp_enqueue_multiple_mem_back(struct qbman_swp *s,
  *
  * Return the number of fd enqueued, or a negative error number.
  */
+static
 int qbman_swp_enqueue_multiple_desc_direct(struct qbman_swp *s,
                                           const struct qbman_eq_desc *d,
                                           const struct dpaa2_fd *fd,
                                           int num_frames)
 {
-       int count = 0;
+       uint32_t *p;
+       const uint32_t *cl;
+       uint32_t eqcr_ci, eqcr_pi, half_mask, full_mask;
+       int i, num_enqueued = 0;
+       uint64_t addr_cena;
+
+       half_mask = (s->eqcr.pi_ci_mask>>1);
+       full_mask = s->eqcr.pi_ci_mask;
+       if (!s->eqcr.available) {
+               eqcr_ci = s->eqcr.ci;
+               p = s->addr_cena + QBMAN_CENA_SWP_EQCR_CI;
+               s->eqcr.ci = qbman_read_register(s, QBMAN_CINH_SWP_EQCR_CI);
+               s->eqcr.available = qm_cyc_diff(s->eqcr.pi_ring_size,
+                                       eqcr_ci, s->eqcr.ci);
+               if (!s->eqcr.available)
+                       return 0;
+       }
+
+       eqcr_pi = s->eqcr.pi;
+       num_enqueued = (s->eqcr.available < num_frames) ?
+                       s->eqcr.available : num_frames;
+       s->eqcr.available -= num_enqueued;
+       /* Fill in the EQCR ring */
+       for (i = 0; i < num_enqueued; i++) {
+               p = (s->addr_cena + QBMAN_CENA_SWP_EQCR(eqcr_pi & half_mask));
+               cl = (uint32_t *)(&d[i]);
+               /* Skip copying the verb */
+               memcpy(&p[1], &cl[1], EQ_DESC_SIZE_WITHOUT_FD - 1);
+               memcpy(&p[EQ_DESC_SIZE_FD_START/sizeof(uint32_t)],
+                      &fd[i], sizeof(*fd));
+               eqcr_pi++;
+       }
+
+       dma_wmb();
 
-       while (count < num_frames) {
-               if (qbman_swp_enqueue_direct(s, &(d[count]), fd) != 0)
-                       break;
-               count++;
+       /* Set the verb byte, have to substitute in the valid-bit */
+       eqcr_pi = s->eqcr.pi;
+       for (i = 0; i < num_enqueued; i++) {
+               p = (s->addr_cena + QBMAN_CENA_SWP_EQCR(eqcr_pi & half_mask));
+               cl = (uint32_t *)(&d[i]);
+               p[0] = cl[0] | s->eqcr.pi_vb;
+               eqcr_pi++;
+               if (!(eqcr_pi & half_mask))
+                       s->eqcr.pi_vb ^= QB_VALID_BIT;
        }
 
-       return count;
+       /* Flush all the cacheline without load/store in between */
+       eqcr_pi = s->eqcr.pi;
+       addr_cena = (uint64_t)s->addr_cena;
+       for (i = 0; i < num_enqueued; i++)
+               eqcr_pi++;
+       s->eqcr.pi = eqcr_pi & full_mask;
+
+       return num_enqueued;
 }
 
 /**
@@ -710,20 +884,62 @@ int qbman_swp_enqueue_multiple_desc_direct(struct qbman_swp *s,
  *
  * Return the number of fd enqueued, or a negative error number.
  */
+static
 int qbman_swp_enqueue_multiple_desc_mem_back(struct qbman_swp *s,
                                           const struct qbman_eq_desc *d,
                                           const struct dpaa2_fd *fd,
                                           int num_frames)
 {
-       int count = 0;
+       uint32_t *p;
+       const uint32_t *cl;
+       uint32_t eqcr_ci, eqcr_pi, half_mask, full_mask;
+       int i, num_enqueued = 0;
+
+       half_mask = (s->eqcr.pi_ci_mask>>1);
+       full_mask = s->eqcr.pi_ci_mask;
+       if (!s->eqcr.available) {
+               eqcr_ci = s->eqcr.ci;
+               p = s->addr_cena + QBMAN_CENA_SWP_EQCR_CI_MEMBACK;
+               s->eqcr.ci = __raw_readl(p) & full_mask;
+               s->eqcr.available = qm_cyc_diff(s->eqcr.pi_ring_size,
+                                       eqcr_ci, s->eqcr.ci);
+               if (!s->eqcr.available)
+                       return 0;
+       }
 
-       while (count < num_frames) {
-               if (qbman_swp_enqueue_mem_back(s, &(d[count]), fd) != 0)
-                       break;
-               count++;
+       eqcr_pi = s->eqcr.pi;
+       num_enqueued = (s->eqcr.available < num_frames) ?
+                       s->eqcr.available : num_frames;
+       s->eqcr.available -= num_enqueued;
+       /* Fill in the EQCR ring */
+       for (i = 0; i < num_enqueued; i++) {
+               p = (s->addr_cena + QBMAN_CENA_SWP_EQCR(eqcr_pi & half_mask));
+               cl = (uint32_t *)(&d[i]);
+               /* Skip copying the verb */
+               memcpy(&p[1], &cl[1], EQ_DESC_SIZE_WITHOUT_FD - 1);
+               memcpy(&p[EQ_DESC_SIZE_FD_START/sizeof(uint32_t)],
+                      &fd[i], sizeof(*fd));
+               eqcr_pi++;
        }
 
-       return count;
+       /* Set the verb byte, have to substitute in the valid-bit */
+       eqcr_pi = s->eqcr.pi;
+       for (i = 0; i < num_enqueued; i++) {
+               p = (s->addr_cena + QBMAN_CENA_SWP_EQCR(eqcr_pi & half_mask));
+               cl = (uint32_t *)(&d[i]);
+               p[0] = cl[0] | s->eqcr.pi_vb;
+               eqcr_pi++;
+               if (!(eqcr_pi & half_mask))
+                       s->eqcr.pi_vb ^= QB_VALID_BIT;
+       }
+
+       s->eqcr.pi = eqcr_pi & full_mask;
+
+       dma_wmb();
+       qbman_write_register(s, QBMAN_CINH_SWP_EQCR_PI,
+                               (QB_RT_BIT)|(s->eqcr.pi)|s->eqcr.pi_vb);
+
+       return num_enqueued;
 }
 
 /* Static (push) dequeue */
@@ -889,6 +1105,7 @@ void qbman_pull_desc_set_channel(struct qbman_pull_desc *d, u32 chid,
  * Return 0 for success, and -EBUSY if the software portal is not ready
  * to do pull dequeue.
  */
+static
 int qbman_swp_pull_direct(struct qbman_swp *s, struct qbman_pull_desc *d)
 {
        struct qbman_pull_desc *p;
@@ -924,6 +1141,7 @@ int qbman_swp_pull_direct(struct qbman_swp *s, struct qbman_pull_desc *d)
  * Return 0 for success, and -EBUSY if the software portal is not ready
  * to do pull dequeue.
  */
+static
 int qbman_swp_pull_mem_back(struct qbman_swp *s, struct qbman_pull_desc *d)
 {
        struct qbman_pull_desc *p;
@@ -942,6 +1160,8 @@ int qbman_swp_pull_mem_back(struct qbman_swp *s, struct qbman_pull_desc *d)
        p->dq_src = d->dq_src;
        p->rsp_addr = d->rsp_addr;
        p->rsp_addr_virt = d->rsp_addr_virt;
+
+       /* Set the verb byte, have to substitute in the valid-bit */
        p->verb = d->verb | s->vdq.valid_bit;
        s->vdq.valid_bit ^= QB_VALID_BIT;
        dma_wmb();
index 3b3fb86d180bc29aef84dbd0ee5b4a877597d1bd..c7c2225b7d914de953a2c9956ce7e68988a534a0 100644 (file)
@@ -143,6 +143,19 @@ struct qbman_swp {
                u8 dqrr_size;
                int reset_bug; /* indicates dqrr reset workaround is needed */
        } dqrr;
+
+       struct {
+               u32 pi;
+               u32 pi_vb;
+               u32 pi_ring_size;
+               u32 pi_ci_mask;
+               u32 ci;
+               int available;
+               u32 pend;
+               u32 no_pfdr;
+       } eqcr;
+
+       spinlock_t access_spinlock;
 };
 
 /* Function pointers */