]> git.proxmox.com Git - ceph.git/blobdiff - ceph/src/isa-l/erasure_code/ppc64le/gf_4vect_dot_prod_vsx.c
Import ceph 15.2.8
[ceph.git] / ceph / src / isa-l / erasure_code / ppc64le / gf_4vect_dot_prod_vsx.c
diff --git a/ceph/src/isa-l/erasure_code/ppc64le/gf_4vect_dot_prod_vsx.c b/ceph/src/isa-l/erasure_code/ppc64le/gf_4vect_dot_prod_vsx.c
new file mode 100644 (file)
index 0000000..e656544
--- /dev/null
@@ -0,0 +1,124 @@
+#include "ec_base_vsx.h"
+
+void gf_4vect_dot_prod_vsx(int len, int vlen, unsigned char *gftbls,
+                          unsigned char **src, unsigned char **dest)
+{
+       unsigned char *s, *t0, *t1, *t2, *t3;
+       vector unsigned char vX1, vX2, vX3, vX4;
+       vector unsigned char vY1, vY2, vY3, vY4, vY5, vY6, vY7, vY8;
+       vector unsigned char vYD, vYE, vYF, vYG, vYH, vYI, vYJ, vYK;
+       vector unsigned char vhi0, vlo0, vhi1, vlo1, vhi2, vlo2, vhi3, vlo3;
+       int i, j, head;
+
+       if (vlen < 128) {
+               gf_vect_mul_vsx(len, &gftbls[0 * 32 * vlen], src[0], (unsigned char *)dest[0]);
+               gf_vect_mul_vsx(len, &gftbls[1 * 32 * vlen], src[0], (unsigned char *)dest[1]);
+               gf_vect_mul_vsx(len, &gftbls[2 * 32 * vlen], src[0], (unsigned char *)dest[2]);
+               gf_vect_mul_vsx(len, &gftbls[3 * 32 * vlen], src[0], (unsigned char *)dest[3]);
+
+               for (j = 1; j < vlen; j++) {
+                       gf_4vect_mad_vsx(len, vlen, j, gftbls, src[j], dest);
+               }
+               return;
+       }
+
+       t0 = (unsigned char *)dest[0];
+       t1 = (unsigned char *)dest[1];
+       t2 = (unsigned char *)dest[2];
+       t3 = (unsigned char *)dest[3];
+
+       head = len % 64;
+       if (head != 0) {
+               gf_vect_dot_prod_base(head, vlen, &gftbls[0 * 32 * vlen], src, t0);
+               gf_vect_dot_prod_base(head, vlen, &gftbls[1 * 32 * vlen], src, t1);
+               gf_vect_dot_prod_base(head, vlen, &gftbls[2 * 32 * vlen], src, t2);
+               gf_vect_dot_prod_base(head, vlen, &gftbls[3 * 32 * vlen], src, t3);
+       }
+
+       for (i = head; i < len - 63; i += 64) {
+               vY1 = vY1 ^ vY1;
+               vY2 = vY2 ^ vY2;
+               vY3 = vY3 ^ vY3;
+               vY4 = vY4 ^ vY4;
+               vY5 = vY5 ^ vY5;
+               vY6 = vY6 ^ vY6;
+               vY7 = vY7 ^ vY7;
+               vY8 = vY8 ^ vY8;
+
+               vYD = vYD ^ vYD;
+               vYE = vYE ^ vYE;
+               vYF = vYF ^ vYF;
+               vYG = vYG ^ vYG;
+               vYH = vYH ^ vYH;
+               vYI = vYI ^ vYI;
+               vYJ = vYJ ^ vYJ;
+               vYK = vYK ^ vYK;
+
+               unsigned char *g0 = &gftbls[0 * 32 * vlen];
+               unsigned char *g1 = &gftbls[1 * 32 * vlen];
+               unsigned char *g2 = &gftbls[2 * 32 * vlen];
+               unsigned char *g3 = &gftbls[3 * 32 * vlen];
+
+               for (j = 0; j < vlen; j++) {
+                       s = (unsigned char *)src[j];
+                       vX1 = vec_xl(0, s + i);
+                       vX2 = vec_xl(16, s + i);
+                       vX3 = vec_xl(32, s + i);
+                       vX4 = vec_xl(48, s + i);
+
+                       vlo0 = EC_vec_xl(0, g0);
+                       vhi0 = EC_vec_xl(16, g0);
+                       vlo1 = EC_vec_xl(0, g1);
+                       vhi1 = EC_vec_xl(16, g1);
+
+                       vY1 = vY1 ^ EC_vec_permxor(vhi0, vlo0, vX1);
+                       vY2 = vY2 ^ EC_vec_permxor(vhi0, vlo0, vX2);
+                       vYD = vYD ^ EC_vec_permxor(vhi0, vlo0, vX3);
+                       vYE = vYE ^ EC_vec_permxor(vhi0, vlo0, vX4);
+
+                       vlo2 = vec_xl(0, g2);
+                       vhi2 = vec_xl(16, g2);
+                       vlo3 = vec_xl(0, g3);
+                       vhi3 = vec_xl(16, g3);
+
+                       vY3 = vY3 ^ EC_vec_permxor(vhi1, vlo1, vX1);
+                       vY4 = vY4 ^ EC_vec_permxor(vhi1, vlo1, vX2);
+                       vYF = vYF ^ EC_vec_permxor(vhi1, vlo1, vX3);
+                       vYG = vYG ^ EC_vec_permxor(vhi1, vlo1, vX4);
+
+                       vY5 = vY5 ^ EC_vec_permxor(vhi2, vlo2, vX1);
+                       vY6 = vY6 ^ EC_vec_permxor(vhi2, vlo2, vX2);
+                       vYH = vYH ^ EC_vec_permxor(vhi2, vlo2, vX3);
+                       vYI = vYI ^ EC_vec_permxor(vhi2, vlo2, vX4);
+
+                       vY7 = vY7 ^ EC_vec_permxor(vhi3, vlo3, vX1);
+                       vY8 = vY8 ^ EC_vec_permxor(vhi3, vlo3, vX2);
+                       vYJ = vYJ ^ EC_vec_permxor(vhi3, vlo3, vX3);
+                       vYK = vYK ^ EC_vec_permxor(vhi3, vlo3, vX4);
+
+                       g0 += 32;
+                       g1 += 32;
+                       g2 += 32;
+                       g3 += 32;
+               }
+
+               vec_xst(vY1, 0, t0 + i);
+               vec_xst(vY2, 16, t0 + i);
+               vec_xst(vY3, 0, t1 + i);
+               vec_xst(vY4, 16, t1 + i);
+               vec_xst(vY5, 0, t2 + i);
+               vec_xst(vY6, 16, t2 + i);
+               vec_xst(vY7, 0, t3 + i);
+               vec_xst(vY8, 16, t3 + i);
+
+               vec_xst(vYD, 32, t0 + i);
+               vec_xst(vYE, 48, t0 + i);
+               vec_xst(vYF, 32, t1 + i);
+               vec_xst(vYG, 48, t1 + i);
+               vec_xst(vYH, 32, t2 + i);
+               vec_xst(vYI, 48, t2 + i);
+               vec_xst(vYJ, 32, t3 + i);
+               vec_xst(vYK, 48, t3 + i);
+       }
+       return;
+}