mirror of
https://github.com/intel/isa-l.git
synced 2024-12-13 09:52:56 +01:00
180c74aefd
1) Implement the ErasureCode function in Altivec Intrinsics 2) Coding style update Change-Id: I2c81d035f4083e9b011dbf3b741f628813b68606 Thanks-to: Daniel Axtens <dja@axtens.net> Signed-off-by: Hong Bo Peng <penghb@cn.ibm.com>
84 lines
2.1 KiB
C
84 lines
2.1 KiB
C
#include "ec_base_vsx.h"
|
|
|
|
void gf_2vect_dot_prod_vsx(int len, int vlen, unsigned char *gftbls,
|
|
unsigned char **src, unsigned char **dest)
|
|
{
|
|
unsigned char *s, *t0, *t1;
|
|
vector unsigned char vX1, vX2, vX3, vX4;
|
|
vector unsigned char vY1, vY2, vY3, vY4;
|
|
vector unsigned char vYD, vYE, vYF, vYG;
|
|
vector unsigned char vhi0, vlo0, vhi1, vlo1;
|
|
int i, j, head;
|
|
|
|
if (vlen < 128) {
|
|
gf_vect_mul_vsx(len, &gftbls[0 * 32 * vlen], src[0], (unsigned char *)dest[0]);
|
|
gf_vect_mul_vsx(len, &gftbls[1 * 32 * vlen], src[0], (unsigned char *)dest[1]);
|
|
|
|
for (j = 1; j < vlen; j++) {
|
|
gf_2vect_mad_vsx(len, vlen, j, gftbls, src[j], dest);
|
|
}
|
|
return;
|
|
}
|
|
|
|
t0 = (unsigned char *)dest[0];
|
|
t1 = (unsigned char *)dest[1];
|
|
|
|
head = len % 64;
|
|
if (head != 0) {
|
|
gf_vect_dot_prod_base(head, vlen, &gftbls[0 * 32 * vlen], src, t0);
|
|
gf_vect_dot_prod_base(head, vlen, &gftbls[1 * 32 * vlen], src, t1);
|
|
}
|
|
|
|
for (i = head; i < len - 63; i += 64) {
|
|
vY1 = vY1 ^ vY1;
|
|
vY2 = vY2 ^ vY2;
|
|
vY3 = vY3 ^ vY3;
|
|
vY4 = vY4 ^ vY4;
|
|
|
|
vYD = vYD ^ vYD;
|
|
vYE = vYE ^ vYE;
|
|
vYF = vYF ^ vYF;
|
|
vYG = vYG ^ vYG;
|
|
|
|
unsigned char *g0 = &gftbls[0 * 32 * vlen];
|
|
unsigned char *g1 = &gftbls[1 * 32 * vlen];
|
|
|
|
for (j = 0; j < vlen; j++) {
|
|
s = (unsigned char *)src[j];
|
|
vX1 = vec_xl(0, s + i);
|
|
vX2 = vec_xl(16, s + i);
|
|
vX3 = vec_xl(32, s + i);
|
|
vX4 = vec_xl(48, s + i);
|
|
|
|
vlo0 = EC_vec_xl(0, g0);
|
|
vhi0 = EC_vec_xl(16, g0);
|
|
vlo1 = EC_vec_xl(0, g1);
|
|
vhi1 = EC_vec_xl(16, g1);
|
|
|
|
vY1 = vY1 ^ EC_vec_permxor(vhi0, vlo0, vX1);
|
|
vY2 = vY2 ^ EC_vec_permxor(vhi0, vlo0, vX2);
|
|
vYD = vYD ^ EC_vec_permxor(vhi0, vlo0, vX3);
|
|
vYE = vYE ^ EC_vec_permxor(vhi0, vlo0, vX4);
|
|
|
|
vY3 = vY3 ^ EC_vec_permxor(vhi1, vlo1, vX1);
|
|
vY4 = vY4 ^ EC_vec_permxor(vhi1, vlo1, vX2);
|
|
vYF = vYF ^ EC_vec_permxor(vhi1, vlo1, vX3);
|
|
vYG = vYG ^ EC_vec_permxor(vhi1, vlo1, vX4);
|
|
|
|
g0 += 32;
|
|
g1 += 32;
|
|
}
|
|
|
|
vec_xst(vY1, 0, t0 + i);
|
|
vec_xst(vY2, 16, t0 + i);
|
|
vec_xst(vY3, 0, t1 + i);
|
|
vec_xst(vY4, 16, t1 + i);
|
|
|
|
vec_xst(vYD, 32, t0 + i);
|
|
vec_xst(vYE, 48, t0 + i);
|
|
vec_xst(vYF, 32, t1 + i);
|
|
vec_xst(vYG, 48, t1 + i);
|
|
}
|
|
return;
|
|
}
|