Merge "Fix warnings reported by -Wshadow: Part1: vpx_dsp directory"
This commit is contained in:
@@ -196,18 +196,18 @@
|
||||
out2, out3) \
|
||||
{ \
|
||||
v8i16 madd_s0_m, madd_s1_m, madd_s2_m, madd_s3_m; \
|
||||
v4i32 tmp0_m, tmp1_m, tmp2_m, tmp3_m; \
|
||||
v4i32 tmp0_madd, tmp1_madd, tmp2_madd, tmp3_madd; \
|
||||
\
|
||||
ILVRL_H2_SH(inp1, inp0, madd_s1_m, madd_s0_m); \
|
||||
ILVRL_H2_SH(inp3, inp2, madd_s3_m, madd_s2_m); \
|
||||
DOTP_SH4_SW(madd_s1_m, madd_s0_m, madd_s1_m, madd_s0_m, cst0, cst0, cst1, \
|
||||
cst1, tmp0_m, tmp1_m, tmp2_m, tmp3_m); \
|
||||
SRARI_W4_SW(tmp0_m, tmp1_m, tmp2_m, tmp3_m, DCT_CONST_BITS); \
|
||||
PCKEV_H2_SH(tmp1_m, tmp0_m, tmp3_m, tmp2_m, out0, out1); \
|
||||
cst1, tmp0_madd, tmp1_madd, tmp2_madd, tmp3_madd); \
|
||||
SRARI_W4_SW(tmp0_madd, tmp1_madd, tmp2_madd, tmp3_madd, DCT_CONST_BITS); \
|
||||
PCKEV_H2_SH(tmp1_madd, tmp0_madd, tmp3_madd, tmp2_madd, out0, out1); \
|
||||
DOTP_SH4_SW(madd_s3_m, madd_s2_m, madd_s3_m, madd_s2_m, cst2, cst2, cst3, \
|
||||
cst3, tmp0_m, tmp1_m, tmp2_m, tmp3_m); \
|
||||
SRARI_W4_SW(tmp0_m, tmp1_m, tmp2_m, tmp3_m, DCT_CONST_BITS); \
|
||||
PCKEV_H2_SH(tmp1_m, tmp0_m, tmp3_m, tmp2_m, out2, out3); \
|
||||
cst3, tmp0_madd, tmp1_madd, tmp2_madd, tmp3_madd); \
|
||||
SRARI_W4_SW(tmp0_madd, tmp1_madd, tmp2_madd, tmp3_madd, DCT_CONST_BITS); \
|
||||
PCKEV_H2_SH(tmp1_madd, tmp0_madd, tmp3_madd, tmp2_madd, out2, out3); \
|
||||
}
|
||||
|
||||
/* idct 8x8 macro */
|
||||
|
||||
@@ -49,35 +49,35 @@
|
||||
p1_out = __msa_xori_b((v16u8)p1_m, 0x80); \
|
||||
}
|
||||
|
||||
#define VP9_FLAT4(p3_in, p2_in, p0_in, q0_in, q2_in, q3_in, flat_out) \
|
||||
{ \
|
||||
v16u8 tmp, p2_a_sub_p0, q2_a_sub_q0, p3_a_sub_p0, q3_a_sub_q0; \
|
||||
v16u8 zero_in = { 0 }; \
|
||||
\
|
||||
tmp = __msa_ori_b(zero_in, 1); \
|
||||
p2_a_sub_p0 = __msa_asub_u_b(p2_in, p0_in); \
|
||||
q2_a_sub_q0 = __msa_asub_u_b(q2_in, q0_in); \
|
||||
p3_a_sub_p0 = __msa_asub_u_b(p3_in, p0_in); \
|
||||
q3_a_sub_q0 = __msa_asub_u_b(q3_in, q0_in); \
|
||||
\
|
||||
p2_a_sub_p0 = __msa_max_u_b(p2_a_sub_p0, q2_a_sub_q0); \
|
||||
flat_out = __msa_max_u_b(p2_a_sub_p0, flat_out); \
|
||||
p3_a_sub_p0 = __msa_max_u_b(p3_a_sub_p0, q3_a_sub_q0); \
|
||||
flat_out = __msa_max_u_b(p3_a_sub_p0, flat_out); \
|
||||
\
|
||||
flat_out = (tmp < (v16u8)flat_out); \
|
||||
flat_out = __msa_xori_b(flat_out, 0xff); \
|
||||
flat_out = flat_out & (mask); \
|
||||
#define VP9_FLAT4(p3_in, p2_in, p0_in, q0_in, q2_in, q3_in, flat_out) \
|
||||
{ \
|
||||
v16u8 tmp_flat4, p2_a_sub_p0, q2_a_sub_q0, p3_a_sub_p0, q3_a_sub_q0; \
|
||||
v16u8 zero_in = { 0 }; \
|
||||
\
|
||||
tmp_flat4 = __msa_ori_b(zero_in, 1); \
|
||||
p2_a_sub_p0 = __msa_asub_u_b(p2_in, p0_in); \
|
||||
q2_a_sub_q0 = __msa_asub_u_b(q2_in, q0_in); \
|
||||
p3_a_sub_p0 = __msa_asub_u_b(p3_in, p0_in); \
|
||||
q3_a_sub_q0 = __msa_asub_u_b(q3_in, q0_in); \
|
||||
\
|
||||
p2_a_sub_p0 = __msa_max_u_b(p2_a_sub_p0, q2_a_sub_q0); \
|
||||
flat_out = __msa_max_u_b(p2_a_sub_p0, flat_out); \
|
||||
p3_a_sub_p0 = __msa_max_u_b(p3_a_sub_p0, q3_a_sub_q0); \
|
||||
flat_out = __msa_max_u_b(p3_a_sub_p0, flat_out); \
|
||||
\
|
||||
flat_out = (tmp_flat4 < (v16u8)flat_out); \
|
||||
flat_out = __msa_xori_b(flat_out, 0xff); \
|
||||
flat_out = flat_out & (mask); \
|
||||
}
|
||||
|
||||
#define VP9_FLAT5(p7_in, p6_in, p5_in, p4_in, p0_in, q0_in, q4_in, q5_in, \
|
||||
q6_in, q7_in, flat_in, flat2_out) \
|
||||
{ \
|
||||
v16u8 tmp, zero_in = { 0 }; \
|
||||
v16u8 tmp_flat5, zero_in = { 0 }; \
|
||||
v16u8 p4_a_sub_p0, q4_a_sub_q0, p5_a_sub_p0, q5_a_sub_q0; \
|
||||
v16u8 p6_a_sub_p0, q6_a_sub_q0, p7_a_sub_p0, q7_a_sub_q0; \
|
||||
\
|
||||
tmp = __msa_ori_b(zero_in, 1); \
|
||||
tmp_flat5 = __msa_ori_b(zero_in, 1); \
|
||||
p4_a_sub_p0 = __msa_asub_u_b(p4_in, p0_in); \
|
||||
q4_a_sub_q0 = __msa_asub_u_b(q4_in, q0_in); \
|
||||
p5_a_sub_p0 = __msa_asub_u_b(p5_in, p0_in); \
|
||||
@@ -95,7 +95,7 @@
|
||||
p7_a_sub_p0 = __msa_max_u_b(p7_a_sub_p0, q7_a_sub_q0); \
|
||||
flat2_out = __msa_max_u_b(p7_a_sub_p0, flat2_out); \
|
||||
\
|
||||
flat2_out = (tmp < (v16u8)flat2_out); \
|
||||
flat2_out = (tmp_flat5 < (v16u8)flat2_out); \
|
||||
flat2_out = __msa_xori_b(flat2_out, 0xff); \
|
||||
flat2_out = flat2_out & flat_in; \
|
||||
}
|
||||
@@ -104,38 +104,38 @@
|
||||
p2_filt8_out, p1_filt8_out, p0_filt8_out, q0_filt8_out, \
|
||||
q1_filt8_out, q2_filt8_out) \
|
||||
{ \
|
||||
v8u16 tmp0, tmp1, tmp2; \
|
||||
v8u16 tmp_filt8_0, tmp_filt8_1, tmp_filt8_2; \
|
||||
\
|
||||
tmp2 = p2_in + p1_in + p0_in; \
|
||||
tmp0 = p3_in << 1; \
|
||||
tmp_filt8_2 = p2_in + p1_in + p0_in; \
|
||||
tmp_filt8_0 = p3_in << 1; \
|
||||
\
|
||||
tmp0 = tmp0 + tmp2 + q0_in; \
|
||||
tmp1 = tmp0 + p3_in + p2_in; \
|
||||
p2_filt8_out = (v8i16)__msa_srari_h((v8i16)tmp1, 3); \
|
||||
tmp_filt8_0 = tmp_filt8_0 + tmp_filt8_2 + q0_in; \
|
||||
tmp_filt8_1 = tmp_filt8_0 + p3_in + p2_in; \
|
||||
p2_filt8_out = (v8i16)__msa_srari_h((v8i16)tmp_filt8_1, 3); \
|
||||
\
|
||||
tmp1 = tmp0 + p1_in + q1_in; \
|
||||
p1_filt8_out = (v8i16)__msa_srari_h((v8i16)tmp1, 3); \
|
||||
tmp_filt8_1 = tmp_filt8_0 + p1_in + q1_in; \
|
||||
p1_filt8_out = (v8i16)__msa_srari_h((v8i16)tmp_filt8_1, 3); \
|
||||
\
|
||||
tmp1 = q2_in + q1_in + q0_in; \
|
||||
tmp2 = tmp2 + tmp1; \
|
||||
tmp0 = tmp2 + (p0_in); \
|
||||
tmp0 = tmp0 + (p3_in); \
|
||||
p0_filt8_out = (v8i16)__msa_srari_h((v8i16)tmp0, 3); \
|
||||
tmp_filt8_1 = q2_in + q1_in + q0_in; \
|
||||
tmp_filt8_2 = tmp_filt8_2 + tmp_filt8_1; \
|
||||
tmp_filt8_0 = tmp_filt8_2 + (p0_in); \
|
||||
tmp_filt8_0 = tmp_filt8_0 + (p3_in); \
|
||||
p0_filt8_out = (v8i16)__msa_srari_h((v8i16)tmp_filt8_0, 3); \
|
||||
\
|
||||
tmp0 = q2_in + q3_in; \
|
||||
tmp0 = p0_in + tmp1 + tmp0; \
|
||||
tmp1 = q3_in + q3_in; \
|
||||
tmp1 = tmp1 + tmp0; \
|
||||
q2_filt8_out = (v8i16)__msa_srari_h((v8i16)tmp1, 3); \
|
||||
tmp_filt8_0 = q2_in + q3_in; \
|
||||
tmp_filt8_0 = p0_in + tmp_filt8_1 + tmp_filt8_0; \
|
||||
tmp_filt8_1 = q3_in + q3_in; \
|
||||
tmp_filt8_1 = tmp_filt8_1 + tmp_filt8_0; \
|
||||
q2_filt8_out = (v8i16)__msa_srari_h((v8i16)tmp_filt8_1, 3); \
|
||||
\
|
||||
tmp0 = tmp2 + q3_in; \
|
||||
tmp1 = tmp0 + q0_in; \
|
||||
q0_filt8_out = (v8i16)__msa_srari_h((v8i16)tmp1, 3); \
|
||||
tmp_filt8_0 = tmp_filt8_2 + q3_in; \
|
||||
tmp_filt8_1 = tmp_filt8_0 + q0_in; \
|
||||
q0_filt8_out = (v8i16)__msa_srari_h((v8i16)tmp_filt8_1, 3); \
|
||||
\
|
||||
tmp1 = tmp0 - p2_in; \
|
||||
tmp0 = q1_in + q3_in; \
|
||||
tmp1 = tmp0 + tmp1; \
|
||||
q1_filt8_out = (v8i16)__msa_srari_h((v8i16)tmp1, 3); \
|
||||
tmp_filt8_1 = tmp_filt8_0 - p2_in; \
|
||||
tmp_filt8_0 = q1_in + q3_in; \
|
||||
tmp_filt8_1 = tmp_filt8_0 + tmp_filt8_1; \
|
||||
q1_filt8_out = (v8i16)__msa_srari_h((v8i16)tmp_filt8_1, 3); \
|
||||
}
|
||||
|
||||
#define LPF_MASK_HEV(p3_in, p2_in, p1_in, p0_in, q0_in, q1_in, q2_in, q3_in, \
|
||||
|
||||
@@ -168,20 +168,20 @@
|
||||
val_m; \
|
||||
})
|
||||
#else // !(__mips == 64)
|
||||
#define LD(psrc) \
|
||||
({ \
|
||||
const uint8_t *psrc_m1 = (const uint8_t *)(psrc); \
|
||||
uint32_t val0_m, val1_m; \
|
||||
uint64_t val_m = 0; \
|
||||
\
|
||||
val0_m = LW(psrc_m1); \
|
||||
val1_m = LW(psrc_m1 + 4); \
|
||||
\
|
||||
val_m = (uint64_t)(val1_m); \
|
||||
val_m = (uint64_t)((val_m << 32) & 0xFFFFFFFF00000000); \
|
||||
val_m = (uint64_t)(val_m | (uint64_t)val0_m); \
|
||||
\
|
||||
val_m; \
|
||||
#define LD(psrc) \
|
||||
({ \
|
||||
const uint8_t *psrc_m1 = (const uint8_t *)(psrc); \
|
||||
uint32_t val0_m, val1_m; \
|
||||
uint64_t val_m_combined = 0; \
|
||||
\
|
||||
val0_m = LW(psrc_m1); \
|
||||
val1_m = LW(psrc_m1 + 4); \
|
||||
\
|
||||
val_m_combined = (uint64_t)(val1_m); \
|
||||
val_m_combined = (uint64_t)((val_m_combined << 32) & 0xFFFFFFFF00000000); \
|
||||
val_m_combined = (uint64_t)(val_m_combined | (uint64_t)val0_m); \
|
||||
\
|
||||
val_m_combined; \
|
||||
})
|
||||
#endif // (__mips == 64)
|
||||
|
||||
@@ -2034,13 +2034,12 @@
|
||||
pdst, stride) \
|
||||
{ \
|
||||
v16u8 tmp0_m, tmp1_m, tmp2_m, tmp3_m; \
|
||||
uint8_t *pdst_m = (uint8_t *)(pdst); \
|
||||
\
|
||||
tmp0_m = PCKEV_XORI128_UB(in0, in1); \
|
||||
tmp1_m = PCKEV_XORI128_UB(in2, in3); \
|
||||
ILVR_D2_UB(dst1, dst0, dst3, dst2, tmp2_m, tmp3_m); \
|
||||
AVER_UB2_UB(tmp0_m, tmp2_m, tmp1_m, tmp3_m, tmp0_m, tmp1_m); \
|
||||
ST8x4_UB(tmp0_m, tmp1_m, pdst_m, stride); \
|
||||
ST8x4_UB(tmp0_m, tmp1_m, pdst, stride); \
|
||||
}
|
||||
|
||||
/* Description : Pack even byte elements and store byte vector in destination
|
||||
|
||||
@@ -16,18 +16,18 @@
|
||||
|
||||
extern const uint8_t mc_filt_mask_arr[16 * 3];
|
||||
|
||||
#define FILT_8TAP_DPADD_S_H(vec0, vec1, vec2, vec3, filt0, filt1, filt2, \
|
||||
filt3) \
|
||||
({ \
|
||||
v8i16 tmp0, tmp1; \
|
||||
\
|
||||
tmp0 = __msa_dotp_s_h((v16i8)vec0, (v16i8)filt0); \
|
||||
tmp0 = __msa_dpadd_s_h(tmp0, (v16i8)vec1, (v16i8)filt1); \
|
||||
tmp1 = __msa_dotp_s_h((v16i8)vec2, (v16i8)filt2); \
|
||||
tmp1 = __msa_dpadd_s_h(tmp1, (v16i8)vec3, (v16i8)filt3); \
|
||||
tmp0 = __msa_adds_s_h(tmp0, tmp1); \
|
||||
\
|
||||
tmp0; \
|
||||
#define FILT_8TAP_DPADD_S_H(vec0, vec1, vec2, vec3, filt0, filt1, filt2, \
|
||||
filt3) \
|
||||
({ \
|
||||
v8i16 tmp_dpadd_0, tmp_dpadd_1; \
|
||||
\
|
||||
tmp_dpadd_0 = __msa_dotp_s_h((v16i8)vec0, (v16i8)filt0); \
|
||||
tmp_dpadd_0 = __msa_dpadd_s_h(tmp_dpadd_0, (v16i8)vec1, (v16i8)filt1); \
|
||||
tmp_dpadd_1 = __msa_dotp_s_h((v16i8)vec2, (v16i8)filt2); \
|
||||
tmp_dpadd_1 = __msa_dpadd_s_h(tmp_dpadd_1, (v16i8)vec3, (v16i8)filt3); \
|
||||
tmp_dpadd_0 = __msa_adds_s_h(tmp_dpadd_0, tmp_dpadd_1); \
|
||||
\
|
||||
tmp_dpadd_0; \
|
||||
})
|
||||
|
||||
#define HORIZ_8TAP_FILT(src0, src1, mask0, mask1, mask2, mask3, filt_h0, \
|
||||
@@ -114,11 +114,10 @@ extern const uint8_t mc_filt_mask_arr[16 * 3];
|
||||
stride) \
|
||||
{ \
|
||||
v16u8 tmp0_m, tmp1_m, tmp2_m, tmp3_m; \
|
||||
uint8_t *pdst_m = (uint8_t *)(pdst); \
|
||||
\
|
||||
PCKEV_B2_UB(in2, in1, in4, in3, tmp0_m, tmp1_m); \
|
||||
PCKEV_D2_UB(dst1, dst0, dst3, dst2, tmp2_m, tmp3_m); \
|
||||
AVER_UB2_UB(tmp0_m, tmp2_m, tmp1_m, tmp3_m, tmp0_m, tmp1_m); \
|
||||
ST8x4_UB(tmp0_m, tmp1_m, pdst_m, stride); \
|
||||
ST8x4_UB(tmp0_m, tmp1_m, pdst, stride); \
|
||||
}
|
||||
#endif /* VPX_DSP_MIPS_VPX_CONVOLVE_MSA_H_ */
|
||||
|
||||
Reference in New Issue
Block a user