From db80c23fd4c25a8b27b2bf3d6b388cf0f8e5c329 Mon Sep 17 00:00:00 2001 From: James Zern Date: Fri, 30 Sep 2016 13:09:25 -0700 Subject: [PATCH] cosmetics,*_neon.c: rm redundant return from void fns + a couple of 'break's after a return Change-Id: Ia21f12ebcef98244feb923c17b689fc8115da015 --- vp9/common/arm/neon/vp9_iht4x4_add_neon.c | 7 ------- vp9/common/arm/neon/vp9_iht8x8_add_neon.c | 4 ---- vpx_dsp/arm/idct16x16_1_add_neon.c | 1 - vpx_dsp/arm/idct16x16_add_neon.c | 4 ---- vpx_dsp/arm/idct16x16_neon.c | 4 ---- vpx_dsp/arm/idct32x32_1_add_neon.c | 5 ----- vpx_dsp/arm/idct32x32_add_neon.c | 7 ------- vpx_dsp/arm/idct4x4_1_add_neon.c | 1 - vpx_dsp/arm/idct4x4_add_neon.c | 1 - vpx_dsp/arm/idct8x8_1_add_neon.c | 1 - vpx_dsp/arm/idct8x8_add_neon.c | 3 --- 11 files changed, 38 deletions(-) diff --git a/vp9/common/arm/neon/vp9_iht4x4_add_neon.c b/vp9/common/arm/neon/vp9_iht4x4_add_neon.c index eca655e72..cbd61de90 100644 --- a/vp9/common/arm/neon/vp9_iht4x4_add_neon.c +++ b/vp9/common/arm/neon/vp9_iht4x4_add_neon.c @@ -37,7 +37,6 @@ static INLINE void TRANSPOSE4X4(int16x8_t *q8s16, int16x8_t *q9s16) { *q8s16 = vreinterpretq_s16_s32(q0x2s32.val[0]); *q9s16 = vreinterpretq_s16_s32(q0x2s32.val[1]); - return; } static INLINE void GENERATE_COSINE_CONSTANTS(int16x4_t *d0s16, int16x4_t *d1s16, @@ -45,7 +44,6 @@ static INLINE void GENERATE_COSINE_CONSTANTS(int16x4_t *d0s16, int16x4_t *d1s16, *d0s16 = vdup_n_s16(cospi_8_64); *d1s16 = vdup_n_s16(cospi_16_64); *d2s16 = vdup_n_s16(cospi_24_64); - return; } static INLINE void GENERATE_SINE_CONSTANTS(int16x4_t *d3s16, int16x4_t *d4s16, @@ -54,7 +52,6 @@ static INLINE void GENERATE_SINE_CONSTANTS(int16x4_t *d3s16, int16x4_t *d4s16, *d4s16 = vdup_n_s16(sinpi_2_9); *q3s16 = vdupq_n_s16(sinpi_3_9); *d5s16 = vdup_n_s16(sinpi_4_9); - return; } static INLINE void IDCT4x4_1D(int16x4_t *d0s16, int16x4_t *d1s16, @@ -90,7 +87,6 @@ static INLINE void IDCT4x4_1D(int16x4_t *d0s16, int16x4_t *d1s16, *q8s16 = vaddq_s16(q13s16, q14s16); *q9s16 = vsubq_s16(q13s16, q14s16); *q9s16 = vcombine_s16(vget_high_s16(*q9s16), vget_low_s16(*q9s16)); // vswp - return; } static INLINE void IADST4x4_1D(int16x4_t *d3s16, int16x4_t *d4s16, @@ -136,7 +132,6 @@ static INLINE void IADST4x4_1D(int16x4_t *d3s16, int16x4_t *d4s16, *q8s16 = vcombine_s16(d16s16, d17s16); *q9s16 = vcombine_s16(d18s16, d19s16); - return; } void vp9_iht4x4_16_add_neon(const tran_low_t *input, uint8_t *dest, @@ -158,7 +153,6 @@ void vp9_iht4x4_16_add_neon(const tran_low_t *input, uint8_t *dest, case 0: // idct_idct is not supported. Fall back to C vp9_iht4x4_16_add_c(input, dest, dest_stride, tx_type); return; - break; case 1: // iadst_idct // generate constants GENERATE_COSINE_CONSTANTS(&d0s16, &d1s16, &d2s16); @@ -229,5 +223,4 @@ void vp9_iht4x4_16_add_neon(const tran_low_t *input, uint8_t *dest, vst1_lane_u32((uint32_t *)dest, vreinterpret_u32_u8(d26u8), 1); dest -= dest_stride; vst1_lane_u32((uint32_t *)dest, vreinterpret_u32_u8(d26u8), 0); - return; } diff --git a/vp9/common/arm/neon/vp9_iht8x8_add_neon.c b/vp9/common/arm/neon/vp9_iht8x8_add_neon.c index 0e3febf15..f7e0a6d98 100644 --- a/vp9/common/arm/neon/vp9_iht8x8_add_neon.c +++ b/vp9/common/arm/neon/vp9_iht8x8_add_neon.c @@ -180,7 +180,6 @@ static INLINE void IDCT8x8_1D(int16x8_t *q8s16, int16x8_t *q9s16, *q13s16 = vsubq_s16(q2s16, q5s16); *q14s16 = vsubq_s16(q1s16, q6s16); *q15s16 = vsubq_s16(q0s16, q7s16); - return; } static INLINE void IADST8X8_1D(int16x8_t *q8s16, int16x8_t *q9s16, @@ -405,7 +404,6 @@ static INLINE void IADST8X8_1D(int16x8_t *q8s16, int16x8_t *q9s16, *q11s16 = vsubq_s16(q5s16, q2s16); *q13s16 = vsubq_s16(q5s16, q6s16); *q15s16 = vsubq_s16(q5s16, q4s16); - return; } void vp9_iht8x8_64_add_neon(const tran_low_t *input, uint8_t *dest, @@ -433,7 +431,6 @@ void vp9_iht8x8_64_add_neon(const tran_low_t *input, uint8_t *dest, case 0: // idct_idct is not supported. Fall back to C vp9_iht8x8_64_add_c(input, dest, dest_stride, tx_type); return; - break; case 1: // iadst_idct // generate IDCT constants // GENERATE_IDCT_CONSTANTS @@ -540,5 +537,4 @@ void vp9_iht8x8_64_add_neon(const tran_low_t *input, uint8_t *dest, vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d3u8)); d2 += dest_stride; } - return; } diff --git a/vpx_dsp/arm/idct16x16_1_add_neon.c b/vpx_dsp/arm/idct16x16_1_add_neon.c index a9e08f911..4035830f3 100644 --- a/vpx_dsp/arm/idct16x16_1_add_neon.c +++ b/vpx_dsp/arm/idct16x16_1_add_neon.c @@ -56,5 +56,4 @@ void vpx_idct16x16_1_add_neon(const tran_low_t *input, uint8_t *dest, d2 += dest_stride; } } - return; } diff --git a/vpx_dsp/arm/idct16x16_add_neon.c b/vpx_dsp/arm/idct16x16_add_neon.c index 5ca2afe48..ce5cbcbcd 100644 --- a/vpx_dsp/arm/idct16x16_add_neon.c +++ b/vpx_dsp/arm/idct16x16_add_neon.c @@ -247,7 +247,6 @@ void vpx_idct16x16_256_add_neon_pass1(int16_t *in, int16_t *out, vst1_u64((uint64_t *)out, d30u64); out += output_stride; vst1_u64((uint64_t *)out, d31u64); - return; } void vpx_idct16x16_256_add_neon_pass2(int16_t *src, int16_t *out, @@ -796,7 +795,6 @@ void vpx_idct16x16_256_add_neon_pass2(int16_t *src, int16_t *out, out += 4; vst1_u64((uint64_t *)out, vreinterpret_u64_s16(vget_high_s16(q15s16))); } - return; } void vpx_idct16x16_10_add_neon_pass1(int16_t *in, int16_t *out, @@ -932,7 +930,6 @@ void vpx_idct16x16_10_add_neon_pass1(int16_t *in, int16_t *out, vst1_u64((uint64_t *)out, d30u64); out += output_stride; vst1_u64((uint64_t *)out, d31u64); - return; } void vpx_idct16x16_10_add_neon_pass2(int16_t *src, int16_t *out, @@ -1227,5 +1224,4 @@ void vpx_idct16x16_10_add_neon_pass2(int16_t *src, int16_t *out, vst1_u64((uint64_t *)out, d30u64); out += 4; vst1_u64((uint64_t *)out, d31u64); - return; } diff --git a/vpx_dsp/arm/idct16x16_neon.c b/vpx_dsp/arm/idct16x16_neon.c index 2b30cd789..bdbbf5193 100644 --- a/vpx_dsp/arm/idct16x16_neon.c +++ b/vpx_dsp/arm/idct16x16_neon.c @@ -90,8 +90,6 @@ void vpx_idct16x16_256_add_neon(const int16_t *input, uint8_t *dest, // restore d8-d15 register values. vpx_pop_neon(store_reg); #endif - - return; } void vpx_idct16x16_10_add_neon(const int16_t *input, uint8_t *dest, @@ -147,6 +145,4 @@ void vpx_idct16x16_10_add_neon(const int16_t *input, uint8_t *dest, // restore d8-d15 register values. vpx_pop_neon(store_reg); #endif - - return; } diff --git a/vpx_dsp/arm/idct32x32_1_add_neon.c b/vpx_dsp/arm/idct32x32_1_add_neon.c index 28ea5ffd6..9dfdf8d69 100644 --- a/vpx_dsp/arm/idct32x32_1_add_neon.c +++ b/vpx_dsp/arm/idct32x32_1_add_neon.c @@ -35,7 +35,6 @@ static INLINE void LD_16x8(uint8_t *d, int d_stride, uint8x16_t *q8u8, *q14u8 = vld1q_u8(d); d += d_stride; *q15u8 = vld1q_u8(d); - return; } static INLINE void ADD_DIFF_16x8(uint8x16_t qdiffu8, uint8x16_t *q8u8, @@ -51,7 +50,6 @@ static INLINE void ADD_DIFF_16x8(uint8x16_t qdiffu8, uint8x16_t *q8u8, *q13u8 = vqaddq_u8(*q13u8, qdiffu8); *q14u8 = vqaddq_u8(*q14u8, qdiffu8); *q15u8 = vqaddq_u8(*q15u8, qdiffu8); - return; } static INLINE void SUB_DIFF_16x8(uint8x16_t qdiffu8, uint8x16_t *q8u8, @@ -67,7 +65,6 @@ static INLINE void SUB_DIFF_16x8(uint8x16_t qdiffu8, uint8x16_t *q8u8, *q13u8 = vqsubq_u8(*q13u8, qdiffu8); *q14u8 = vqsubq_u8(*q14u8, qdiffu8); *q15u8 = vqsubq_u8(*q15u8, qdiffu8); - return; } static INLINE void ST_16x8(uint8_t *d, int d_stride, uint8x16_t *q8u8, @@ -90,7 +87,6 @@ static INLINE void ST_16x8(uint8_t *d, int d_stride, uint8x16_t *q8u8, vst1q_u8(d, *q14u8); d += d_stride; vst1q_u8(d, *q15u8); - return; } void vpx_idct32x32_1_add_neon(const tran_low_t *input, uint8_t *dest, @@ -137,5 +133,4 @@ void vpx_idct32x32_1_add_neon(const tran_low_t *input, uint8_t *dest, } } } - return; } diff --git a/vpx_dsp/arm/idct32x32_add_neon.c b/vpx_dsp/arm/idct32x32_add_neon.c index c4580e016..4eff9b970 100644 --- a/vpx_dsp/arm/idct32x32_add_neon.c +++ b/vpx_dsp/arm/idct32x32_add_neon.c @@ -68,7 +68,6 @@ static INLINE void __STORE_COMBINE_CENTER_RESULTS(uint8_t *p1, uint8_t *p2, p2 += stride; vst1_s16((int16_t *)p1, d8s16); vst1_s16((int16_t *)p2, d11s16); - return; } #define STORE_COMBINE_EXTREME_RESULTS(r7, r6) \ @@ -112,7 +111,6 @@ static INLINE void __STORE_COMBINE_EXTREME_RESULTS(uint8_t *p1, uint8_t *p2, p2 += stride; vst1_s16((int16_t *)p2, d7s16); vst1_s16((int16_t *)p1, d4s16); - return; } #define DO_BUTTERFLY_STD(const_1, const_2, qA, qB) \ @@ -150,7 +148,6 @@ static INLINE void DO_BUTTERFLY(int16x8_t q14s16, int16x8_t q13s16, *qAs16 = vcombine_s16(vqrshrn_n_s32(q8s32, 14), vqrshrn_n_s32(q9s32, 14)); *qBs16 = vcombine_s16(vqrshrn_n_s32(q11s32, 14), vqrshrn_n_s32(q10s32, 14)); - return; } static INLINE void idct32_transpose_pair(const int16_t *input, int16_t *t_buf) { @@ -197,7 +194,6 @@ static INLINE void idct32_transpose_pair(const int16_t *input, int16_t *t_buf) { vst1q_s16(t_buf, q15s16); t_buf += 8; } - return; } static INLINE void idct32_bands_end_1st_pass(int16_t *out, int16x8_t q2s16, @@ -285,7 +281,6 @@ static INLINE void idct32_bands_end_1st_pass(int16_t *out, int16x8_t q2s16, q7s16 = vsubq_s16(q2s16, q1s16); STORE_IN_OUTPUT(25, 24, 25, q6s16, q7s16); STORE_IN_OUTPUT(25, 6, 7, q4s16, q5s16); - return; } static INLINE void idct32_bands_end_2nd_pass( @@ -380,7 +375,6 @@ static INLINE void idct32_bands_end_2nd_pass( q6s16 = vsubq_s16(q3s16, q0s16); q7s16 = vsubq_s16(q2s16, q1s16); STORE_COMBINE_EXTREME_RESULTS(r7, r6); - return; } void vpx_idct32x32_1024_add_neon(const tran_low_t *input, uint8_t *dest, @@ -638,5 +632,4 @@ void vpx_idct32x32_1024_add_neon(const tran_low_t *input, uint8_t *dest, } } } - return; } diff --git a/vpx_dsp/arm/idct4x4_1_add_neon.c b/vpx_dsp/arm/idct4x4_1_add_neon.c index 97c1dc258..b37cb51a1 100644 --- a/vpx_dsp/arm/idct4x4_1_add_neon.c +++ b/vpx_dsp/arm/idct4x4_1_add_neon.c @@ -44,5 +44,4 @@ void vpx_idct4x4_1_add_neon(const tran_low_t *input, uint8_t *dest, vst1_lane_u32((uint32_t *)d2, vreinterpret_u32_u8(d6u8), 1); d2 += dest_stride; } - return; } diff --git a/vpx_dsp/arm/idct4x4_add_neon.c b/vpx_dsp/arm/idct4x4_add_neon.c index 5afd8ec61..1caa45698 100644 --- a/vpx_dsp/arm/idct4x4_add_neon.c +++ b/vpx_dsp/arm/idct4x4_add_neon.c @@ -145,5 +145,4 @@ void vpx_idct4x4_16_add_neon(const tran_low_t *input, uint8_t *dest, vst1_lane_u32((uint32_t *)d, vreinterpret_u32_u8(d27u8), 1); d += dest_stride; vst1_lane_u32((uint32_t *)d, vreinterpret_u32_u8(d27u8), 0); - return; } diff --git a/vpx_dsp/arm/idct8x8_1_add_neon.c b/vpx_dsp/arm/idct8x8_1_add_neon.c index 31f87b945..df557de81 100644 --- a/vpx_dsp/arm/idct8x8_1_add_neon.c +++ b/vpx_dsp/arm/idct8x8_1_add_neon.c @@ -59,5 +59,4 @@ void vpx_idct8x8_1_add_neon(const tran_low_t *input, uint8_t *dest, vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d31u8)); d2 += dest_stride; } - return; } diff --git a/vpx_dsp/arm/idct8x8_add_neon.c b/vpx_dsp/arm/idct8x8_add_neon.c index c623ed4a5..7d6561241 100644 --- a/vpx_dsp/arm/idct8x8_add_neon.c +++ b/vpx_dsp/arm/idct8x8_add_neon.c @@ -163,7 +163,6 @@ static INLINE void IDCT8x8_1D(int16x8_t *q8s16, int16x8_t *q9s16, *q13s16 = vsubq_s16(q2s16, q5s16); *q14s16 = vsubq_s16(q1s16, q6s16); *q15s16 = vsubq_s16(q0s16, q7s16); - return; } void vpx_idct8x8_64_add_neon(const tran_low_t *input, uint8_t *dest, @@ -266,7 +265,6 @@ void vpx_idct8x8_64_add_neon(const tran_low_t *input, uint8_t *dest, d2 += dest_stride; vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d3u8)); d2 += dest_stride; - return; } void vpx_idct8x8_12_add_neon(const tran_low_t *input, uint8_t *dest, @@ -444,5 +442,4 @@ void vpx_idct8x8_12_add_neon(const tran_low_t *input, uint8_t *dest, d2 += dest_stride; vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d3u8)); d2 += dest_stride; - return; }