add Neon optimized add constant residual functions
- Add add_constant_residual_8x8 16x16 32x32 functions - Tested under RealView debugger enviroment Change-Id: I5c3a432f651b49bf375de6496353706a33e3e68e
This commit is contained in:
parent
d6264f9adf
commit
a83cfd4da1
@ -59,13 +59,13 @@ specialize void vp9_build_intra_predictors
|
||||
|
||||
if [ "$CONFIG_VP9_DECODER" = "yes" ]; then
|
||||
prototype void vp9_add_constant_residual_8x8 "const int16_t diff, uint8_t *dest, int stride"
|
||||
specialize vp9_add_constant_residual_8x8 sse2
|
||||
specialize vp9_add_constant_residual_8x8 sse2 neon
|
||||
|
||||
prototype void vp9_add_constant_residual_16x16 "const int16_t diff, uint8_t *dest, int stride"
|
||||
specialize vp9_add_constant_residual_16x16 sse2
|
||||
specialize vp9_add_constant_residual_16x16 sse2 neon
|
||||
|
||||
prototype void vp9_add_constant_residual_32x32 "const int16_t diff, uint8_t *dest, int stride"
|
||||
specialize vp9_add_constant_residual_32x32 sse2
|
||||
specialize vp9_add_constant_residual_32x32 sse2 neon
|
||||
fi
|
||||
|
||||
#
|
||||
|
230
vp9/decoder/arm/neon/vp9_add_constant_residual_neon.asm
Normal file
230
vp9/decoder/arm/neon/vp9_add_constant_residual_neon.asm
Normal file
@ -0,0 +1,230 @@
|
||||
;
|
||||
; Copyright (c) 2013 The WebM project authors. All Rights Reserved.
|
||||
;
|
||||
; Use of this source code is governed by a BSD-style license
|
||||
; that can be found in the LICENSE file in the root of the source
|
||||
; tree. An additional intellectual property rights grant can be found
|
||||
; in the file PATENTS. All contributing project authors may
|
||||
; be found in the AUTHORS file in the root of the source tree.
|
||||
;
|
||||
|
||||
EXPORT |vp9_add_constant_residual_8x8_neon|
|
||||
EXPORT |vp9_add_constant_residual_16x16_neon|
|
||||
EXPORT |vp9_add_constant_residual_32x32_neon|
|
||||
ARM
|
||||
|
||||
AREA ||.text||, CODE, READONLY, ALIGN=2
|
||||
|
||||
MACRO
|
||||
LD_16x8 $src, $stride
|
||||
vld1.8 {q8}, [$src], $stride
|
||||
vld1.8 {q9}, [$src], $stride
|
||||
vld1.8 {q10}, [$src], $stride
|
||||
vld1.8 {q11}, [$src], $stride
|
||||
vld1.8 {q12}, [$src], $stride
|
||||
vld1.8 {q13}, [$src], $stride
|
||||
vld1.8 {q14}, [$src], $stride
|
||||
vld1.8 {q15}, [$src], $stride
|
||||
MEND
|
||||
|
||||
MACRO
|
||||
ADD_DIFF_16x8 $diff
|
||||
vqadd.u8 q8, q8, $diff
|
||||
vqadd.u8 q9, q9, $diff
|
||||
vqadd.u8 q10, q10, $diff
|
||||
vqadd.u8 q11, q11, $diff
|
||||
vqadd.u8 q12, q12, $diff
|
||||
vqadd.u8 q13, q13, $diff
|
||||
vqadd.u8 q14, q14, $diff
|
||||
vqadd.u8 q15, q15, $diff
|
||||
MEND
|
||||
|
||||
MACRO
|
||||
SUB_DIFF_16x8 $diff
|
||||
vqsub.u8 q8, q8, $diff
|
||||
vqsub.u8 q9, q9, $diff
|
||||
vqsub.u8 q10, q10, $diff
|
||||
vqsub.u8 q11, q11, $diff
|
||||
vqsub.u8 q12, q12, $diff
|
||||
vqsub.u8 q13, q13, $diff
|
||||
vqsub.u8 q14, q14, $diff
|
||||
vqsub.u8 q15, q15, $diff
|
||||
MEND
|
||||
|
||||
MACRO
|
||||
ST_16x8 $dst, $stride
|
||||
vst1.8 {q8}, [$dst], $stride
|
||||
vst1.8 {q9}, [$dst], $stride
|
||||
vst1.8 {q10}, [$dst], $stride
|
||||
vst1.8 {q11}, [$dst], $stride
|
||||
vst1.8 {q12}, [$dst], $stride
|
||||
vst1.8 {q13}, [$dst], $stride
|
||||
vst1.8 {q14}, [$dst], $stride
|
||||
vst1.8 {q15}, [$dst], $stride
|
||||
MEND
|
||||
|
||||
; void add_constant_residual(const int16_t diff, uint8_t *dest, int stride,
|
||||
; int width, int height) {
|
||||
; int r, c;
|
||||
;
|
||||
; for (r = 0; r < height; r++) {
|
||||
; for (c = 0; c < width; c++)
|
||||
; dest[c] = clip_pixel(diff + dest[c]);
|
||||
;
|
||||
; dest += stride;
|
||||
; }
|
||||
;}
|
||||
;void vp9_add_constant_residual_8x8_c(const int16_t diff, uint8_t *dest,
|
||||
; int stride) {
|
||||
; add_constant_residual(diff, dest, stride, 8, 8);
|
||||
;}
|
||||
; r0 : const int16_t diff
|
||||
; r1 : const uint8_t *dest
|
||||
; r2 : int stride
|
||||
;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
|
||||
|vp9_add_constant_residual_8x8_neon| PROC
|
||||
mov r3, r1 ; r3: save dest to r3
|
||||
vld1.8 {d0}, [r1], r2
|
||||
vld1.8 {d1}, [r1], r2
|
||||
vld1.8 {d2}, [r1], r2
|
||||
vld1.8 {d3}, [r1], r2
|
||||
vld1.8 {d4}, [r1], r2
|
||||
vld1.8 {d5}, [r1], r2
|
||||
vld1.8 {d6}, [r1], r2
|
||||
vld1.8 {d7}, [r1], r2
|
||||
cmp r0, #0
|
||||
bge DIFF_POSITIVE_8x8
|
||||
|
||||
DIFF_NEGATIVE_8x8 ; diff < 0
|
||||
neg r0, r0
|
||||
usat r0, #8, r0
|
||||
vdup.u8 q8, r0
|
||||
|
||||
vqsub.u8 q0, q0, q8
|
||||
vqsub.u8 q1, q1, q8
|
||||
vqsub.u8 q2, q2, q8
|
||||
vqsub.u8 q3, q3, q8
|
||||
b DIFF_SAVE_8x8
|
||||
|
||||
DIFF_POSITIVE_8x8 ; diff >= 0
|
||||
usat r0, #8, r0
|
||||
vdup.u8 q8, r0
|
||||
|
||||
vqadd.u8 q0, q0, q8
|
||||
vqadd.u8 q1, q1, q8
|
||||
vqadd.u8 q2, q2, q8
|
||||
vqadd.u8 q3, q3, q8
|
||||
|
||||
DIFF_SAVE_8x8
|
||||
vst1.8 {d0}, [r3], r2
|
||||
vst1.8 {d1}, [r3], r2
|
||||
vst1.8 {d2}, [r3], r2
|
||||
vst1.8 {d3}, [r3], r2
|
||||
vst1.8 {d4}, [r3], r2
|
||||
vst1.8 {d5}, [r3], r2
|
||||
vst1.8 {d6}, [r3], r2
|
||||
vst1.8 {d7}, [r3], r2
|
||||
|
||||
bx lr
|
||||
ENDP
|
||||
|
||||
;void vp9_add_constant_residual_16x16_c(const int16_t diff, uint8_t *dest,
|
||||
; int stride) {
|
||||
; add_constant_residual(diff, dest, stride, 16, 16);
|
||||
;}
|
||||
; r0 : const int16_t diff
|
||||
; r1 : const uint8_t *dest
|
||||
; r2 : int stride
|
||||
;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
|
||||
|vp9_add_constant_residual_16x16_neon| PROC
|
||||
mov r3, r1
|
||||
LD_16x8 r1, r2
|
||||
cmp r0, #0
|
||||
bge DIFF_POSITIVE_16x16
|
||||
|
||||
|DIFF_NEGATIVE_16x16|
|
||||
neg r0, r0
|
||||
usat r0, #8, r0
|
||||
vdup.u8 q0, r0
|
||||
|
||||
SUB_DIFF_16x8 q0
|
||||
ST_16x8 r3, r2
|
||||
LD_16x8 r1, r2
|
||||
SUB_DIFF_16x8 q0
|
||||
b DIFF_SAVE_16x16
|
||||
|
||||
|DIFF_POSITIVE_16x16|
|
||||
usat r0, #8, r0
|
||||
vdup.u8 q0, r0
|
||||
|
||||
ADD_DIFF_16x8 q0
|
||||
ST_16x8 r3, r2
|
||||
LD_16x8 r1, r2
|
||||
ADD_DIFF_16x8 q0
|
||||
|
||||
|DIFF_SAVE_16x16|
|
||||
ST_16x8 r3, r2
|
||||
bx lr
|
||||
ENDP
|
||||
|
||||
;void vp9_add_constant_residual_32x32_c(const int16_t diff, uint8_t *dest,
|
||||
; int stride) {
|
||||
; add_constant_residual(diff, dest, stride, 32, 32);
|
||||
;}
|
||||
; r0 : const int16_t diff
|
||||
; r1 : const uint8_t *dest
|
||||
; r2 : int stride
|
||||
;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
|
||||
|vp9_add_constant_residual_32x32_neon| PROC
|
||||
push {r4,lr}
|
||||
pld [r1]
|
||||
mov r3, r1
|
||||
add r4, r1, #16 ; r4 dest + 16 for second loop
|
||||
cmp r0, #0
|
||||
bge DIFF_POSITIVE_32x32
|
||||
|
||||
|DIFF_NEGATIVE_32x32|
|
||||
neg r0, r0
|
||||
usat r0, #8, r0
|
||||
vdup.u8 q0, r0
|
||||
mov r0, #4
|
||||
|
||||
|DIFF_NEGATIVE_32x32_LOOP|
|
||||
sub r0, #1
|
||||
LD_16x8 r1, r2
|
||||
SUB_DIFF_16x8 q0
|
||||
ST_16x8 r3, r2
|
||||
|
||||
LD_16x8 r1, r2
|
||||
SUB_DIFF_16x8 q0
|
||||
ST_16x8 r3, r2
|
||||
cmp r0, #2
|
||||
moveq r1, r4
|
||||
moveq r3, r4
|
||||
cmp r0, #0
|
||||
bne DIFF_NEGATIVE_32x32_LOOP
|
||||
pop {r4,pc}
|
||||
|
||||
|DIFF_POSITIVE_32x32|
|
||||
usat r0, #8, r0
|
||||
vdup.u8 q0, r0
|
||||
mov r0, #4
|
||||
|
||||
|DIFF_POSITIVE_32x32_LOOP|
|
||||
sub r0, #1
|
||||
LD_16x8 r1, r2
|
||||
ADD_DIFF_16x8 q0
|
||||
ST_16x8 r3, r2
|
||||
|
||||
LD_16x8 r1, r2
|
||||
ADD_DIFF_16x8 q0
|
||||
ST_16x8 r3, r2
|
||||
cmp r0, #2
|
||||
moveq r1, r4
|
||||
moveq r3, r4
|
||||
cmp r0, #0
|
||||
bne DIFF_POSITIVE_32x32_LOOP
|
||||
pop {r4,pc}
|
||||
ENDP
|
||||
|
||||
END
|
@ -39,6 +39,7 @@ VP9_DX_SRCS-yes += decoder/vp9_dsubexp.h
|
||||
VP9_DX_SRCS-yes := $(filter-out $(VP9_DX_SRCS_REMOVE-yes),$(VP9_DX_SRCS-yes))
|
||||
|
||||
VP9_DX_SRCS-$(HAVE_SSE2) += decoder/x86/vp9_dequantize_sse2.c
|
||||
VP9_DX_SRCS-$(HAVE_NEON) += decoder/arm/neon/vp9_add_constant_residual_neon$(ASM)
|
||||
|
||||
$(eval $(call asm_offsets_template,\
|
||||
vp9_asm_dec_offsets.asm, $(VP9_PREFIX)decoder/vp9_asm_dec_offsets.c))
|
||||
|
Loading…
x
Reference in New Issue
Block a user