vpx/vp9/decoder/arm/neon/vp9_add_constant_residual_neon.asm

231 lines
8.2 KiB
NASM
Raw Normal View History

;
; Copyright (c) 2013 The WebM project authors. All Rights Reserved.
;
; Use of this source code is governed by a BSD-style license
; that can be found in the LICENSE file in the root of the source
; tree. An additional intellectual property rights grant can be found
; in the file PATENTS. All contributing project authors may
; be found in the AUTHORS file in the root of the source tree.
;
EXPORT |vp9_add_constant_residual_8x8_neon|
EXPORT |vp9_add_constant_residual_16x16_neon|
EXPORT |vp9_add_constant_residual_32x32_neon|
ARM
AREA ||.text||, CODE, READONLY, ALIGN=2
MACRO
LD_16x8 $src, $stride
vld1.8 {q8}, [$src], $stride
vld1.8 {q9}, [$src], $stride
vld1.8 {q10}, [$src], $stride
vld1.8 {q11}, [$src], $stride
vld1.8 {q12}, [$src], $stride
vld1.8 {q13}, [$src], $stride
vld1.8 {q14}, [$src], $stride
vld1.8 {q15}, [$src], $stride
MEND
MACRO
ADD_DIFF_16x8 $diff
vqadd.u8 q8, q8, $diff
vqadd.u8 q9, q9, $diff
vqadd.u8 q10, q10, $diff
vqadd.u8 q11, q11, $diff
vqadd.u8 q12, q12, $diff
vqadd.u8 q13, q13, $diff
vqadd.u8 q14, q14, $diff
vqadd.u8 q15, q15, $diff
MEND
MACRO
SUB_DIFF_16x8 $diff
vqsub.u8 q8, q8, $diff
vqsub.u8 q9, q9, $diff
vqsub.u8 q10, q10, $diff
vqsub.u8 q11, q11, $diff
vqsub.u8 q12, q12, $diff
vqsub.u8 q13, q13, $diff
vqsub.u8 q14, q14, $diff
vqsub.u8 q15, q15, $diff
MEND
MACRO
ST_16x8 $dst, $stride
vst1.8 {q8}, [$dst], $stride
vst1.8 {q9}, [$dst], $stride
vst1.8 {q10}, [$dst], $stride
vst1.8 {q11}, [$dst], $stride
vst1.8 {q12}, [$dst], $stride
vst1.8 {q13}, [$dst], $stride
vst1.8 {q14}, [$dst], $stride
vst1.8 {q15}, [$dst], $stride
MEND
; void add_constant_residual(const int16_t diff, uint8_t *dest, int stride,
; int width, int height) {
; int r, c;
;
; for (r = 0; r < height; r++) {
; for (c = 0; c < width; c++)
; dest[c] = clip_pixel(diff + dest[c]);
;
; dest += stride;
; }
;}
;void vp9_add_constant_residual_8x8_c(const int16_t diff, uint8_t *dest,
; int stride) {
; add_constant_residual(diff, dest, stride, 8, 8);
;}
; r0 : const int16_t diff
; r1 : const uint8_t *dest
; r2 : int stride
;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
|vp9_add_constant_residual_8x8_neon| PROC
mov r3, r1 ; r3: save dest to r3
vld1.8 {d0}, [r1], r2
vld1.8 {d1}, [r1], r2
vld1.8 {d2}, [r1], r2
vld1.8 {d3}, [r1], r2
vld1.8 {d4}, [r1], r2
vld1.8 {d5}, [r1], r2
vld1.8 {d6}, [r1], r2
vld1.8 {d7}, [r1], r2
cmp r0, #0
bge DIFF_POSITIVE_8x8
DIFF_NEGATIVE_8x8 ; diff < 0
neg r0, r0
usat r0, #8, r0
vdup.u8 q8, r0
vqsub.u8 q0, q0, q8
vqsub.u8 q1, q1, q8
vqsub.u8 q2, q2, q8
vqsub.u8 q3, q3, q8
b DIFF_SAVE_8x8
DIFF_POSITIVE_8x8 ; diff >= 0
usat r0, #8, r0
vdup.u8 q8, r0
vqadd.u8 q0, q0, q8
vqadd.u8 q1, q1, q8
vqadd.u8 q2, q2, q8
vqadd.u8 q3, q3, q8
DIFF_SAVE_8x8
vst1.8 {d0}, [r3], r2
vst1.8 {d1}, [r3], r2
vst1.8 {d2}, [r3], r2
vst1.8 {d3}, [r3], r2
vst1.8 {d4}, [r3], r2
vst1.8 {d5}, [r3], r2
vst1.8 {d6}, [r3], r2
vst1.8 {d7}, [r3], r2
bx lr
ENDP
;void vp9_add_constant_residual_16x16_c(const int16_t diff, uint8_t *dest,
; int stride) {
; add_constant_residual(diff, dest, stride, 16, 16);
;}
; r0 : const int16_t diff
; r1 : const uint8_t *dest
; r2 : int stride
;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
|vp9_add_constant_residual_16x16_neon| PROC
mov r3, r1
LD_16x8 r1, r2
cmp r0, #0
bge DIFF_POSITIVE_16x16
|DIFF_NEGATIVE_16x16|
neg r0, r0
usat r0, #8, r0
vdup.u8 q0, r0
SUB_DIFF_16x8 q0
ST_16x8 r3, r2
LD_16x8 r1, r2
SUB_DIFF_16x8 q0
b DIFF_SAVE_16x16
|DIFF_POSITIVE_16x16|
usat r0, #8, r0
vdup.u8 q0, r0
ADD_DIFF_16x8 q0
ST_16x8 r3, r2
LD_16x8 r1, r2
ADD_DIFF_16x8 q0
|DIFF_SAVE_16x16|
ST_16x8 r3, r2
bx lr
ENDP
;void vp9_add_constant_residual_32x32_c(const int16_t diff, uint8_t *dest,
; int stride) {
; add_constant_residual(diff, dest, stride, 32, 32);
;}
; r0 : const int16_t diff
; r1 : const uint8_t *dest
; r2 : int stride
;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
|vp9_add_constant_residual_32x32_neon| PROC
push {r4,lr}
pld [r1]
mov r3, r1
add r4, r1, #16 ; r4 dest + 16 for second loop
cmp r0, #0
bge DIFF_POSITIVE_32x32
|DIFF_NEGATIVE_32x32|
neg r0, r0
usat r0, #8, r0
vdup.u8 q0, r0
mov r0, #4
|DIFF_NEGATIVE_32x32_LOOP|
sub r0, #1
LD_16x8 r1, r2
SUB_DIFF_16x8 q0
ST_16x8 r3, r2
LD_16x8 r1, r2
SUB_DIFF_16x8 q0
ST_16x8 r3, r2
cmp r0, #2
moveq r1, r4
moveq r3, r4
cmp r0, #0
bne DIFF_NEGATIVE_32x32_LOOP
pop {r4,pc}
|DIFF_POSITIVE_32x32|
usat r0, #8, r0
vdup.u8 q0, r0
mov r0, #4
|DIFF_POSITIVE_32x32_LOOP|
sub r0, #1
LD_16x8 r1, r2
ADD_DIFF_16x8 q0
ST_16x8 r3, r2
LD_16x8 r1, r2
ADD_DIFF_16x8 q0
ST_16x8 r3, r2
cmp r0, #2
moveq r1, r4
moveq r3, r4
cmp r0, #0
bne DIFF_POSITIVE_32x32_LOOP
pop {r4,pc}
ENDP
END