2010-05-18 17:58:33 +02:00
|
|
|
;
|
2010-09-09 14:16:39 +02:00
|
|
|
; Copyright (c) 2010 The WebM project authors. All Rights Reserved.
|
2010-05-18 17:58:33 +02:00
|
|
|
;
|
2010-06-18 18:39:21 +02:00
|
|
|
; Use of this source code is governed by a BSD-style license
|
2010-06-04 22:19:40 +02:00
|
|
|
; that can be found in the LICENSE file in the root of the source
|
|
|
|
; tree. An additional intellectual property rights grant can be found
|
2010-06-18 18:39:21 +02:00
|
|
|
; in the file PATENTS. All contributing project authors may
|
2010-06-04 22:19:40 +02:00
|
|
|
; be found in the AUTHORS file in the root of the source tree.
|
2010-05-18 17:58:33 +02:00
|
|
|
;
|
|
|
|
|
|
|
|
|
|
|
|
EXPORT |vp8_variance16x16_neon|
|
|
|
|
EXPORT |vp8_variance16x8_neon|
|
|
|
|
EXPORT |vp8_variance8x16_neon|
|
|
|
|
EXPORT |vp8_variance8x8_neon|
|
|
|
|
|
|
|
|
ARM
|
|
|
|
REQUIRE8
|
|
|
|
PRESERVE8
|
|
|
|
|
|
|
|
AREA ||.text||, CODE, READONLY, ALIGN=2
|
|
|
|
|
|
|
|
; r0 unsigned char *src_ptr
|
|
|
|
; r1 int source_stride
|
|
|
|
; r2 unsigned char *ref_ptr
|
|
|
|
; r3 int recon_stride
|
|
|
|
; stack unsigned int *sse
|
|
|
|
|vp8_variance16x16_neon| PROC
|
|
|
|
vmov.i8 q8, #0 ;q8 - sum
|
|
|
|
vmov.i8 q9, #0 ;q9, q10 - sse
|
|
|
|
vmov.i8 q10, #0
|
|
|
|
|
|
|
|
mov r12, #8
|
|
|
|
|
|
|
|
variance16x16_neon_loop
|
|
|
|
vld1.8 {q0}, [r0], r1 ;Load up source and reference
|
|
|
|
vld1.8 {q2}, [r2], r3
|
|
|
|
vld1.8 {q1}, [r0], r1
|
|
|
|
vld1.8 {q3}, [r2], r3
|
|
|
|
|
|
|
|
vsubl.u8 q11, d0, d4 ;calculate diff
|
|
|
|
vsubl.u8 q12, d1, d5
|
|
|
|
vsubl.u8 q13, d2, d6
|
|
|
|
vsubl.u8 q14, d3, d7
|
|
|
|
|
|
|
|
;VPADAL adds adjacent pairs of elements of a vector, and accumulates
|
|
|
|
;the results into the elements of the destination vector. The explanation
|
|
|
|
;in ARM guide is wrong.
|
|
|
|
vpadal.s16 q8, q11 ;calculate sum
|
|
|
|
vmlal.s16 q9, d22, d22 ;calculate sse
|
|
|
|
vmlal.s16 q10, d23, d23
|
|
|
|
|
|
|
|
subs r12, r12, #1
|
|
|
|
|
|
|
|
vpadal.s16 q8, q12
|
|
|
|
vmlal.s16 q9, d24, d24
|
|
|
|
vmlal.s16 q10, d25, d25
|
|
|
|
vpadal.s16 q8, q13
|
|
|
|
vmlal.s16 q9, d26, d26
|
|
|
|
vmlal.s16 q10, d27, d27
|
|
|
|
vpadal.s16 q8, q14
|
|
|
|
vmlal.s16 q9, d28, d28
|
|
|
|
vmlal.s16 q10, d29, d29
|
|
|
|
|
|
|
|
bne variance16x16_neon_loop
|
|
|
|
|
|
|
|
vadd.u32 q10, q9, q10 ;accumulate sse
|
|
|
|
vpaddl.s32 q0, q8 ;accumulate sum
|
|
|
|
|
|
|
|
ldr r12, [sp] ;load *sse from stack
|
|
|
|
|
|
|
|
vpaddl.u32 q1, q10
|
|
|
|
vadd.s64 d0, d0, d1
|
|
|
|
vadd.u64 d1, d2, d3
|
|
|
|
|
|
|
|
;vmov.32 r0, d0[0] ;this instruction costs a lot
|
|
|
|
;vmov.32 r1, d1[0]
|
|
|
|
;mul r0, r0, r0
|
|
|
|
;str r1, [r12]
|
|
|
|
;sub r0, r1, r0, asr #8
|
|
|
|
|
|
|
|
;sum is in [-255x256, 255x256]. sumxsum is 32-bit. Shift to right should
|
|
|
|
;have sign-bit exension, which is vshr.s. Have to use s32 to make it right.
|
|
|
|
vmull.s32 q5, d0, d0
|
|
|
|
vst1.32 {d1[0]}, [r12] ;store sse
|
|
|
|
vshr.s32 d10, d10, #8
|
|
|
|
vsub.s32 d0, d1, d10
|
|
|
|
|
|
|
|
vmov.32 r0, d0[0] ;return
|
|
|
|
bx lr
|
|
|
|
|
|
|
|
ENDP
|
|
|
|
|
|
|
|
;================================
|
|
|
|
;unsigned int vp8_variance16x8_c(
|
|
|
|
; unsigned char *src_ptr,
|
|
|
|
; int source_stride,
|
|
|
|
; unsigned char *ref_ptr,
|
|
|
|
; int recon_stride,
|
|
|
|
; unsigned int *sse)
|
|
|
|
|vp8_variance16x8_neon| PROC
|
|
|
|
vmov.i8 q8, #0 ;q8 - sum
|
|
|
|
vmov.i8 q9, #0 ;q9, q10 - sse
|
|
|
|
vmov.i8 q10, #0
|
|
|
|
|
|
|
|
mov r12, #4
|
|
|
|
|
|
|
|
variance16x8_neon_loop
|
|
|
|
vld1.8 {q0}, [r0], r1 ;Load up source and reference
|
|
|
|
vld1.8 {q2}, [r2], r3
|
|
|
|
vld1.8 {q1}, [r0], r1
|
|
|
|
vld1.8 {q3}, [r2], r3
|
|
|
|
|
|
|
|
vsubl.u8 q11, d0, d4 ;calculate diff
|
|
|
|
vsubl.u8 q12, d1, d5
|
|
|
|
vsubl.u8 q13, d2, d6
|
|
|
|
vsubl.u8 q14, d3, d7
|
|
|
|
|
|
|
|
vpadal.s16 q8, q11 ;calculate sum
|
|
|
|
vmlal.s16 q9, d22, d22 ;calculate sse
|
|
|
|
vmlal.s16 q10, d23, d23
|
|
|
|
|
|
|
|
subs r12, r12, #1
|
|
|
|
|
|
|
|
vpadal.s16 q8, q12
|
|
|
|
vmlal.s16 q9, d24, d24
|
|
|
|
vmlal.s16 q10, d25, d25
|
|
|
|
vpadal.s16 q8, q13
|
|
|
|
vmlal.s16 q9, d26, d26
|
|
|
|
vmlal.s16 q10, d27, d27
|
|
|
|
vpadal.s16 q8, q14
|
|
|
|
vmlal.s16 q9, d28, d28
|
|
|
|
vmlal.s16 q10, d29, d29
|
|
|
|
|
|
|
|
bne variance16x8_neon_loop
|
|
|
|
|
|
|
|
vadd.u32 q10, q9, q10 ;accumulate sse
|
|
|
|
vpaddl.s32 q0, q8 ;accumulate sum
|
|
|
|
|
|
|
|
ldr r12, [sp] ;load *sse from stack
|
|
|
|
|
|
|
|
vpaddl.u32 q1, q10
|
|
|
|
vadd.s64 d0, d0, d1
|
|
|
|
vadd.u64 d1, d2, d3
|
|
|
|
|
|
|
|
vmull.s32 q5, d0, d0
|
|
|
|
vst1.32 {d1[0]}, [r12] ;store sse
|
|
|
|
vshr.s32 d10, d10, #7
|
|
|
|
vsub.s32 d0, d1, d10
|
|
|
|
|
|
|
|
vmov.32 r0, d0[0] ;return
|
|
|
|
bx lr
|
|
|
|
|
|
|
|
ENDP
|
|
|
|
|
|
|
|
;=================================
|
|
|
|
;unsigned int vp8_variance8x16_c(
|
|
|
|
; unsigned char *src_ptr,
|
|
|
|
; int source_stride,
|
|
|
|
; unsigned char *ref_ptr,
|
|
|
|
; int recon_stride,
|
|
|
|
; unsigned int *sse)
|
|
|
|
|
|
|
|
|vp8_variance8x16_neon| PROC
|
|
|
|
vmov.i8 q8, #0 ;q8 - sum
|
|
|
|
vmov.i8 q9, #0 ;q9, q10 - sse
|
|
|
|
vmov.i8 q10, #0
|
|
|
|
|
|
|
|
mov r12, #8
|
|
|
|
|
|
|
|
variance8x16_neon_loop
|
|
|
|
vld1.8 {d0}, [r0], r1 ;Load up source and reference
|
|
|
|
vld1.8 {d4}, [r2], r3
|
|
|
|
vld1.8 {d2}, [r0], r1
|
|
|
|
vld1.8 {d6}, [r2], r3
|
|
|
|
|
|
|
|
vsubl.u8 q11, d0, d4 ;calculate diff
|
|
|
|
vsubl.u8 q12, d2, d6
|
|
|
|
|
|
|
|
vpadal.s16 q8, q11 ;calculate sum
|
|
|
|
vmlal.s16 q9, d22, d22 ;calculate sse
|
|
|
|
vmlal.s16 q10, d23, d23
|
|
|
|
|
|
|
|
subs r12, r12, #1
|
|
|
|
|
|
|
|
vpadal.s16 q8, q12
|
|
|
|
vmlal.s16 q9, d24, d24
|
|
|
|
vmlal.s16 q10, d25, d25
|
|
|
|
|
|
|
|
bne variance8x16_neon_loop
|
|
|
|
|
|
|
|
vadd.u32 q10, q9, q10 ;accumulate sse
|
|
|
|
vpaddl.s32 q0, q8 ;accumulate sum
|
|
|
|
|
|
|
|
ldr r12, [sp] ;load *sse from stack
|
|
|
|
|
|
|
|
vpaddl.u32 q1, q10
|
|
|
|
vadd.s64 d0, d0, d1
|
|
|
|
vadd.u64 d1, d2, d3
|
|
|
|
|
|
|
|
vmull.s32 q5, d0, d0
|
|
|
|
vst1.32 {d1[0]}, [r12] ;store sse
|
|
|
|
vshr.s32 d10, d10, #7
|
|
|
|
vsub.s32 d0, d1, d10
|
|
|
|
|
|
|
|
vmov.32 r0, d0[0] ;return
|
|
|
|
bx lr
|
|
|
|
|
|
|
|
ENDP
|
|
|
|
|
|
|
|
;==================================
|
|
|
|
; r0 unsigned char *src_ptr
|
|
|
|
; r1 int source_stride
|
|
|
|
; r2 unsigned char *ref_ptr
|
|
|
|
; r3 int recon_stride
|
|
|
|
; stack unsigned int *sse
|
|
|
|
|vp8_variance8x8_neon| PROC
|
|
|
|
vmov.i8 q8, #0 ;q8 - sum
|
|
|
|
vmov.i8 q9, #0 ;q9, q10 - sse
|
|
|
|
vmov.i8 q10, #0
|
|
|
|
|
|
|
|
mov r12, #2
|
|
|
|
|
|
|
|
variance8x8_neon_loop
|
|
|
|
vld1.8 {d0}, [r0], r1 ;Load up source and reference
|
|
|
|
vld1.8 {d4}, [r2], r3
|
|
|
|
vld1.8 {d1}, [r0], r1
|
|
|
|
vld1.8 {d5}, [r2], r3
|
|
|
|
vld1.8 {d2}, [r0], r1
|
|
|
|
vld1.8 {d6}, [r2], r3
|
|
|
|
vld1.8 {d3}, [r0], r1
|
|
|
|
vld1.8 {d7}, [r2], r3
|
|
|
|
|
|
|
|
vsubl.u8 q11, d0, d4 ;calculate diff
|
|
|
|
vsubl.u8 q12, d1, d5
|
|
|
|
vsubl.u8 q13, d2, d6
|
|
|
|
vsubl.u8 q14, d3, d7
|
|
|
|
|
|
|
|
vpadal.s16 q8, q11 ;calculate sum
|
|
|
|
vmlal.s16 q9, d22, d22 ;calculate sse
|
|
|
|
vmlal.s16 q10, d23, d23
|
|
|
|
|
|
|
|
subs r12, r12, #1
|
|
|
|
|
|
|
|
vpadal.s16 q8, q12
|
|
|
|
vmlal.s16 q9, d24, d24
|
|
|
|
vmlal.s16 q10, d25, d25
|
|
|
|
vpadal.s16 q8, q13
|
|
|
|
vmlal.s16 q9, d26, d26
|
|
|
|
vmlal.s16 q10, d27, d27
|
|
|
|
vpadal.s16 q8, q14
|
|
|
|
vmlal.s16 q9, d28, d28
|
|
|
|
vmlal.s16 q10, d29, d29
|
|
|
|
|
|
|
|
bne variance8x8_neon_loop
|
|
|
|
|
|
|
|
vadd.u32 q10, q9, q10 ;accumulate sse
|
|
|
|
vpaddl.s32 q0, q8 ;accumulate sum
|
|
|
|
|
|
|
|
ldr r12, [sp] ;load *sse from stack
|
|
|
|
|
|
|
|
vpaddl.u32 q1, q10
|
|
|
|
vadd.s64 d0, d0, d1
|
|
|
|
vadd.u64 d1, d2, d3
|
|
|
|
|
|
|
|
vmull.s32 q5, d0, d0
|
|
|
|
vst1.32 {d1[0]}, [r12] ;store sse
|
|
|
|
vshr.s32 d10, d10, #6
|
|
|
|
vsub.s32 d0, d1, d10
|
|
|
|
|
|
|
|
vmov.32 r0, d0[0] ;return
|
|
|
|
bx lr
|
|
|
|
|
|
|
|
ENDP
|
|
|
|
|
|
|
|
END
|