;****************************************************************************** ;* VP9 loop filter SIMD optimizations ;* ;* Copyright (C) 2013-2014 Clément Bœsch ;* ;* This file is part of FFmpeg. ;* ;* FFmpeg is free software; you can redistribute it and/or ;* modify it under the terms of the GNU Lesser General Public ;* License as published by the Free Software Foundation; either ;* version 2.1 of the License, or (at your option) any later version. ;* ;* FFmpeg is distributed in the hope that it will be useful, ;* but WITHOUT ANY WARRANTY; without even the implied warranty of ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU ;* Lesser General Public License for more details. ;* ;* You should have received a copy of the GNU Lesser General Public ;* License along with FFmpeg; if not, write to the Free Software ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA ;****************************************************************************** %include "libavutil/x86/x86util.asm" SECTION_RODATA cextern pb_3 cextern pb_80 pb_4: times 16 db 0x04 pb_10: times 16 db 0x10 pb_40: times 16 db 0x40 pb_81: times 16 db 0x81 pb_f8: times 16 db 0xf8 pb_fe: times 16 db 0xfe pb_ff: times 16 db 0xff cextern pw_4 cextern pw_8 ; with mix functions, two 8-bit thresholds are stored in a 16-bit storage, ; the following mask is used to splat both in the same register mask_mix: times 8 db 0 times 8 db 1 mask_mix84: times 8 db 0xff times 8 db 0x00 mask_mix48: times 8 db 0x00 times 8 db 0xff SECTION .text %macro SCRATCH 3 %if ARCH_X86_64 SWAP %1, %2 %else mova [%3], m%1 %endif %endmacro %macro UNSCRATCH 3 %if ARCH_X86_64 SWAP %1, %2 %else mova m%1, [%3] %endif %endmacro ; %1 = abs(%2-%3) %macro ABSSUB 4 ; dst, src1 (RO), src2 (RO), tmp %if ARCH_X86_64 psubusb %1, %3, %2 psubusb %4, %2, %3 %else mova %1, %3 mova %4, %2 psubusb %1, %2 psubusb %4, %3 %endif por %1, %4 %endmacro ; %1 = %1>%2 %macro CMP_GT 2-3 ; src/dst, cmp, pb_80 %if %0 == 3 pxor %1, %3 %endif pcmpgtb %1, %2 %endmacro ; %1 = abs(%2-%3) > %4 %macro ABSSUB_GT 5-6 [pb_80]; dst, src1, src2, cmp, tmp, [pb_80] ABSSUB %1, %2, %3, %5 ; dst = abs(src1-src2) CMP_GT %1, %4, %6 ; dst > cmp %endmacro %macro MASK_APPLY 4 ; %1=new_data/dst %2=old_data %3=mask %4=tmp pand %1, %3 ; new &= mask pandn %4, %3, %2 ; tmp = ~mask & old por %1, %4 ; new&mask | old&~mask %endmacro %macro UNPACK 4 %if ARCH_X86_64 punpck%1bw %2, %3, %4 %else mova %2, %3 punpck%1bw %2, %4 %endif %endmacro %macro FILTER_SUBx2_ADDx2 11 ; %1=dst %2=h/l %3=cache %4=stack_off %5=sub1 %6=sub2 %7=add1 ; %8=add2 %9=rshift, [unpack], [unpack_is_mem_on_x86_32] psubw %3, [rsp+%4+%5*32] psubw %3, [rsp+%4+%6*32] paddw %3, [rsp+%4+%7*32] %ifnidn %10, "" %if %11 == 0 punpck%2bw %1, %10, m0 %else UNPACK %2, %1, %10, m0 %endif mova [rsp+%4+%8*32], %1 paddw %3, %1 %else paddw %3, [rsp+%4+%8*32] %endif psraw %1, %3, %9 %endmacro ; FIXME interleave l/h better (for instruction pairing) %macro FILTER_INIT 9 ; tmp1, tmp2, cacheL, cacheH, dstp, stack_off, filterid, mask, source FILTER%7_INIT %1, l, %3, %6 + 0 FILTER%7_INIT %2, h, %4, %6 + 16 packuswb %1, %2 MASK_APPLY %1, %9, %8, %2 mova %5, %1 %endmacro %macro FILTER_UPDATE 12-16 "", "", "", 0 ; tmp1, tmp2, cacheL, cacheH, dstp, stack_off, -, -, +, +, rshift, ; mask, [source], [unpack + src], [unpack_is_mem_on_x86_32] ; FIXME interleave this properly with the subx2/addx2 %ifnidn %15, "" %if %16 == 0 || ARCH_X86_64 mova %14, %15 %endif %endif FILTER_SUBx2_ADDx2 %1, l, %3, %6 + 0, %7, %8, %9, %10, %11, %14, %16 FILTER_SUBx2_ADDx2 %2, h, %4, %6 + 16, %7, %8, %9, %10, %11, %14, %16 packuswb %1, %2 %ifnidn %13, "" MASK_APPLY %1, %13, %12, %2 %else MASK_APPLY %1, %5, %12, %2 %endif mova %5, %1 %endmacro %macro SRSHIFT3B_2X 4 ; reg1, reg2, [pb_10], tmp mova %4, [pb_f8] pand %1, %4 pand %2, %4 psrlq %1, 3 psrlq %2, 3 pxor %1, %3 pxor %2, %3 psubb %1, %3 psubb %2, %3 %endmacro %macro EXTRACT_POS_NEG 3 ; i8, neg, pos pxor %3, %3 pxor %2, %2 pcmpgtb %3, %1 ; i8 < 0 mask psubb %2, %1 ; neg values (only the originally - will be kept) pand %2, %3 ; negative values of i8 (but stored as +) pandn %3, %1 ; positive values of i8 %endmacro ; clip_u8(u8 + i8) %macro SIGN_ADD 4 ; dst, u8, i8, tmp1 EXTRACT_POS_NEG %3, %4, %1 paddusb %1, %2 ; add the positives psubusb %1, %4 ; sub the negatives %endmacro ; clip_u8(u8 - i8) %macro SIGN_SUB 4 ; dst, u8, i8, tmp1 EXTRACT_POS_NEG %3, %1, %4 paddusb %1, %2 ; add the negatives psubusb %1, %4 ; sub the positives %endmacro %macro FILTER6_INIT 4 ; %1=dst %2=h/l %3=cache, %4=stack_off UNPACK %2, %1, rp3, m0 ; p3: B->W mova [rsp+%4+0*32], %1 paddw %3, %1, %1 ; p3*2 paddw %3, %1 ; p3*3 punpck%2bw %1, m1, m0 ; p2: B->W mova [rsp+%4+1*32], %1 paddw %3, %1 ; p3*3 + p2 paddw %3, %1 ; p3*3 + p2*2 UNPACK %2, %1, rp1, m0 ; p1: B->W mova [rsp+%4+2*32], %1 paddw %3, %1 ; p3*3 + p2*2 + p1 UNPACK %2, %1, rp0, m0 ; p0: B->W mova [rsp+%4+3*32], %1 paddw %3, %1 ; p3*3 + p2*2 + p1 + p0 UNPACK %2, %1, rq0, m0 ; q0: B->W mova [rsp+%4+4*32], %1 paddw %3, %1 ; p3*3 + p2*2 + p1 + p0 + q0 paddw %3, [pw_4] ; p3*3 + p2*2 + p1 + p0 + q0 + 4 psraw %1, %3, 3 ; (p3*3 + p2*2 + p1 + p0 + q0 + 4) >> 3 %endmacro %macro FILTER14_INIT 4 ; %1=dst %2=h/l %3=cache, %4=stack_off punpck%2bw %1, m2, m0 ; p7: B->W mova [rsp+%4+ 8*32], %1 psllw %3, %1, 3 ; p7*8 psubw %3, %1 ; p7*7 punpck%2bw %1, m3, m0 ; p6: B->W mova [rsp+%4+ 9*32], %1 paddw %3, %1 ; p7*7 + p6 paddw %3, %1 ; p7*7 + p6*2 UNPACK %2, %1, rp5, m0 ; p5: B->W mova [rsp+%4+10*32], %1 paddw %3, %1 ; p7*7 + p6*2 + p5 UNPACK %2, %1, rp4, m0 ; p4: B->W mova [rsp+%4+11*32], %1 paddw %3, %1 ; p7*7 + p6*2 + p5 + p4 paddw %3, [rsp+%4+ 0*32] ; p7*7 + p6*2 + p5 + p4 + p3 paddw %3, [rsp+%4+ 1*32] ; p7*7 + p6*2 + p5 + .. + p2 paddw %3, [rsp+%4+ 2*32] ; p7*7 + p6*2 + p5 + .. + p1 paddw %3, [rsp+%4+ 3*32] ; p7*7 + p6*2 + p5 + .. + p0 paddw %3, [rsp+%4+ 4*32] ; p7*7 + p6*2 + p5 + .. + p0 + q0 paddw %3, [pw_8] ; p7*7 + p6*2 + p5 + .. + p0 + q0 + 8 psraw %1, %3, 4 ; (p7*7 + p6*2 + p5 + .. + p0 + q0 + 8) >> 4 %endmacro %macro TRANSPOSE16x16B 17 mova %17, m%16 SBUTTERFLY bw, %1, %2, %16 SBUTTERFLY bw, %3, %4, %16 SBUTTERFLY bw, %5, %6, %16 SBUTTERFLY bw, %7, %8, %16 SBUTTERFLY bw, %9, %10, %16 SBUTTERFLY bw, %11, %12, %16 SBUTTERFLY bw, %13, %14, %16 mova m%16, %17 mova %17, m%14 SBUTTERFLY bw, %15, %16, %14 SBUTTERFLY wd, %1, %3, %14 SBUTTERFLY wd, %2, %4, %14 SBUTTERFLY wd, %5, %7, %14 SBUTTERFLY wd, %6, %8, %14 SBUTTERFLY wd, %9, %11, %14 SBUTTERFLY wd, %10, %12, %14 SBUTTERFLY wd, %13, %15, %14 mova m%14, %17 mova %17, m%12 SBUTTERFLY wd, %14, %16, %12 SBUTTERFLY dq, %1, %5, %12 SBUTTERFLY dq, %2, %6, %12 SBUTTERFLY dq, %3, %7, %12 SBUTTERFLY dq, %4, %8, %12 SBUTTERFLY dq, %9, %13, %12 SBUTTERFLY dq, %10, %14, %12 SBUTTERFLY dq, %11, %15, %12 mova m%12, %17 mova %17, m%8 SBUTTERFLY dq, %12, %16, %8 SBUTTERFLY qdq, %1, %9, %8 SBUTTERFLY qdq, %2, %10, %8 SBUTTERFLY qdq, %3, %11, %8 SBUTTERFLY qdq, %4, %12, %8 SBUTTERFLY qdq, %5, %13, %8 SBUTTERFLY qdq, %6, %14, %8 SBUTTERFLY qdq, %7, %15, %8 mova m%8, %17 mova %17, m%1 SBUTTERFLY qdq, %8, %16, %1 mova m%1, %17 SWAP %2, %9 SWAP %3, %5 SWAP %4, %13 SWAP %6, %11 SWAP %8, %15 SWAP %12, %14 %endmacro %macro DEFINE_REAL_P7_TO_Q7 0-1 0 %define P7 dstq + 4*mstrideq + %1 %define P6 dstq + mstride3q + %1 %define P5 dstq + 2*mstrideq + %1 %define P4 dstq + mstrideq + %1 %define P3 dstq + %1 %define P2 dstq + strideq + %1 %define P1 dstq + 2* strideq + %1 %define P0 dstq + stride3q + %1 %define Q0 dstq + 4* strideq + %1 %define Q1 dst2q + mstride3q + %1 %define Q2 dst2q + 2*mstrideq + %1 %define Q3 dst2q + mstrideq + %1 %define Q4 dst2q + %1 %define Q5 dst2q + strideq + %1 %define Q6 dst2q + 2* strideq + %1 %define Q7 dst2q + stride3q + %1 %endmacro ; ..............AB -> AAAAAAAABBBBBBBB %macro SPLATB_MIX 1-2 [mask_mix] %if cpuflag(ssse3) pshufb %1, %2 %else punpcklbw %1, %1 punpcklwd %1, %1 punpckldq %1, %1 %endif %endmacro %macro LOOPFILTER 5 ; %1=v/h %2=size1 %3+%4=stack, %5=32bit stack only %if UNIX64 cglobal vp9_loop_filter_%1_%2_16, 5, 9, 16, %3 + %4, dst, stride, E, I, H, mstride, dst2, stride3, mstride3 %else %if WIN64 cglobal vp9_loop_filter_%1_%2_16, 4, 8, 16, %3 + %4, dst, stride, E, I, mstride, dst2, stride3, mstride3 %else cglobal vp9_loop_filter_%1_%2_16, 2, 6, 16, %3 + %4 + %5, dst, stride, mstride, dst2, stride3, mstride3 %define Ed dword r2m %define Id dword r3m %endif %define Hd dword r4m %endif mov mstrideq, strideq neg mstrideq lea stride3q, [strideq*3] lea mstride3q, [mstrideq*3] %ifidn %1, h %if %2 > 16 %define movx movh lea dstq, [dstq + 4*strideq - 4] %else %define movx movu lea dstq, [dstq + 4*strideq - 8] ; go from top center (h pos) to center left (v pos) %endif lea dst2q, [dstq + 8*strideq] %else lea dstq, [dstq + 4*mstrideq] lea dst2q, [dstq + 8*strideq] %endif DEFINE_REAL_P7_TO_Q7 %ifidn %1, h movx m0, [P7] movx m1, [P6] movx m2, [P5] movx m3, [P4] movx m4, [P3] movx m5, [P2] movx m6, [P1] movx m7, [P0] %if ARCH_X86_64 movx m8, [Q0] movx m9, [Q1] movx m10, [Q2] movx m11, [Q3] movx m12, [Q4] movx m13, [Q5] movx m14, [Q6] movx m15, [Q7] %if %2 == 16 TRANSPOSE16x16B 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, [rsp] %define P7 rsp + 128 %define P6 rsp + 144 %define P5 rsp + 160 %define P4 rsp + 176 %define Q4 rsp + 192 %define Q5 rsp + 208 %define Q6 rsp + 224 %define Q7 rsp + 240 mova [P7], m0 mova [P6], m1 mova [P5], m2 mova [P4], m3 %else ; 8x16 transpose punpcklbw m0, m1 punpcklbw m2, m3 punpcklbw m4, m5 punpcklbw m6, m7 punpcklbw m8, m9 punpcklbw m10, m11 punpcklbw m12, m13 punpcklbw m14, m15 TRANSPOSE8x8W 0, 2, 4, 6, 8, 10, 12, 14, 15 SWAP 0, 4 SWAP 2, 5 SWAP 0, 6 SWAP 0, 7 SWAP 10, 9 SWAP 12, 10 SWAP 14, 11 %endif %else ; x86-32 punpcklbw m0, m1 punpcklbw m2, m3 punpcklbw m4, m5 punpcklbw m6, m7 movx m1, [Q0] movx m3, [Q1] movx m5, [Q2] movx m7, [Q3] punpcklbw m1, m3 punpcklbw m5, m7 movx m3, [Q4] movx m7, [Q5] punpcklbw m3, m7 mova [rsp], m3 movx m3, [Q6] movx m7, [Q7] punpcklbw m3, m7 %endif %define P3 rsp + 0 %define P2 rsp + 16 %define P1 rsp + 32 %define P0 rsp + 48 %define Q0 rsp + 64 %define Q1 rsp + 80 %define Q2 rsp + 96 %define Q3 rsp + 112 %if ARCH_X86_64 mova [P3], m4 mova [P2], m5 mova [P1], m6 mova [P0], m7 mova [Q0], m8 mova [Q1], m9 mova [Q2], m10 mova [Q3], m11 %if %2 == 16 mova [Q4], m12 mova [Q5], m13 mova [Q6], m14 mova [Q7], m15 %endif %else ; x86-32 TRANSPOSE8x8W 0, 2, 4, 6, 1, 5, 7, 3, [rsp], [Q0], 1 mova [P3], m0 mova [P2], m2 mova [P1], m4 mova [P0], m6 mova [Q1], m5 mova [Q2], m7 mova [Q3], m3 %endif %endif ; %1 == h ; calc fm mask %if %2 == 16 %if cpuflag(ssse3) pxor m0, m0 %endif SPLATB_REG m2, I, m0 ; I I I I ... SPLATB_REG m3, E, m0 ; E E E E ... %else %if cpuflag(ssse3) mova m0, [mask_mix] %endif movd m2, Id movd m3, Ed SPLATB_MIX m2, m0 SPLATB_MIX m3, m0 %endif mova m0, [pb_80] pxor m2, m0 pxor m3, m0 %if ARCH_X86_64 %ifidn %1, v mova m8, [P3] mova m9, [P2] mova m10, [P1] mova m11, [P0] mova m12, [Q0] mova m13, [Q1] mova m14, [Q2] mova m15, [Q3] %else ; In case of horizontal, P3..Q3 are already present in some registers due ; to the previous transpose, so we just swap registers. SWAP 8, 4, 12 SWAP 9, 5, 13 SWAP 10, 6, 14 SWAP 11, 7, 15 %endif %define rp3 m8 %define rp2 m9 %define rp1 m10 %define rp0 m11 %define rq0 m12 %define rq1 m13 %define rq2 m14 %define rq3 m15 %else %define rp3 [P3] %define rp2 [P2] %define rp1 [P1] %define rp0 [P0] %define rq0 [Q0] %define rq1 [Q1] %define rq2 [Q2] %define rq3 [Q3] %endif ABSSUB_GT m5, rp3, rp2, m2, m7, m0 ; m5 = abs(p3-p2) <= I ABSSUB_GT m1, rp2, rp1, m2, m7, m0 ; m1 = abs(p2-p1) <= I por m5, m1 ABSSUB_GT m1, rp1, rp0, m2, m7, m0 ; m1 = abs(p1-p0) <= I por m5, m1 ABSSUB_GT m1, rq0, rq1, m2, m7, m0 ; m1 = abs(q1-q0) <= I por m5, m1 ABSSUB_GT m1, rq1, rq2, m2, m7, m0 ; m1 = abs(q2-q1) <= I por m5, m1 ABSSUB_GT m1, rq2, rq3, m2, m7, m0 ; m1 = abs(q3-q2) <= I por m5, m1 ABSSUB m1, rp0, rq0, m7 ; abs(p0-q0) paddusb m1, m1 ; abs(p0-q0) * 2 ABSSUB m2, rp1, rq1, m7 ; abs(p1-q1) pand m2, [pb_fe] ; drop lsb so shift can work psrlq m2, 1 ; abs(p1-q1)/2 paddusb m1, m2 ; abs(p0-q0)*2 + abs(p1-q1)/2 pxor m1, m0 pcmpgtb m1, m3 por m1, m5 ; fm final value SWAP 1, 3 pxor m3, [pb_ff] ; (m3: fm, m8..15: p3 p2 p1 p0 q0 q1 q2 q3) ; calc flat8in (if not 44_16) and hev masks %if %2 != 44 mova m6, [pb_81] ; [1 1 1 1 ...] ^ 0x80 ABSSUB_GT m2, rp3, rp0, m6, m5 ; abs(p3 - p0) <= 1 %if ARCH_X86_64 mova m8, [pb_80] %define rb80 m8 %else %define rb80 [pb_80] %endif ABSSUB_GT m1, rp2, rp0, m6, m5, rb80 ; abs(p2 - p0) <= 1 por m2, m1 ABSSUB m4, rp1, rp0, m5 ; abs(p1 - p0) %if %2 == 16 %if cpuflag(ssse3) pxor m0, m0 %endif SPLATB_REG m7, H, m0 ; H H H H ... %else movd m7, Hd SPLATB_MIX m7 %endif pxor m7, rb80 pxor m4, rb80 pcmpgtb m0, m4, m7 ; abs(p1 - p0) > H (1/2 hev condition) CMP_GT m4, m6 ; abs(p1 - p0) <= 1 por m2, m4 ; (flat8in) ABSSUB m4, rq1, rq0, m1 ; abs(q1 - q0) pxor m4, rb80 pcmpgtb m5, m4, m7 ; abs(q1 - q0) > H (2/2 hev condition) por m0, m5 ; hev final value CMP_GT m4, m6 ; abs(q1 - q0) <= 1 por m2, m4 ; (flat8in) ABSSUB_GT m1, rq2, rq0, m6, m5, rb80 ; abs(q2 - q0) <= 1 por m2, m1 ABSSUB_GT m1, rq3, rq0, m6, m5, rb80 ; abs(q3 - q0) <= 1 por m2, m1 ; flat8in final value pxor m2, [pb_ff] %if %2 == 84 || %2 == 48 pand m2, [mask_mix%2] %endif %else mova m6, [pb_80] movd m7, Hd SPLATB_MIX m7 pxor m7, m6 ABSSUB m4, rp1, rp0, m1 ; abs(p1 - p0) pxor m4, m6 pcmpgtb m0, m4, m7 ; abs(p1 - p0) > H (1/2 hev condition) ABSSUB m4, rq1, rq0, m1 ; abs(q1 - q0) pxor m4, m6 pcmpgtb m5, m4, m7 ; abs(q1 - q0) > H (2/2 hev condition) por m0, m5 ; hev final value %endif %if %2 == 16 ; (m0: hev, m2: flat8in, m3: fm, m6: pb_81, m9..15: p2 p1 p0 q0 q1 q2 q3) ; calc flat8out mask %if ARCH_X86_64 mova m8, [P7] mova m9, [P6] %define rp7 m8 %define rp6 m9 %else %define rp7 [P7] %define rp6 [P6] %endif ABSSUB_GT m1, rp7, rp0, m6, m5 ; abs(p7 - p0) <= 1 ABSSUB_GT m7, rp6, rp0, m6, m5 ; abs(p6 - p0) <= 1 por m1, m7 %if ARCH_X86_64 mova m8, [P5] mova m9, [P4] %define rp5 m8 %define rp4 m9 %else %define rp5 [P5] %define rp4 [P4] %endif ABSSUB_GT m7, rp5, rp0, m6, m5 ; abs(p5 - p0) <= 1 por m1, m7 ABSSUB_GT m7, rp4, rp0, m6, m5 ; abs(p4 - p0) <= 1 por m1, m7 %if ARCH_X86_64 mova m14, [Q4] mova m15, [Q5] %define rq4 m14 %define rq5 m15 %else %define rq4 [Q4] %define rq5 [Q5] %endif ABSSUB_GT m7, rq4, rq0, m6, m5 ; abs(q4 - q0) <= 1 por m1, m7 ABSSUB_GT m7, rq5, rq0, m6, m5 ; abs(q5 - q0) <= 1 por m1, m7 %if ARCH_X86_64 mova m14, [Q6] mova m15, [Q7] %define rq6 m14 %define rq7 m15 %else %define rq6 [Q6] %define rq7 [Q7] %endif ABSSUB_GT m7, rq6, rq0, m6, m5 ; abs(q4 - q0) <= 1 por m1, m7 ABSSUB_GT m7, rq7, rq0, m6, m5 ; abs(q5 - q0) <= 1 por m1, m7 ; flat8out final value pxor m1, [pb_ff] %endif ; if (fm) { ; if (out && in) filter_14() ; else if (in) filter_6() ; else if (hev) filter_2() ; else filter_4() ; } ; ; f14: fm & out & in ; f6: fm & ~f14 & in => fm & ~(out & in) & in => fm & ~out & in ; f2: fm & ~f14 & ~f6 & hev => fm & ~(out & in) & ~(~out & in) & hev => fm & ~in & hev ; f4: fm & ~f14 & ~f6 & ~f2 => fm & ~(out & in) & ~(~out & in) & ~(~in & hev) => fm & ~in & ~hev ; (m0: hev, [m1: flat8out], [m2: flat8in], m3: fm, m8..15: p5 p4 p1 p0 q0 q1 q6 q7) ; filter2() %if %2 != 44 mova m6, [pb_80] ; already in m6 if 44_16 SCRATCH 2, 15, rsp+%3+%4 %if %2 == 16 SCRATCH 1, 8, rsp+%3+%4+16 %endif %endif pxor m2, m6, rq0 ; q0 ^ 0x80 pxor m4, m6, rp0 ; p0 ^ 0x80 psubsb m2, m4 ; (signed) q0 - p0 pxor m4, m6, rp1 ; p1 ^ 0x80 pxor m5, m6, rq1 ; q1 ^ 0x80 psubsb m4, m5 ; (signed) p1 - q1 paddsb m4, m2 ; (q0 - p0) + (p1 - q1) paddsb m4, m2 ; 2*(q0 - p0) + (p1 - q1) paddsb m4, m2 ; 3*(q0 - p0) + (p1 - q1) paddsb m6, m4, [pb_4] ; m6: f1 = clip(f + 4, 127) paddsb m4, [pb_3] ; m4: f2 = clip(f + 3, 127) %if ARCH_X86_64 mova m14, [pb_10] ; will be reused in filter4() %define rb10 m14 %else %define rb10 [pb_10] %endif SRSHIFT3B_2X m6, m4, rb10, m7 ; f1 and f2 sign byte shift by 3 SIGN_SUB m7, rq0, m6, m5 ; m7 = q0 - f1 SIGN_ADD m1, rp0, m4, m5 ; m1 = p0 + f2 %if %2 != 44 %if ARCH_X86_64 pandn m6, m15, m3 ; ~mask(in) & mask(fm) %else mova m6, [rsp+%3+%4] pandn m6, m3 %endif pand m6, m0 ; (~mask(in) & mask(fm)) & mask(hev) %else pand m6, m3, m0 %endif MASK_APPLY m7, rq0, m6, m5 ; m7 = filter2(q0) & mask / we write it in filter4() MASK_APPLY m1, rp0, m6, m5 ; m1 = filter2(p0) & mask / we write it in filter4() ; (m0: hev, m1: p0', m2: q0-p0, m3: fm, m7: q0', [m8: flat8out], m10..13: p1 p0 q0 q1, m14: pb_10, [m15: flat8in], ) ; filter4() mova m4, m2 paddsb m2, m4 ; 2 * (q0 - p0) paddsb m2, m4 ; 3 * (q0 - p0) paddsb m6, m2, [pb_4] ; m6: f1 = clip(f + 4, 127) paddsb m2, [pb_3] ; m2: f2 = clip(f + 3, 127) SRSHIFT3B_2X m6, m2, rb10, m4 ; f1 and f2 sign byte shift by 3 %if %2 != 44 %if ARCH_X86_64 pandn m5, m15, m3 ; ~mask(in) & mask(fm) %else mova m5, [rsp+%3+%4] pandn m5, m3 %endif pandn m0, m5 ; ~mask(hev) & (~mask(in) & mask(fm)) %else pandn m0, m3 %endif SIGN_SUB m5, rq0, m6, m4 ; q0 - f1 MASK_APPLY m5, m7, m0, m4 ; filter4(q0) & mask mova [Q0], m5 SIGN_ADD m7, rp0, m2, m4 ; p0 + f2 MASK_APPLY m7, m1, m0, m4 ; filter4(p0) & mask mova [P0], m7 paddb m6, [pb_80] ; pxor m1, m1 ; f=(f1+1)>>1 pavgb m6, m1 ; psubb m6, [pb_40] ; SIGN_ADD m1, rp1, m6, m2 ; p1 + f SIGN_SUB m4, rq1, m6, m2 ; q1 - f MASK_APPLY m1, rp1, m0, m2 ; m1 = filter4(p1) MASK_APPLY m4, rq1, m0, m2 ; m4 = filter4(q1) mova [P1], m1 mova [Q1], m4 %if %2 != 44 UNSCRATCH 2, 15, rsp+%3+%4 %endif ; ([m1: flat8out], m2: flat8in, m3: fm, m10..13: p1 p0 q0 q1) ; filter6() %if %2 != 44 pxor m0, m0 %if %2 > 16 pand m3, m2 %else pand m2, m3 ; mask(fm) & mask(in) %if ARCH_X86_64 pandn m3, m8, m2 ; ~mask(out) & (mask(fm) & mask(in)) %else mova m3, [rsp+%3+%4+16] pandn m3, m2 %endif %endif %if ARCH_X86_64 mova m14, [P3] mova m9, [Q3] %define rp3 m14 %define rq3 m9 %else %define rp3 [P3] %define rq3 [Q3] %endif mova m1, [P2] FILTER_INIT m4, m5, m6, m7, [P2], %4, 6, m3, m1 ; [p2] mova m1, [Q2] FILTER_UPDATE m4, m5, m6, m7, [P1], %4, 0, 1, 2, 5, 3, m3, "", rq1, "", 1 ; [p1] -p3 -p2 +p1 +q1 FILTER_UPDATE m4, m5, m6, m7, [P0], %4, 0, 2, 3, 6, 3, m3, "", m1 ; [p0] -p3 -p1 +p0 +q2 FILTER_UPDATE m4, m5, m6, m7, [Q0], %4, 0, 3, 4, 7, 3, m3, "", rq3, "", 1 ; [q0] -p3 -p0 +q0 +q3 FILTER_UPDATE m4, m5, m6, m7, [Q1], %4, 1, 4, 5, 7, 3, m3, "" ; [q1] -p2 -q0 +q1 +q3 FILTER_UPDATE m4, m5, m6, m7, [Q2], %4, 2, 5, 6, 7, 3, m3, m1 ; [q2] -p1 -q1 +q2 +q3 %endif %if %2 == 16 UNSCRATCH 1, 8, rsp+%3+%4+16 %endif ; (m0: 0, [m1: flat8out], m2: fm & flat8in, m8..15: q2 q3 p1 p0 q0 q1 p3 p2) ; filter14() ; ; m2 m3 m8 m9 m14 m15 m10 m11 m12 m13 ; ; q2 q3 p3 p2 p1 p0 q0 q1 ; p6 -7 p7 p6 p5 p4 . . . . . ; p5 -6 -p7 -p6 +p5 +q1 . . . . ; p4 -5 -p7 -p5 +p4 +q2 . . . q2 ; p3 -4 -p7 -p4 +p3 +q3 . . . q3 ; p2 -3 -p7 -p3 +p2 +q4 . . . q4 ; p1 -2 -p7 -p2 +p1 +q5 . . . q5 ; p0 -1 -p7 -p1 +p0 +q6 . . . q6 ; q0 +0 -p7 -p0 +q0 +q7 . . . q7 ; q1 +1 -p6 -q0 +q1 +q7 q1 . . . ; q2 +2 -p5 -q1 +q2 +q7 . q2 . . ; q3 +3 -p4 -q2 +q3 +q7 . q3 . . ; q4 +4 -p3 -q3 +q4 +q7 . q4 . . ; q5 +5 -p2 -q4 +q5 +q7 . q5 . . ; q6 +6 -p1 -q5 +q6 +q7 . q6 . . %if %2 == 16 pand m1, m2 ; mask(out) & (mask(fm) & mask(in)) mova m2, [P7] mova m3, [P6] %if ARCH_X86_64 mova m8, [P5] mova m9, [P4] %define rp5 m8 %define rp4 m9 %define rp5s m8 %define rp4s m9 %define rp3s m14 %define rq4 m8 %define rq5 m9 %define rq6 m14 %define rq7 m15 %define rq4s m8 %define rq5s m9 %define rq6s m14 %else %define rp5 [P5] %define rp4 [P4] %define rp5s "" %define rp4s "" %define rp3s "" %define rq4 [Q4] %define rq5 [Q5] %define rq6 [Q6] %define rq7 [Q7] %define rq4s "" %define rq5s "" %define rq6s "" %endif FILTER_INIT m4, m5, m6, m7, [P6], %4, 14, m1, m3 ; [p6] FILTER_UPDATE m4, m5, m6, m7, [P5], %4, 8, 9, 10, 5, 4, m1, rp5s ; [p5] -p7 -p6 +p5 +q1 FILTER_UPDATE m4, m5, m6, m7, [P4], %4, 8, 10, 11, 6, 4, m1, rp4s ; [p4] -p7 -p5 +p4 +q2 FILTER_UPDATE m4, m5, m6, m7, [P3], %4, 8, 11, 0, 7, 4, m1, rp3s ; [p3] -p7 -p4 +p3 +q3 FILTER_UPDATE m4, m5, m6, m7, [P2], %4, 8, 0, 1, 12, 4, m1, "", rq4, [Q4], 1 ; [p2] -p7 -p3 +p2 +q4 FILTER_UPDATE m4, m5, m6, m7, [P1], %4, 8, 1, 2, 13, 4, m1, "", rq5, [Q5], 1 ; [p1] -p7 -p2 +p1 +q5 FILTER_UPDATE m4, m5, m6, m7, [P0], %4, 8, 2, 3, 14, 4, m1, "", rq6, [Q6], 1 ; [p0] -p7 -p1 +p0 +q6 FILTER_UPDATE m4, m5, m6, m7, [Q0], %4, 8, 3, 4, 15, 4, m1, "", rq7, [Q7], 1 ; [q0] -p7 -p0 +q0 +q7 FILTER_UPDATE m4, m5, m6, m7, [Q1], %4, 9, 4, 5, 15, 4, m1, "" ; [q1] -p6 -q0 +q1 +q7 FILTER_UPDATE m4, m5, m6, m7, [Q2], %4, 10, 5, 6, 15, 4, m1, "" ; [q2] -p5 -q1 +q2 +q7 FILTER_UPDATE m4, m5, m6, m7, [Q3], %4, 11, 6, 7, 15, 4, m1, "" ; [q3] -p4 -q2 +q3 +q7 FILTER_UPDATE m4, m5, m6, m7, [Q4], %4, 0, 7, 12, 15, 4, m1, rq4s ; [q4] -p3 -q3 +q4 +q7 FILTER_UPDATE m4, m5, m6, m7, [Q5], %4, 1, 12, 13, 15, 4, m1, rq5s ; [q5] -p2 -q4 +q5 +q7 FILTER_UPDATE m4, m5, m6, m7, [Q6], %4, 2, 13, 14, 15, 4, m1, rq6s ; [q6] -p1 -q5 +q6 +q7 %endif %ifidn %1, h %if %2 == 16 mova m0, [P7] mova m1, [P6] mova m2, [P5] mova m3, [P4] mova m4, [P3] mova m5, [P2] mova m6, [P1] mova m7, [P0] mova m8, [Q0] mova m9, [Q1] mova m10, [Q2] mova m11, [Q3] mova m12, [Q4] mova m13, [Q5] mova m14, [Q6] mova m15, [Q7] TRANSPOSE16x16B 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, [rsp] DEFINE_REAL_P7_TO_Q7 movu [P7], m0 movu [P6], m1 movu [P5], m2 movu [P4], m3 movu [P3], m4 movu [P2], m5 movu [P1], m6 movu [P0], m7 movu [Q0], m8 movu [Q1], m9 movu [Q2], m10 movu [Q3], m11 movu [Q4], m12 movu [Q5], m13 movu [Q6], m14 movu [Q7], m15 %elif %2 == 44 SWAP 0, 1 ; m0 = p1 SWAP 1, 7 ; m1 = p0 SWAP 2, 5 ; m2 = q0 SWAP 3, 4 ; m3 = q1 DEFINE_REAL_P7_TO_Q7 2 SBUTTERFLY bw, 0, 1, 4 SBUTTERFLY bw, 2, 3, 4 SBUTTERFLY wd, 0, 2, 4 SBUTTERFLY wd, 1, 3, 4 movd [P7], m0 movd [P3], m2 movd [Q0], m1 movd [Q4], m3 psrldq m0, 4 psrldq m1, 4 psrldq m2, 4 psrldq m3, 4 movd [P6], m0 movd [P2], m2 movd [Q1], m1 movd [Q5], m3 psrldq m0, 4 psrldq m1, 4 psrldq m2, 4 psrldq m3, 4 movd [P5], m0 movd [P1], m2 movd [Q2], m1 movd [Q6], m3 psrldq m0, 4 psrldq m1, 4 psrldq m2, 4 psrldq m3, 4 movd [P4], m0 movd [P0], m2 movd [Q3], m1 movd [Q7], m3 %else ; the following code do a transpose of 8 full lines to 16 half ; lines (high part). It is inlined to avoid the need of a staging area mova m0, [P3] mova m1, [P2] mova m2, [P1] mova m3, [P0] mova m4, [Q0] mova m5, [Q1] %if ARCH_X86_64 mova m6, [Q2] %endif mova m7, [Q3] DEFINE_REAL_P7_TO_Q7 %if ARCH_X86_64 SBUTTERFLY bw, 0, 1, 8 SBUTTERFLY bw, 2, 3, 8 SBUTTERFLY bw, 4, 5, 8 SBUTTERFLY bw, 6, 7, 8 SBUTTERFLY wd, 0, 2, 8 SBUTTERFLY wd, 1, 3, 8 SBUTTERFLY wd, 4, 6, 8 SBUTTERFLY wd, 5, 7, 8 SBUTTERFLY dq, 0, 4, 8 SBUTTERFLY dq, 1, 5, 8 SBUTTERFLY dq, 2, 6, 8 SBUTTERFLY dq, 3, 7, 8 %else SBUTTERFLY bw, 0, 1, 6 mova [rsp+64], m1 mova m6, [rsp+96] SBUTTERFLY bw, 2, 3, 1 SBUTTERFLY bw, 4, 5, 1 SBUTTERFLY bw, 6, 7, 1 SBUTTERFLY wd, 0, 2, 1 mova [rsp+96], m2 mova m1, [rsp+64] SBUTTERFLY wd, 1, 3, 2 SBUTTERFLY wd, 4, 6, 2 SBUTTERFLY wd, 5, 7, 2 SBUTTERFLY dq, 0, 4, 2 SBUTTERFLY dq, 1, 5, 2 movh [Q0], m1 movhps [Q1], m1 mova m2, [rsp+96] SBUTTERFLY dq, 2, 6, 1 SBUTTERFLY dq, 3, 7, 1 %endif SWAP 3, 6 SWAP 1, 4 movh [P7], m0 movhps [P6], m0 movh [P5], m1 movhps [P4], m1 movh [P3], m2 movhps [P2], m2 movh [P1], m3 movhps [P0], m3 %if ARCH_X86_64 movh [Q0], m4 movhps [Q1], m4 %endif movh [Q2], m5 movhps [Q3], m5 movh [Q4], m6 movhps [Q5], m6 movh [Q6], m7 movhps [Q7], m7 %endif %endif RET %endmacro %macro LPF_16_VH 5 INIT_XMM %5 LOOPFILTER v, %1, %2, 0, %4 %if ARCH_X86_64 || %1 != 16 LOOPFILTER h, %1, %2, %3, %4 %endif %endmacro %macro LPF_16_VH_ALL_OPTS 4 LPF_16_VH %1, %2, %3, %4, sse2 LPF_16_VH %1, %2, %3, %4, ssse3 LPF_16_VH %1, %2, %3, %4, avx %endmacro LPF_16_VH_ALL_OPTS 16, 512, 256, 32 LPF_16_VH_ALL_OPTS 44, 0, 128, 0 LPF_16_VH_ALL_OPTS 48, 256, 128, 16 LPF_16_VH_ALL_OPTS 84, 256, 128, 16 LPF_16_VH_ALL_OPTS 88, 256, 128, 16