15574c505b
Original x86 intrinsics code by Pierre-Edouard Lepere. Yasm port, refactoring and optimizations by James Almer. Benchmarks of BQTerrace_1920x1080_60_qp22.bin with an Intel Core i5-4200U Width 32 342694 decicycles in sao_edge_filter_10, 16384 runs, 0 skips 29476 decicycles in ff_hevc_sao_edge_filter_32_10_ssse3, 16384 runs, 0 skips 13996 decicycles in ff_hevc_sao_edge_filter_32_10_avx2, 16381 runs, 3 skips Width 64 581163 decicycles in sao_edge_filter_10, 8192 runs, 0 skips 59774 decicycles in ff_hevc_sao_edge_filter_64_10_ssse3, 8192 runs, 0 skips 28383 decicycles in ff_hevc_sao_edge_filter_64_10_avx2, 8191 runs, 1 skips Signed-off-by: James Almer <jamrial@gmail.com>
593 lines
17 KiB
NASM
593 lines
17 KiB
NASM
;******************************************************************************
|
|
;* SIMD optimized SAO functions for HEVC decoding
|
|
;*
|
|
;* Copyright (c) 2013 Pierre-Edouard LEPERE
|
|
;* Copyright (c) 2014 James Almer
|
|
;*
|
|
;* This file is part of FFmpeg.
|
|
;*
|
|
;* FFmpeg is free software; you can redistribute it and/or
|
|
;* modify it under the terms of the GNU Lesser General Public
|
|
;* License as published by the Free Software Foundation; either
|
|
;* version 2.1 of the License, or (at your option) any later version.
|
|
;*
|
|
;* FFmpeg is distributed in the hope that it will be useful,
|
|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
;* Lesser General Public License for more details.
|
|
;*
|
|
;* You should have received a copy of the GNU Lesser General Public
|
|
;* License along with FFmpeg; if not, write to the Free Software
|
|
;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
;******************************************************************************
|
|
|
|
%include "libavutil/x86/x86util.asm"
|
|
|
|
SECTION_RODATA 32
|
|
|
|
pw_mask10: times 16 dw 0x03FF
|
|
pw_mask12: times 16 dw 0x0FFF
|
|
pb_2: times 32 db 2
|
|
pw_m2: times 16 dw -2
|
|
pb_edge_shuffle: times 2 db 1, 2, 0, 3, 4, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1
|
|
pb_eo: db -1, 0, 1, 0, 0, -1, 0, 1, -1, -1, 1, 1, 1, -1, -1, 1
|
|
cextern pw_m1
|
|
cextern pw_1
|
|
cextern pw_2
|
|
cextern pb_1
|
|
|
|
SECTION_TEXT
|
|
|
|
%define MAX_PB_SIZE 64
|
|
%define PADDING_SIZE 32 ; FF_INPUT_BUFFER_PADDING_SIZE
|
|
|
|
;******************************************************************************
|
|
;SAO Band Filter
|
|
;******************************************************************************
|
|
|
|
%if ARCH_X86_64
|
|
%macro HEVC_SAO_BAND_FILTER_INIT 1
|
|
and leftq, 31
|
|
movd xm0, leftd
|
|
add leftq, 1
|
|
and leftq, 31
|
|
movd xm1, leftd
|
|
add leftq, 1
|
|
and leftq, 31
|
|
movd xm2, leftd
|
|
add leftq, 1
|
|
and leftq, 31
|
|
movd xm3, leftd
|
|
|
|
SPLATW m0, xm0
|
|
SPLATW m1, xm1
|
|
SPLATW m2, xm2
|
|
SPLATW m3, xm3
|
|
%if mmsize > 16
|
|
SPLATW m4, [offsetq + 2]
|
|
SPLATW m5, [offsetq + 4]
|
|
SPLATW m6, [offsetq + 6]
|
|
SPLATW m7, [offsetq + 8]
|
|
%else
|
|
movq m7, [offsetq + 2]
|
|
SPLATW m4, m7, 0
|
|
SPLATW m5, m7, 1
|
|
SPLATW m6, m7, 2
|
|
SPLATW m7, m7, 3
|
|
%endif
|
|
|
|
%if %1 > 8
|
|
mova m13, [pw_mask %+ %1]
|
|
%endif
|
|
pxor m14, m14
|
|
|
|
DEFINE_ARGS dst, src, dststride, srcstride, offset, height
|
|
mov heightd, r7m
|
|
%endmacro
|
|
|
|
%macro HEVC_SAO_BAND_FILTER_COMPUTE 3
|
|
psraw %2, %3, %1-5
|
|
pcmpeqw m10, %2, m0
|
|
pcmpeqw m11, %2, m1
|
|
pcmpeqw m12, %2, m2
|
|
pcmpeqw %2, m3
|
|
pand m10, m4
|
|
pand m11, m5
|
|
pand m12, m6
|
|
pand %2, m7
|
|
por m10, m11
|
|
por m12, %2
|
|
por m10, m12
|
|
paddw %3, m10
|
|
%endmacro
|
|
|
|
;void ff_hevc_sao_band_filter_<width>_8_<opt>(uint8_t *_dst, uint8_t *_src, ptrdiff_t _stride_dst, ptrdiff_t _stride_src,
|
|
; int16_t *sao_offset_val, int sao_left_class, int width, int height);
|
|
%macro HEVC_SAO_BAND_FILTER_8 2
|
|
cglobal hevc_sao_band_filter_%1_8, 6, 6, 15, dst, src, dststride, srcstride, offset, left
|
|
HEVC_SAO_BAND_FILTER_INIT 8
|
|
|
|
align 16
|
|
.loop
|
|
%if %1 == 8
|
|
movq m8, [srcq]
|
|
punpcklbw m8, m14
|
|
HEVC_SAO_BAND_FILTER_COMPUTE 8, m9, m8
|
|
packuswb m8, m14
|
|
movq [dstq], m8
|
|
%endif ; %1 == 8
|
|
|
|
%assign i 0
|
|
%rep %2
|
|
mova m13, [srcq + i]
|
|
punpcklbw m8, m13, m14
|
|
HEVC_SAO_BAND_FILTER_COMPUTE 8, m9, m8
|
|
punpckhbw m13, m14
|
|
HEVC_SAO_BAND_FILTER_COMPUTE 8, m9, m13
|
|
packuswb m8, m13
|
|
mova [dstq + i], m8
|
|
%assign i i+mmsize
|
|
%endrep
|
|
|
|
%if %1 == 48
|
|
INIT_XMM cpuname
|
|
|
|
mova m13, [srcq + i]
|
|
punpcklbw m8, m13, m14
|
|
HEVC_SAO_BAND_FILTER_COMPUTE 8, m9, m8
|
|
punpckhbw m13, m14
|
|
HEVC_SAO_BAND_FILTER_COMPUTE 8, m9, m13
|
|
packuswb m8, m13
|
|
mova [dstq + i], m8
|
|
%if cpuflag(avx2)
|
|
INIT_YMM cpuname
|
|
%endif
|
|
%endif ; %1 == 48
|
|
|
|
add dstq, dststrideq ; dst += dststride
|
|
add srcq, srcstrideq ; src += srcstride
|
|
dec heightd ; cmp height
|
|
jnz .loop ; height loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
;void ff_hevc_sao_band_filter_<width>_<depth>_<opt>(uint8_t *_dst, uint8_t *_src, ptrdiff_t _stride_dst, ptrdiff_t _stride_src,
|
|
; int16_t *sao_offset_val, int sao_left_class, int width, int height);
|
|
%macro HEVC_SAO_BAND_FILTER_16 3
|
|
cglobal hevc_sao_band_filter_%2_%1, 6, 6, 15, dst, src, dststride, srcstride, offset, left
|
|
HEVC_SAO_BAND_FILTER_INIT %1
|
|
|
|
align 16
|
|
.loop
|
|
%if %2 == 8
|
|
mova m8, [srcq]
|
|
HEVC_SAO_BAND_FILTER_COMPUTE %1, m9, m8
|
|
CLIPW m8, m14, m13
|
|
mova [dstq], m8
|
|
%endif
|
|
|
|
%assign i 0
|
|
%rep %3
|
|
mova m8, [srcq + i]
|
|
HEVC_SAO_BAND_FILTER_COMPUTE %1, m9, m8
|
|
CLIPW m8, m14, m13
|
|
mova [dstq + i], m8
|
|
|
|
mova m9, [srcq + i + mmsize]
|
|
HEVC_SAO_BAND_FILTER_COMPUTE %1, m8, m9
|
|
CLIPW m9, m14, m13
|
|
mova [dstq + i + mmsize], m9
|
|
%assign i i+mmsize*2
|
|
%endrep
|
|
|
|
%if %2 == 48
|
|
INIT_XMM cpuname
|
|
mova m8, [srcq + i]
|
|
HEVC_SAO_BAND_FILTER_COMPUTE %1, m9, m8
|
|
CLIPW m8, m14, m13
|
|
mova [dstq + i], m8
|
|
|
|
mova m9, [srcq + i + mmsize]
|
|
HEVC_SAO_BAND_FILTER_COMPUTE %1, m8, m9
|
|
CLIPW m9, m14, m13
|
|
mova [dstq + i + mmsize], m9
|
|
%if cpuflag(avx2)
|
|
INIT_YMM cpuname
|
|
%endif
|
|
%endif ; %1 == 48
|
|
|
|
add dstq, dststrideq
|
|
add srcq, srcstrideq
|
|
dec heightd
|
|
jg .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
%macro HEVC_SAO_BAND_FILTER_FUNCS 0
|
|
HEVC_SAO_BAND_FILTER_8 8, 0
|
|
HEVC_SAO_BAND_FILTER_8 16, 1
|
|
HEVC_SAO_BAND_FILTER_8 32, 2
|
|
HEVC_SAO_BAND_FILTER_8 48, 2
|
|
HEVC_SAO_BAND_FILTER_8 64, 4
|
|
|
|
HEVC_SAO_BAND_FILTER_16 10, 8, 0
|
|
HEVC_SAO_BAND_FILTER_16 10, 16, 1
|
|
HEVC_SAO_BAND_FILTER_16 10, 32, 2
|
|
HEVC_SAO_BAND_FILTER_16 10, 48, 2
|
|
HEVC_SAO_BAND_FILTER_16 10, 64, 4
|
|
|
|
HEVC_SAO_BAND_FILTER_16 12, 8, 0
|
|
HEVC_SAO_BAND_FILTER_16 12, 16, 1
|
|
HEVC_SAO_BAND_FILTER_16 12, 32, 2
|
|
HEVC_SAO_BAND_FILTER_16 12, 48, 2
|
|
HEVC_SAO_BAND_FILTER_16 12, 64, 4
|
|
%endmacro
|
|
|
|
INIT_XMM sse2
|
|
HEVC_SAO_BAND_FILTER_FUNCS
|
|
INIT_XMM avx
|
|
HEVC_SAO_BAND_FILTER_FUNCS
|
|
|
|
%if HAVE_AVX2_EXTERNAL
|
|
INIT_XMM avx2
|
|
HEVC_SAO_BAND_FILTER_8 8, 0
|
|
HEVC_SAO_BAND_FILTER_8 16, 1
|
|
INIT_YMM avx2
|
|
HEVC_SAO_BAND_FILTER_8 32, 1
|
|
HEVC_SAO_BAND_FILTER_8 48, 1
|
|
HEVC_SAO_BAND_FILTER_8 64, 2
|
|
|
|
INIT_XMM avx2
|
|
HEVC_SAO_BAND_FILTER_16 10, 8, 0
|
|
HEVC_SAO_BAND_FILTER_16 10, 16, 1
|
|
INIT_YMM avx2
|
|
HEVC_SAO_BAND_FILTER_16 10, 32, 1
|
|
HEVC_SAO_BAND_FILTER_16 10, 48, 1
|
|
HEVC_SAO_BAND_FILTER_16 10, 64, 2
|
|
|
|
INIT_XMM avx2
|
|
HEVC_SAO_BAND_FILTER_16 12, 8, 0
|
|
HEVC_SAO_BAND_FILTER_16 12, 16, 1
|
|
INIT_YMM avx2
|
|
HEVC_SAO_BAND_FILTER_16 12, 32, 1
|
|
HEVC_SAO_BAND_FILTER_16 12, 48, 1
|
|
HEVC_SAO_BAND_FILTER_16 12, 64, 2
|
|
%endif
|
|
%endif
|
|
|
|
;******************************************************************************
|
|
;SAO Edge Filter
|
|
;******************************************************************************
|
|
|
|
%define EDGE_SRCSTRIDE 2 * MAX_PB_SIZE + PADDING_SIZE
|
|
|
|
%macro HEVC_SAO_EDGE_FILTER_COMPUTE_8 1
|
|
pminub m4, m1, m2
|
|
pminub m5, m1, m3
|
|
pcmpeqb m2, m4
|
|
pcmpeqb m3, m5
|
|
pcmpeqb m4, m1
|
|
pcmpeqb m5, m1
|
|
psubb m4, m2
|
|
psubb m5, m3
|
|
paddb m4, m6
|
|
paddb m4, m5
|
|
|
|
pshufb m2, m0, m4
|
|
%if %1 > 8
|
|
punpckhbw m5, m7, m1
|
|
punpckhbw m4, m2, m7
|
|
punpcklbw m3, m7, m1
|
|
punpcklbw m2, m7
|
|
pmaddubsw m5, m4
|
|
pmaddubsw m3, m2
|
|
packuswb m3, m5
|
|
%else
|
|
punpcklbw m3, m7, m1
|
|
punpcklbw m2, m7
|
|
pmaddubsw m3, m2
|
|
packuswb m3, m3
|
|
%endif
|
|
%endmacro
|
|
|
|
;void ff_hevc_sao_edge_filter_<width>_8_<opt>(uint8_t *_dst, uint8_t *_src, ptrdiff_t stride_dst, int16_t *sao_offset_val,
|
|
; int eo, int width, int height);
|
|
%macro HEVC_SAO_EDGE_FILTER_8 2-3
|
|
%if WIN64
|
|
cglobal hevc_sao_edge_filter_%1_8, 4, 8, 8, dst, src, dststride, offset, a_stride, b_stride, height, tmp
|
|
%define eoq heightq
|
|
movsxd eoq, dword r4m
|
|
movsx a_strideq, byte [pb_eo+eoq*4+1]
|
|
movsx b_strideq, byte [pb_eo+eoq*4+3]
|
|
imul a_strideq, EDGE_SRCSTRIDE
|
|
imul b_strideq, EDGE_SRCSTRIDE
|
|
movsx tmpq, byte [pb_eo+eoq*4]
|
|
add a_strideq, tmpq
|
|
movsx tmpq, byte [pb_eo+eoq*4+2]
|
|
add b_strideq, tmpq
|
|
mov heightd, r6m
|
|
|
|
%elif ARCH_X86_64
|
|
cglobal hevc_sao_edge_filter_%1_8, 5, 9, 8, dst, src, dststride, offset, eo, a_stride, b_stride, height, tmp
|
|
%define tmp2q heightq
|
|
movsxd eoq, eod
|
|
lea tmp2q, [pb_eo]
|
|
movsx a_strideq, byte [tmp2q+eoq*4+1]
|
|
movsx b_strideq, byte [tmp2q+eoq*4+3]
|
|
imul a_strideq, EDGE_SRCSTRIDE
|
|
imul b_strideq, EDGE_SRCSTRIDE
|
|
movsx tmpq, byte [tmp2q+eoq*4]
|
|
add a_strideq, tmpq
|
|
movsx tmpq, byte [tmp2q+eoq*4+2]
|
|
add b_strideq, tmpq
|
|
mov heightd, r6m
|
|
|
|
%else ; ARCH_X86_32
|
|
cglobal hevc_sao_edge_filter_%1_8, 1, 6, 8, dst, src, dststride, a_stride, b_stride, height
|
|
%define eoq srcq
|
|
%define tmpq heightq
|
|
%define tmp2q dststrideq
|
|
%define offsetq heightq
|
|
mov eoq, r4m
|
|
lea tmp2q, [pb_eo]
|
|
movsx a_strideq, byte [tmp2q+eoq*4+1]
|
|
movsx b_strideq, byte [tmp2q+eoq*4+3]
|
|
imul a_strideq, EDGE_SRCSTRIDE
|
|
imul b_strideq, EDGE_SRCSTRIDE
|
|
movsx tmpq, byte [tmp2q+eoq*4]
|
|
add a_strideq, tmpq
|
|
movsx tmpq, byte [tmp2q+eoq*4+2]
|
|
add b_strideq, tmpq
|
|
|
|
mov srcq, srcm
|
|
mov offsetq, r3m
|
|
mov dststrideq, dststridem
|
|
%endif ; ARCH
|
|
|
|
%if mmsize > 16
|
|
vbroadcasti128 m0, [offsetq]
|
|
%else
|
|
movu m0, [offsetq]
|
|
%endif
|
|
mova m1, [pb_edge_shuffle]
|
|
packsswb m0, m0
|
|
mova m7, [pb_1]
|
|
pshufb m0, m1
|
|
mova m6, [pb_2]
|
|
%if ARCH_X86_32
|
|
mov heightd, r6m
|
|
%endif
|
|
|
|
align 16
|
|
.loop:
|
|
|
|
%if %1 == 8
|
|
movq m1, [srcq]
|
|
movq m2, [srcq + a_strideq]
|
|
movq m3, [srcq + b_strideq]
|
|
HEVC_SAO_EDGE_FILTER_COMPUTE_8 %1
|
|
movq [dstq], m3
|
|
%endif
|
|
|
|
%assign i 0
|
|
%rep %2
|
|
mova m1, [srcq + i]
|
|
movu m2, [srcq + a_strideq + i]
|
|
movu m3, [srcq + b_strideq + i]
|
|
HEVC_SAO_EDGE_FILTER_COMPUTE_8 %1
|
|
mov%3 [dstq + i], m3
|
|
%assign i i+mmsize
|
|
%endrep
|
|
|
|
%if %1 == 48
|
|
INIT_XMM cpuname
|
|
|
|
mova m1, [srcq + i]
|
|
movu m2, [srcq + a_strideq + i]
|
|
movu m3, [srcq + b_strideq + i]
|
|
HEVC_SAO_EDGE_FILTER_COMPUTE_8 %1
|
|
mova [dstq + i], m3
|
|
%if cpuflag(avx2)
|
|
INIT_YMM cpuname
|
|
%endif
|
|
%endif
|
|
|
|
add dstq, dststrideq
|
|
add srcq, EDGE_SRCSTRIDE
|
|
dec heightd
|
|
jg .loop
|
|
RET
|
|
%endmacro
|
|
|
|
%macro PMINUW 4
|
|
%if cpuflag(sse4)
|
|
pminuw %1, %2, %3
|
|
%else
|
|
psubusw %4, %2, %3
|
|
psubw %1, %2, %4
|
|
%endif
|
|
%endmacro
|
|
|
|
%macro HEVC_SAO_EDGE_FILTER_COMPUTE_10 0
|
|
PMINUW m4, m1, m2, m6
|
|
PMINUW m5, m1, m3, m7
|
|
pcmpeqw m2, m4
|
|
pcmpeqw m3, m5
|
|
pcmpeqw m4, m1
|
|
pcmpeqw m5, m1
|
|
psubw m4, m2
|
|
psubw m5, m3
|
|
|
|
paddw m4, m5
|
|
pcmpeqw m2, m4, [pw_m2]
|
|
pcmpeqw m3, m4, m13
|
|
pcmpeqw m5, m4, m0
|
|
pcmpeqw m6, m4, m14
|
|
pcmpeqw m7, m4, m15
|
|
pand m2, m8
|
|
pand m3, m9
|
|
pand m5, m10
|
|
pand m6, m11
|
|
pand m7, m12
|
|
paddw m2, m3
|
|
paddw m5, m6
|
|
paddw m2, m7
|
|
paddw m2, m1
|
|
paddw m2, m5
|
|
%endmacro
|
|
|
|
;void ff_hevc_sao_edge_filter_<width>_<depth>_<opt>(uint8_t *_dst, uint8_t *_src, ptrdiff_t stride_dst, int16_t *sao_offset_val,
|
|
; int eo, int width, int height);
|
|
%macro HEVC_SAO_EDGE_FILTER_16 3
|
|
%if WIN64
|
|
cglobal hevc_sao_edge_filter_%2_%1, 4, 8, 16, dst, src, dststride, offset, a_stride, b_stride, height, tmp
|
|
%define eoq heightq
|
|
movsxd eoq, dword r4m
|
|
movsx a_strideq, byte [pb_eo+eoq*4+1]
|
|
movsx b_strideq, byte [pb_eo+eoq*4+3]
|
|
imul a_strideq, EDGE_SRCSTRIDE>>1
|
|
imul b_strideq, EDGE_SRCSTRIDE>>1
|
|
movsx tmpq, byte [pb_eo+eoq*4]
|
|
add a_strideq, tmpq
|
|
movsx tmpq, byte [pb_eo+eoq*4+2]
|
|
add b_strideq, tmpq
|
|
mov heightd, r6m
|
|
add a_strideq, a_strideq
|
|
add b_strideq, b_strideq
|
|
|
|
%else ; UNIX64
|
|
cglobal hevc_sao_edge_filter_%2_%1, 5, 9, 16, dst, src, dststride, offset, eo, a_stride, b_stride, height, tmp
|
|
%define tmp2q heightq
|
|
movsxd eoq, eod
|
|
lea tmp2q, [pb_eo]
|
|
movsx a_strideq, byte [tmp2q+eoq*4+1]
|
|
movsx b_strideq, byte [tmp2q+eoq*4+3]
|
|
imul a_strideq, EDGE_SRCSTRIDE>>1
|
|
imul b_strideq, EDGE_SRCSTRIDE>>1
|
|
movsx tmpq, byte [tmp2q+eoq*4]
|
|
add a_strideq, tmpq
|
|
movsx tmpq, byte [tmp2q+eoq*4+2]
|
|
add b_strideq, tmpq
|
|
mov heightd, r6m
|
|
add a_strideq, a_strideq
|
|
add b_strideq, b_strideq
|
|
%endif ; ARCH
|
|
|
|
%if cpuflag(avx2)
|
|
SPLATW m8, [offsetq+2]
|
|
SPLATW m9, [offsetq+4]
|
|
SPLATW m10, [offsetq+0]
|
|
SPLATW m11, [offsetq+6]
|
|
SPLATW m12, [offsetq+8]
|
|
%else
|
|
movq m10, [offsetq+0]
|
|
movd m12, [offsetq+6]
|
|
SPLATW m8, xm10, 1
|
|
SPLATW m9, xm10, 2
|
|
SPLATW m10, xm10, 0
|
|
SPLATW m11, xm12, 0
|
|
SPLATW m12, xm12, 1
|
|
%endif
|
|
pxor m0, m0
|
|
mova m13, [pw_m1]
|
|
mova m14, [pw_1]
|
|
mova m15, [pw_2]
|
|
|
|
align 16
|
|
.loop
|
|
|
|
%if %2 == 8
|
|
mova m1, [srcq]
|
|
movu m2, [srcq+a_strideq]
|
|
movu m3, [srcq+b_strideq]
|
|
|
|
HEVC_SAO_EDGE_FILTER_COMPUTE_10
|
|
CLIPW m2, m0, [pw_mask %+ %1]
|
|
mova [dstq], m2
|
|
%endif
|
|
|
|
%assign i 0
|
|
%rep %3
|
|
mova m1, [srcq + i]
|
|
movu m2, [srcq+a_strideq + i]
|
|
movu m3, [srcq+b_strideq + i]
|
|
HEVC_SAO_EDGE_FILTER_COMPUTE_10
|
|
CLIPW m2, m0, [pw_mask %+ %1]
|
|
mova [dstq + i], m2
|
|
|
|
mova m1, [srcq + i + mmsize]
|
|
movu m2, [srcq+a_strideq + i + mmsize]
|
|
movu m3, [srcq+b_strideq + i + mmsize]
|
|
HEVC_SAO_EDGE_FILTER_COMPUTE_10
|
|
CLIPW m2, m0, [pw_mask %+ %1]
|
|
mova [dstq + i + mmsize], m2
|
|
%assign i i+mmsize*2
|
|
%endrep
|
|
|
|
%if %2 == 48
|
|
INIT_XMM cpuname
|
|
mova m1, [srcq + i]
|
|
movu m2, [srcq+a_strideq + i]
|
|
movu m3, [srcq+b_strideq + i]
|
|
HEVC_SAO_EDGE_FILTER_COMPUTE_10
|
|
CLIPW m2, m0, [pw_mask %+ %1]
|
|
mova [dstq + i], m2
|
|
|
|
mova m1, [srcq + i + mmsize]
|
|
movu m2, [srcq+a_strideq + i + mmsize]
|
|
movu m3, [srcq+b_strideq + i + mmsize]
|
|
HEVC_SAO_EDGE_FILTER_COMPUTE_10
|
|
CLIPW m2, m0, [pw_mask %+ %1]
|
|
mova [dstq + i + mmsize], m2
|
|
%if cpuflag(avx2)
|
|
INIT_YMM cpuname
|
|
%endif
|
|
%endif
|
|
|
|
add dstq, dststrideq
|
|
add srcq, EDGE_SRCSTRIDE
|
|
dec heightd
|
|
jg .loop
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_XMM ssse3
|
|
HEVC_SAO_EDGE_FILTER_8 8, 0
|
|
HEVC_SAO_EDGE_FILTER_8 16, 1, a
|
|
HEVC_SAO_EDGE_FILTER_8 32, 2, a
|
|
HEVC_SAO_EDGE_FILTER_8 48, 2, a
|
|
HEVC_SAO_EDGE_FILTER_8 64, 4, a
|
|
|
|
%if HAVE_AVX2_EXTERNAL
|
|
INIT_YMM avx2
|
|
HEVC_SAO_EDGE_FILTER_8 32, 1, a
|
|
HEVC_SAO_EDGE_FILTER_8 48, 1, u
|
|
HEVC_SAO_EDGE_FILTER_8 64, 2, a
|
|
%endif
|
|
|
|
%if ARCH_X86_64
|
|
INIT_XMM sse2
|
|
HEVC_SAO_EDGE_FILTER_16 10, 8, 0
|
|
HEVC_SAO_EDGE_FILTER_16 10, 16, 1
|
|
HEVC_SAO_EDGE_FILTER_16 10, 32, 2
|
|
HEVC_SAO_EDGE_FILTER_16 10, 48, 2
|
|
HEVC_SAO_EDGE_FILTER_16 10, 64, 4
|
|
|
|
HEVC_SAO_EDGE_FILTER_16 12, 8, 0
|
|
HEVC_SAO_EDGE_FILTER_16 12, 16, 1
|
|
HEVC_SAO_EDGE_FILTER_16 12, 32, 2
|
|
HEVC_SAO_EDGE_FILTER_16 12, 48, 2
|
|
HEVC_SAO_EDGE_FILTER_16 12, 64, 4
|
|
|
|
%if HAVE_AVX2_EXTERNAL
|
|
INIT_YMM avx2
|
|
HEVC_SAO_EDGE_FILTER_16 10, 32, 1
|
|
HEVC_SAO_EDGE_FILTER_16 10, 48, 1
|
|
HEVC_SAO_EDGE_FILTER_16 10, 64, 2
|
|
|
|
HEVC_SAO_EDGE_FILTER_16 12, 32, 1
|
|
HEVC_SAO_EDGE_FILTER_16 12, 48, 1
|
|
HEVC_SAO_EDGE_FILTER_16 12, 64, 2
|
|
%endif
|
|
%endif
|