2015-10-06 17:01:48 +02:00
|
|
|
;******************************************************************************
|
|
|
|
;* VP9 inverse transform x86 SIMD optimizations
|
|
|
|
;*
|
|
|
|
;* Copyright (C) 2015 Ronald S. Bultje <rsbultje gmail com>
|
|
|
|
;*
|
|
|
|
;* This file is part of FFmpeg.
|
|
|
|
;*
|
|
|
|
;* FFmpeg is free software; you can redistribute it and/or
|
|
|
|
;* modify it under the terms of the GNU Lesser General Public
|
|
|
|
;* License as published by the Free Software Foundation; either
|
|
|
|
;* version 2.1 of the License, or (at your option) any later version.
|
|
|
|
;*
|
|
|
|
;* FFmpeg is distributed in the hope that it will be useful,
|
|
|
|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
;* Lesser General Public License for more details.
|
|
|
|
;*
|
|
|
|
;* You should have received a copy of the GNU Lesser General Public
|
|
|
|
;* License along with FFmpeg; if not, write to the Free Software
|
|
|
|
;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
|
|
;******************************************************************************
|
|
|
|
|
|
|
|
%include "libavutil/x86/x86util.asm"
|
|
|
|
%include "vp9itxfm_template.asm"
|
|
|
|
|
|
|
|
SECTION_RODATA
|
|
|
|
|
2015-10-06 17:03:45 +02:00
|
|
|
cextern pw_8
|
2015-10-06 17:01:48 +02:00
|
|
|
cextern pw_1023
|
2015-10-06 17:03:45 +02:00
|
|
|
cextern pw_2048
|
2015-10-06 17:01:48 +02:00
|
|
|
cextern pw_4095
|
2015-10-07 21:42:52 +02:00
|
|
|
cextern pw_m1
|
|
|
|
cextern pd_1
|
2015-10-12 16:17:23 +02:00
|
|
|
cextern pd_16
|
2015-10-08 16:16:36 +02:00
|
|
|
cextern pd_32
|
2015-10-06 17:03:45 +02:00
|
|
|
cextern pd_8192
|
|
|
|
|
2015-10-12 16:16:56 +02:00
|
|
|
pd_8: times 4 dd 8
|
|
|
|
pd_3fff: times 4 dd 0x3fff
|
|
|
|
|
2015-10-06 17:03:45 +02:00
|
|
|
; FIXME these should probably be shared between 8bpp and 10/12bpp
|
|
|
|
pw_m11585_11585: times 4 dw -11585, 11585
|
|
|
|
pw_11585_11585: times 8 dw 11585
|
|
|
|
pw_m15137_6270: times 4 dw -15137, 6270
|
|
|
|
pw_6270_15137: times 4 dw 6270, 15137
|
|
|
|
pw_11585x2: times 8 dw 11585*2
|
2015-10-06 17:01:48 +02:00
|
|
|
|
2015-10-06 17:42:03 +02:00
|
|
|
pw_5283_13377: times 4 dw 5283, 13377
|
|
|
|
pw_9929_13377: times 4 dw 9929, 13377
|
|
|
|
pw_15212_m13377: times 4 dw 15212, -13377
|
|
|
|
pw_15212_9929: times 4 dw 15212, 9929
|
|
|
|
pw_m5283_m15212: times 4 dw -5283, -15212
|
|
|
|
pw_13377x2: times 8 dw 13377*2
|
|
|
|
pw_m13377_13377: times 4 dw -13377, 13377
|
|
|
|
pw_13377_0: times 4 dw 13377, 0
|
2015-10-07 03:59:10 +02:00
|
|
|
pw_9929_m5283: times 4 dw 9929, -5283
|
2015-10-06 17:42:03 +02:00
|
|
|
|
2015-10-12 16:17:23 +02:00
|
|
|
pw_3196_16069: times 4 dw 3196, 16069
|
|
|
|
pw_m16069_3196: times 4 dw -16069, 3196
|
|
|
|
pw_13623_9102: times 4 dw 13623, 9102
|
|
|
|
pw_m9102_13623: times 4 dw -9102, 13623
|
|
|
|
|
2015-10-07 21:42:52 +02:00
|
|
|
pw_1606_16305: times 4 dw 1606, 16305
|
|
|
|
pw_m16305_1606: times 4 dw -16305, 1606
|
|
|
|
pw_12665_10394: times 4 dw 12665, 10394
|
|
|
|
pw_m10394_12665: times 4 dw -10394, 12665
|
|
|
|
pw_7723_14449: times 4 dw 7723, 14449
|
|
|
|
pw_m14449_7723: times 4 dw -14449, 7723
|
|
|
|
pw_15679_4756: times 4 dw 15679, 4756
|
|
|
|
pw_m4756_15679: times 4 dw -4756, 15679
|
|
|
|
pw_15137_6270: times 4 dw 15137, 6270
|
|
|
|
pw_m6270_15137: times 4 dw -6270, 15137
|
|
|
|
|
2015-10-06 17:01:48 +02:00
|
|
|
SECTION .text
|
|
|
|
|
|
|
|
%macro VP9_STORE_2X 6-7 dstq ; reg1, reg2, tmp1, tmp2, min, max, dst
|
|
|
|
mova m%3, [%7]
|
|
|
|
mova m%4, [%7+strideq]
|
|
|
|
paddw m%3, m%1
|
|
|
|
paddw m%4, m%2
|
|
|
|
pmaxsw m%3, m%5
|
|
|
|
pmaxsw m%4, m%5
|
|
|
|
pminsw m%3, m%6
|
|
|
|
pminsw m%4, m%6
|
|
|
|
mova [%7], m%3
|
|
|
|
mova [%7+strideq], m%4
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
%macro ZERO_BLOCK 4 ; mem, stride, nnzcpl, zero_reg
|
|
|
|
%assign %%y 0
|
|
|
|
%rep %3
|
|
|
|
%assign %%x 0
|
|
|
|
%rep %3*4/mmsize
|
|
|
|
mova [%1+%%y+%%x], %4
|
|
|
|
%assign %%x (%%x+mmsize)
|
|
|
|
%endrep
|
|
|
|
%assign %%y (%%y+%2)
|
|
|
|
%endrep
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
; the input coefficients are scaled up by 2 bit (which we downscale immediately
|
|
|
|
; in the iwht), and is otherwise orthonormally increased by 1 bit per iwht_1d.
|
|
|
|
; therefore, a diff of 10-12+sign bit will fit in 12-14+sign bit after scaling,
|
|
|
|
; i.e. everything can be done in 15+1bpp words. Since the quant fractional bits
|
|
|
|
; add 2 bits, we need to scale before converting to word in 12bpp, since the
|
|
|
|
; input will be 16+sign bit which doesn't fit in 15+sign words, but in 10bpp
|
|
|
|
; we can scale after converting to words (which is half the instructions),
|
|
|
|
; since the input is only 14+sign bit, which fits in 15+sign words directly.
|
|
|
|
|
|
|
|
%macro IWHT4_FN 2 ; bpp, max
|
|
|
|
cglobal vp9_iwht_iwht_4x4_add_%1, 3, 3, 8, dst, stride, block, eob
|
|
|
|
mova m7, [pw_%2]
|
|
|
|
mova m0, [blockq+0*16+0]
|
|
|
|
mova m1, [blockq+1*16+0]
|
|
|
|
%if %1 >= 12
|
|
|
|
mova m4, [blockq+0*16+8]
|
|
|
|
mova m5, [blockq+1*16+8]
|
|
|
|
psrad m0, 2
|
|
|
|
psrad m1, 2
|
|
|
|
psrad m4, 2
|
|
|
|
psrad m5, 2
|
|
|
|
packssdw m0, m4
|
|
|
|
packssdw m1, m5
|
|
|
|
%else
|
|
|
|
packssdw m0, [blockq+0*16+8]
|
|
|
|
packssdw m1, [blockq+1*16+8]
|
|
|
|
psraw m0, 2
|
|
|
|
psraw m1, 2
|
|
|
|
%endif
|
|
|
|
mova m2, [blockq+2*16+0]
|
|
|
|
mova m3, [blockq+3*16+0]
|
|
|
|
%if %1 >= 12
|
|
|
|
mova m4, [blockq+2*16+8]
|
|
|
|
mova m5, [blockq+3*16+8]
|
|
|
|
psrad m2, 2
|
|
|
|
psrad m3, 2
|
|
|
|
psrad m4, 2
|
|
|
|
psrad m5, 2
|
|
|
|
packssdw m2, m4
|
|
|
|
packssdw m3, m5
|
|
|
|
%else
|
|
|
|
packssdw m2, [blockq+2*16+8]
|
|
|
|
packssdw m3, [blockq+3*16+8]
|
|
|
|
psraw m2, 2
|
|
|
|
psraw m3, 2
|
|
|
|
%endif
|
|
|
|
|
|
|
|
VP9_IWHT4_1D
|
|
|
|
TRANSPOSE4x4W 0, 1, 2, 3, 4
|
|
|
|
VP9_IWHT4_1D
|
|
|
|
|
|
|
|
pxor m6, m6
|
|
|
|
VP9_STORE_2X 0, 1, 4, 5, 6, 7
|
|
|
|
lea dstq, [dstq+strideq*2]
|
|
|
|
VP9_STORE_2X 2, 3, 4, 5, 6, 7
|
|
|
|
ZERO_BLOCK blockq, 16, 4, m6
|
|
|
|
RET
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
INIT_MMX mmxext
|
|
|
|
IWHT4_FN 10, 1023
|
|
|
|
INIT_MMX mmxext
|
|
|
|
IWHT4_FN 12, 4095
|
2015-10-06 17:03:45 +02:00
|
|
|
|
2015-10-06 17:42:03 +02:00
|
|
|
%macro VP9_IDCT4_WRITEOUT 0
|
|
|
|
%if cpuflag(ssse3)
|
|
|
|
mova m5, [pw_2048]
|
|
|
|
pmulhrsw m0, m5
|
|
|
|
pmulhrsw m1, m5
|
|
|
|
pmulhrsw m2, m5
|
|
|
|
pmulhrsw m3, m5
|
|
|
|
%else
|
|
|
|
mova m5, [pw_8]
|
|
|
|
paddw m0, m5
|
|
|
|
paddw m1, m5
|
|
|
|
paddw m2, m5
|
|
|
|
paddw m3, m5
|
|
|
|
psraw m0, 4
|
|
|
|
psraw m1, 4
|
|
|
|
psraw m2, 4
|
|
|
|
psraw m3, 4
|
|
|
|
%endif
|
|
|
|
mova m5, [pw_1023]
|
|
|
|
VP9_STORE_2X 0, 1, 6, 7, 4, 5
|
|
|
|
lea dstq, [dstq+2*strideq]
|
|
|
|
VP9_STORE_2X 2, 3, 6, 7, 4, 5
|
|
|
|
%endmacro
|
|
|
|
|
2015-10-06 17:03:45 +02:00
|
|
|
; 4x4 coefficients are 5+depth+sign bits, so for 10bpp, everything still fits
|
|
|
|
; in 15+1 words without additional effort, since the coefficients are 15bpp.
|
|
|
|
|
|
|
|
%macro IDCT4_10_FN 0
|
|
|
|
cglobal vp9_idct_idct_4x4_add_10, 4, 4, 8, dst, stride, block, eob
|
|
|
|
cmp eobd, 1
|
|
|
|
jg .idctfull
|
|
|
|
|
|
|
|
; dc-only
|
|
|
|
%if cpuflag(ssse3)
|
|
|
|
movd m0, [blockq]
|
|
|
|
mova m5, [pw_11585x2]
|
|
|
|
pmulhrsw m0, m5
|
|
|
|
pmulhrsw m0, m5
|
|
|
|
%else
|
|
|
|
DEFINE_ARGS dst, stride, block, coef
|
|
|
|
mov coefd, dword [blockq]
|
|
|
|
imul coefd, 11585
|
|
|
|
add coefd, 8192
|
|
|
|
sar coefd, 14
|
|
|
|
imul coefd, 11585
|
|
|
|
add coefd, (8 << 14) + 8192
|
|
|
|
sar coefd, 14 + 4
|
|
|
|
movd m0, coefd
|
|
|
|
%endif
|
|
|
|
pshufw m0, m0, 0
|
|
|
|
pxor m4, m4
|
|
|
|
mova m5, [pw_1023]
|
|
|
|
movh [blockq], m4
|
|
|
|
%if cpuflag(ssse3)
|
|
|
|
pmulhrsw m0, [pw_2048] ; (x*2048 + (1<<14))>>15 <=> (x+8)>>4
|
|
|
|
%endif
|
|
|
|
VP9_STORE_2X 0, 0, 6, 7, 4, 5
|
|
|
|
lea dstq, [dstq+2*strideq]
|
|
|
|
VP9_STORE_2X 0, 0, 6, 7, 4, 5
|
|
|
|
RET
|
|
|
|
|
|
|
|
.idctfull:
|
|
|
|
mova m0, [blockq+0*16+0]
|
|
|
|
mova m1, [blockq+1*16+0]
|
|
|
|
packssdw m0, [blockq+0*16+8]
|
|
|
|
packssdw m1, [blockq+1*16+8]
|
|
|
|
mova m2, [blockq+2*16+0]
|
|
|
|
mova m3, [blockq+3*16+0]
|
|
|
|
packssdw m2, [blockq+2*16+8]
|
|
|
|
packssdw m3, [blockq+3*16+8]
|
|
|
|
|
|
|
|
%if cpuflag(ssse3)
|
|
|
|
mova m6, [pw_11585x2]
|
|
|
|
%endif
|
|
|
|
mova m7, [pd_8192] ; rounding
|
|
|
|
VP9_IDCT4_1D
|
|
|
|
TRANSPOSE4x4W 0, 1, 2, 3, 4
|
|
|
|
VP9_IDCT4_1D
|
|
|
|
|
|
|
|
pxor m4, m4
|
|
|
|
ZERO_BLOCK blockq, 16, 4, m4
|
2015-10-06 17:42:03 +02:00
|
|
|
VP9_IDCT4_WRITEOUT
|
2015-10-06 17:03:45 +02:00
|
|
|
RET
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
INIT_MMX mmxext
|
|
|
|
IDCT4_10_FN
|
|
|
|
INIT_MMX ssse3
|
|
|
|
IDCT4_10_FN
|
2015-10-06 17:42:03 +02:00
|
|
|
|
|
|
|
%macro IADST4_FN 4
|
|
|
|
cglobal vp9_%1_%3_4x4_add_10, 3, 3, 0, dst, stride, block, eob
|
|
|
|
%if WIN64 && notcpuflag(ssse3)
|
|
|
|
WIN64_SPILL_XMM 8
|
|
|
|
%endif
|
|
|
|
movdqa xmm5, [pd_8192]
|
|
|
|
mova m0, [blockq+0*16+0]
|
|
|
|
mova m1, [blockq+1*16+0]
|
|
|
|
packssdw m0, [blockq+0*16+8]
|
|
|
|
packssdw m1, [blockq+1*16+8]
|
|
|
|
mova m2, [blockq+2*16+0]
|
|
|
|
mova m3, [blockq+3*16+0]
|
|
|
|
packssdw m2, [blockq+2*16+8]
|
|
|
|
packssdw m3, [blockq+3*16+8]
|
|
|
|
|
|
|
|
%if cpuflag(ssse3)
|
|
|
|
mova m6, [pw_11585x2]
|
|
|
|
%endif
|
|
|
|
%ifnidn %1%3, iadstiadst
|
|
|
|
movdq2q m7, xmm5
|
|
|
|
%endif
|
|
|
|
VP9_%2_1D
|
|
|
|
TRANSPOSE4x4W 0, 1, 2, 3, 4
|
|
|
|
VP9_%4_1D
|
|
|
|
|
|
|
|
pxor m4, m4
|
|
|
|
ZERO_BLOCK blockq, 16, 4, m4
|
|
|
|
VP9_IDCT4_WRITEOUT
|
|
|
|
RET
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
INIT_MMX sse2
|
|
|
|
IADST4_FN idct, IDCT4, iadst, IADST4
|
|
|
|
IADST4_FN iadst, IADST4, idct, IDCT4
|
|
|
|
IADST4_FN iadst, IADST4, iadst, IADST4
|
|
|
|
|
|
|
|
INIT_MMX ssse3
|
|
|
|
IADST4_FN idct, IDCT4, iadst, IADST4
|
|
|
|
IADST4_FN iadst, IADST4, idct, IDCT4
|
|
|
|
IADST4_FN iadst, IADST4, iadst, IADST4
|
2015-10-12 16:16:56 +02:00
|
|
|
|
|
|
|
; inputs and outputs are dwords, coefficients are words
|
|
|
|
;
|
|
|
|
; dst1 = src1 * coef1 + src2 * coef2 + rnd >> 14
|
|
|
|
; dst2 = src1 * coef2 - src2 * coef1 + rnd >> 14
|
|
|
|
%macro SUMSUB_MUL 6 ; src/dst 1-2, tmp1-2, coef1-2
|
|
|
|
pand m%3, m%1, [pd_3fff]
|
|
|
|
pand m%4, m%2, [pd_3fff]
|
|
|
|
psrad m%1, 14
|
|
|
|
psrad m%2, 14
|
|
|
|
packssdw m%4, m%2
|
|
|
|
packssdw m%3, m%1
|
|
|
|
punpckhwd m%2, m%4, m%3
|
|
|
|
punpcklwd m%4, m%3
|
|
|
|
pmaddwd m%3, m%4, [pw_%6_%5]
|
|
|
|
pmaddwd m%1, m%2, [pw_%6_%5]
|
|
|
|
pmaddwd m%4, [pw_m%5_%6]
|
|
|
|
pmaddwd m%2, [pw_m%5_%6]
|
|
|
|
paddd m%3, [pd_8192]
|
|
|
|
paddd m%4, [pd_8192]
|
|
|
|
psrad m%3, 14
|
|
|
|
psrad m%4, 14
|
|
|
|
paddd m%1, m%3
|
|
|
|
paddd m%2, m%4
|
|
|
|
%endmacro
|
|
|
|
|
2015-10-12 16:17:23 +02:00
|
|
|
%macro IDCT4_12BPP_1D 0-6 0, 1, 2, 3, 4, 5
|
|
|
|
SUMSUB_MUL %1, %3, %5, %6, 11585, 11585
|
|
|
|
SUMSUB_MUL %2, %4, %5, %6, 15137, 6270
|
|
|
|
SUMSUB_BA d, %2, %1, %5
|
|
|
|
SUMSUB_BA d, %4, %3, %5
|
|
|
|
SWAP %2, %4, %1
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
%macro STORE_4x4 6 ; tmp1-2, reg1-2, min, max
|
|
|
|
movh m%1, [dstq+strideq*0]
|
|
|
|
movh m%2, [dstq+strideq*2]
|
|
|
|
movhps m%1, [dstq+strideq*1]
|
|
|
|
movhps m%2, [dstq+stride3q ]
|
|
|
|
paddw m%1, m%3
|
|
|
|
paddw m%2, m%4
|
|
|
|
pmaxsw m%1, %5
|
|
|
|
pmaxsw m%2, %5
|
|
|
|
pminsw m%1, %6
|
|
|
|
pminsw m%2, %6
|
|
|
|
movh [dstq+strideq*0], m%1
|
|
|
|
movhps [dstq+strideq*1], m%1
|
|
|
|
movh [dstq+strideq*2], m%2
|
|
|
|
movhps [dstq+stride3q ], m%2
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
%macro ROUND_AND_STORE_4x4 8 ; reg1-4, min, max, rnd, shift
|
|
|
|
paddd m%1, %7
|
|
|
|
paddd m%2, %7
|
|
|
|
paddd m%3, %7
|
|
|
|
paddd m%4, %7
|
|
|
|
psrad m%1, %8
|
|
|
|
psrad m%2, %8
|
|
|
|
psrad m%3, %8
|
|
|
|
psrad m%4, %8
|
|
|
|
packssdw m%1, m%2
|
|
|
|
packssdw m%3, m%4
|
|
|
|
STORE_4x4 %2, %4, %1, %3, %5, %6
|
2015-10-12 16:16:56 +02:00
|
|
|
%endmacro
|
|
|
|
|
|
|
|
INIT_XMM sse2
|
|
|
|
cglobal vp9_idct_idct_4x4_add_12, 4, 4, 6, dst, stride, block, eob
|
|
|
|
cmp eobd, 1
|
|
|
|
jg .idctfull
|
|
|
|
|
|
|
|
; dc-only - this is special, since for 4x4 12bpp, the max coef size is
|
|
|
|
; 17+sign bpp. Since the multiply is with 11585, which is 14bpp, the
|
|
|
|
; result of each multiply is 31+sign bit, i.e. it _exactly_ fits in a
|
|
|
|
; dword. After the final shift (4), the result is 13+sign bits, so we
|
|
|
|
; don't need any additional processing to fit it in a word
|
|
|
|
DEFINE_ARGS dst, stride, block, coef
|
|
|
|
mov coefd, dword [blockq]
|
|
|
|
imul coefd, 11585
|
|
|
|
add coefd, 8192
|
|
|
|
sar coefd, 14
|
|
|
|
imul coefd, 11585
|
|
|
|
add coefd, (8 << 14) + 8192
|
|
|
|
sar coefd, 14 + 4
|
|
|
|
movd m0, coefd
|
|
|
|
pshuflw m0, m0, q0000
|
|
|
|
punpcklqdq m0, m0
|
|
|
|
pxor m4, m4
|
|
|
|
mova m5, [pw_4095]
|
|
|
|
movd [blockq], m4
|
|
|
|
DEFINE_ARGS dst, stride, stride3
|
|
|
|
lea stride3q, [strideq*3]
|
2015-10-12 16:17:23 +02:00
|
|
|
STORE_4x4 1, 3, 0, 0, m4, m5
|
2015-10-12 16:16:56 +02:00
|
|
|
RET
|
|
|
|
|
|
|
|
.idctfull:
|
|
|
|
DEFINE_ARGS dst, stride, block, eob
|
|
|
|
mova m0, [blockq+0*16]
|
|
|
|
mova m1, [blockq+1*16]
|
|
|
|
mova m2, [blockq+2*16]
|
|
|
|
mova m3, [blockq+3*16]
|
|
|
|
|
|
|
|
IDCT4_12BPP_1D
|
|
|
|
TRANSPOSE4x4D 0, 1, 2, 3, 4
|
|
|
|
IDCT4_12BPP_1D
|
|
|
|
|
|
|
|
pxor m4, m4
|
|
|
|
ZERO_BLOCK blockq, 16, 4, m4
|
|
|
|
|
|
|
|
; writeout
|
|
|
|
DEFINE_ARGS dst, stride, stride3
|
|
|
|
lea stride3q, [strideq*3]
|
|
|
|
mova m5, [pw_4095]
|
2015-10-12 16:17:23 +02:00
|
|
|
ROUND_AND_STORE_4x4 0, 1, 2, 3, m4, m5, [pd_8], 4
|
2015-10-12 16:16:56 +02:00
|
|
|
RET
|
2015-10-07 03:59:10 +02:00
|
|
|
|
|
|
|
%macro SCRATCH 3-4
|
|
|
|
%if ARCH_X86_64
|
|
|
|
SWAP %1, %2
|
|
|
|
%if %0 == 4
|
|
|
|
%define reg_%4 m%2
|
|
|
|
%endif
|
|
|
|
%else
|
|
|
|
mova [%3], m%1
|
|
|
|
%if %0 == 4
|
|
|
|
%define reg_%4 [%3]
|
|
|
|
%endif
|
|
|
|
%endif
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
%macro UNSCRATCH 3-4
|
|
|
|
%if ARCH_X86_64
|
|
|
|
SWAP %1, %2
|
|
|
|
%else
|
|
|
|
mova m%1, [%3]
|
|
|
|
%endif
|
|
|
|
%if %0 == 4
|
|
|
|
%undef reg_%4
|
|
|
|
%endif
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
; out0 = 5283 * in0 + 13377 + in1 + 15212 * in2 + 9929 * in3 + rnd >> 14
|
|
|
|
; out1 = 9929 * in0 + 13377 * in1 - 5283 * in2 - 15282 * in3 + rnd >> 14
|
|
|
|
; out2 = 13377 * in0 - 13377 * in2 + 13377 * in3 + rnd >> 14
|
|
|
|
; out3 = 15212 * in0 - 13377 * in1 + 9929 * in2 - 5283 * in3 + rnd >> 14
|
|
|
|
%macro IADST4_12BPP_1D 0
|
|
|
|
pand m4, m0, [pd_3fff]
|
|
|
|
pand m5, m1, [pd_3fff]
|
|
|
|
psrad m0, 14
|
|
|
|
psrad m1, 14
|
|
|
|
packssdw m5, m1
|
|
|
|
packssdw m4, m0
|
|
|
|
punpckhwd m1, m4, m5
|
|
|
|
punpcklwd m4, m5
|
|
|
|
pand m5, m2, [pd_3fff]
|
|
|
|
pand m6, m3, [pd_3fff]
|
|
|
|
psrad m2, 14
|
|
|
|
psrad m3, 14
|
|
|
|
packssdw m6, m3
|
|
|
|
packssdw m5, m2
|
|
|
|
punpckhwd m3, m5, m6
|
|
|
|
punpcklwd m5, m6
|
|
|
|
SCRATCH 1, 8, rsp+0*mmsize, a
|
|
|
|
SCRATCH 5, 9, rsp+1*mmsize, b
|
|
|
|
|
|
|
|
; m1/3 have the high bits of 0,1,2,3
|
|
|
|
; m4/5 have the low bits of 0,1,2,3
|
|
|
|
; m0/2/6/7 are free
|
|
|
|
|
|
|
|
pmaddwd m7, reg_b, [pw_15212_9929]
|
|
|
|
pmaddwd m6, m4, [pw_5283_13377]
|
|
|
|
pmaddwd m2, m3, [pw_15212_9929]
|
|
|
|
pmaddwd m0, reg_a, [pw_5283_13377]
|
|
|
|
paddd m6, m7
|
|
|
|
paddd m0, m2
|
|
|
|
pmaddwd m7, reg_b, [pw_m13377_13377]
|
|
|
|
pmaddwd m2, m4, [pw_13377_0]
|
|
|
|
pmaddwd m1, m3, [pw_m13377_13377]
|
|
|
|
pmaddwd m5, reg_a, [pw_13377_0]
|
|
|
|
paddd m2, m7
|
|
|
|
paddd m1, m5
|
|
|
|
paddd m6, [pd_8192]
|
|
|
|
paddd m2, [pd_8192]
|
|
|
|
psrad m6, 14
|
|
|
|
psrad m2, 14
|
|
|
|
paddd m0, m6 ; t0
|
|
|
|
paddd m2, m1 ; t2
|
|
|
|
|
|
|
|
pmaddwd m1, reg_b, [pw_m5283_m15212]
|
|
|
|
pmaddwd m6, m4, [pw_9929_13377]
|
|
|
|
pmaddwd m7, m3, [pw_m5283_m15212]
|
|
|
|
pmaddwd m5, reg_a, [pw_9929_13377]
|
|
|
|
paddd m6, m1
|
|
|
|
paddd m7, m5
|
|
|
|
UNSCRATCH 5, 9, rsp+1*mmsize, b
|
|
|
|
pmaddwd m5, [pw_9929_m5283]
|
|
|
|
pmaddwd m4, [pw_15212_m13377]
|
|
|
|
pmaddwd m3, [pw_9929_m5283]
|
|
|
|
UNSCRATCH 1, 8, rsp+0*mmsize, a
|
|
|
|
pmaddwd m1, [pw_15212_m13377]
|
|
|
|
paddd m4, m5
|
|
|
|
paddd m3, m1
|
|
|
|
paddd m6, [pd_8192]
|
|
|
|
paddd m4, [pd_8192]
|
|
|
|
psrad m6, 14
|
|
|
|
psrad m4, 14
|
|
|
|
paddd m7, m6 ; t1
|
|
|
|
paddd m3, m4 ; t3
|
|
|
|
|
|
|
|
SWAP 1, 7
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
%macro IADST4_12BPP_FN 4
|
|
|
|
cglobal vp9_%1_%3_4x4_add_12, 3, 3, 10, 2 * ARCH_X86_32 * mmsize, dst, stride, block, eob
|
|
|
|
mova m0, [blockq+0*16]
|
|
|
|
mova m1, [blockq+1*16]
|
|
|
|
mova m2, [blockq+2*16]
|
|
|
|
mova m3, [blockq+3*16]
|
|
|
|
|
|
|
|
%2_12BPP_1D
|
|
|
|
TRANSPOSE4x4D 0, 1, 2, 3, 4
|
|
|
|
%4_12BPP_1D
|
|
|
|
|
|
|
|
pxor m4, m4
|
|
|
|
ZERO_BLOCK blockq, 16, 4, m4
|
|
|
|
|
|
|
|
; writeout
|
|
|
|
DEFINE_ARGS dst, stride, stride3
|
|
|
|
lea stride3q, [strideq*3]
|
|
|
|
mova m5, [pw_4095]
|
2015-10-12 16:17:23 +02:00
|
|
|
ROUND_AND_STORE_4x4 0, 1, 2, 3, m4, m5, [pd_8], 4
|
2015-10-07 03:59:10 +02:00
|
|
|
RET
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
INIT_XMM sse2
|
|
|
|
IADST4_12BPP_FN idct, IDCT4, iadst, IADST4
|
|
|
|
IADST4_12BPP_FN iadst, IADST4, idct, IDCT4
|
|
|
|
IADST4_12BPP_FN iadst, IADST4, iadst, IADST4
|
2015-10-12 16:17:23 +02:00
|
|
|
|
|
|
|
; the following line has not been executed at the end of this macro:
|
2015-10-08 16:16:36 +02:00
|
|
|
; UNSCRATCH 6, 8, rsp+%3*mmsize
|
|
|
|
%macro IDCT8_1D 1-3 2 * mmsize, 17 ; src, src_stride, stack_offset
|
|
|
|
mova m0, [%1+0*%2]
|
|
|
|
mova m2, [%1+2*%2]
|
|
|
|
mova m4, [%1+4*%2]
|
|
|
|
mova m6, [%1+6*%2]
|
2015-10-12 16:17:23 +02:00
|
|
|
IDCT4_12BPP_1D 0, 2, 4, 6, 1, 3 ; m0/2/4/6 have t0/1/2/3
|
2015-10-08 16:16:36 +02:00
|
|
|
SCRATCH 4, 8, rsp+(%3+0)*mmsize
|
|
|
|
SCRATCH 6, 9, rsp+(%3+1)*mmsize
|
|
|
|
mova m1, [%1+1*%2]
|
|
|
|
mova m3, [%1+3*%2]
|
|
|
|
mova m5, [%1+5*%2]
|
|
|
|
mova m7, [%1+7*%2]
|
2015-10-12 16:17:23 +02:00
|
|
|
SUMSUB_MUL 1, 7, 4, 6, 16069, 3196 ; m1=t7a, m7=t4a
|
|
|
|
SUMSUB_MUL 5, 3, 4, 6, 9102, 13623 ; m5=t6a, m3=t5a
|
|
|
|
SUMSUB_BA d, 3, 7, 4 ; m3=t4, m7=t5a
|
|
|
|
SUMSUB_BA d, 5, 1, 4 ; m5=t7, m1=t6a
|
|
|
|
SUMSUB_MUL 1, 7, 4, 6, 11585, 11585 ; m1=t6, m7=t5
|
|
|
|
SUMSUB_BA d, 5, 0, 4 ; m5=out0, m0=out7
|
|
|
|
SUMSUB_BA d, 1, 2, 4 ; m1=out1, m2=out6
|
2015-10-08 16:16:36 +02:00
|
|
|
UNSCRATCH 4, 8, rsp+(%3+0)*mmsize
|
|
|
|
UNSCRATCH 6, 9, rsp+(%3+1)*mmsize
|
|
|
|
SCRATCH 2, 8, rsp+(%3+0)*mmsize
|
2015-10-07 21:42:52 +02:00
|
|
|
SUMSUB_BA d, 7, 4, 2 ; m7=out2, m4=out5
|
|
|
|
SUMSUB_BA d, 3, 6, 2 ; m3=out3, m6=out4
|
2015-10-12 16:17:23 +02:00
|
|
|
SWAP 0, 5, 4, 6, 2, 7
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
%macro STORE_2x8 5 ; tmp1-2, reg, min, max
|
|
|
|
mova m%1, [dstq+strideq*0]
|
|
|
|
mova m%2, [dstq+strideq*1]
|
|
|
|
paddw m%1, m%3
|
|
|
|
paddw m%2, m%3
|
|
|
|
pmaxsw m%1, %4
|
|
|
|
pmaxsw m%2, %4
|
|
|
|
pminsw m%1, %5
|
|
|
|
pminsw m%2, %5
|
|
|
|
mova [dstq+strideq*0], m%1
|
|
|
|
mova [dstq+strideq*1], m%2
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
%macro PRELOAD 2-3
|
|
|
|
%if ARCH_X86_64
|
|
|
|
mova m%1, [%2]
|
|
|
|
%if %0 == 3
|
|
|
|
%define reg_%3 m%1
|
|
|
|
%endif
|
|
|
|
%elif %0 == 3
|
|
|
|
%define reg_%3 [%2]
|
|
|
|
%endif
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
; FIXME we can use the intermediate storage (rsp[0-15]) on x86-32 for temp
|
|
|
|
; storage also instead of allocating two more stack spaces. This doesn't
|
|
|
|
; matter much but it's something...
|
|
|
|
INIT_XMM sse2
|
|
|
|
cglobal vp9_idct_idct_8x8_add_10, 4, 6 + ARCH_X86_64, 10, \
|
|
|
|
17 * mmsize + 2 * ARCH_X86_32 * mmsize, \
|
|
|
|
dst, stride, block, eob
|
|
|
|
mova m0, [pw_1023]
|
|
|
|
cmp eobd, 1
|
|
|
|
jg .idctfull
|
|
|
|
|
|
|
|
; dc-only - the 10bit version can be done entirely in 32bit, since the max
|
|
|
|
; coef values are 16+sign bit, and the coef is 14bit, so 30+sign easily
|
|
|
|
; fits in 32bit
|
|
|
|
DEFINE_ARGS dst, stride, block, coef
|
|
|
|
mov coefd, dword [blockq]
|
|
|
|
imul coefd, 11585
|
|
|
|
add coefd, 8192
|
|
|
|
sar coefd, 14
|
|
|
|
imul coefd, 11585
|
|
|
|
add coefd, (16 << 14) + 8192
|
|
|
|
sar coefd, 14 + 5
|
|
|
|
movd m1, coefd
|
|
|
|
pshuflw m1, m1, q0000
|
|
|
|
punpcklqdq m1, m1
|
|
|
|
pxor m2, m2
|
|
|
|
movd [blockq], m2
|
|
|
|
DEFINE_ARGS dst, stride, cnt
|
|
|
|
mov cntd, 4
|
|
|
|
.loop_dc:
|
|
|
|
STORE_2x8 3, 4, 1, m2, m0
|
|
|
|
lea dstq, [dstq+strideq*2]
|
|
|
|
dec cntd
|
|
|
|
jg .loop_dc
|
|
|
|
RET
|
|
|
|
|
|
|
|
; FIXME a sub-idct for the top-left 4x4 coefficients would save 1 loop
|
|
|
|
; iteration in the first idct (2->1) and thus probably a lot of time.
|
|
|
|
; I haven't implemented that yet, though
|
|
|
|
|
|
|
|
.idctfull:
|
|
|
|
mova [rsp+16*mmsize], m0
|
|
|
|
DEFINE_ARGS dst, stride, block, cnt, ptr, stride3, dstbak
|
|
|
|
%if ARCH_X86_64
|
|
|
|
mov dstbakq, dstq
|
|
|
|
%endif
|
|
|
|
lea stride3q, [strideq*3]
|
|
|
|
mov cntd, 2
|
|
|
|
mov ptrq, rsp
|
|
|
|
.loop_1:
|
|
|
|
IDCT8_1D blockq
|
|
|
|
|
2015-10-07 21:42:52 +02:00
|
|
|
TRANSPOSE4x4D 0, 1, 2, 3, 6
|
2015-10-12 16:17:23 +02:00
|
|
|
mova [ptrq+ 0*mmsize], m0
|
|
|
|
mova [ptrq+ 2*mmsize], m1
|
|
|
|
mova [ptrq+ 4*mmsize], m2
|
|
|
|
mova [ptrq+ 6*mmsize], m3
|
2015-10-07 21:42:52 +02:00
|
|
|
UNSCRATCH 6, 8, rsp+17*mmsize
|
2015-10-12 16:17:23 +02:00
|
|
|
TRANSPOSE4x4D 4, 5, 6, 7, 0
|
|
|
|
mova [ptrq+ 1*mmsize], m4
|
|
|
|
mova [ptrq+ 3*mmsize], m5
|
|
|
|
mova [ptrq+ 5*mmsize], m6
|
|
|
|
mova [ptrq+ 7*mmsize], m7
|
|
|
|
add ptrq, 8 * mmsize
|
|
|
|
add blockq, mmsize
|
|
|
|
dec cntd
|
|
|
|
jg .loop_1
|
|
|
|
|
|
|
|
mov cntd, 2
|
|
|
|
mov ptrq, rsp
|
|
|
|
.loop_2:
|
|
|
|
IDCT8_1D ptrq
|
|
|
|
|
2015-10-07 21:42:52 +02:00
|
|
|
pxor m6, m6
|
2015-10-12 16:17:23 +02:00
|
|
|
PRELOAD 9, rsp+16*mmsize, max
|
2015-10-07 21:42:52 +02:00
|
|
|
ROUND_AND_STORE_4x4 0, 1, 2, 3, m6, reg_max, [pd_16], 5
|
2015-10-12 16:17:23 +02:00
|
|
|
lea dstq, [dstq+strideq*4]
|
|
|
|
UNSCRATCH 0, 8, rsp+17*mmsize
|
2015-10-07 21:42:52 +02:00
|
|
|
ROUND_AND_STORE_4x4 4, 5, 0, 7, m6, reg_max, [pd_16], 5
|
2015-10-12 16:17:23 +02:00
|
|
|
add ptrq, 16
|
|
|
|
%if ARCH_X86_64
|
|
|
|
lea dstq, [dstbakq+8]
|
|
|
|
%else
|
|
|
|
mov dstq, dstm
|
|
|
|
add dstq, 8
|
|
|
|
%endif
|
|
|
|
dec cntd
|
|
|
|
jg .loop_2
|
|
|
|
|
2015-10-07 21:42:52 +02:00
|
|
|
; m6 is still zero
|
|
|
|
ZERO_BLOCK blockq-2*mmsize, 32, 8, m6
|
2015-10-12 16:17:23 +02:00
|
|
|
RET
|
|
|
|
|
|
|
|
INIT_XMM sse2
|
|
|
|
cglobal vp9_idct_idct_8x8_add_12, 4, 6 + ARCH_X86_64, 10, \
|
|
|
|
17 * mmsize + 2 * ARCH_X86_32 * mmsize, \
|
|
|
|
dst, stride, block, eob
|
|
|
|
mova m0, [pw_4095]
|
|
|
|
cmp eobd, 1
|
|
|
|
jg mangle(private_prefix %+ _ %+ vp9_idct_idct_8x8_add_10 %+ SUFFIX).idctfull
|
|
|
|
|
|
|
|
; dc-only - unfortunately, this one can overflow, since coefs are 18+sign
|
|
|
|
; bpp, and 18+14+sign does not fit in 32bit, so we do 2-stage multiplies
|
|
|
|
DEFINE_ARGS dst, stride, block, coef, coefl
|
|
|
|
%if ARCH_X86_64
|
|
|
|
DEFINE_ARGS dst, stride, block, coef
|
|
|
|
movsxd coefq, dword [blockq]
|
|
|
|
pxor m2, m2
|
|
|
|
movd [blockq], m2
|
|
|
|
imul coefq, 11585
|
|
|
|
add coefq, 8192
|
|
|
|
sar coefq, 14
|
|
|
|
imul coefq, 11585
|
|
|
|
add coefq, (16 << 14) + 8192
|
|
|
|
sar coefq, 14 + 5
|
|
|
|
%else
|
|
|
|
mov coefd, dword [blockq]
|
|
|
|
pxor m2, m2
|
|
|
|
movd [blockq], m2
|
|
|
|
DEFINE_ARGS dst, stride, cnt, coef, coefl
|
|
|
|
mov cntd, 2
|
|
|
|
.loop_dc_calc:
|
|
|
|
mov coefld, coefd
|
|
|
|
sar coefd, 14
|
|
|
|
and coefld, 0x3fff
|
|
|
|
imul coefd, 11585
|
|
|
|
imul coefld, 11585
|
|
|
|
add coefld, 8192
|
|
|
|
sar coefld, 14
|
|
|
|
add coefd, coefld
|
|
|
|
dec cntd
|
|
|
|
jg .loop_dc_calc
|
|
|
|
add coefd, 16
|
|
|
|
sar coefd, 5
|
|
|
|
%endif
|
|
|
|
movd m1, coefd
|
|
|
|
pshuflw m1, m1, q0000
|
|
|
|
punpcklqdq m1, m1
|
|
|
|
DEFINE_ARGS dst, stride, cnt
|
|
|
|
mov cntd, 4
|
|
|
|
.loop_dc:
|
|
|
|
STORE_2x8 3, 4, 1, m2, m0
|
|
|
|
lea dstq, [dstq+strideq*2]
|
|
|
|
dec cntd
|
|
|
|
jg .loop_dc
|
|
|
|
RET
|
2015-10-07 21:42:52 +02:00
|
|
|
|
|
|
|
; inputs and outputs are dwords, coefficients are words
|
|
|
|
;
|
|
|
|
; dst1[hi]:dst3[lo] = src1 * coef1 + src2 * coef2
|
|
|
|
; dst2[hi]:dst4[lo] = src1 * coef2 - src2 * coef1
|
|
|
|
%macro SUMSUB_MUL_D 6 ; src/dst 1-2, dst3-4, coef1-2
|
|
|
|
pand m%3, m%1, [pd_3fff]
|
|
|
|
pand m%4, m%2, [pd_3fff]
|
|
|
|
psrad m%1, 14
|
|
|
|
psrad m%2, 14
|
|
|
|
packssdw m%4, m%2
|
|
|
|
packssdw m%3, m%1
|
|
|
|
punpckhwd m%2, m%4, m%3
|
|
|
|
punpcklwd m%4, m%3
|
|
|
|
pmaddwd m%3, m%4, [pw_%6_%5]
|
|
|
|
pmaddwd m%1, m%2, [pw_%6_%5]
|
|
|
|
pmaddwd m%4, [pw_m%5_%6]
|
|
|
|
pmaddwd m%2, [pw_m%5_%6]
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
; dst1 = src2[hi]:src4[lo] + src1[hi]:src3[lo] + rnd >> 14
|
|
|
|
; dst2 = src2[hi]:src4[lo] - src1[hi]:src3[lo] + rnd >> 14
|
|
|
|
%macro SUMSUB_PACK_D 5 ; src/dst 1-2, src3-4, tmp
|
|
|
|
SUMSUB_BA d, %1, %2, %5
|
|
|
|
SUMSUB_BA d, %3, %4, %5
|
|
|
|
paddd m%3, [pd_8192]
|
|
|
|
paddd m%4, [pd_8192]
|
|
|
|
psrad m%3, 14
|
|
|
|
psrad m%4, 14
|
|
|
|
paddd m%1, m%3
|
|
|
|
paddd m%2, m%4
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
%macro NEGD 1
|
|
|
|
%if cpuflag(ssse3)
|
|
|
|
psignd %1, [pw_m1]
|
|
|
|
%else
|
|
|
|
pxor %1, [pw_m1]
|
|
|
|
paddd %1, [pd_1]
|
|
|
|
%endif
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
; the following line has not been executed at the end of this macro:
|
|
|
|
; UNSCRATCH 6, 8, rsp+17*mmsize
|
2015-10-08 16:16:36 +02:00
|
|
|
%macro IADST8_1D 1 ; src
|
2015-10-07 21:42:52 +02:00
|
|
|
mova m0, [%1+ 0*mmsize]
|
|
|
|
mova m3, [%1+ 6*mmsize]
|
|
|
|
mova m4, [%1+ 8*mmsize]
|
|
|
|
mova m7, [%1+14*mmsize]
|
|
|
|
SUMSUB_MUL_D 7, 0, 1, 2, 16305, 1606 ; m7/1=t0a, m0/2=t1a
|
|
|
|
SUMSUB_MUL_D 3, 4, 5, 6, 10394, 12665 ; m3/5=t4a, m4/6=t5a
|
|
|
|
SCRATCH 0, 8, rsp+17*mmsize
|
|
|
|
SUMSUB_PACK_D 3, 7, 5, 1, 0 ; m3=t0, m7=t4
|
|
|
|
UNSCRATCH 0, 8, rsp+17*mmsize
|
|
|
|
SUMSUB_PACK_D 4, 0, 6, 2, 1 ; m4=t1, m0=t5
|
|
|
|
|
|
|
|
SCRATCH 3, 8, rsp+17*mmsize
|
|
|
|
SCRATCH 4, 9, rsp+18*mmsize
|
|
|
|
SCRATCH 7, 10, rsp+19*mmsize
|
|
|
|
SCRATCH 0, 11, rsp+20*mmsize
|
|
|
|
|
|
|
|
mova m1, [%1+ 2*mmsize]
|
|
|
|
mova m2, [%1+ 4*mmsize]
|
|
|
|
mova m5, [%1+10*mmsize]
|
|
|
|
mova m6, [%1+12*mmsize]
|
|
|
|
SUMSUB_MUL_D 5, 2, 3, 4, 14449, 7723 ; m5/8=t2a, m2/9=t3a
|
|
|
|
SUMSUB_MUL_D 1, 6, 7, 0, 4756, 15679 ; m1/10=t6a, m6/11=t7a
|
|
|
|
SCRATCH 2, 12, rsp+21*mmsize
|
|
|
|
SUMSUB_PACK_D 1, 5, 7, 3, 2 ; m1=t2, m5=t6
|
|
|
|
UNSCRATCH 2, 12, rsp+21*mmsize
|
|
|
|
SUMSUB_PACK_D 6, 2, 0, 4, 3 ; m6=t3, m2=t7
|
|
|
|
|
|
|
|
UNSCRATCH 7, 10, rsp+19*mmsize
|
|
|
|
UNSCRATCH 0, 11, rsp+20*mmsize
|
|
|
|
SCRATCH 1, 10, rsp+19*mmsize
|
|
|
|
SCRATCH 6, 11, rsp+20*mmsize
|
|
|
|
|
|
|
|
SUMSUB_MUL_D 7, 0, 3, 4, 15137, 6270 ; m7/8=t4a, m0/9=t5a
|
|
|
|
SUMSUB_MUL_D 2, 5, 1, 6, 6270, 15137 ; m2/10=t7a, m5/11=t6a
|
|
|
|
SCRATCH 2, 12, rsp+21*mmsize
|
|
|
|
SUMSUB_PACK_D 5, 7, 6, 3, 2 ; m5=-out1, m7=t6
|
|
|
|
UNSCRATCH 2, 12, rsp+21*mmsize
|
|
|
|
NEGD m5 ; m5=out1
|
|
|
|
SUMSUB_PACK_D 2, 0, 1, 4, 3 ; m2=out6, m0=t7
|
|
|
|
SUMSUB_MUL 7, 0, 3, 4, 11585, 11585 ; m7=out2, m0=-out5
|
|
|
|
NEGD m0 ; m0=out5
|
|
|
|
|
|
|
|
UNSCRATCH 3, 8, rsp+17*mmsize
|
|
|
|
UNSCRATCH 4, 9, rsp+18*mmsize
|
|
|
|
UNSCRATCH 1, 10, rsp+19*mmsize
|
|
|
|
UNSCRATCH 6, 11, rsp+20*mmsize
|
|
|
|
SCRATCH 2, 8, rsp+17*mmsize
|
|
|
|
SCRATCH 0, 9, rsp+18*mmsize
|
|
|
|
|
|
|
|
SUMSUB_BA d, 1, 3, 2 ; m1=out0, m3=t2
|
|
|
|
SUMSUB_BA d, 6, 4, 2 ; m6=-out7, m4=t3
|
|
|
|
NEGD m6 ; m6=out7
|
|
|
|
SUMSUB_MUL 3, 4, 2, 0, 11585, 11585 ; m3=-out3, m4=out4
|
|
|
|
NEGD m3 ; m3=out3
|
|
|
|
|
|
|
|
UNSCRATCH 0, 9, rsp+18*mmsize
|
|
|
|
|
|
|
|
SWAP 0, 1, 5
|
|
|
|
SWAP 2, 7, 6
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
%macro IADST8_FN 4
|
|
|
|
cglobal vp9_%1_%3_8x8_add_10, 3, 6 + ARCH_X86_64, 13, \
|
|
|
|
17 * mmsize + ARCH_X86_32 * 5 * mmsize, \
|
|
|
|
dst, stride, block, eob
|
|
|
|
mova m0, [pw_1023]
|
|
|
|
|
|
|
|
.body:
|
|
|
|
mova [rsp+16*mmsize], m0
|
|
|
|
DEFINE_ARGS dst, stride, block, cnt, ptr, stride3, dstbak
|
|
|
|
%if ARCH_X86_64
|
|
|
|
mov dstbakq, dstq
|
|
|
|
%endif
|
|
|
|
lea stride3q, [strideq*3]
|
|
|
|
mov cntd, 2
|
|
|
|
mov ptrq, rsp
|
|
|
|
.loop_1:
|
|
|
|
%2_1D blockq
|
|
|
|
|
|
|
|
TRANSPOSE4x4D 0, 1, 2, 3, 6
|
|
|
|
mova [ptrq+ 0*mmsize], m0
|
|
|
|
mova [ptrq+ 2*mmsize], m1
|
|
|
|
mova [ptrq+ 4*mmsize], m2
|
|
|
|
mova [ptrq+ 6*mmsize], m3
|
|
|
|
UNSCRATCH 6, 8, rsp+17*mmsize
|
|
|
|
TRANSPOSE4x4D 4, 5, 6, 7, 0
|
|
|
|
mova [ptrq+ 1*mmsize], m4
|
|
|
|
mova [ptrq+ 3*mmsize], m5
|
|
|
|
mova [ptrq+ 5*mmsize], m6
|
|
|
|
mova [ptrq+ 7*mmsize], m7
|
|
|
|
add ptrq, 8 * mmsize
|
|
|
|
add blockq, mmsize
|
|
|
|
dec cntd
|
|
|
|
jg .loop_1
|
|
|
|
|
|
|
|
mov cntd, 2
|
|
|
|
mov ptrq, rsp
|
|
|
|
.loop_2:
|
|
|
|
%4_1D ptrq
|
|
|
|
|
|
|
|
pxor m6, m6
|
|
|
|
PRELOAD 9, rsp+16*mmsize, max
|
|
|
|
ROUND_AND_STORE_4x4 0, 1, 2, 3, m6, reg_max, [pd_16], 5
|
|
|
|
lea dstq, [dstq+strideq*4]
|
|
|
|
UNSCRATCH 0, 8, rsp+17*mmsize
|
|
|
|
ROUND_AND_STORE_4x4 4, 5, 0, 7, m6, reg_max, [pd_16], 5
|
|
|
|
add ptrq, 16
|
|
|
|
%if ARCH_X86_64
|
|
|
|
lea dstq, [dstbakq+8]
|
|
|
|
%else
|
|
|
|
mov dstq, dstm
|
|
|
|
add dstq, 8
|
|
|
|
%endif
|
|
|
|
dec cntd
|
|
|
|
jg .loop_2
|
|
|
|
|
|
|
|
; m6 is still zero
|
|
|
|
ZERO_BLOCK blockq-2*mmsize, 32, 8, m6
|
|
|
|
RET
|
|
|
|
|
|
|
|
cglobal vp9_%1_%3_8x8_add_12, 3, 6 + ARCH_X86_64, 13, \
|
|
|
|
17 * mmsize + ARCH_X86_32 * 5 * mmsize, \
|
|
|
|
dst, stride, block, eob
|
|
|
|
mova m0, [pw_4095]
|
|
|
|
jmp mangle(private_prefix %+ _ %+ vp9_%1_%3_8x8_add_10 %+ SUFFIX).body
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
INIT_XMM sse2
|
|
|
|
IADST8_FN idct, IDCT8, iadst, IADST8
|
|
|
|
IADST8_FN iadst, IADST8, idct, IDCT8
|
|
|
|
IADST8_FN iadst, IADST8, iadst, IADST8
|
2015-10-08 16:16:36 +02:00
|
|
|
|
|
|
|
%macro IDCT16_1D 1 ; src
|
|
|
|
IDCT8_1D %1, 8 * mmsize, 67 ; m0-3=t0-3a, m4-5/m8|r67/m7=t4-7
|
|
|
|
; SCRATCH 6, 8, rsp+67*mmsize ; t6
|
|
|
|
SCRATCH 0, 15, rsp+74*mmsize ; t0a
|
|
|
|
SCRATCH 1, 14, rsp+73*mmsize ; t1a
|
|
|
|
SCRATCH 2, 13, rsp+72*mmsize ; t2a
|
|
|
|
SCRATCH 3, 12, rsp+71*mmsize ; t3a
|
|
|
|
SCRATCH 4, 11, rsp+70*mmsize ; t4
|
|
|
|
mova [rsp+65*mmsize], m5 ; t5
|
|
|
|
mova [rsp+66*mmsize], m7 ; t7
|
|
|
|
|
|
|
|
mova m0, [%1+ 1*4*mmsize] ; in1
|
|
|
|
mova m3, [%1+ 7*4*mmsize] ; in7
|
|
|
|
mova m4, [%1+ 9*4*mmsize] ; in9
|
|
|
|
mova m7, [%1+15*4*mmsize] ; in15
|
|
|
|
|
|
|
|
SUMSUB_MUL 0, 7, 1, 2, 16305, 1606 ; m0=t15a, m7=t8a
|
|
|
|
SUMSUB_MUL 4, 3, 1, 2, 10394, 12665 ; m4=t14a, m3=t9a
|
|
|
|
SUMSUB_BA d, 3, 7, 1 ; m3=t8, m7=t9
|
|
|
|
SUMSUB_BA d, 4, 0, 1 ; m4=t15,m0=t14
|
|
|
|
SUMSUB_MUL 0, 7, 1, 2, 15137, 6270 ; m0=t14a, m7=t9a
|
|
|
|
|
|
|
|
mova m1, [%1+ 3*4*mmsize] ; in3
|
|
|
|
mova m2, [%1+ 5*4*mmsize] ; in5
|
|
|
|
mova m5, [%1+11*4*mmsize] ; in11
|
|
|
|
mova m6, [%1+13*4*mmsize] ; in13
|
|
|
|
|
|
|
|
SCRATCH 0, 9, rsp+68*mmsize
|
|
|
|
SCRATCH 7, 10, rsp+69*mmsize
|
|
|
|
|
|
|
|
SUMSUB_MUL 2, 5, 0, 7, 14449, 7723 ; m2=t13a, m5=t10a
|
|
|
|
SUMSUB_MUL 6, 1, 0, 7, 4756, 15679 ; m6=t12a, m1=t11a
|
|
|
|
SUMSUB_BA d, 5, 1, 0 ; m5=t11,m1=t10
|
|
|
|
SUMSUB_BA d, 2, 6, 0 ; m2=t12,m6=t13
|
|
|
|
NEGD m1 ; m1=-t10
|
|
|
|
SUMSUB_MUL 1, 6, 0, 7, 15137, 6270 ; m1=t13a, m6=t10a
|
|
|
|
|
|
|
|
UNSCRATCH 7, 10, rsp+69*mmsize
|
|
|
|
SUMSUB_BA d, 5, 3, 0 ; m5=t8a, m3=t11a
|
|
|
|
SUMSUB_BA d, 6, 7, 0 ; m6=t9, m7=t10
|
|
|
|
SUMSUB_BA d, 2, 4, 0 ; m2=t15a,m4=t12a
|
|
|
|
SCRATCH 5, 10, rsp+69*mmsize
|
|
|
|
SUMSUB_MUL 4, 3, 0, 5, 11585, 11585 ; m4=t12, m3=t11
|
|
|
|
UNSCRATCH 0, 9, rsp+68*mmsize
|
|
|
|
SUMSUB_BA d, 1, 0, 5 ; m1=t14, m0=t13
|
|
|
|
SCRATCH 6, 9, rsp+68*mmsize
|
|
|
|
SUMSUB_MUL 0, 7, 6, 5, 11585, 11585 ; m0=t13a,m7=t10a
|
|
|
|
|
|
|
|
; order: 15|r74,14|r73,13|r72,12|r71,11|r70,r65,8|r67,r66,10|r69,9|r68,7,3,4,0,1,2
|
|
|
|
; free: 6,5
|
|
|
|
|
|
|
|
UNSCRATCH 5, 15, rsp+74*mmsize
|
|
|
|
SUMSUB_BA d, 2, 5, 6 ; m2=out0, m5=out15
|
|
|
|
SCRATCH 5, 15, rsp+74*mmsize
|
|
|
|
UNSCRATCH 5, 14, rsp+73*mmsize
|
|
|
|
SUMSUB_BA d, 1, 5, 6 ; m1=out1, m5=out14
|
|
|
|
SCRATCH 5, 14, rsp+73*mmsize
|
|
|
|
UNSCRATCH 5, 13, rsp+72*mmsize
|
|
|
|
SUMSUB_BA d, 0, 5, 6 ; m0=out2, m5=out13
|
|
|
|
SCRATCH 5, 13, rsp+72*mmsize
|
|
|
|
UNSCRATCH 5, 12, rsp+71*mmsize
|
|
|
|
SUMSUB_BA d, 4, 5, 6 ; m4=out3, m5=out12
|
|
|
|
SCRATCH 5, 12, rsp+71*mmsize
|
|
|
|
UNSCRATCH 5, 11, rsp+70*mmsize
|
|
|
|
SUMSUB_BA d, 3, 5, 6 ; m3=out4, m5=out11
|
|
|
|
SCRATCH 4, 11, rsp+70*mmsize
|
|
|
|
mova m4, [rsp+65*mmsize]
|
|
|
|
SUMSUB_BA d, 7, 4, 6 ; m7=out5, m4=out10
|
|
|
|
mova [rsp+65*mmsize], m5
|
|
|
|
UNSCRATCH 5, 8, rsp+67*mmsize
|
|
|
|
UNSCRATCH 6, 9, rsp+68*mmsize
|
|
|
|
SCRATCH 2, 8, rsp+67*mmsize
|
|
|
|
SCRATCH 1, 9, rsp+68*mmsize
|
|
|
|
UNSCRATCH 1, 10, rsp+69*mmsize
|
|
|
|
SCRATCH 0, 10, rsp+69*mmsize
|
|
|
|
mova m0, [rsp+66*mmsize]
|
|
|
|
SUMSUB_BA d, 6, 5, 2 ; m6=out6, m5=out9
|
|
|
|
SUMSUB_BA d, 1, 0, 2 ; m1=out7, m0=out8
|
|
|
|
|
|
|
|
SWAP 0, 3, 1, 7, 2, 6, 4
|
|
|
|
|
|
|
|
; output order: 8-11|r67-70=out0-3
|
|
|
|
; 0-6,r65=out4-11
|
|
|
|
; 12-15|r71-74=out12-15
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
INIT_XMM sse2
|
|
|
|
cglobal vp9_idct_idct_16x16_add_10, 4, 6 + ARCH_X86_64, 16, \
|
|
|
|
67 * mmsize + ARCH_X86_32 * 8 * mmsize, \
|
|
|
|
dst, stride, block, eob
|
|
|
|
mova m0, [pw_1023]
|
|
|
|
cmp eobd, 1
|
|
|
|
jg .idctfull
|
|
|
|
|
|
|
|
; dc-only
|
|
|
|
|
|
|
|
.idctfull:
|
|
|
|
mova [rsp+64*mmsize], m0
|
|
|
|
DEFINE_ARGS dst, stride, block, cnt, ptr, stride3, dstbak
|
|
|
|
%if ARCH_X86_64
|
|
|
|
mov dstbakq, dstq
|
|
|
|
%endif
|
|
|
|
lea stride3q, [strideq*3]
|
|
|
|
mov cntd, 4
|
|
|
|
mov ptrq, rsp
|
|
|
|
.loop_1:
|
|
|
|
IDCT16_1D blockq
|
|
|
|
; order: 2,1,0,11,3,7,9,10,6,8,4,5,12,13,r65,15
|
|
|
|
|
|
|
|
TRANSPOSE4x4D 0, 1, 2, 3, 7
|
|
|
|
mova [ptrq+ 1*mmsize], m0
|
|
|
|
mova [ptrq+ 5*mmsize], m1
|
|
|
|
mova [ptrq+ 9*mmsize], m2
|
|
|
|
mova [ptrq+13*mmsize], m3
|
|
|
|
mova m7, [rsp+65*mmsize]
|
|
|
|
TRANSPOSE4x4D 4, 5, 6, 7, 0
|
|
|
|
mova [ptrq+ 2*mmsize], m4
|
|
|
|
mova [ptrq+ 6*mmsize], m5
|
|
|
|
mova [ptrq+10*mmsize], m6
|
|
|
|
mova [ptrq+14*mmsize], m7
|
|
|
|
UNSCRATCH 0, 8, rsp+67*mmsize
|
|
|
|
UNSCRATCH 1, 9, rsp+68*mmsize
|
|
|
|
UNSCRATCH 2, 10, rsp+69*mmsize
|
|
|
|
UNSCRATCH 3, 11, rsp+70*mmsize
|
|
|
|
TRANSPOSE4x4D 0, 1, 2, 3, 7
|
|
|
|
mova [ptrq+ 0*mmsize], m0
|
|
|
|
mova [ptrq+ 4*mmsize], m1
|
|
|
|
mova [ptrq+ 8*mmsize], m2
|
|
|
|
mova [ptrq+12*mmsize], m3
|
|
|
|
UNSCRATCH 4, 12, rsp+71*mmsize
|
|
|
|
UNSCRATCH 5, 13, rsp+72*mmsize
|
|
|
|
UNSCRATCH 6, 14, rsp+73*mmsize
|
|
|
|
UNSCRATCH 7, 15, rsp+74*mmsize
|
|
|
|
TRANSPOSE4x4D 4, 5, 6, 7, 0
|
|
|
|
mova [ptrq+ 3*mmsize], m4
|
|
|
|
mova [ptrq+ 7*mmsize], m5
|
|
|
|
mova [ptrq+11*mmsize], m6
|
|
|
|
mova [ptrq+15*mmsize], m7
|
|
|
|
add ptrq, 16 * mmsize
|
|
|
|
add blockq, mmsize
|
|
|
|
dec cntd
|
|
|
|
jg .loop_1
|
|
|
|
|
|
|
|
mov cntd, 4
|
|
|
|
mov ptrq, rsp
|
|
|
|
.loop_2:
|
|
|
|
IDCT16_1D ptrq
|
|
|
|
|
|
|
|
pxor m7, m7
|
|
|
|
lea dstq, [dstq+strideq*4]
|
|
|
|
ROUND_AND_STORE_4x4 0, 1, 2, 3, m7, [rsp+64*mmsize], [pd_32], 6
|
|
|
|
lea dstq, [dstq+strideq*4]
|
|
|
|
mova m0, [rsp+65*mmsize]
|
|
|
|
ROUND_AND_STORE_4x4 4, 5, 6, 0, m7, [rsp+64*mmsize], [pd_32], 6
|
|
|
|
|
|
|
|
%if ARCH_X86_64
|
|
|
|
DEFINE_ARGS dstbak, stride, block, cnt, ptr, stride3, dst
|
|
|
|
%else
|
|
|
|
mov dstq, dstm
|
|
|
|
%endif
|
|
|
|
UNSCRATCH 0, 8, rsp+67*mmsize
|
|
|
|
UNSCRATCH 1, 9, rsp+68*mmsize
|
|
|
|
UNSCRATCH 2, 10, rsp+69*mmsize
|
|
|
|
UNSCRATCH 3, 11, rsp+70*mmsize
|
|
|
|
ROUND_AND_STORE_4x4 0, 1, 2, 3, m7, [rsp+64*mmsize], [pd_32], 6
|
|
|
|
%if ARCH_X86_64
|
|
|
|
DEFINE_ARGS dst, stride, block, cnt, ptr, stride3, dstbak
|
|
|
|
lea dstq, [dstbakq+stride3q*4]
|
|
|
|
%else
|
|
|
|
lea dstq, [dstq+stride3q*4]
|
|
|
|
%endif
|
|
|
|
UNSCRATCH 4, 12, rsp+71*mmsize
|
|
|
|
UNSCRATCH 5, 13, rsp+72*mmsize
|
|
|
|
UNSCRATCH 6, 14, rsp+73*mmsize
|
|
|
|
UNSCRATCH 0, 15, rsp+74*mmsize
|
|
|
|
ROUND_AND_STORE_4x4 4, 5, 6, 0, m7, [rsp+64*mmsize], [pd_32], 6
|
|
|
|
|
|
|
|
add ptrq, mmsize
|
|
|
|
%if ARCH_X86_64
|
|
|
|
add dstbakq, 8
|
|
|
|
mov dstq, dstbakq
|
|
|
|
%else
|
|
|
|
add dword dstm, 8
|
|
|
|
mov dstq, dstm
|
|
|
|
%endif
|
|
|
|
dec cntd
|
|
|
|
jg .loop_2
|
|
|
|
|
|
|
|
; m7 is still zero
|
|
|
|
ZERO_BLOCK blockq-4*mmsize, 64, 16, m7
|
|
|
|
RET
|
|
|
|
|
|
|
|
INIT_XMM sse2
|
|
|
|
cglobal vp9_idct_idct_16x16_add_12, 4, 6 + ARCH_X86_64, 16, \
|
|
|
|
67 * mmsize + ARCH_X86_32 * 8 * mmsize, \
|
|
|
|
dst, stride, block, eob
|
|
|
|
mova m0, [pw_4095]
|
|
|
|
cmp eobd, 1
|
|
|
|
jg mangle(private_prefix %+ _ %+ vp9_idct_idct_16x16_add_10 %+ SUFFIX).idctfull
|
|
|
|
|
|
|
|
; dc-only
|
|
|
|
jmp mangle(private_prefix %+ _ %+ vp9_idct_idct_16x16_add_10 %+ SUFFIX).idctfull
|
|
|
|
RET
|