use asm_offsets with vp8_regular_quantize_b_sse2
remove helper function and avoid shadowing all the arguments to the stack on 64bit systems when running with --good --cpu-used=0: ~2% on linux x86 and x86_64 ~2% on win32 x86 msys and visual studio more on darwin10 x86_64 significantly more on x86_64-win64-vs9 Change-Id: Ib7be12edf511fbf2922f191afd5b33b19a0c4ae6
This commit is contained in:
parent
edfc93aeba
commit
8edaf6e2f2
@ -331,11 +331,8 @@ ifneq ($(call enabled,DIST-SRCS),)
|
|||||||
DIST-SRCS-$(CONFIG_MSVS) += build/make/gen_msvs_sln.sh
|
DIST-SRCS-$(CONFIG_MSVS) += build/make/gen_msvs_sln.sh
|
||||||
DIST-SRCS-$(CONFIG_MSVS) += build/x86-msvs/yasm.rules
|
DIST-SRCS-$(CONFIG_MSVS) += build/x86-msvs/yasm.rules
|
||||||
DIST-SRCS-$(CONFIG_RVCT) += build/make/armlink_adapter.sh
|
DIST-SRCS-$(CONFIG_RVCT) += build/make/armlink_adapter.sh
|
||||||
#
|
# Include obj_int_extract if we use offsets from asm_*_offsets
|
||||||
# This isn't really ARCH_ARM dependent, it's dependent on whether we're
|
DIST-SRCS-$(ARCH_ARM)$(ARCH_X86)$(ARCH_X86_64) += build/make/obj_int_extract.c
|
||||||
# using assembly code or not (CONFIG_OPTIMIZATIONS maybe). Just use
|
|
||||||
# this for now.
|
|
||||||
DIST-SRCS-$(ARCH_ARM) += build/make/obj_int_extract.c
|
|
||||||
DIST-SRCS-$(ARCH_ARM) += build/make/ads2gas.pl
|
DIST-SRCS-$(ARCH_ARM) += build/make/ads2gas.pl
|
||||||
DIST-SRCS-yes += $(target:-$(TOOLCHAIN)=).mk
|
DIST-SRCS-yes += $(target:-$(TOOLCHAIN)=).mk
|
||||||
endif
|
endif
|
||||||
|
4
libs.mk
4
libs.mk
@ -245,7 +245,9 @@ ifeq ($(CONFIG_EXTERNAL_BUILD),) # Visual Studio uses obj_int_extract.bat
|
|||||||
OBJS-yes += $(VP8_PREFIX)common/asm_com_offsets.c.o
|
OBJS-yes += $(VP8_PREFIX)common/asm_com_offsets.c.o
|
||||||
CLEAN-OBJS += asm_com_offsets.asm
|
CLEAN-OBJS += asm_com_offsets.asm
|
||||||
$(filter %$(ASM).o,$(OBJS-yes)): $(BUILD_PFX)asm_com_offsets.asm
|
$(filter %$(ASM).o,$(OBJS-yes)): $(BUILD_PFX)asm_com_offsets.asm
|
||||||
|
endif
|
||||||
|
|
||||||
|
ifeq ($(ARCH_ARM)$(ARCH_X86)$(ARCH_X86_64), yes)
|
||||||
ifeq ($(CONFIG_VP8_ENCODER), yes)
|
ifeq ($(CONFIG_VP8_ENCODER), yes)
|
||||||
asm_enc_offsets.asm: obj_int_extract
|
asm_enc_offsets.asm: obj_int_extract
|
||||||
asm_enc_offsets.asm: $(VP8_PREFIX)encoder/asm_enc_offsets.c.o
|
asm_enc_offsets.asm: $(VP8_PREFIX)encoder/asm_enc_offsets.c.o
|
||||||
@ -254,7 +256,9 @@ ifeq ($(CONFIG_EXTERNAL_BUILD),) # Visual Studio uses obj_int_extract.bat
|
|||||||
CLEAN-OBJS += asm_enc_offsets.asm
|
CLEAN-OBJS += asm_enc_offsets.asm
|
||||||
$(filter %$(ASM).o,$(OBJS-yes)): $(BUILD_PFX)asm_enc_offsets.asm
|
$(filter %$(ASM).o,$(OBJS-yes)): $(BUILD_PFX)asm_enc_offsets.asm
|
||||||
endif
|
endif
|
||||||
|
endif
|
||||||
|
|
||||||
|
ifeq ($(ARCH_ARM), yes)
|
||||||
ifeq ($(CONFIG_VP8_DECODER), yes)
|
ifeq ($(CONFIG_VP8_DECODER), yes)
|
||||||
asm_dec_offsets.asm: obj_int_extract
|
asm_dec_offsets.asm: obj_int_extract
|
||||||
asm_dec_offsets.asm: $(VP8_PREFIX)decoder/asm_dec_offsets.c.o
|
asm_dec_offsets.asm: $(VP8_PREFIX)decoder/asm_dec_offsets.c.o
|
||||||
|
@ -12,9 +12,11 @@
|
|||||||
#include "vpx_ports/config.h"
|
#include "vpx_ports/config.h"
|
||||||
#include <stddef.h>
|
#include <stddef.h>
|
||||||
|
|
||||||
|
#include "block.h"
|
||||||
|
#include "vp8/common/blockd.h"
|
||||||
|
#include "onyx_int.h"
|
||||||
#include "treewriter.h"
|
#include "treewriter.h"
|
||||||
#include "tokenize.h"
|
#include "tokenize.h"
|
||||||
#include "onyx_int.h"
|
|
||||||
|
|
||||||
#define ct_assert(name,cond) \
|
#define ct_assert(name,cond) \
|
||||||
static void assert_##name(void) UNUSED;\
|
static void assert_##name(void) UNUSED;\
|
||||||
@ -31,6 +33,21 @@
|
|||||||
* {
|
* {
|
||||||
*/
|
*/
|
||||||
|
|
||||||
|
//regular quantize
|
||||||
|
DEFINE(vp8_block_coeff, offsetof(BLOCK, coeff));
|
||||||
|
DEFINE(vp8_block_zbin, offsetof(BLOCK, zbin));
|
||||||
|
DEFINE(vp8_block_round, offsetof(BLOCK, round));
|
||||||
|
DEFINE(vp8_block_quant, offsetof(BLOCK, quant));
|
||||||
|
DEFINE(vp8_block_quant_fast, offsetof(BLOCK, quant_fast));
|
||||||
|
DEFINE(vp8_block_zbin_extra, offsetof(BLOCK, zbin_extra));
|
||||||
|
DEFINE(vp8_block_zrun_zbin_boost, offsetof(BLOCK, zrun_zbin_boost));
|
||||||
|
DEFINE(vp8_block_quant_shift, offsetof(BLOCK, quant_shift));
|
||||||
|
|
||||||
|
DEFINE(vp8_blockd_qcoeff, offsetof(BLOCKD, qcoeff));
|
||||||
|
DEFINE(vp8_blockd_dequant, offsetof(BLOCKD, dequant));
|
||||||
|
DEFINE(vp8_blockd_dqcoeff, offsetof(BLOCKD, dqcoeff));
|
||||||
|
DEFINE(vp8_blockd_eob, offsetof(BLOCKD, eob));
|
||||||
|
|
||||||
//pack tokens
|
//pack tokens
|
||||||
DEFINE(vp8_writer_lowvalue, offsetof(vp8_writer, lowvalue));
|
DEFINE(vp8_writer_lowvalue, offsetof(vp8_writer, lowvalue));
|
||||||
DEFINE(vp8_writer_range, offsetof(vp8_writer, range));
|
DEFINE(vp8_writer_range, offsetof(vp8_writer, range));
|
||||||
@ -65,17 +82,6 @@ DEFINE(TOKENLIST_SZ, sizeof(TOKENLIST));
|
|||||||
|
|
||||||
DEFINE(vp8_common_mb_rows, offsetof(VP8_COMMON, mb_rows));
|
DEFINE(vp8_common_mb_rows, offsetof(VP8_COMMON, mb_rows));
|
||||||
|
|
||||||
// offsets from BLOCK structure
|
|
||||||
DEFINE(vp8_block_coeff, offsetof(BLOCK, coeff));
|
|
||||||
DEFINE(vp8_block_quant_fast, offsetof(BLOCK, quant_fast));
|
|
||||||
DEFINE(vp8_block_round, offsetof(BLOCK, round));
|
|
||||||
|
|
||||||
// offsets from BLOCKD structure
|
|
||||||
DEFINE(vp8_blockd_qcoeff, offsetof(BLOCKD, qcoeff));
|
|
||||||
DEFINE(vp8_blockd_dqcoeff, offsetof(BLOCKD, dqcoeff));
|
|
||||||
DEFINE(vp8_blockd_dequant, offsetof(BLOCKD, dequant));
|
|
||||||
DEFINE(vp8_blockd_eob, offsetof(BLOCKD, eob));
|
|
||||||
|
|
||||||
// These two sizes are used in vp8cx_pack_tokens. They are hard coded
|
// These two sizes are used in vp8cx_pack_tokens. They are hard coded
|
||||||
// so if the size changes this will have to be adjusted.
|
// so if the size changes this will have to be adjusted.
|
||||||
#if HAVE_ARMV5TE
|
#if HAVE_ARMV5TE
|
||||||
|
@ -9,48 +9,59 @@
|
|||||||
|
|
||||||
|
|
||||||
%include "vpx_ports/x86_abi_support.asm"
|
%include "vpx_ports/x86_abi_support.asm"
|
||||||
|
%include "asm_enc_offsets.asm"
|
||||||
|
|
||||||
|
|
||||||
;int vp8_regular_quantize_b_impl_sse2(
|
; void vp8_regular_quantize_b_sse2 | arg
|
||||||
; short *coeff_ptr,
|
; (BLOCK *b, | 0
|
||||||
; short *zbin_ptr,
|
; BLOCKD *d) | 1
|
||||||
; short *qcoeff_ptr,
|
|
||||||
; short *dequant_ptr,
|
global sym(vp8_regular_quantize_b_sse2)
|
||||||
; const int *default_zig_zag,
|
sym(vp8_regular_quantize_b_sse2):
|
||||||
; short *round_ptr,
|
|
||||||
; short *quant_ptr,
|
|
||||||
; short *dqcoeff_ptr,
|
|
||||||
; unsigned short zbin_oq_value,
|
|
||||||
; short *zbin_boost_ptr,
|
|
||||||
; short *quant_shift);
|
|
||||||
;
|
|
||||||
global sym(vp8_regular_quantize_b_impl_sse2)
|
|
||||||
sym(vp8_regular_quantize_b_impl_sse2):
|
|
||||||
push rbp
|
push rbp
|
||||||
mov rbp, rsp
|
mov rbp, rsp
|
||||||
SHADOW_ARGS_TO_STACK 11
|
|
||||||
SAVE_XMM
|
SAVE_XMM
|
||||||
|
GET_GOT rbx
|
||||||
push rsi
|
push rsi
|
||||||
|
|
||||||
|
%if ABI_IS_32BIT
|
||||||
push rdi
|
push rdi
|
||||||
push rbx
|
%else
|
||||||
|
%ifidn __OUTPUT_FORMAT__,x64
|
||||||
|
push rdi
|
||||||
|
%endif
|
||||||
|
%endif
|
||||||
|
|
||||||
ALIGN_STACK 16, rax
|
ALIGN_STACK 16, rax
|
||||||
%define abs_minus_zbin 0
|
%define BLOCKD_d 0 ; 8
|
||||||
%define temp_qcoeff 32
|
%define zrun_zbin_boost 8 ; 8
|
||||||
%define qcoeff 64
|
%define abs_minus_zbin 16 ; 32
|
||||||
%define eob_tmp 96
|
%define temp_qcoeff 48 ; 32
|
||||||
|
%define qcoeff 80 ; 32
|
||||||
%define stack_size 112
|
%define stack_size 112
|
||||||
sub rsp, stack_size
|
sub rsp, stack_size
|
||||||
; end prolog
|
; end prolog
|
||||||
|
|
||||||
mov rdx, arg(0) ; coeff_ptr
|
%if ABI_IS_32BIT
|
||||||
mov rcx, arg(1) ; zbin_ptr
|
mov rdi, arg(0)
|
||||||
movd xmm7, arg(8) ; zbin_oq_value
|
%else
|
||||||
mov rdi, arg(5) ; round_ptr
|
%ifidn __OUTPUT_FORMAT__,x64
|
||||||
mov rsi, arg(6) ; quant_ptr
|
mov rdi, rcx ; BLOCK *b
|
||||||
|
mov [rsp + BLOCKD_d], rdx
|
||||||
|
%else
|
||||||
|
;mov rdi, rdi ; BLOCK *b
|
||||||
|
mov [rsp + BLOCKD_d], rsi
|
||||||
|
%endif
|
||||||
|
%endif
|
||||||
|
|
||||||
|
mov rdx, [rdi + vp8_block_coeff] ; coeff_ptr
|
||||||
|
mov rcx, [rdi + vp8_block_zbin] ; zbin_ptr
|
||||||
|
movd xmm7, [rdi + vp8_block_zbin_extra] ; zbin_oq_value
|
||||||
|
|
||||||
; z
|
; z
|
||||||
movdqa xmm0, OWORD PTR[rdx]
|
movdqa xmm0, [rdx]
|
||||||
movdqa xmm4, OWORD PTR[rdx + 16]
|
movdqa xmm4, [rdx + 16]
|
||||||
|
mov rdx, [rdi + vp8_block_round] ; round_ptr
|
||||||
|
|
||||||
pshuflw xmm7, xmm7, 0
|
pshuflw xmm7, xmm7, 0
|
||||||
punpcklwd xmm7, xmm7 ; duplicated zbin_oq_value
|
punpcklwd xmm7, xmm7 ; duplicated zbin_oq_value
|
||||||
@ -70,8 +81,9 @@ sym(vp8_regular_quantize_b_impl_sse2):
|
|||||||
psubw xmm1, xmm0
|
psubw xmm1, xmm0
|
||||||
psubw xmm5, xmm4
|
psubw xmm5, xmm4
|
||||||
|
|
||||||
movdqa xmm2, OWORD PTR[rcx]
|
movdqa xmm2, [rcx]
|
||||||
movdqa xmm3, OWORD PTR[rcx + 16]
|
movdqa xmm3, [rcx + 16]
|
||||||
|
mov rcx, [rdi + vp8_block_quant] ; quant_ptr
|
||||||
|
|
||||||
; *zbin_ptr + zbin_oq_value
|
; *zbin_ptr + zbin_oq_value
|
||||||
paddw xmm2, xmm7
|
paddw xmm2, xmm7
|
||||||
@ -80,18 +92,18 @@ sym(vp8_regular_quantize_b_impl_sse2):
|
|||||||
; x - (*zbin_ptr + zbin_oq_value)
|
; x - (*zbin_ptr + zbin_oq_value)
|
||||||
psubw xmm1, xmm2
|
psubw xmm1, xmm2
|
||||||
psubw xmm5, xmm3
|
psubw xmm5, xmm3
|
||||||
movdqa OWORD PTR[rsp + abs_minus_zbin], xmm1
|
movdqa [rsp + abs_minus_zbin], xmm1
|
||||||
movdqa OWORD PTR[rsp + abs_minus_zbin + 16], xmm5
|
movdqa [rsp + abs_minus_zbin + 16], xmm5
|
||||||
|
|
||||||
; add (zbin_ptr + zbin_oq_value) back
|
; add (zbin_ptr + zbin_oq_value) back
|
||||||
paddw xmm1, xmm2
|
paddw xmm1, xmm2
|
||||||
paddw xmm5, xmm3
|
paddw xmm5, xmm3
|
||||||
|
|
||||||
movdqa xmm2, OWORD PTR[rdi]
|
movdqa xmm2, [rdx]
|
||||||
movdqa xmm6, OWORD PTR[rdi + 16]
|
movdqa xmm6, [rdx + 16]
|
||||||
|
|
||||||
movdqa xmm3, OWORD PTR[rsi]
|
movdqa xmm3, [rcx]
|
||||||
movdqa xmm7, OWORD PTR[rsi + 16]
|
movdqa xmm7, [rcx + 16]
|
||||||
|
|
||||||
; x + round
|
; x + round
|
||||||
paddw xmm1, xmm2
|
paddw xmm1, xmm2
|
||||||
@ -105,68 +117,67 @@ sym(vp8_regular_quantize_b_impl_sse2):
|
|||||||
paddw xmm1, xmm3
|
paddw xmm1, xmm3
|
||||||
paddw xmm5, xmm7
|
paddw xmm5, xmm7
|
||||||
|
|
||||||
movdqa OWORD PTR[rsp + temp_qcoeff], xmm1
|
movdqa [rsp + temp_qcoeff], xmm1
|
||||||
movdqa OWORD PTR[rsp + temp_qcoeff + 16], xmm5
|
movdqa [rsp + temp_qcoeff + 16], xmm5
|
||||||
|
|
||||||
pxor xmm6, xmm6
|
pxor xmm6, xmm6
|
||||||
; zero qcoeff
|
; zero qcoeff
|
||||||
movdqa OWORD PTR[rsp + qcoeff], xmm6
|
movdqa [rsp + qcoeff], xmm6
|
||||||
movdqa OWORD PTR[rsp + qcoeff + 16], xmm6
|
movdqa [rsp + qcoeff + 16], xmm6
|
||||||
|
|
||||||
mov [rsp + eob_tmp], DWORD -1 ; eob
|
mov rsi, [rdi + vp8_block_zrun_zbin_boost] ; zbin_boost_ptr
|
||||||
mov rsi, arg(9) ; zbin_boost_ptr
|
mov rax, [rdi + vp8_block_quant_shift] ; quant_shift_ptr
|
||||||
mov rdi, arg(4) ; default_zig_zag
|
mov [rsp + zrun_zbin_boost], rsi
|
||||||
mov rax, arg(10) ; quant_shift_ptr
|
|
||||||
|
|
||||||
%macro ZIGZAG_LOOP 2
|
%macro ZIGZAG_LOOP 1
|
||||||
rq_zigzag_loop_%1:
|
movsx edx, WORD PTR[GLOBAL(zig_zag) + (%1 * 2)] ; rc
|
||||||
movsxd rdx, DWORD PTR[rdi + (%1 * 4)] ; rc
|
|
||||||
movsx ebx, WORD PTR [rsi] ; *zbin_boost_ptr
|
|
||||||
lea rsi, [rsi + 2] ; zbin_boost_ptr++
|
|
||||||
|
|
||||||
; x
|
; x
|
||||||
movsx ecx, WORD PTR[rsp + abs_minus_zbin + rdx *2]
|
movsx ecx, WORD PTR[rsp + abs_minus_zbin + rdx *2]
|
||||||
|
|
||||||
; if (x >= zbin)
|
; if (x >= zbin)
|
||||||
sub ecx, ebx ; x - zbin
|
sub cx, WORD PTR[rsi] ; x - zbin
|
||||||
jl rq_zigzag_loop_%2 ; x < zbin
|
lea rsi, [rsi + 2] ; zbin_boost_ptr++
|
||||||
|
jl rq_zigzag_loop_%1 ; x < zbin
|
||||||
|
|
||||||
movsx ebx, WORD PTR[rsp + temp_qcoeff + rdx *2]
|
movsx edi, WORD PTR[rsp + temp_qcoeff + rdx *2]
|
||||||
|
|
||||||
; downshift by quant_shift[rdx]
|
; downshift by quant_shift[rdx]
|
||||||
movsx ecx, WORD PTR[rax + rdx*2] ; quant_shift_ptr[rc]
|
movsx ecx, WORD PTR[rax + rdx*2] ; quant_shift_ptr[rc]
|
||||||
sar ebx, cl ; also sets Z bit
|
sar edi, cl ; also sets Z bit
|
||||||
je rq_zigzag_loop_%2 ; !y
|
je rq_zigzag_loop_%1 ; !y
|
||||||
mov WORD PTR[rsp + qcoeff + rdx * 2], bx ;qcoeff_ptr[rc] = temp_qcoeff[rc]
|
mov WORD PTR[rsp + qcoeff + rdx*2], di ;qcoeff_ptr[rc] = temp_qcoeff[rc]
|
||||||
|
mov rsi, [rsp + zrun_zbin_boost] ; reset to b->zrun_zbin_boost
|
||||||
mov rsi, arg(9) ; reset to b->zrun_zbin_boost
|
rq_zigzag_loop_%1:
|
||||||
mov [rsp + eob_tmp], DWORD %1 ; eob = i
|
|
||||||
%endmacro
|
%endmacro
|
||||||
ZIGZAG_LOOP 0, 1
|
ZIGZAG_LOOP 0
|
||||||
ZIGZAG_LOOP 1, 2
|
ZIGZAG_LOOP 1
|
||||||
ZIGZAG_LOOP 2, 3
|
ZIGZAG_LOOP 2
|
||||||
ZIGZAG_LOOP 3, 4
|
ZIGZAG_LOOP 3
|
||||||
ZIGZAG_LOOP 4, 5
|
ZIGZAG_LOOP 4
|
||||||
ZIGZAG_LOOP 5, 6
|
ZIGZAG_LOOP 5
|
||||||
ZIGZAG_LOOP 6, 7
|
ZIGZAG_LOOP 6
|
||||||
ZIGZAG_LOOP 7, 8
|
ZIGZAG_LOOP 7
|
||||||
ZIGZAG_LOOP 8, 9
|
ZIGZAG_LOOP 8
|
||||||
ZIGZAG_LOOP 9, 10
|
ZIGZAG_LOOP 9
|
||||||
ZIGZAG_LOOP 10, 11
|
ZIGZAG_LOOP 10
|
||||||
ZIGZAG_LOOP 11, 12
|
ZIGZAG_LOOP 11
|
||||||
ZIGZAG_LOOP 12, 13
|
ZIGZAG_LOOP 12
|
||||||
ZIGZAG_LOOP 13, 14
|
ZIGZAG_LOOP 13
|
||||||
ZIGZAG_LOOP 14, 15
|
ZIGZAG_LOOP 14
|
||||||
ZIGZAG_LOOP 15, end
|
ZIGZAG_LOOP 15
|
||||||
rq_zigzag_loop_end:
|
|
||||||
|
|
||||||
mov rbx, arg(2) ; qcoeff_ptr
|
movdqa xmm2, [rsp + qcoeff]
|
||||||
mov rcx, arg(3) ; dequant_ptr
|
movdqa xmm3, [rsp + qcoeff + 16]
|
||||||
mov rsi, arg(7) ; dqcoeff_ptr
|
|
||||||
mov rax, [rsp + eob_tmp] ; eob
|
|
||||||
|
|
||||||
movdqa xmm2, OWORD PTR[rsp + qcoeff]
|
%if ABI_IS_32BIT
|
||||||
movdqa xmm3, OWORD PTR[rsp + qcoeff + 16]
|
mov rdi, arg(1)
|
||||||
|
%else
|
||||||
|
mov rdi, [rsp + BLOCKD_d]
|
||||||
|
%endif
|
||||||
|
|
||||||
|
mov rcx, [rdi + vp8_blockd_dequant] ; dequant_ptr
|
||||||
|
mov rsi, [rdi + vp8_blockd_dqcoeff] ; dqcoeff_ptr
|
||||||
|
|
||||||
; y ^ sz
|
; y ^ sz
|
||||||
pxor xmm2, xmm0
|
pxor xmm2, xmm0
|
||||||
@ -175,34 +186,67 @@ rq_zigzag_loop_end:
|
|||||||
psubw xmm2, xmm0
|
psubw xmm2, xmm0
|
||||||
psubw xmm3, xmm4
|
psubw xmm3, xmm4
|
||||||
|
|
||||||
movdqa xmm0, OWORD PTR[rcx]
|
; dequant
|
||||||
movdqa xmm1, OWORD PTR[rcx + 16]
|
movdqa xmm0, [rcx]
|
||||||
|
movdqa xmm1, [rcx + 16]
|
||||||
|
|
||||||
|
mov rcx, [rdi + vp8_blockd_qcoeff] ; qcoeff_ptr
|
||||||
|
|
||||||
pmullw xmm0, xmm2
|
pmullw xmm0, xmm2
|
||||||
pmullw xmm1, xmm3
|
pmullw xmm1, xmm3
|
||||||
|
|
||||||
movdqa OWORD PTR[rbx], xmm2
|
movdqa [rcx], xmm2 ; store qcoeff
|
||||||
movdqa OWORD PTR[rbx + 16], xmm3
|
movdqa [rcx + 16], xmm3
|
||||||
movdqa OWORD PTR[rsi], xmm0 ; store dqcoeff
|
movdqa [rsi], xmm0 ; store dqcoeff
|
||||||
movdqa OWORD PTR[rsi + 16], xmm1 ; store dqcoeff
|
movdqa [rsi + 16], xmm1
|
||||||
|
|
||||||
add rax, 1
|
; select the last value (in zig_zag order) for EOB
|
||||||
|
pcmpeqw xmm2, xmm6
|
||||||
|
pcmpeqw xmm3, xmm6
|
||||||
|
; !
|
||||||
|
pcmpeqw xmm6, xmm6
|
||||||
|
pxor xmm2, xmm6
|
||||||
|
pxor xmm3, xmm6
|
||||||
|
; mask inv_zig_zag
|
||||||
|
pand xmm2, [GLOBAL(inv_zig_zag)]
|
||||||
|
pand xmm3, [GLOBAL(inv_zig_zag) + 16]
|
||||||
|
; select the max value
|
||||||
|
pmaxsw xmm2, xmm3
|
||||||
|
pshufd xmm3, xmm2, 00001110b
|
||||||
|
pmaxsw xmm2, xmm3
|
||||||
|
pshuflw xmm3, xmm2, 00001110b
|
||||||
|
pmaxsw xmm2, xmm3
|
||||||
|
pshuflw xmm3, xmm2, 00000001b
|
||||||
|
pmaxsw xmm2, xmm3
|
||||||
|
movd eax, xmm2
|
||||||
|
and eax, 0xff
|
||||||
|
mov [rdi + vp8_blockd_eob], eax
|
||||||
|
|
||||||
; begin epilog
|
; begin epilog
|
||||||
add rsp, stack_size
|
add rsp, stack_size
|
||||||
pop rsp
|
pop rsp
|
||||||
pop rbx
|
%if ABI_IS_32BIT
|
||||||
pop rdi
|
pop rdi
|
||||||
|
%else
|
||||||
|
%ifidn __OUTPUT_FORMAT__,x64
|
||||||
|
pop rdi
|
||||||
|
%endif
|
||||||
|
%endif
|
||||||
pop rsi
|
pop rsi
|
||||||
|
RESTORE_GOT
|
||||||
RESTORE_XMM
|
RESTORE_XMM
|
||||||
UNSHADOW_ARGS
|
|
||||||
pop rbp
|
pop rbp
|
||||||
ret
|
ret
|
||||||
|
|
||||||
;int vp8_fast_quantize_b_impl_sse2(short *coeff_ptr,
|
; int vp8_fast_quantize_b_impl_sse2 | arg
|
||||||
; short *qcoeff_ptr,short *dequant_ptr,
|
; (short *coeff_ptr, | 0
|
||||||
; short *inv_scan_order, short *round_ptr,
|
; short *qcoeff_ptr, | 1
|
||||||
; short *quant_ptr, short *dqcoeff_ptr);
|
; short *dequant_ptr, | 2
|
||||||
|
; short *inv_scan_order, | 3
|
||||||
|
; short *round_ptr, | 4
|
||||||
|
; short *quant_ptr, | 5
|
||||||
|
; short *dqcoeff_ptr) | 6
|
||||||
|
|
||||||
global sym(vp8_fast_quantize_b_impl_sse2)
|
global sym(vp8_fast_quantize_b_impl_sse2)
|
||||||
sym(vp8_fast_quantize_b_impl_sse2):
|
sym(vp8_fast_quantize_b_impl_sse2):
|
||||||
push rbp
|
push rbp
|
||||||
@ -300,3 +344,16 @@ sym(vp8_fast_quantize_b_impl_sse2):
|
|||||||
UNSHADOW_ARGS
|
UNSHADOW_ARGS
|
||||||
pop rbp
|
pop rbp
|
||||||
ret
|
ret
|
||||||
|
|
||||||
|
SECTION_RODATA
|
||||||
|
align 16
|
||||||
|
zig_zag:
|
||||||
|
dw 0x0000, 0x0001, 0x0004, 0x0008
|
||||||
|
dw 0x0005, 0x0002, 0x0003, 0x0006
|
||||||
|
dw 0x0009, 0x000c, 0x000d, 0x000a
|
||||||
|
dw 0x0007, 0x000b, 0x000e, 0x000f
|
||||||
|
inv_zig_zag:
|
||||||
|
dw 0x0001, 0x0002, 0x0006, 0x0007
|
||||||
|
dw 0x0003, 0x0005, 0x0008, 0x000d
|
||||||
|
dw 0x0004, 0x0009, 0x000c, 0x000e
|
||||||
|
dw 0x000a, 0x000b, 0x000f, 0x0010
|
||||||
|
@ -27,11 +27,8 @@ extern prototype_quantize_block(vp8_regular_quantize_b_sse2);
|
|||||||
|
|
||||||
#if !CONFIG_RUNTIME_CPU_DETECT
|
#if !CONFIG_RUNTIME_CPU_DETECT
|
||||||
|
|
||||||
// Currently, this function realizes a gain on x86 and a loss on x86_64
|
|
||||||
#if ARCH_X86
|
|
||||||
#undef vp8_quantize_quantb
|
#undef vp8_quantize_quantb
|
||||||
#define vp8_quantize_quantb vp8_regular_quantize_b_sse2
|
#define vp8_quantize_quantb vp8_regular_quantize_b_sse2
|
||||||
#endif
|
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
@ -106,30 +106,6 @@ static void fast_quantize_b_sse2(BLOCK *b, BLOCKD *d)
|
|||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
int vp8_regular_quantize_b_impl_sse2(short *coeff_ptr, short *zbin_ptr,
|
|
||||||
short *qcoeff_ptr,short *dequant_ptr,
|
|
||||||
const int *default_zig_zag, short *round_ptr,
|
|
||||||
short *quant_ptr, short *dqcoeff_ptr,
|
|
||||||
unsigned short zbin_oq_value,
|
|
||||||
short *zbin_boost_ptr,
|
|
||||||
short *quant_shift_ptr);
|
|
||||||
|
|
||||||
static void regular_quantize_b_sse2(BLOCK *b,BLOCKD *d)
|
|
||||||
{
|
|
||||||
d->eob = vp8_regular_quantize_b_impl_sse2(b->coeff,
|
|
||||||
b->zbin,
|
|
||||||
d->qcoeff,
|
|
||||||
d->dequant,
|
|
||||||
vp8_default_zig_zag1d,
|
|
||||||
b->round,
|
|
||||||
b->quant,
|
|
||||||
d->dqcoeff,
|
|
||||||
b->zbin_extra,
|
|
||||||
b->zrun_zbin_boost,
|
|
||||||
b->quant_shift);
|
|
||||||
}
|
|
||||||
|
|
||||||
int vp8_mbblock_error_xmm_impl(short *coeff_ptr, short *dcoef_ptr, int dc);
|
int vp8_mbblock_error_xmm_impl(short *coeff_ptr, short *dcoef_ptr, int dc);
|
||||||
static int mbblock_error_xmm(MACROBLOCK *mb, int dc)
|
static int mbblock_error_xmm(MACROBLOCK *mb, int dc)
|
||||||
{
|
{
|
||||||
@ -317,9 +293,7 @@ void vp8_arch_x86_encoder_init(VP8_COMP *cpi)
|
|||||||
cpi->rtcd.encodemb.submby = vp8_subtract_mby_sse2;
|
cpi->rtcd.encodemb.submby = vp8_subtract_mby_sse2;
|
||||||
cpi->rtcd.encodemb.submbuv = vp8_subtract_mbuv_sse2;
|
cpi->rtcd.encodemb.submbuv = vp8_subtract_mbuv_sse2;
|
||||||
|
|
||||||
#if ARCH_X86
|
cpi->rtcd.quantize.quantb = vp8_regular_quantize_b_sse2;
|
||||||
cpi->rtcd.quantize.quantb = regular_quantize_b_sse2;
|
|
||||||
#endif
|
|
||||||
cpi->rtcd.quantize.fastquantb = fast_quantize_b_sse2;
|
cpi->rtcd.quantize.fastquantb = fast_quantize_b_sse2;
|
||||||
|
|
||||||
#if !(CONFIG_REALTIME_ONLY)
|
#if !(CONFIG_REALTIME_ONLY)
|
||||||
|
Loading…
x
Reference in New Issue
Block a user