~15% faster. Also add an mmxext version that takes advantage of the new code, and build it alongside with the mmx version only on x86_32. Reviewed-by: Michael Niedermayer <michaelni@gmx.at> Signed-off-by: James Almer <jamrial@gmail.com>
		
			
				
	
	
		
			155 lines
		
	
	
		
			3.6 KiB
		
	
	
	
		
			NASM
		
	
	
	
	
	
			
		
		
	
	
			155 lines
		
	
	
		
			3.6 KiB
		
	
	
	
		
			NASM
		
	
	
	
	
	
;*****************************************************************************
 | 
						|
;* SIMD-optimized MPEG encoding functions
 | 
						|
;*****************************************************************************
 | 
						|
;* Copyright (c) 2000, 2001 Fabrice Bellard
 | 
						|
;* Copyright (c) 2002-2004 Michael Niedermayer <michaelni@gmx.at>
 | 
						|
;*
 | 
						|
;* This file is part of FFmpeg.
 | 
						|
;*
 | 
						|
;* FFmpeg is free software; you can redistribute it and/or
 | 
						|
;* modify it under the terms of the GNU Lesser General Public
 | 
						|
;* License as published by the Free Software Foundation; either
 | 
						|
;* version 2.1 of the License, or (at your option) any later version.
 | 
						|
;*
 | 
						|
;* FFmpeg is distributed in the hope that it will be useful,
 | 
						|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
						|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 | 
						|
;* Lesser General Public License for more details.
 | 
						|
;*
 | 
						|
;* You should have received a copy of the GNU Lesser General Public
 | 
						|
;* License along with FFmpeg; if not, write to the Free Software
 | 
						|
;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
 | 
						|
;*****************************************************************************
 | 
						|
 | 
						|
%include "libavutil/x86/x86util.asm"
 | 
						|
 | 
						|
SECTION_RODATA
 | 
						|
 | 
						|
cextern pw_1
 | 
						|
 | 
						|
SECTION .text
 | 
						|
; int ff_pix_sum16_mmx(uint8_t *pix, int line_size)
 | 
						|
; %1 = number of loops
 | 
						|
; %2 = number of GPRs used
 | 
						|
%macro PIX_SUM16 3
 | 
						|
cglobal pix_sum16, 2, %2, 6
 | 
						|
    movsxdifnidn r1, r1d
 | 
						|
    mov          r2, %1
 | 
						|
%if mmsize == 16
 | 
						|
    lea          r3, [r1*3]
 | 
						|
%endif
 | 
						|
%if notcpuflag(xop)
 | 
						|
    pxor         m5, m5
 | 
						|
%endif
 | 
						|
    pxor         m4, m4
 | 
						|
.loop:
 | 
						|
%if cpuflag(xop)
 | 
						|
    vphaddubq    m0, [r0]
 | 
						|
    vphaddubq    m1, [r0+r1]
 | 
						|
    vphaddubq    m2, [r0+r1*2]
 | 
						|
    vphaddubq    m3, [r0+r3]
 | 
						|
%else
 | 
						|
    mova         m0, [r0]
 | 
						|
%if mmsize == 8
 | 
						|
    mova         m1, [r0+8]
 | 
						|
%if cpuflag(mmxext)
 | 
						|
    mova         m2, [r0+r1]
 | 
						|
    mova         m3, [r0+r1+8]
 | 
						|
%endif
 | 
						|
%else ; sse2
 | 
						|
    mova         m1, [r0+r1]
 | 
						|
    mova         m2, [r0+r1*2]
 | 
						|
    mova         m3, [r0+r3]
 | 
						|
%endif
 | 
						|
%if cpuflag(mmxext)
 | 
						|
    psadbw       m0, m5
 | 
						|
    psadbw       m1, m5
 | 
						|
    psadbw       m2, m5
 | 
						|
    psadbw       m3, m5
 | 
						|
%else ; mmx
 | 
						|
    punpckhbw    m2, m0, m5
 | 
						|
    punpcklbw    m0, m5
 | 
						|
    punpckhbw    m3, m1, m5
 | 
						|
    punpcklbw    m1, m5
 | 
						|
%endif ; cpuflag(mmxext)
 | 
						|
%endif ; cpuflag(xop)
 | 
						|
    paddw        m1, m0
 | 
						|
    paddw        m3, m2
 | 
						|
    paddw        m3, m1
 | 
						|
    paddw        m4, m3
 | 
						|
%if cpuflag(mmxext)
 | 
						|
    lea          r0, [r0+r1*%3]
 | 
						|
%else
 | 
						|
    add          r0, r1
 | 
						|
%endif
 | 
						|
    dec r2
 | 
						|
    jne .loop
 | 
						|
%if mmsize == 16
 | 
						|
    pshufd       m0, m4, q0032
 | 
						|
    paddd        m4, m0
 | 
						|
%elif notcpuflag(mmxext)
 | 
						|
    HADDW        m4, m5
 | 
						|
%endif
 | 
						|
    movd        eax, m4
 | 
						|
    RET
 | 
						|
%endmacro
 | 
						|
 | 
						|
%if ARCH_X86_32
 | 
						|
INIT_MMX mmx
 | 
						|
PIX_SUM16 16, 3, 0
 | 
						|
INIT_MMX mmxext
 | 
						|
PIX_SUM16  8, 4, 2
 | 
						|
%endif
 | 
						|
INIT_XMM sse2
 | 
						|
PIX_SUM16  4, 4, 4
 | 
						|
%if HAVE_XOP_EXTERNAL
 | 
						|
INIT_XMM xop
 | 
						|
PIX_SUM16  4, 4, 4
 | 
						|
%endif
 | 
						|
 | 
						|
; int ff_pix_norm1_mmx(uint8_t *pix, int line_size)
 | 
						|
; %1 = number of xmm registers used
 | 
						|
; %2 = number of loops
 | 
						|
%macro PIX_NORM1 2
 | 
						|
cglobal pix_norm1, 2, 3, %1
 | 
						|
    movsxdifnidn r1, r1d
 | 
						|
    mov          r2, %2
 | 
						|
    pxor         m0, m0
 | 
						|
    pxor         m5, m5
 | 
						|
.loop:
 | 
						|
    mova         m2, [r0+0]
 | 
						|
%if mmsize == 8
 | 
						|
    mova         m3, [r0+8]
 | 
						|
%else
 | 
						|
    mova         m3, [r0+r1]
 | 
						|
%endif
 | 
						|
    punpckhbw    m1, m2, m0
 | 
						|
    punpcklbw    m2, m0
 | 
						|
    punpckhbw    m4, m3, m0
 | 
						|
    punpcklbw    m3, m0
 | 
						|
    pmaddwd      m1, m1
 | 
						|
    pmaddwd      m2, m2
 | 
						|
    pmaddwd      m3, m3
 | 
						|
    pmaddwd      m4, m4
 | 
						|
    paddd        m2, m1
 | 
						|
    paddd        m4, m3
 | 
						|
    paddd        m5, m2
 | 
						|
    paddd        m5, m4
 | 
						|
%if mmsize == 8
 | 
						|
    add          r0, r1
 | 
						|
%else
 | 
						|
    lea          r0, [r0+r1*2]
 | 
						|
%endif
 | 
						|
    dec r2
 | 
						|
    jne .loop
 | 
						|
    HADDD        m5, m1
 | 
						|
    movd        eax, m5
 | 
						|
    RET
 | 
						|
%endmacro
 | 
						|
 | 
						|
INIT_MMX mmx
 | 
						|
PIX_NORM1 0, 16
 | 
						|
INIT_XMM sse2
 | 
						|
PIX_NORM1 6, 8
 | 
						|
 |