* commit '2d91abade29e43bb45c881d45909b8ee77e904e2': x86: h264_intrapred: Don't treat 32-bit integers as 64-bit Merged-by: Michael Niedermayer <michaelni@gmx.at>
		
			
				
	
	
		
			2718 lines
		
	
	
		
			69 KiB
		
	
	
	
		
			NASM
		
	
	
	
	
	
			
		
		
	
	
			2718 lines
		
	
	
		
			69 KiB
		
	
	
	
		
			NASM
		
	
	
	
	
	
;******************************************************************************
 | 
						|
;* H.264 intra prediction asm optimizations
 | 
						|
;* Copyright (c) 2010 Fiona Glaser
 | 
						|
;* Copyright (c) 2010 Holger Lubitz
 | 
						|
;* Copyright (c) 2010 Loren Merritt
 | 
						|
;* Copyright (c) 2010 Ronald S. Bultje
 | 
						|
;*
 | 
						|
;* This file is part of FFmpeg.
 | 
						|
;*
 | 
						|
;* FFmpeg is free software; you can redistribute it and/or
 | 
						|
;* modify it under the terms of the GNU Lesser General Public
 | 
						|
;* License as published by the Free Software Foundation; either
 | 
						|
;* version 2.1 of the License, or (at your option) any later version.
 | 
						|
;*
 | 
						|
;* FFmpeg is distributed in the hope that it will be useful,
 | 
						|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
						|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 | 
						|
;* Lesser General Public License for more details.
 | 
						|
;*
 | 
						|
;* You should have received a copy of the GNU Lesser General Public
 | 
						|
;* License along with FFmpeg; if not, write to the Free Software
 | 
						|
;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
 | 
						|
;******************************************************************************
 | 
						|
 | 
						|
%include "libavutil/x86/x86util.asm"
 | 
						|
 | 
						|
SECTION_RODATA
 | 
						|
 | 
						|
tm_shuf: times 8 db 0x03, 0x80
 | 
						|
pw_ff00: times 8 dw 0xff00
 | 
						|
plane_shuf:  db -8, -7, -6, -5, -4, -3, -2, -1
 | 
						|
             db  1,  2,  3,  4,  5,  6,  7,  8
 | 
						|
plane8_shuf: db -4, -3, -2, -1,  0,  0,  0,  0
 | 
						|
             db  1,  2,  3,  4,  0,  0,  0,  0
 | 
						|
pw_0to7:     dw  0,  1,  2,  3,  4,  5,  6,  7
 | 
						|
pw_1to8:     dw  1,  2,  3,  4,  5,  6,  7,  8
 | 
						|
pw_m8tom1:   dw -8, -7, -6, -5, -4, -3, -2, -1
 | 
						|
pw_m4to4:    dw -4, -3, -2, -1,  1,  2,  3,  4
 | 
						|
 | 
						|
SECTION .text
 | 
						|
 | 
						|
cextern pb_1
 | 
						|
cextern pb_3
 | 
						|
cextern pw_4
 | 
						|
cextern pw_5
 | 
						|
cextern pw_8
 | 
						|
cextern pw_16
 | 
						|
cextern pw_17
 | 
						|
cextern pw_32
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred16x16_vertical_8(uint8_t *src, int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
INIT_MMX mmx
 | 
						|
cglobal pred16x16_vertical_8, 2,3
 | 
						|
    sub   r0, r1
 | 
						|
    mov   r2, 8
 | 
						|
    movq mm0, [r0+0]
 | 
						|
    movq mm1, [r0+8]
 | 
						|
.loop:
 | 
						|
    movq [r0+r1*1+0], mm0
 | 
						|
    movq [r0+r1*1+8], mm1
 | 
						|
    movq [r0+r1*2+0], mm0
 | 
						|
    movq [r0+r1*2+8], mm1
 | 
						|
    lea   r0, [r0+r1*2]
 | 
						|
    dec   r2
 | 
						|
    jg .loop
 | 
						|
    REP_RET
 | 
						|
 | 
						|
INIT_XMM sse
 | 
						|
cglobal pred16x16_vertical_8, 2,3
 | 
						|
    sub   r0, r1
 | 
						|
    mov   r2, 4
 | 
						|
    movaps xmm0, [r0]
 | 
						|
.loop:
 | 
						|
    movaps [r0+r1*1], xmm0
 | 
						|
    movaps [r0+r1*2], xmm0
 | 
						|
    lea   r0, [r0+r1*2]
 | 
						|
    movaps [r0+r1*1], xmm0
 | 
						|
    movaps [r0+r1*2], xmm0
 | 
						|
    lea   r0, [r0+r1*2]
 | 
						|
    dec   r2
 | 
						|
    jg .loop
 | 
						|
    REP_RET
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred16x16_horizontal_8(uint8_t *src, int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
%macro PRED16x16_H 0
 | 
						|
cglobal pred16x16_horizontal_8, 2,3
 | 
						|
    mov       r2, 8
 | 
						|
%if cpuflag(ssse3)
 | 
						|
    mova      m2, [pb_3]
 | 
						|
%endif
 | 
						|
.loop:
 | 
						|
    movd      m0, [r0+r1*0-4]
 | 
						|
    movd      m1, [r0+r1*1-4]
 | 
						|
 | 
						|
%if cpuflag(ssse3)
 | 
						|
    pshufb    m0, m2
 | 
						|
    pshufb    m1, m2
 | 
						|
%else
 | 
						|
    punpcklbw m0, m0
 | 
						|
    punpcklbw m1, m1
 | 
						|
    SPLATW    m0, m0, 3
 | 
						|
    SPLATW    m1, m1, 3
 | 
						|
    mova [r0+r1*0+8], m0
 | 
						|
    mova [r0+r1*1+8], m1
 | 
						|
%endif
 | 
						|
 | 
						|
    mova [r0+r1*0], m0
 | 
						|
    mova [r0+r1*1], m1
 | 
						|
    lea       r0, [r0+r1*2]
 | 
						|
    dec       r2
 | 
						|
    jg .loop
 | 
						|
    REP_RET
 | 
						|
%endmacro
 | 
						|
 | 
						|
INIT_MMX mmx
 | 
						|
PRED16x16_H
 | 
						|
INIT_MMX mmxext
 | 
						|
PRED16x16_H
 | 
						|
INIT_XMM ssse3
 | 
						|
PRED16x16_H
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred16x16_dc_8(uint8_t *src, int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
%macro PRED16x16_DC 0
 | 
						|
cglobal pred16x16_dc_8, 2,7
 | 
						|
    mov       r4, r0
 | 
						|
    sub       r0, r1
 | 
						|
    pxor      mm0, mm0
 | 
						|
    pxor      mm1, mm1
 | 
						|
    psadbw    mm0, [r0+0]
 | 
						|
    psadbw    mm1, [r0+8]
 | 
						|
    dec        r0
 | 
						|
    movzx     r5d, byte [r0+r1*1]
 | 
						|
    paddw     mm0, mm1
 | 
						|
    movd      r6d, mm0
 | 
						|
    lea        r0, [r0+r1*2]
 | 
						|
%rep 7
 | 
						|
    movzx     r2d, byte [r0+r1*0]
 | 
						|
    movzx     r3d, byte [r0+r1*1]
 | 
						|
    add       r5d, r2d
 | 
						|
    add       r6d, r3d
 | 
						|
    lea        r0, [r0+r1*2]
 | 
						|
%endrep
 | 
						|
    movzx     r2d, byte [r0+r1*0]
 | 
						|
    add       r5d, r6d
 | 
						|
    lea       r2d, [r2+r5+16]
 | 
						|
    shr       r2d, 5
 | 
						|
%if cpuflag(ssse3)
 | 
						|
    pxor       m1, m1
 | 
						|
%endif
 | 
						|
    SPLATB_REG m0, r2, m1
 | 
						|
 | 
						|
%if mmsize==8
 | 
						|
    mov       r3d, 8
 | 
						|
.loop:
 | 
						|
    mova [r4+r1*0+0], m0
 | 
						|
    mova [r4+r1*0+8], m0
 | 
						|
    mova [r4+r1*1+0], m0
 | 
						|
    mova [r4+r1*1+8], m0
 | 
						|
%else
 | 
						|
    mov       r3d, 4
 | 
						|
.loop:
 | 
						|
    mova [r4+r1*0], m0
 | 
						|
    mova [r4+r1*1], m0
 | 
						|
    lea   r4, [r4+r1*2]
 | 
						|
    mova [r4+r1*0], m0
 | 
						|
    mova [r4+r1*1], m0
 | 
						|
%endif
 | 
						|
    lea   r4, [r4+r1*2]
 | 
						|
    dec   r3d
 | 
						|
    jg .loop
 | 
						|
    REP_RET
 | 
						|
%endmacro
 | 
						|
 | 
						|
INIT_MMX mmxext
 | 
						|
PRED16x16_DC
 | 
						|
INIT_XMM sse2
 | 
						|
PRED16x16_DC
 | 
						|
INIT_XMM ssse3
 | 
						|
PRED16x16_DC
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred16x16_tm_vp8_8(uint8_t *src, int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
%macro PRED16x16_TM 0
 | 
						|
cglobal pred16x16_tm_vp8_8, 2,5
 | 
						|
    sub        r0, r1
 | 
						|
    pxor      mm7, mm7
 | 
						|
    movq      mm0, [r0+0]
 | 
						|
    movq      mm2, [r0+8]
 | 
						|
    movq      mm1, mm0
 | 
						|
    movq      mm3, mm2
 | 
						|
    punpcklbw mm0, mm7
 | 
						|
    punpckhbw mm1, mm7
 | 
						|
    punpcklbw mm2, mm7
 | 
						|
    punpckhbw mm3, mm7
 | 
						|
    movzx     r3d, byte [r0-1]
 | 
						|
    mov       r4d, 16
 | 
						|
.loop:
 | 
						|
    movzx     r2d, byte [r0+r1-1]
 | 
						|
    sub       r2d, r3d
 | 
						|
    movd      mm4, r2d
 | 
						|
    SPLATW    mm4, mm4, 0
 | 
						|
    movq      mm5, mm4
 | 
						|
    movq      mm6, mm4
 | 
						|
    movq      mm7, mm4
 | 
						|
    paddw     mm4, mm0
 | 
						|
    paddw     mm5, mm1
 | 
						|
    paddw     mm6, mm2
 | 
						|
    paddw     mm7, mm3
 | 
						|
    packuswb  mm4, mm5
 | 
						|
    packuswb  mm6, mm7
 | 
						|
    movq [r0+r1+0], mm4
 | 
						|
    movq [r0+r1+8], mm6
 | 
						|
    add        r0, r1
 | 
						|
    dec       r4d
 | 
						|
    jg .loop
 | 
						|
    REP_RET
 | 
						|
%endmacro
 | 
						|
 | 
						|
INIT_MMX mmx
 | 
						|
PRED16x16_TM
 | 
						|
INIT_MMX mmxext
 | 
						|
PRED16x16_TM
 | 
						|
 | 
						|
INIT_XMM sse2
 | 
						|
cglobal pred16x16_tm_vp8_8, 2,6,6
 | 
						|
    sub          r0, r1
 | 
						|
    pxor       xmm2, xmm2
 | 
						|
    movdqa     xmm0, [r0]
 | 
						|
    movdqa     xmm1, xmm0
 | 
						|
    punpcklbw  xmm0, xmm2
 | 
						|
    punpckhbw  xmm1, xmm2
 | 
						|
    movzx       r4d, byte [r0-1]
 | 
						|
    mov         r5d, 8
 | 
						|
.loop:
 | 
						|
    movzx       r2d, byte [r0+r1*1-1]
 | 
						|
    movzx       r3d, byte [r0+r1*2-1]
 | 
						|
    sub         r2d, r4d
 | 
						|
    sub         r3d, r4d
 | 
						|
    movd       xmm2, r2d
 | 
						|
    movd       xmm4, r3d
 | 
						|
    pshuflw    xmm2, xmm2, 0
 | 
						|
    pshuflw    xmm4, xmm4, 0
 | 
						|
    punpcklqdq xmm2, xmm2
 | 
						|
    punpcklqdq xmm4, xmm4
 | 
						|
    movdqa     xmm3, xmm2
 | 
						|
    movdqa     xmm5, xmm4
 | 
						|
    paddw      xmm2, xmm0
 | 
						|
    paddw      xmm3, xmm1
 | 
						|
    paddw      xmm4, xmm0
 | 
						|
    paddw      xmm5, xmm1
 | 
						|
    packuswb   xmm2, xmm3
 | 
						|
    packuswb   xmm4, xmm5
 | 
						|
    movdqa [r0+r1*1], xmm2
 | 
						|
    movdqa [r0+r1*2], xmm4
 | 
						|
    lea          r0, [r0+r1*2]
 | 
						|
    dec         r5d
 | 
						|
    jg .loop
 | 
						|
    REP_RET
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred16x16_plane_*_8(uint8_t *src, int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
%macro H264_PRED16x16_PLANE 1
 | 
						|
cglobal pred16x16_plane_%1_8, 2,9,7
 | 
						|
    mov          r2, r1           ; +stride
 | 
						|
    neg          r1               ; -stride
 | 
						|
 | 
						|
    movh         m0, [r0+r1  -1]
 | 
						|
%if mmsize == 8
 | 
						|
    pxor         m4, m4
 | 
						|
    movh         m1, [r0+r1  +3 ]
 | 
						|
    movh         m2, [r0+r1  +8 ]
 | 
						|
    movh         m3, [r0+r1  +12]
 | 
						|
    punpcklbw    m0, m4
 | 
						|
    punpcklbw    m1, m4
 | 
						|
    punpcklbw    m2, m4
 | 
						|
    punpcklbw    m3, m4
 | 
						|
    pmullw       m0, [pw_m8tom1  ]
 | 
						|
    pmullw       m1, [pw_m8tom1+8]
 | 
						|
    pmullw       m2, [pw_1to8    ]
 | 
						|
    pmullw       m3, [pw_1to8  +8]
 | 
						|
    paddw        m0, m2
 | 
						|
    paddw        m1, m3
 | 
						|
%else ; mmsize == 16
 | 
						|
%if cpuflag(ssse3)
 | 
						|
    movhps       m0, [r0+r1  +8]
 | 
						|
    pmaddubsw    m0, [plane_shuf] ; H coefficients
 | 
						|
%else ; sse2
 | 
						|
    pxor         m2, m2
 | 
						|
    movh         m1, [r0+r1  +8]
 | 
						|
    punpcklbw    m0, m2
 | 
						|
    punpcklbw    m1, m2
 | 
						|
    pmullw       m0, [pw_m8tom1]
 | 
						|
    pmullw       m1, [pw_1to8]
 | 
						|
    paddw        m0, m1
 | 
						|
%endif
 | 
						|
    movhlps      m1, m0
 | 
						|
%endif
 | 
						|
    paddw        m0, m1
 | 
						|
%if cpuflag(mmxext)
 | 
						|
    PSHUFLW      m1, m0, 0xE
 | 
						|
%elif cpuflag(mmx)
 | 
						|
    mova         m1, m0
 | 
						|
    psrlq        m1, 32
 | 
						|
%endif
 | 
						|
    paddw        m0, m1
 | 
						|
%if cpuflag(mmxext)
 | 
						|
    PSHUFLW      m1, m0, 0x1
 | 
						|
%elif cpuflag(mmx)
 | 
						|
    mova         m1, m0
 | 
						|
    psrlq        m1, 16
 | 
						|
%endif
 | 
						|
    paddw        m0, m1           ; sum of H coefficients
 | 
						|
 | 
						|
    lea          r4, [r0+r2*8-1]
 | 
						|
    lea          r3, [r0+r2*4-1]
 | 
						|
    add          r4, r2
 | 
						|
 | 
						|
%if ARCH_X86_64
 | 
						|
%define e_reg r8
 | 
						|
%else
 | 
						|
%define e_reg r0
 | 
						|
%endif
 | 
						|
 | 
						|
    movzx     e_reg, byte [r3+r2*2   ]
 | 
						|
    movzx        r5, byte [r4+r1     ]
 | 
						|
    sub          r5, e_reg
 | 
						|
 | 
						|
    movzx     e_reg, byte [r3+r2     ]
 | 
						|
    movzx        r6, byte [r4        ]
 | 
						|
    sub          r6, e_reg
 | 
						|
    lea          r5, [r5+r6*2]
 | 
						|
 | 
						|
    movzx     e_reg, byte [r3+r1     ]
 | 
						|
    movzx        r6, byte [r4+r2*2   ]
 | 
						|
    sub          r6, e_reg
 | 
						|
    lea          r5, [r5+r6*4]
 | 
						|
 | 
						|
    movzx     e_reg, byte [r3        ]
 | 
						|
%if ARCH_X86_64
 | 
						|
    movzx        r7, byte [r4+r2     ]
 | 
						|
    sub          r7, e_reg
 | 
						|
%else
 | 
						|
    movzx        r6, byte [r4+r2     ]
 | 
						|
    sub          r6, e_reg
 | 
						|
    lea          r5, [r5+r6*4]
 | 
						|
    sub          r5, r6
 | 
						|
%endif
 | 
						|
 | 
						|
    lea       e_reg, [r3+r1*4]
 | 
						|
    lea          r3, [r4+r2*4]
 | 
						|
 | 
						|
    movzx        r4, byte [e_reg+r2  ]
 | 
						|
    movzx        r6, byte [r3        ]
 | 
						|
    sub          r6, r4
 | 
						|
%if ARCH_X86_64
 | 
						|
    lea          r6, [r7+r6*2]
 | 
						|
    lea          r5, [r5+r6*2]
 | 
						|
    add          r5, r6
 | 
						|
%else
 | 
						|
    lea          r5, [r5+r6*4]
 | 
						|
    lea          r5, [r5+r6*2]
 | 
						|
%endif
 | 
						|
 | 
						|
    movzx        r4, byte [e_reg     ]
 | 
						|
%if ARCH_X86_64
 | 
						|
    movzx        r7, byte [r3   +r2  ]
 | 
						|
    sub          r7, r4
 | 
						|
    sub          r5, r7
 | 
						|
%else
 | 
						|
    movzx        r6, byte [r3   +r2  ]
 | 
						|
    sub          r6, r4
 | 
						|
    lea          r5, [r5+r6*8]
 | 
						|
    sub          r5, r6
 | 
						|
%endif
 | 
						|
 | 
						|
    movzx        r4, byte [e_reg+r1  ]
 | 
						|
    movzx        r6, byte [r3   +r2*2]
 | 
						|
    sub          r6, r4
 | 
						|
%if ARCH_X86_64
 | 
						|
    add          r6, r7
 | 
						|
%endif
 | 
						|
    lea          r5, [r5+r6*8]
 | 
						|
 | 
						|
    movzx        r4, byte [e_reg+r2*2]
 | 
						|
    movzx        r6, byte [r3   +r1  ]
 | 
						|
    sub          r6, r4
 | 
						|
    lea          r5, [r5+r6*4]
 | 
						|
    add          r5, r6           ; sum of V coefficients
 | 
						|
 | 
						|
%if ARCH_X86_64 == 0
 | 
						|
    mov          r0, r0m
 | 
						|
%endif
 | 
						|
 | 
						|
%ifidn %1, h264
 | 
						|
    lea          r5, [r5*5+32]
 | 
						|
    sar          r5, 6
 | 
						|
%elifidn %1, rv40
 | 
						|
    lea          r5, [r5*5]
 | 
						|
    sar          r5, 6
 | 
						|
%elifidn %1, svq3
 | 
						|
    test         r5, r5
 | 
						|
    lea          r6, [r5+3]
 | 
						|
    cmovs        r5, r6
 | 
						|
    sar          r5, 2            ; V/4
 | 
						|
    lea          r5, [r5*5]       ; 5*(V/4)
 | 
						|
    test         r5, r5
 | 
						|
    lea          r6, [r5+15]
 | 
						|
    cmovs        r5, r6
 | 
						|
    sar          r5, 4            ; (5*(V/4))/16
 | 
						|
%endif
 | 
						|
 | 
						|
    movzx        r4, byte [r0+r1  +15]
 | 
						|
    movzx        r3, byte [r3+r2*2   ]
 | 
						|
    lea          r3, [r3+r4+1]
 | 
						|
    shl          r3, 4
 | 
						|
 | 
						|
    movd        r1d, m0
 | 
						|
    movsx       r1d, r1w
 | 
						|
%ifnidn %1, svq3
 | 
						|
%ifidn %1, h264
 | 
						|
    lea         r1d, [r1d*5+32]
 | 
						|
%else ; rv40
 | 
						|
    lea         r1d, [r1d*5]
 | 
						|
%endif
 | 
						|
    sar         r1d, 6
 | 
						|
%else ; svq3
 | 
						|
    test        r1d, r1d
 | 
						|
    lea         r4d, [r1d+3]
 | 
						|
    cmovs       r1d, r4d
 | 
						|
    sar         r1d, 2           ; H/4
 | 
						|
    lea         r1d, [r1d*5]     ; 5*(H/4)
 | 
						|
    test        r1d, r1d
 | 
						|
    lea         r4d, [r1d+15]
 | 
						|
    cmovs       r1d, r4d
 | 
						|
    sar         r1d, 4           ; (5*(H/4))/16
 | 
						|
%endif
 | 
						|
    movd         m0, r1d
 | 
						|
 | 
						|
    add         r1d, r5d
 | 
						|
    add         r3d, r1d
 | 
						|
    shl         r1d, 3
 | 
						|
    sub         r3d, r1d          ; a
 | 
						|
 | 
						|
    movd         m1, r5d
 | 
						|
    movd         m3, r3d
 | 
						|
    SPLATW       m0, m0, 0        ; H
 | 
						|
    SPLATW       m1, m1, 0        ; V
 | 
						|
    SPLATW       m3, m3, 0        ; a
 | 
						|
%ifidn %1, svq3
 | 
						|
    SWAP          0, 1
 | 
						|
%endif
 | 
						|
    mova         m2, m0
 | 
						|
%if mmsize == 8
 | 
						|
    mova         m5, m0
 | 
						|
%endif
 | 
						|
    pmullw       m0, [pw_0to7]    ; 0*H, 1*H, ..., 7*H  (words)
 | 
						|
%if mmsize == 16
 | 
						|
    psllw        m2, 3
 | 
						|
%else
 | 
						|
    psllw        m5, 3
 | 
						|
    psllw        m2, 2
 | 
						|
    mova         m6, m5
 | 
						|
    paddw        m6, m2
 | 
						|
%endif
 | 
						|
    paddw        m0, m3           ; a + {0,1,2,3,4,5,6,7}*H
 | 
						|
    paddw        m2, m0           ; a + {8,9,10,11,12,13,14,15}*H
 | 
						|
%if mmsize == 8
 | 
						|
    paddw        m5, m0           ; a + {8,9,10,11}*H
 | 
						|
    paddw        m6, m0           ; a + {12,13,14,15}*H
 | 
						|
%endif
 | 
						|
 | 
						|
    mov          r4, 8
 | 
						|
.loop:
 | 
						|
    mova         m3, m0           ; b[0..7]
 | 
						|
    mova         m4, m2           ; b[8..15]
 | 
						|
    psraw        m3, 5
 | 
						|
    psraw        m4, 5
 | 
						|
    packuswb     m3, m4
 | 
						|
    mova       [r0], m3
 | 
						|
%if mmsize == 8
 | 
						|
    mova         m3, m5           ; b[8..11]
 | 
						|
    mova         m4, m6           ; b[12..15]
 | 
						|
    psraw        m3, 5
 | 
						|
    psraw        m4, 5
 | 
						|
    packuswb     m3, m4
 | 
						|
    mova     [r0+8], m3
 | 
						|
%endif
 | 
						|
    paddw        m0, m1
 | 
						|
    paddw        m2, m1
 | 
						|
%if mmsize == 8
 | 
						|
    paddw        m5, m1
 | 
						|
    paddw        m6, m1
 | 
						|
%endif
 | 
						|
 | 
						|
    mova         m3, m0           ; b[0..7]
 | 
						|
    mova         m4, m2           ; b[8..15]
 | 
						|
    psraw        m3, 5
 | 
						|
    psraw        m4, 5
 | 
						|
    packuswb     m3, m4
 | 
						|
    mova    [r0+r2], m3
 | 
						|
%if mmsize == 8
 | 
						|
    mova         m3, m5           ; b[8..11]
 | 
						|
    mova         m4, m6           ; b[12..15]
 | 
						|
    psraw        m3, 5
 | 
						|
    psraw        m4, 5
 | 
						|
    packuswb     m3, m4
 | 
						|
    mova  [r0+r2+8], m3
 | 
						|
%endif
 | 
						|
    paddw        m0, m1
 | 
						|
    paddw        m2, m1
 | 
						|
%if mmsize == 8
 | 
						|
    paddw        m5, m1
 | 
						|
    paddw        m6, m1
 | 
						|
%endif
 | 
						|
 | 
						|
    lea          r0, [r0+r2*2]
 | 
						|
    dec          r4
 | 
						|
    jg .loop
 | 
						|
    REP_RET
 | 
						|
%endmacro
 | 
						|
 | 
						|
INIT_MMX mmx
 | 
						|
H264_PRED16x16_PLANE h264
 | 
						|
H264_PRED16x16_PLANE rv40
 | 
						|
H264_PRED16x16_PLANE svq3
 | 
						|
INIT_MMX mmxext
 | 
						|
H264_PRED16x16_PLANE h264
 | 
						|
H264_PRED16x16_PLANE rv40
 | 
						|
H264_PRED16x16_PLANE svq3
 | 
						|
INIT_XMM sse2
 | 
						|
H264_PRED16x16_PLANE h264
 | 
						|
H264_PRED16x16_PLANE rv40
 | 
						|
H264_PRED16x16_PLANE svq3
 | 
						|
INIT_XMM ssse3
 | 
						|
H264_PRED16x16_PLANE h264
 | 
						|
H264_PRED16x16_PLANE rv40
 | 
						|
H264_PRED16x16_PLANE svq3
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred8x8_plane_8(uint8_t *src, int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
%macro H264_PRED8x8_PLANE 0
 | 
						|
cglobal pred8x8_plane_8, 2,9,7
 | 
						|
    mov          r2, r1           ; +stride
 | 
						|
    neg          r1               ; -stride
 | 
						|
 | 
						|
    movd         m0, [r0+r1  -1]
 | 
						|
%if mmsize == 8
 | 
						|
    pxor         m2, m2
 | 
						|
    movh         m1, [r0+r1  +4 ]
 | 
						|
    punpcklbw    m0, m2
 | 
						|
    punpcklbw    m1, m2
 | 
						|
    pmullw       m0, [pw_m4to4]
 | 
						|
    pmullw       m1, [pw_m4to4+8]
 | 
						|
%else ; mmsize == 16
 | 
						|
%if cpuflag(ssse3)
 | 
						|
    movhps       m0, [r0+r1  +4]   ; this reads 4 bytes more than necessary
 | 
						|
    pmaddubsw    m0, [plane8_shuf] ; H coefficients
 | 
						|
%else ; sse2
 | 
						|
    pxor         m2, m2
 | 
						|
    movd         m1, [r0+r1  +4]
 | 
						|
    punpckldq    m0, m1
 | 
						|
    punpcklbw    m0, m2
 | 
						|
    pmullw       m0, [pw_m4to4]
 | 
						|
%endif
 | 
						|
    movhlps      m1, m0
 | 
						|
%endif
 | 
						|
    paddw        m0, m1
 | 
						|
 | 
						|
%if notcpuflag(ssse3)
 | 
						|
%if cpuflag(mmxext)
 | 
						|
    PSHUFLW      m1, m0, 0xE
 | 
						|
%elif cpuflag(mmx)
 | 
						|
    mova         m1, m0
 | 
						|
    psrlq        m1, 32
 | 
						|
%endif
 | 
						|
    paddw        m0, m1
 | 
						|
%endif ; !ssse3
 | 
						|
 | 
						|
%if cpuflag(mmxext)
 | 
						|
    PSHUFLW      m1, m0, 0x1
 | 
						|
%elif cpuflag(mmx)
 | 
						|
    mova         m1, m0
 | 
						|
    psrlq        m1, 16
 | 
						|
%endif
 | 
						|
    paddw        m0, m1           ; sum of H coefficients
 | 
						|
 | 
						|
    lea          r4, [r0+r2*4-1]
 | 
						|
    lea          r3, [r0     -1]
 | 
						|
    add          r4, r2
 | 
						|
 | 
						|
%if ARCH_X86_64
 | 
						|
%define e_reg r8
 | 
						|
%else
 | 
						|
%define e_reg r0
 | 
						|
%endif
 | 
						|
 | 
						|
    movzx     e_reg, byte [r3+r2*2   ]
 | 
						|
    movzx        r5, byte [r4+r1     ]
 | 
						|
    sub          r5, e_reg
 | 
						|
 | 
						|
    movzx     e_reg, byte [r3        ]
 | 
						|
%if ARCH_X86_64
 | 
						|
    movzx        r7, byte [r4+r2     ]
 | 
						|
    sub          r7, e_reg
 | 
						|
    sub          r5, r7
 | 
						|
%else
 | 
						|
    movzx        r6, byte [r4+r2     ]
 | 
						|
    sub          r6, e_reg
 | 
						|
    lea          r5, [r5+r6*4]
 | 
						|
    sub          r5, r6
 | 
						|
%endif
 | 
						|
 | 
						|
    movzx     e_reg, byte [r3+r1     ]
 | 
						|
    movzx        r6, byte [r4+r2*2   ]
 | 
						|
    sub          r6, e_reg
 | 
						|
%if ARCH_X86_64
 | 
						|
    add          r6, r7
 | 
						|
%endif
 | 
						|
    lea          r5, [r5+r6*4]
 | 
						|
 | 
						|
    movzx     e_reg, byte [r3+r2     ]
 | 
						|
    movzx        r6, byte [r4        ]
 | 
						|
    sub          r6, e_reg
 | 
						|
    lea          r6, [r5+r6*2]
 | 
						|
 | 
						|
    lea          r5, [r6*9+16]
 | 
						|
    lea          r5, [r5+r6*8]
 | 
						|
    sar          r5, 5
 | 
						|
 | 
						|
%if ARCH_X86_64 == 0
 | 
						|
    mov          r0, r0m
 | 
						|
%endif
 | 
						|
 | 
						|
    movzx        r3, byte [r4+r2*2  ]
 | 
						|
    movzx        r4, byte [r0+r1  +7]
 | 
						|
    lea          r3, [r3+r4+1]
 | 
						|
    shl          r3, 4
 | 
						|
    movd        r1d, m0
 | 
						|
    movsx       r1d, r1w
 | 
						|
    imul        r1d, 17
 | 
						|
    add         r1d, 16
 | 
						|
    sar         r1d, 5
 | 
						|
    movd         m0, r1d
 | 
						|
    add         r1d, r5d
 | 
						|
    sub         r3d, r1d
 | 
						|
    add         r1d, r1d
 | 
						|
    sub         r3d, r1d          ; a
 | 
						|
 | 
						|
    movd         m1, r5d
 | 
						|
    movd         m3, r3d
 | 
						|
    SPLATW       m0, m0, 0        ; H
 | 
						|
    SPLATW       m1, m1, 0        ; V
 | 
						|
    SPLATW       m3, m3, 0        ; a
 | 
						|
%if mmsize == 8
 | 
						|
    mova         m2, m0
 | 
						|
%endif
 | 
						|
    pmullw       m0, [pw_0to7]    ; 0*H, 1*H, ..., 7*H  (words)
 | 
						|
    paddw        m0, m3           ; a + {0,1,2,3,4,5,6,7}*H
 | 
						|
%if mmsize == 8
 | 
						|
    psllw        m2, 2
 | 
						|
    paddw        m2, m0           ; a + {4,5,6,7}*H
 | 
						|
%endif
 | 
						|
 | 
						|
    mov          r4, 4
 | 
						|
ALIGN 16
 | 
						|
.loop:
 | 
						|
%if mmsize == 16
 | 
						|
    mova         m3, m0           ; b[0..7]
 | 
						|
    paddw        m0, m1
 | 
						|
    psraw        m3, 5
 | 
						|
    mova         m4, m0           ; V+b[0..7]
 | 
						|
    paddw        m0, m1
 | 
						|
    psraw        m4, 5
 | 
						|
    packuswb     m3, m4
 | 
						|
    movh       [r0], m3
 | 
						|
    movhps  [r0+r2], m3
 | 
						|
%else ; mmsize == 8
 | 
						|
    mova         m3, m0           ; b[0..3]
 | 
						|
    mova         m4, m2           ; b[4..7]
 | 
						|
    paddw        m0, m1
 | 
						|
    paddw        m2, m1
 | 
						|
    psraw        m3, 5
 | 
						|
    psraw        m4, 5
 | 
						|
    mova         m5, m0           ; V+b[0..3]
 | 
						|
    mova         m6, m2           ; V+b[4..7]
 | 
						|
    paddw        m0, m1
 | 
						|
    paddw        m2, m1
 | 
						|
    psraw        m5, 5
 | 
						|
    psraw        m6, 5
 | 
						|
    packuswb     m3, m4
 | 
						|
    packuswb     m5, m6
 | 
						|
    mova       [r0], m3
 | 
						|
    mova    [r0+r2], m5
 | 
						|
%endif
 | 
						|
 | 
						|
    lea          r0, [r0+r2*2]
 | 
						|
    dec          r4
 | 
						|
    jg .loop
 | 
						|
    REP_RET
 | 
						|
%endmacro
 | 
						|
 | 
						|
INIT_MMX mmx
 | 
						|
H264_PRED8x8_PLANE
 | 
						|
INIT_MMX mmxext
 | 
						|
H264_PRED8x8_PLANE
 | 
						|
INIT_XMM sse2
 | 
						|
H264_PRED8x8_PLANE
 | 
						|
INIT_XMM ssse3
 | 
						|
H264_PRED8x8_PLANE
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred8x8_vertical_8(uint8_t *src, int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
INIT_MMX mmx
 | 
						|
cglobal pred8x8_vertical_8, 2,2
 | 
						|
    sub    r0, r1
 | 
						|
    movq  mm0, [r0]
 | 
						|
%rep 3
 | 
						|
    movq [r0+r1*1], mm0
 | 
						|
    movq [r0+r1*2], mm0
 | 
						|
    lea    r0, [r0+r1*2]
 | 
						|
%endrep
 | 
						|
    movq [r0+r1*1], mm0
 | 
						|
    movq [r0+r1*2], mm0
 | 
						|
    RET
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred8x8_horizontal_8(uint8_t *src, int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
%macro PRED8x8_H 0
 | 
						|
cglobal pred8x8_horizontal_8, 2,3
 | 
						|
    mov       r2, 4
 | 
						|
%if cpuflag(ssse3)
 | 
						|
    mova      m2, [pb_3]
 | 
						|
%endif
 | 
						|
.loop:
 | 
						|
    SPLATB_LOAD m0, r0+r1*0-1, m2
 | 
						|
    SPLATB_LOAD m1, r0+r1*1-1, m2
 | 
						|
    mova [r0+r1*0], m0
 | 
						|
    mova [r0+r1*1], m1
 | 
						|
    lea       r0, [r0+r1*2]
 | 
						|
    dec       r2
 | 
						|
    jg .loop
 | 
						|
    REP_RET
 | 
						|
%endmacro
 | 
						|
 | 
						|
INIT_MMX mmx
 | 
						|
PRED8x8_H
 | 
						|
INIT_MMX mmxext
 | 
						|
PRED8x8_H
 | 
						|
INIT_MMX ssse3
 | 
						|
PRED8x8_H
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred8x8_top_dc_8_mmxext(uint8_t *src, int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
INIT_MMX mmxext
 | 
						|
cglobal pred8x8_top_dc_8, 2,5
 | 
						|
    sub         r0, r1
 | 
						|
    movq       mm0, [r0]
 | 
						|
    pxor       mm1, mm1
 | 
						|
    pxor       mm2, mm2
 | 
						|
    lea         r2, [r0+r1*2]
 | 
						|
    punpckhbw  mm1, mm0
 | 
						|
    punpcklbw  mm0, mm2
 | 
						|
    psadbw     mm1, mm2        ; s1
 | 
						|
    lea         r3, [r2+r1*2]
 | 
						|
    psadbw     mm0, mm2        ; s0
 | 
						|
    psrlw      mm1, 1
 | 
						|
    psrlw      mm0, 1
 | 
						|
    pavgw      mm1, mm2
 | 
						|
    lea         r4, [r3+r1*2]
 | 
						|
    pavgw      mm0, mm2
 | 
						|
    pshufw     mm1, mm1, 0
 | 
						|
    pshufw     mm0, mm0, 0     ; dc0 (w)
 | 
						|
    packuswb   mm0, mm1        ; dc0,dc1 (b)
 | 
						|
    movq [r0+r1*1], mm0
 | 
						|
    movq [r0+r1*2], mm0
 | 
						|
    lea         r0, [r3+r1*2]
 | 
						|
    movq [r2+r1*1], mm0
 | 
						|
    movq [r2+r1*2], mm0
 | 
						|
    movq [r3+r1*1], mm0
 | 
						|
    movq [r3+r1*2], mm0
 | 
						|
    movq [r0+r1*1], mm0
 | 
						|
    movq [r0+r1*2], mm0
 | 
						|
    RET
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred8x8_dc_8_mmxext(uint8_t *src, int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
INIT_MMX mmxext
 | 
						|
cglobal pred8x8_dc_8, 2,5
 | 
						|
    sub       r0, r1
 | 
						|
    pxor      m7, m7
 | 
						|
    movd      m0, [r0+0]
 | 
						|
    movd      m1, [r0+4]
 | 
						|
    psadbw    m0, m7            ; s0
 | 
						|
    mov       r4, r0
 | 
						|
    psadbw    m1, m7            ; s1
 | 
						|
 | 
						|
    movzx    r2d, byte [r0+r1*1-1]
 | 
						|
    movzx    r3d, byte [r0+r1*2-1]
 | 
						|
    lea       r0, [r0+r1*2]
 | 
						|
    add      r2d, r3d
 | 
						|
    movzx    r3d, byte [r0+r1*1-1]
 | 
						|
    add      r2d, r3d
 | 
						|
    movzx    r3d, byte [r0+r1*2-1]
 | 
						|
    add      r2d, r3d
 | 
						|
    lea       r0, [r0+r1*2]
 | 
						|
    movd      m2, r2d            ; s2
 | 
						|
    movzx    r2d, byte [r0+r1*1-1]
 | 
						|
    movzx    r3d, byte [r0+r1*2-1]
 | 
						|
    lea       r0, [r0+r1*2]
 | 
						|
    add      r2d, r3d
 | 
						|
    movzx    r3d, byte [r0+r1*1-1]
 | 
						|
    add      r2d, r3d
 | 
						|
    movzx    r3d, byte [r0+r1*2-1]
 | 
						|
    add      r2d, r3d
 | 
						|
    movd      m3, r2d            ; s3
 | 
						|
 | 
						|
    punpcklwd m0, m1
 | 
						|
    mov       r0, r4
 | 
						|
    punpcklwd m2, m3
 | 
						|
    punpckldq m0, m2            ; s0, s1, s2, s3
 | 
						|
    pshufw    m3, m0, 11110110b ; s2, s1, s3, s3
 | 
						|
    lea       r2, [r0+r1*2]
 | 
						|
    pshufw    m0, m0, 01110100b ; s0, s1, s3, s1
 | 
						|
    paddw     m0, m3
 | 
						|
    lea       r3, [r2+r1*2]
 | 
						|
    psrlw     m0, 2
 | 
						|
    pavgw     m0, m7            ; s0+s2, s1, s3, s1+s3
 | 
						|
    lea       r4, [r3+r1*2]
 | 
						|
    packuswb  m0, m0
 | 
						|
    punpcklbw m0, m0
 | 
						|
    movq      m1, m0
 | 
						|
    punpcklbw m0, m0
 | 
						|
    punpckhbw m1, m1
 | 
						|
    movq [r0+r1*1], m0
 | 
						|
    movq [r0+r1*2], m0
 | 
						|
    movq [r2+r1*1], m0
 | 
						|
    movq [r2+r1*2], m0
 | 
						|
    movq [r3+r1*1], m1
 | 
						|
    movq [r3+r1*2], m1
 | 
						|
    movq [r4+r1*1], m1
 | 
						|
    movq [r4+r1*2], m1
 | 
						|
    RET
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred8x8_dc_rv40_8(uint8_t *src, int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
INIT_MMX mmxext
 | 
						|
cglobal pred8x8_dc_rv40_8, 2,7
 | 
						|
    mov       r4, r0
 | 
						|
    sub       r0, r1
 | 
						|
    pxor      mm0, mm0
 | 
						|
    psadbw    mm0, [r0]
 | 
						|
    dec        r0
 | 
						|
    movzx     r5d, byte [r0+r1*1]
 | 
						|
    movd      r6d, mm0
 | 
						|
    lea        r0, [r0+r1*2]
 | 
						|
%rep 3
 | 
						|
    movzx     r2d, byte [r0+r1*0]
 | 
						|
    movzx     r3d, byte [r0+r1*1]
 | 
						|
    add       r5d, r2d
 | 
						|
    add       r6d, r3d
 | 
						|
    lea        r0, [r0+r1*2]
 | 
						|
%endrep
 | 
						|
    movzx     r2d, byte [r0+r1*0]
 | 
						|
    add       r5d, r6d
 | 
						|
    lea       r2d, [r2+r5+8]
 | 
						|
    shr       r2d, 4
 | 
						|
    movd      mm0, r2d
 | 
						|
    punpcklbw mm0, mm0
 | 
						|
    pshufw    mm0, mm0, 0
 | 
						|
    mov       r3d, 4
 | 
						|
.loop:
 | 
						|
    movq [r4+r1*0], mm0
 | 
						|
    movq [r4+r1*1], mm0
 | 
						|
    lea   r4, [r4+r1*2]
 | 
						|
    dec   r3d
 | 
						|
    jg .loop
 | 
						|
    REP_RET
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred8x8_tm_vp8_8(uint8_t *src, int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
%macro PRED8x8_TM 0
 | 
						|
cglobal pred8x8_tm_vp8_8, 2,6
 | 
						|
    sub        r0, r1
 | 
						|
    pxor      mm7, mm7
 | 
						|
    movq      mm0, [r0]
 | 
						|
    movq      mm1, mm0
 | 
						|
    punpcklbw mm0, mm7
 | 
						|
    punpckhbw mm1, mm7
 | 
						|
    movzx     r4d, byte [r0-1]
 | 
						|
    mov       r5d, 4
 | 
						|
.loop:
 | 
						|
    movzx     r2d, byte [r0+r1*1-1]
 | 
						|
    movzx     r3d, byte [r0+r1*2-1]
 | 
						|
    sub       r2d, r4d
 | 
						|
    sub       r3d, r4d
 | 
						|
    movd      mm2, r2d
 | 
						|
    movd      mm4, r3d
 | 
						|
    SPLATW    mm2, mm2, 0
 | 
						|
    SPLATW    mm4, mm4, 0
 | 
						|
    movq      mm3, mm2
 | 
						|
    movq      mm5, mm4
 | 
						|
    paddw     mm2, mm0
 | 
						|
    paddw     mm3, mm1
 | 
						|
    paddw     mm4, mm0
 | 
						|
    paddw     mm5, mm1
 | 
						|
    packuswb  mm2, mm3
 | 
						|
    packuswb  mm4, mm5
 | 
						|
    movq [r0+r1*1], mm2
 | 
						|
    movq [r0+r1*2], mm4
 | 
						|
    lea        r0, [r0+r1*2]
 | 
						|
    dec       r5d
 | 
						|
    jg .loop
 | 
						|
    REP_RET
 | 
						|
%endmacro
 | 
						|
 | 
						|
INIT_MMX mmx
 | 
						|
PRED8x8_TM
 | 
						|
INIT_MMX mmxext
 | 
						|
PRED8x8_TM
 | 
						|
 | 
						|
INIT_XMM sse2
 | 
						|
cglobal pred8x8_tm_vp8_8, 2,6,4
 | 
						|
    sub          r0, r1
 | 
						|
    pxor       xmm1, xmm1
 | 
						|
    movq       xmm0, [r0]
 | 
						|
    punpcklbw  xmm0, xmm1
 | 
						|
    movzx       r4d, byte [r0-1]
 | 
						|
    mov         r5d, 4
 | 
						|
.loop:
 | 
						|
    movzx       r2d, byte [r0+r1*1-1]
 | 
						|
    movzx       r3d, byte [r0+r1*2-1]
 | 
						|
    sub         r2d, r4d
 | 
						|
    sub         r3d, r4d
 | 
						|
    movd       xmm2, r2d
 | 
						|
    movd       xmm3, r3d
 | 
						|
    pshuflw    xmm2, xmm2, 0
 | 
						|
    pshuflw    xmm3, xmm3, 0
 | 
						|
    punpcklqdq xmm2, xmm2
 | 
						|
    punpcklqdq xmm3, xmm3
 | 
						|
    paddw      xmm2, xmm0
 | 
						|
    paddw      xmm3, xmm0
 | 
						|
    packuswb   xmm2, xmm3
 | 
						|
    movq   [r0+r1*1], xmm2
 | 
						|
    movhps [r0+r1*2], xmm2
 | 
						|
    lea          r0, [r0+r1*2]
 | 
						|
    dec         r5d
 | 
						|
    jg .loop
 | 
						|
    REP_RET
 | 
						|
 | 
						|
INIT_XMM ssse3
 | 
						|
cglobal pred8x8_tm_vp8_8, 2,3,6
 | 
						|
    sub          r0, r1
 | 
						|
    movdqa     xmm4, [tm_shuf]
 | 
						|
    pxor       xmm1, xmm1
 | 
						|
    movq       xmm0, [r0]
 | 
						|
    punpcklbw  xmm0, xmm1
 | 
						|
    movd       xmm5, [r0-4]
 | 
						|
    pshufb     xmm5, xmm4
 | 
						|
    mov         r2d, 4
 | 
						|
.loop:
 | 
						|
    movd       xmm2, [r0+r1*1-4]
 | 
						|
    movd       xmm3, [r0+r1*2-4]
 | 
						|
    pshufb     xmm2, xmm4
 | 
						|
    pshufb     xmm3, xmm4
 | 
						|
    psubw      xmm2, xmm5
 | 
						|
    psubw      xmm3, xmm5
 | 
						|
    paddw      xmm2, xmm0
 | 
						|
    paddw      xmm3, xmm0
 | 
						|
    packuswb   xmm2, xmm3
 | 
						|
    movq   [r0+r1*1], xmm2
 | 
						|
    movhps [r0+r1*2], xmm2
 | 
						|
    lea          r0, [r0+r1*2]
 | 
						|
    dec         r2d
 | 
						|
    jg .loop
 | 
						|
    REP_RET
 | 
						|
 | 
						|
; dest, left, right, src, tmp
 | 
						|
; output: %1 = (t[n-1] + t[n]*2 + t[n+1] + 2) >> 2
 | 
						|
%macro PRED4x4_LOWPASS 5
 | 
						|
    mova    %5, %2
 | 
						|
    pavgb   %2, %3
 | 
						|
    pxor    %3, %5
 | 
						|
    mova    %1, %4
 | 
						|
    pand    %3, [pb_1]
 | 
						|
    psubusb %2, %3
 | 
						|
    pavgb   %1, %2
 | 
						|
%endmacro
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred8x8l_top_dc_8(uint8_t *src, int has_topleft, int has_topright,
 | 
						|
;                           int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
%macro PRED8x8L_TOP_DC 0
 | 
						|
cglobal pred8x8l_top_dc_8, 4,4
 | 
						|
    sub          r0, r3
 | 
						|
    pxor        mm7, mm7
 | 
						|
    movq        mm0, [r0-8]
 | 
						|
    movq        mm3, [r0]
 | 
						|
    movq        mm1, [r0+8]
 | 
						|
    movq        mm2, mm3
 | 
						|
    movq        mm4, mm3
 | 
						|
    PALIGNR     mm2, mm0, 7, mm0
 | 
						|
    PALIGNR     mm1, mm4, 1, mm4
 | 
						|
    test        r1d, r1d ; top_left
 | 
						|
    jz .fix_lt_2
 | 
						|
    test        r2d, r2d ; top_right
 | 
						|
    jz .fix_tr_1
 | 
						|
    jmp .body
 | 
						|
.fix_lt_2:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm2
 | 
						|
    psllq       mm5, 56
 | 
						|
    psrlq       mm5, 56
 | 
						|
    pxor        mm2, mm5
 | 
						|
    test        r2d, r2d ; top_right
 | 
						|
    jnz .body
 | 
						|
.fix_tr_1:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm1
 | 
						|
    psrlq       mm5, 56
 | 
						|
    psllq       mm5, 56
 | 
						|
    pxor        mm1, mm5
 | 
						|
.body:
 | 
						|
    PRED4x4_LOWPASS mm0, mm2, mm1, mm3, mm5
 | 
						|
    psadbw   mm7, mm0
 | 
						|
    paddw    mm7, [pw_4]
 | 
						|
    psrlw    mm7, 3
 | 
						|
    pshufw   mm7, mm7, 0
 | 
						|
    packuswb mm7, mm7
 | 
						|
%rep 3
 | 
						|
    movq [r0+r3*1], mm7
 | 
						|
    movq [r0+r3*2], mm7
 | 
						|
    lea    r0, [r0+r3*2]
 | 
						|
%endrep
 | 
						|
    movq [r0+r3*1], mm7
 | 
						|
    movq [r0+r3*2], mm7
 | 
						|
    RET
 | 
						|
%endmacro
 | 
						|
 | 
						|
INIT_MMX mmxext
 | 
						|
PRED8x8L_TOP_DC
 | 
						|
INIT_MMX ssse3
 | 
						|
PRED8x8L_TOP_DC
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred8x8l_dc_8(uint8_t *src, int has_topleft, int has_topright,
 | 
						|
;                       int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
%macro PRED8x8L_DC 0
 | 
						|
cglobal pred8x8l_dc_8, 4,5
 | 
						|
    sub          r0, r3
 | 
						|
    lea          r4, [r0+r3*2]
 | 
						|
    movq        mm0, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm0, [r0+r3*0-8]
 | 
						|
    movq        mm1, [r4+r3*1-8]
 | 
						|
    punpckhbw   mm1, [r0+r3*2-8]
 | 
						|
    mov          r4, r0
 | 
						|
    punpckhwd   mm1, mm0
 | 
						|
    lea          r0, [r0+r3*4]
 | 
						|
    movq        mm2, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm2, [r0+r3*0-8]
 | 
						|
    lea          r0, [r0+r3*2]
 | 
						|
    movq        mm3, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm3, [r0+r3*0-8]
 | 
						|
    punpckhwd   mm3, mm2
 | 
						|
    punpckhdq   mm3, mm1
 | 
						|
    lea          r0, [r0+r3*2]
 | 
						|
    movq        mm0, [r0+r3*0-8]
 | 
						|
    movq        mm1, [r4]
 | 
						|
    mov          r0, r4
 | 
						|
    movq        mm4, mm3
 | 
						|
    movq        mm2, mm3
 | 
						|
    PALIGNR     mm4, mm0, 7, mm0
 | 
						|
    PALIGNR     mm1, mm2, 1, mm2
 | 
						|
    test        r1d, r1d
 | 
						|
    jnz .do_left
 | 
						|
.fix_lt_1:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm4
 | 
						|
    psrlq       mm5, 56
 | 
						|
    psllq       mm5, 48
 | 
						|
    pxor        mm1, mm5
 | 
						|
    jmp .do_left
 | 
						|
.fix_lt_2:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm2
 | 
						|
    psllq       mm5, 56
 | 
						|
    psrlq       mm5, 56
 | 
						|
    pxor        mm2, mm5
 | 
						|
    test        r2d, r2d
 | 
						|
    jnz .body
 | 
						|
.fix_tr_1:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm1
 | 
						|
    psrlq       mm5, 56
 | 
						|
    psllq       mm5, 56
 | 
						|
    pxor        mm1, mm5
 | 
						|
    jmp .body
 | 
						|
.do_left:
 | 
						|
    movq        mm0, mm4
 | 
						|
    PRED4x4_LOWPASS mm2, mm1, mm4, mm3, mm5
 | 
						|
    movq        mm4, mm0
 | 
						|
    movq        mm7, mm2
 | 
						|
    PRED4x4_LOWPASS mm1, mm3, mm0, mm4, mm5
 | 
						|
    psllq       mm1, 56
 | 
						|
    PALIGNR     mm7, mm1, 7, mm3
 | 
						|
    movq        mm0, [r0-8]
 | 
						|
    movq        mm3, [r0]
 | 
						|
    movq        mm1, [r0+8]
 | 
						|
    movq        mm2, mm3
 | 
						|
    movq        mm4, mm3
 | 
						|
    PALIGNR     mm2, mm0, 7, mm0
 | 
						|
    PALIGNR     mm1, mm4, 1, mm4
 | 
						|
    test        r1d, r1d
 | 
						|
    jz .fix_lt_2
 | 
						|
    test        r2d, r2d
 | 
						|
    jz .fix_tr_1
 | 
						|
.body:
 | 
						|
    lea          r1, [r0+r3*2]
 | 
						|
    PRED4x4_LOWPASS mm6, mm2, mm1, mm3, mm5
 | 
						|
    pxor        mm0, mm0
 | 
						|
    pxor        mm1, mm1
 | 
						|
    lea          r2, [r1+r3*2]
 | 
						|
    psadbw      mm0, mm7
 | 
						|
    psadbw      mm1, mm6
 | 
						|
    paddw       mm0, [pw_8]
 | 
						|
    paddw       mm0, mm1
 | 
						|
    lea          r4, [r2+r3*2]
 | 
						|
    psrlw       mm0, 4
 | 
						|
    pshufw      mm0, mm0, 0
 | 
						|
    packuswb    mm0, mm0
 | 
						|
    movq [r0+r3*1], mm0
 | 
						|
    movq [r0+r3*2], mm0
 | 
						|
    movq [r1+r3*1], mm0
 | 
						|
    movq [r1+r3*2], mm0
 | 
						|
    movq [r2+r3*1], mm0
 | 
						|
    movq [r2+r3*2], mm0
 | 
						|
    movq [r4+r3*1], mm0
 | 
						|
    movq [r4+r3*2], mm0
 | 
						|
    RET
 | 
						|
%endmacro
 | 
						|
 | 
						|
INIT_MMX mmxext
 | 
						|
PRED8x8L_DC
 | 
						|
INIT_MMX ssse3
 | 
						|
PRED8x8L_DC
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred8x8l_horizontal_8(uint8_t *src, int has_topleft,
 | 
						|
;                               int has_topright, int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
%macro PRED8x8L_HORIZONTAL 0
 | 
						|
cglobal pred8x8l_horizontal_8, 4,4
 | 
						|
    sub          r0, r3
 | 
						|
    lea          r2, [r0+r3*2]
 | 
						|
    movq        mm0, [r0+r3*1-8]
 | 
						|
    test        r1d, r1d
 | 
						|
    lea          r1, [r0+r3]
 | 
						|
    cmovnz       r1, r0
 | 
						|
    punpckhbw   mm0, [r1+r3*0-8]
 | 
						|
    movq        mm1, [r2+r3*1-8]
 | 
						|
    punpckhbw   mm1, [r0+r3*2-8]
 | 
						|
    mov          r2, r0
 | 
						|
    punpckhwd   mm1, mm0
 | 
						|
    lea          r0, [r0+r3*4]
 | 
						|
    movq        mm2, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm2, [r0+r3*0-8]
 | 
						|
    lea          r0, [r0+r3*2]
 | 
						|
    movq        mm3, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm3, [r0+r3*0-8]
 | 
						|
    punpckhwd   mm3, mm2
 | 
						|
    punpckhdq   mm3, mm1
 | 
						|
    lea          r0, [r0+r3*2]
 | 
						|
    movq        mm0, [r0+r3*0-8]
 | 
						|
    movq        mm1, [r1+r3*0-8]
 | 
						|
    mov          r0, r2
 | 
						|
    movq        mm4, mm3
 | 
						|
    movq        mm2, mm3
 | 
						|
    PALIGNR     mm4, mm0, 7, mm0
 | 
						|
    PALIGNR     mm1, mm2, 1, mm2
 | 
						|
    movq        mm0, mm4
 | 
						|
    PRED4x4_LOWPASS mm2, mm1, mm4, mm3, mm5
 | 
						|
    movq        mm4, mm0
 | 
						|
    movq        mm7, mm2
 | 
						|
    PRED4x4_LOWPASS mm1, mm3, mm0, mm4, mm5
 | 
						|
    psllq       mm1, 56
 | 
						|
    PALIGNR     mm7, mm1, 7, mm3
 | 
						|
    movq        mm3, mm7
 | 
						|
    lea         r1, [r0+r3*2]
 | 
						|
    movq       mm7, mm3
 | 
						|
    punpckhbw  mm3, mm3
 | 
						|
    punpcklbw  mm7, mm7
 | 
						|
    pshufw     mm0, mm3, 0xff
 | 
						|
    pshufw     mm1, mm3, 0xaa
 | 
						|
    lea         r2, [r1+r3*2]
 | 
						|
    pshufw     mm2, mm3, 0x55
 | 
						|
    pshufw     mm3, mm3, 0x00
 | 
						|
    pshufw     mm4, mm7, 0xff
 | 
						|
    pshufw     mm5, mm7, 0xaa
 | 
						|
    pshufw     mm6, mm7, 0x55
 | 
						|
    pshufw     mm7, mm7, 0x00
 | 
						|
    movq [r0+r3*1], mm0
 | 
						|
    movq [r0+r3*2], mm1
 | 
						|
    movq [r1+r3*1], mm2
 | 
						|
    movq [r1+r3*2], mm3
 | 
						|
    movq [r2+r3*1], mm4
 | 
						|
    movq [r2+r3*2], mm5
 | 
						|
    lea         r0, [r2+r3*2]
 | 
						|
    movq [r0+r3*1], mm6
 | 
						|
    movq [r0+r3*2], mm7
 | 
						|
    RET
 | 
						|
%endmacro
 | 
						|
 | 
						|
INIT_MMX mmxext
 | 
						|
PRED8x8L_HORIZONTAL
 | 
						|
INIT_MMX ssse3
 | 
						|
PRED8x8L_HORIZONTAL
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred8x8l_vertical_8(uint8_t *src, int has_topleft, int has_topright,
 | 
						|
;                             int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
%macro PRED8x8L_VERTICAL 0
 | 
						|
cglobal pred8x8l_vertical_8, 4,4
 | 
						|
    sub          r0, r3
 | 
						|
    movq        mm0, [r0-8]
 | 
						|
    movq        mm3, [r0]
 | 
						|
    movq        mm1, [r0+8]
 | 
						|
    movq        mm2, mm3
 | 
						|
    movq        mm4, mm3
 | 
						|
    PALIGNR     mm2, mm0, 7, mm0
 | 
						|
    PALIGNR     mm1, mm4, 1, mm4
 | 
						|
    test        r1d, r1d ; top_left
 | 
						|
    jz .fix_lt_2
 | 
						|
    test        r2d, r2d ; top_right
 | 
						|
    jz .fix_tr_1
 | 
						|
    jmp .body
 | 
						|
.fix_lt_2:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm2
 | 
						|
    psllq       mm5, 56
 | 
						|
    psrlq       mm5, 56
 | 
						|
    pxor        mm2, mm5
 | 
						|
    test        r2d, r2d ; top_right
 | 
						|
    jnz .body
 | 
						|
.fix_tr_1:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm1
 | 
						|
    psrlq       mm5, 56
 | 
						|
    psllq       mm5, 56
 | 
						|
    pxor        mm1, mm5
 | 
						|
.body:
 | 
						|
    PRED4x4_LOWPASS mm0, mm2, mm1, mm3, mm5
 | 
						|
%rep 3
 | 
						|
    movq [r0+r3*1], mm0
 | 
						|
    movq [r0+r3*2], mm0
 | 
						|
    lea    r0, [r0+r3*2]
 | 
						|
%endrep
 | 
						|
    movq [r0+r3*1], mm0
 | 
						|
    movq [r0+r3*2], mm0
 | 
						|
    RET
 | 
						|
%endmacro
 | 
						|
 | 
						|
INIT_MMX mmxext
 | 
						|
PRED8x8L_VERTICAL
 | 
						|
INIT_MMX ssse3
 | 
						|
PRED8x8L_VERTICAL
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred8x8l_down_left_8(uint8_t *src, int has_topleft,
 | 
						|
;                              int has_topright, int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
INIT_MMX mmxext
 | 
						|
cglobal pred8x8l_down_left_8, 4,5
 | 
						|
    sub          r0, r3
 | 
						|
    movq        mm0, [r0-8]
 | 
						|
    movq        mm3, [r0]
 | 
						|
    movq        mm1, [r0+8]
 | 
						|
    movq        mm2, mm3
 | 
						|
    movq        mm4, mm3
 | 
						|
    PALIGNR     mm2, mm0, 7, mm0
 | 
						|
    PALIGNR     mm1, mm4, 1, mm4
 | 
						|
    test        r1d, r1d
 | 
						|
    jz .fix_lt_2
 | 
						|
    test        r2d, r2d
 | 
						|
    jz .fix_tr_1
 | 
						|
    jmp .do_top
 | 
						|
.fix_lt_2:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm2
 | 
						|
    psllq       mm5, 56
 | 
						|
    psrlq       mm5, 56
 | 
						|
    pxor        mm2, mm5
 | 
						|
    test        r2d, r2d
 | 
						|
    jnz .do_top
 | 
						|
.fix_tr_1:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm1
 | 
						|
    psrlq       mm5, 56
 | 
						|
    psllq       mm5, 56
 | 
						|
    pxor        mm1, mm5
 | 
						|
    jmp .do_top
 | 
						|
.fix_tr_2:
 | 
						|
    punpckhbw   mm3, mm3
 | 
						|
    pshufw      mm1, mm3, 0xFF
 | 
						|
    jmp .do_topright
 | 
						|
.do_top:
 | 
						|
    PRED4x4_LOWPASS mm4, mm2, mm1, mm3, mm5
 | 
						|
    movq        mm7, mm4
 | 
						|
    test        r2d, r2d
 | 
						|
    jz .fix_tr_2
 | 
						|
    movq        mm0, [r0+8]
 | 
						|
    movq        mm5, mm0
 | 
						|
    movq        mm2, mm0
 | 
						|
    movq        mm4, mm0
 | 
						|
    psrlq       mm5, 56
 | 
						|
    PALIGNR     mm2, mm3, 7, mm3
 | 
						|
    PALIGNR     mm5, mm4, 1, mm4
 | 
						|
    PRED4x4_LOWPASS mm1, mm2, mm5, mm0, mm4
 | 
						|
.do_topright:
 | 
						|
    lea          r1, [r0+r3*2]
 | 
						|
    movq        mm6, mm1
 | 
						|
    psrlq       mm1, 56
 | 
						|
    movq        mm4, mm1
 | 
						|
    lea          r2, [r1+r3*2]
 | 
						|
    movq        mm2, mm6
 | 
						|
    PALIGNR     mm2, mm7, 1, mm0
 | 
						|
    movq        mm3, mm6
 | 
						|
    PALIGNR     mm3, mm7, 7, mm0
 | 
						|
    PALIGNR     mm4, mm6, 1, mm0
 | 
						|
    movq        mm5, mm7
 | 
						|
    movq        mm1, mm7
 | 
						|
    movq        mm7, mm6
 | 
						|
    lea          r4, [r2+r3*2]
 | 
						|
    psllq       mm1, 8
 | 
						|
    PRED4x4_LOWPASS mm0, mm1, mm2, mm5, mm6
 | 
						|
    PRED4x4_LOWPASS mm1, mm3, mm4, mm7, mm6
 | 
						|
    movq  [r4+r3*2], mm1
 | 
						|
    movq        mm2, mm0
 | 
						|
    psllq       mm1, 8
 | 
						|
    psrlq       mm2, 56
 | 
						|
    psllq       mm0, 8
 | 
						|
    por         mm1, mm2
 | 
						|
    movq  [r4+r3*1], mm1
 | 
						|
    movq        mm2, mm0
 | 
						|
    psllq       mm1, 8
 | 
						|
    psrlq       mm2, 56
 | 
						|
    psllq       mm0, 8
 | 
						|
    por         mm1, mm2
 | 
						|
    movq  [r2+r3*2], mm1
 | 
						|
    movq        mm2, mm0
 | 
						|
    psllq       mm1, 8
 | 
						|
    psrlq       mm2, 56
 | 
						|
    psllq       mm0, 8
 | 
						|
    por         mm1, mm2
 | 
						|
    movq  [r2+r3*1], mm1
 | 
						|
    movq        mm2, mm0
 | 
						|
    psllq       mm1, 8
 | 
						|
    psrlq       mm2, 56
 | 
						|
    psllq       mm0, 8
 | 
						|
    por         mm1, mm2
 | 
						|
    movq  [r1+r3*2], mm1
 | 
						|
    movq        mm2, mm0
 | 
						|
    psllq       mm1, 8
 | 
						|
    psrlq       mm2, 56
 | 
						|
    psllq       mm0, 8
 | 
						|
    por         mm1, mm2
 | 
						|
    movq  [r1+r3*1], mm1
 | 
						|
    movq        mm2, mm0
 | 
						|
    psllq       mm1, 8
 | 
						|
    psrlq       mm2, 56
 | 
						|
    psllq       mm0, 8
 | 
						|
    por         mm1, mm2
 | 
						|
    movq  [r0+r3*2], mm1
 | 
						|
    psllq       mm1, 8
 | 
						|
    psrlq       mm0, 56
 | 
						|
    por         mm1, mm0
 | 
						|
    movq  [r0+r3*1], mm1
 | 
						|
    RET
 | 
						|
 | 
						|
%macro PRED8x8L_DOWN_LEFT 0
 | 
						|
cglobal pred8x8l_down_left_8, 4,4
 | 
						|
    sub          r0, r3
 | 
						|
    movq        mm0, [r0-8]
 | 
						|
    movq        mm3, [r0]
 | 
						|
    movq        mm1, [r0+8]
 | 
						|
    movq        mm2, mm3
 | 
						|
    movq        mm4, mm3
 | 
						|
    PALIGNR     mm2, mm0, 7, mm0
 | 
						|
    PALIGNR     mm1, mm4, 1, mm4
 | 
						|
    test        r1d, r1d ; top_left
 | 
						|
    jz .fix_lt_2
 | 
						|
    test        r2d, r2d ; top_right
 | 
						|
    jz .fix_tr_1
 | 
						|
    jmp .do_top
 | 
						|
.fix_lt_2:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm2
 | 
						|
    psllq       mm5, 56
 | 
						|
    psrlq       mm5, 56
 | 
						|
    pxor        mm2, mm5
 | 
						|
    test        r2d, r2d ; top_right
 | 
						|
    jnz .do_top
 | 
						|
.fix_tr_1:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm1
 | 
						|
    psrlq       mm5, 56
 | 
						|
    psllq       mm5, 56
 | 
						|
    pxor        mm1, mm5
 | 
						|
    jmp .do_top
 | 
						|
.fix_tr_2:
 | 
						|
    punpckhbw   mm3, mm3
 | 
						|
    pshufw      mm1, mm3, 0xFF
 | 
						|
    jmp .do_topright
 | 
						|
.do_top:
 | 
						|
    PRED4x4_LOWPASS mm4, mm2, mm1, mm3, mm5
 | 
						|
    movq2dq    xmm3, mm4
 | 
						|
    test        r2d, r2d ; top_right
 | 
						|
    jz .fix_tr_2
 | 
						|
    movq        mm0, [r0+8]
 | 
						|
    movq        mm5, mm0
 | 
						|
    movq        mm2, mm0
 | 
						|
    movq        mm4, mm0
 | 
						|
    psrlq       mm5, 56
 | 
						|
    PALIGNR     mm2, mm3, 7, mm3
 | 
						|
    PALIGNR     mm5, mm4, 1, mm4
 | 
						|
    PRED4x4_LOWPASS mm1, mm2, mm5, mm0, mm4
 | 
						|
.do_topright:
 | 
						|
    movq2dq    xmm4, mm1
 | 
						|
    psrlq       mm1, 56
 | 
						|
    movq2dq    xmm5, mm1
 | 
						|
    lea         r1, [r0+r3*2]
 | 
						|
    pslldq    xmm4, 8
 | 
						|
    por       xmm3, xmm4
 | 
						|
    movdqa    xmm2, xmm3
 | 
						|
    psrldq    xmm2, 1
 | 
						|
    pslldq    xmm5, 15
 | 
						|
    por       xmm2, xmm5
 | 
						|
    lea         r2, [r1+r3*2]
 | 
						|
    movdqa    xmm1, xmm3
 | 
						|
    pslldq    xmm1, 1
 | 
						|
INIT_XMM cpuname
 | 
						|
    PRED4x4_LOWPASS xmm0, xmm1, xmm2, xmm3, xmm4
 | 
						|
    psrldq    xmm0, 1
 | 
						|
    movq [r0+r3*1], xmm0
 | 
						|
    psrldq    xmm0, 1
 | 
						|
    movq [r0+r3*2], xmm0
 | 
						|
    psrldq    xmm0, 1
 | 
						|
    lea         r0, [r2+r3*2]
 | 
						|
    movq [r1+r3*1], xmm0
 | 
						|
    psrldq    xmm0, 1
 | 
						|
    movq [r1+r3*2], xmm0
 | 
						|
    psrldq    xmm0, 1
 | 
						|
    movq [r2+r3*1], xmm0
 | 
						|
    psrldq    xmm0, 1
 | 
						|
    movq [r2+r3*2], xmm0
 | 
						|
    psrldq    xmm0, 1
 | 
						|
    movq [r0+r3*1], xmm0
 | 
						|
    psrldq    xmm0, 1
 | 
						|
    movq [r0+r3*2], xmm0
 | 
						|
    RET
 | 
						|
%endmacro
 | 
						|
 | 
						|
INIT_MMX sse2
 | 
						|
PRED8x8L_DOWN_LEFT
 | 
						|
INIT_MMX ssse3
 | 
						|
PRED8x8L_DOWN_LEFT
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred8x8l_down_right_8_mmxext(uint8_t *src, int has_topleft,
 | 
						|
;                                      int has_topright, int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
INIT_MMX mmxext
 | 
						|
cglobal pred8x8l_down_right_8, 4,5
 | 
						|
    sub          r0, r3
 | 
						|
    lea          r4, [r0+r3*2]
 | 
						|
    movq        mm0, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm0, [r0+r3*0-8]
 | 
						|
    movq        mm1, [r4+r3*1-8]
 | 
						|
    punpckhbw   mm1, [r0+r3*2-8]
 | 
						|
    mov          r4, r0
 | 
						|
    punpckhwd   mm1, mm0
 | 
						|
    lea          r0, [r0+r3*4]
 | 
						|
    movq        mm2, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm2, [r0+r3*0-8]
 | 
						|
    lea          r0, [r0+r3*2]
 | 
						|
    movq        mm3, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm3, [r0+r3*0-8]
 | 
						|
    punpckhwd   mm3, mm2
 | 
						|
    punpckhdq   mm3, mm1
 | 
						|
    lea          r0, [r0+r3*2]
 | 
						|
    movq        mm0, [r0+r3*0-8]
 | 
						|
    movq        mm1, [r4]
 | 
						|
    mov          r0, r4
 | 
						|
    movq        mm4, mm3
 | 
						|
    movq        mm2, mm3
 | 
						|
    PALIGNR     mm4, mm0, 7, mm0
 | 
						|
    PALIGNR     mm1, mm2, 1, mm2
 | 
						|
    test        r1d, r1d ; top_left
 | 
						|
    jz .fix_lt_1
 | 
						|
.do_left:
 | 
						|
    movq        mm0, mm4
 | 
						|
    PRED4x4_LOWPASS mm2, mm1, mm4, mm3, mm5
 | 
						|
    movq        mm4, mm0
 | 
						|
    movq        mm7, mm2
 | 
						|
    movq        mm6, mm2
 | 
						|
    PRED4x4_LOWPASS mm1, mm3, mm0, mm4, mm5
 | 
						|
    psllq       mm1, 56
 | 
						|
    PALIGNR     mm7, mm1, 7, mm3
 | 
						|
    movq        mm0, [r0-8]
 | 
						|
    movq        mm3, [r0]
 | 
						|
    movq        mm1, [r0+8]
 | 
						|
    movq        mm2, mm3
 | 
						|
    movq        mm4, mm3
 | 
						|
    PALIGNR     mm2, mm0, 7, mm0
 | 
						|
    PALIGNR     mm1, mm4, 1, mm4
 | 
						|
    test        r1d, r1d ; top_left
 | 
						|
    jz .fix_lt_2
 | 
						|
    test        r2d, r2d ; top_right
 | 
						|
    jz .fix_tr_1
 | 
						|
.do_top:
 | 
						|
    PRED4x4_LOWPASS mm4, mm2, mm1, mm3, mm5
 | 
						|
    movq        mm5, mm4
 | 
						|
    jmp .body
 | 
						|
.fix_lt_1:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm4
 | 
						|
    psrlq       mm5, 56
 | 
						|
    psllq       mm5, 48
 | 
						|
    pxor        mm1, mm5
 | 
						|
    jmp .do_left
 | 
						|
.fix_lt_2:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm2
 | 
						|
    psllq       mm5, 56
 | 
						|
    psrlq       mm5, 56
 | 
						|
    pxor        mm2, mm5
 | 
						|
    test        r2d, r2d ; top_right
 | 
						|
    jnz .do_top
 | 
						|
.fix_tr_1:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm1
 | 
						|
    psrlq       mm5, 56
 | 
						|
    psllq       mm5, 56
 | 
						|
    pxor        mm1, mm5
 | 
						|
    jmp .do_top
 | 
						|
.body:
 | 
						|
    lea         r1, [r0+r3*2]
 | 
						|
    movq       mm1, mm7
 | 
						|
    movq       mm7, mm5
 | 
						|
    movq       mm5, mm6
 | 
						|
    movq       mm2, mm7
 | 
						|
    lea         r2, [r1+r3*2]
 | 
						|
    PALIGNR    mm2, mm6, 1, mm0
 | 
						|
    movq       mm3, mm7
 | 
						|
    PALIGNR    mm3, mm6, 7, mm0
 | 
						|
    movq       mm4, mm7
 | 
						|
    lea         r4, [r2+r3*2]
 | 
						|
    psrlq      mm4, 8
 | 
						|
    PRED4x4_LOWPASS mm0, mm1, mm2, mm5, mm6
 | 
						|
    PRED4x4_LOWPASS mm1, mm3, mm4, mm7, mm6
 | 
						|
    movq [r4+r3*2], mm0
 | 
						|
    movq       mm2, mm1
 | 
						|
    psrlq      mm0, 8
 | 
						|
    psllq      mm2, 56
 | 
						|
    psrlq      mm1, 8
 | 
						|
    por        mm0, mm2
 | 
						|
    movq [r4+r3*1], mm0
 | 
						|
    movq       mm2, mm1
 | 
						|
    psrlq      mm0, 8
 | 
						|
    psllq      mm2, 56
 | 
						|
    psrlq      mm1, 8
 | 
						|
    por        mm0, mm2
 | 
						|
    movq [r2+r3*2], mm0
 | 
						|
    movq       mm2, mm1
 | 
						|
    psrlq      mm0, 8
 | 
						|
    psllq      mm2, 56
 | 
						|
    psrlq      mm1, 8
 | 
						|
    por        mm0, mm2
 | 
						|
    movq [r2+r3*1], mm0
 | 
						|
    movq       mm2, mm1
 | 
						|
    psrlq      mm0, 8
 | 
						|
    psllq      mm2, 56
 | 
						|
    psrlq      mm1, 8
 | 
						|
    por        mm0, mm2
 | 
						|
    movq [r1+r3*2], mm0
 | 
						|
    movq       mm2, mm1
 | 
						|
    psrlq      mm0, 8
 | 
						|
    psllq      mm2, 56
 | 
						|
    psrlq      mm1, 8
 | 
						|
    por        mm0, mm2
 | 
						|
    movq [r1+r3*1], mm0
 | 
						|
    movq       mm2, mm1
 | 
						|
    psrlq      mm0, 8
 | 
						|
    psllq      mm2, 56
 | 
						|
    psrlq      mm1, 8
 | 
						|
    por        mm0, mm2
 | 
						|
    movq [r0+r3*2], mm0
 | 
						|
    psrlq      mm0, 8
 | 
						|
    psllq      mm1, 56
 | 
						|
    por        mm0, mm1
 | 
						|
    movq [r0+r3*1], mm0
 | 
						|
    RET
 | 
						|
 | 
						|
%macro PRED8x8L_DOWN_RIGHT 0
 | 
						|
cglobal pred8x8l_down_right_8, 4,5
 | 
						|
    sub          r0, r3
 | 
						|
    lea          r4, [r0+r3*2]
 | 
						|
    movq        mm0, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm0, [r0+r3*0-8]
 | 
						|
    movq        mm1, [r4+r3*1-8]
 | 
						|
    punpckhbw   mm1, [r0+r3*2-8]
 | 
						|
    mov          r4, r0
 | 
						|
    punpckhwd   mm1, mm0
 | 
						|
    lea          r0, [r0+r3*4]
 | 
						|
    movq        mm2, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm2, [r0+r3*0-8]
 | 
						|
    lea          r0, [r0+r3*2]
 | 
						|
    movq        mm3, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm3, [r0+r3*0-8]
 | 
						|
    punpckhwd   mm3, mm2
 | 
						|
    punpckhdq   mm3, mm1
 | 
						|
    lea          r0, [r0+r3*2]
 | 
						|
    movq        mm0, [r0+r3*0-8]
 | 
						|
    movq        mm1, [r4]
 | 
						|
    mov          r0, r4
 | 
						|
    movq        mm4, mm3
 | 
						|
    movq        mm2, mm3
 | 
						|
    PALIGNR     mm4, mm0, 7, mm0
 | 
						|
    PALIGNR     mm1, mm2, 1, mm2
 | 
						|
    test        r1d, r1d
 | 
						|
    jz .fix_lt_1
 | 
						|
    jmp .do_left
 | 
						|
.fix_lt_1:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm4
 | 
						|
    psrlq       mm5, 56
 | 
						|
    psllq       mm5, 48
 | 
						|
    pxor        mm1, mm5
 | 
						|
    jmp .do_left
 | 
						|
.fix_lt_2:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm2
 | 
						|
    psllq       mm5, 56
 | 
						|
    psrlq       mm5, 56
 | 
						|
    pxor        mm2, mm5
 | 
						|
    test        r2d, r2d
 | 
						|
    jnz .do_top
 | 
						|
.fix_tr_1:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm1
 | 
						|
    psrlq       mm5, 56
 | 
						|
    psllq       mm5, 56
 | 
						|
    pxor        mm1, mm5
 | 
						|
    jmp .do_top
 | 
						|
.do_left:
 | 
						|
    movq        mm0, mm4
 | 
						|
    PRED4x4_LOWPASS mm2, mm1, mm4, mm3, mm5
 | 
						|
    movq        mm4, mm0
 | 
						|
    movq        mm7, mm2
 | 
						|
    movq2dq    xmm3, mm2
 | 
						|
    PRED4x4_LOWPASS mm1, mm3, mm0, mm4, mm5
 | 
						|
    psllq       mm1, 56
 | 
						|
    PALIGNR     mm7, mm1, 7, mm3
 | 
						|
    movq2dq    xmm1, mm7
 | 
						|
    movq        mm0, [r0-8]
 | 
						|
    movq        mm3, [r0]
 | 
						|
    movq        mm1, [r0+8]
 | 
						|
    movq        mm2, mm3
 | 
						|
    movq        mm4, mm3
 | 
						|
    PALIGNR     mm2, mm0, 7, mm0
 | 
						|
    PALIGNR     mm1, mm4, 1, mm4
 | 
						|
    test        r1d, r1d
 | 
						|
    jz .fix_lt_2
 | 
						|
    test        r2d, r2d
 | 
						|
    jz .fix_tr_1
 | 
						|
.do_top:
 | 
						|
    PRED4x4_LOWPASS mm4, mm2, mm1, mm3, mm5
 | 
						|
    movq2dq   xmm4, mm4
 | 
						|
    lea         r1, [r0+r3*2]
 | 
						|
    movdqa    xmm0, xmm3
 | 
						|
    pslldq    xmm4, 8
 | 
						|
    por       xmm3, xmm4
 | 
						|
    lea         r2, [r1+r3*2]
 | 
						|
    pslldq    xmm4, 1
 | 
						|
    por       xmm1, xmm4
 | 
						|
    psrldq    xmm0, 7
 | 
						|
    pslldq    xmm0, 15
 | 
						|
    psrldq    xmm0, 7
 | 
						|
    por       xmm1, xmm0
 | 
						|
    lea         r0, [r2+r3*2]
 | 
						|
    movdqa    xmm2, xmm3
 | 
						|
    psrldq    xmm2, 1
 | 
						|
INIT_XMM cpuname
 | 
						|
    PRED4x4_LOWPASS xmm0, xmm1, xmm2, xmm3, xmm4
 | 
						|
    movdqa    xmm1, xmm0
 | 
						|
    psrldq    xmm1, 1
 | 
						|
    movq [r0+r3*2], xmm0
 | 
						|
    movq [r0+r3*1], xmm1
 | 
						|
    psrldq    xmm0, 2
 | 
						|
    psrldq    xmm1, 2
 | 
						|
    movq [r2+r3*2], xmm0
 | 
						|
    movq [r2+r3*1], xmm1
 | 
						|
    psrldq    xmm0, 2
 | 
						|
    psrldq    xmm1, 2
 | 
						|
    movq [r1+r3*2], xmm0
 | 
						|
    movq [r1+r3*1], xmm1
 | 
						|
    psrldq    xmm0, 2
 | 
						|
    psrldq    xmm1, 2
 | 
						|
    movq [r4+r3*2], xmm0
 | 
						|
    movq [r4+r3*1], xmm1
 | 
						|
    RET
 | 
						|
%endmacro
 | 
						|
 | 
						|
INIT_MMX sse2
 | 
						|
PRED8x8L_DOWN_RIGHT
 | 
						|
INIT_MMX ssse3
 | 
						|
PRED8x8L_DOWN_RIGHT
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred8x8l_vertical_right_8(uint8_t *src, int has_topleft,
 | 
						|
;                                   int has_topright, int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
INIT_MMX mmxext
 | 
						|
cglobal pred8x8l_vertical_right_8, 4,5
 | 
						|
    sub          r0, r3
 | 
						|
    lea          r4, [r0+r3*2]
 | 
						|
    movq        mm0, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm0, [r0+r3*0-8]
 | 
						|
    movq        mm1, [r4+r3*1-8]
 | 
						|
    punpckhbw   mm1, [r0+r3*2-8]
 | 
						|
    mov          r4, r0
 | 
						|
    punpckhwd   mm1, mm0
 | 
						|
    lea          r0, [r0+r3*4]
 | 
						|
    movq        mm2, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm2, [r0+r3*0-8]
 | 
						|
    lea          r0, [r0+r3*2]
 | 
						|
    movq        mm3, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm3, [r0+r3*0-8]
 | 
						|
    punpckhwd   mm3, mm2
 | 
						|
    punpckhdq   mm3, mm1
 | 
						|
    lea          r0, [r0+r3*2]
 | 
						|
    movq        mm0, [r0+r3*0-8]
 | 
						|
    movq        mm1, [r4]
 | 
						|
    mov          r0, r4
 | 
						|
    movq        mm4, mm3
 | 
						|
    movq        mm2, mm3
 | 
						|
    PALIGNR     mm4, mm0, 7, mm0
 | 
						|
    PALIGNR     mm1, mm2, 1, mm2
 | 
						|
    test        r1d, r1d
 | 
						|
    jz .fix_lt_1
 | 
						|
    jmp .do_left
 | 
						|
.fix_lt_1:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm4
 | 
						|
    psrlq       mm5, 56
 | 
						|
    psllq       mm5, 48
 | 
						|
    pxor        mm1, mm5
 | 
						|
    jmp .do_left
 | 
						|
.fix_lt_2:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm2
 | 
						|
    psllq       mm5, 56
 | 
						|
    psrlq       mm5, 56
 | 
						|
    pxor        mm2, mm5
 | 
						|
    test        r2d, r2d
 | 
						|
    jnz .do_top
 | 
						|
.fix_tr_1:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm1
 | 
						|
    psrlq       mm5, 56
 | 
						|
    psllq       mm5, 56
 | 
						|
    pxor        mm1, mm5
 | 
						|
    jmp .do_top
 | 
						|
.do_left:
 | 
						|
    movq        mm0, mm4
 | 
						|
    PRED4x4_LOWPASS mm2, mm1, mm4, mm3, mm5
 | 
						|
    movq        mm7, mm2
 | 
						|
    movq        mm0, [r0-8]
 | 
						|
    movq        mm3, [r0]
 | 
						|
    movq        mm1, [r0+8]
 | 
						|
    movq        mm2, mm3
 | 
						|
    movq        mm4, mm3
 | 
						|
    PALIGNR     mm2, mm0, 7, mm0
 | 
						|
    PALIGNR     mm1, mm4, 1, mm4
 | 
						|
    test        r1d, r1d
 | 
						|
    jz .fix_lt_2
 | 
						|
    test        r2d, r2d
 | 
						|
    jz .fix_tr_1
 | 
						|
.do_top:
 | 
						|
    PRED4x4_LOWPASS mm6, mm2, mm1, mm3, mm5
 | 
						|
    lea         r1, [r0+r3*2]
 | 
						|
    movq       mm2, mm6
 | 
						|
    movq       mm3, mm6
 | 
						|
    PALIGNR    mm3, mm7, 7, mm0
 | 
						|
    PALIGNR    mm6, mm7, 6, mm1
 | 
						|
    movq       mm4, mm3
 | 
						|
    pavgb      mm3, mm2
 | 
						|
    lea         r2, [r1+r3*2]
 | 
						|
    PRED4x4_LOWPASS mm0, mm6, mm2, mm4, mm5
 | 
						|
    movq [r0+r3*1], mm3
 | 
						|
    movq [r0+r3*2], mm0
 | 
						|
    movq       mm5, mm0
 | 
						|
    movq       mm6, mm3
 | 
						|
    movq       mm1, mm7
 | 
						|
    movq       mm2, mm1
 | 
						|
    psllq      mm2, 8
 | 
						|
    movq       mm3, mm1
 | 
						|
    psllq      mm3, 16
 | 
						|
    lea         r4, [r2+r3*2]
 | 
						|
    PRED4x4_LOWPASS mm0, mm1, mm3, mm2, mm4
 | 
						|
    PALIGNR    mm6, mm0, 7, mm2
 | 
						|
    movq [r1+r3*1], mm6
 | 
						|
    psllq      mm0, 8
 | 
						|
    PALIGNR    mm5, mm0, 7, mm1
 | 
						|
    movq [r1+r3*2], mm5
 | 
						|
    psllq      mm0, 8
 | 
						|
    PALIGNR    mm6, mm0, 7, mm2
 | 
						|
    movq [r2+r3*1], mm6
 | 
						|
    psllq      mm0, 8
 | 
						|
    PALIGNR    mm5, mm0, 7, mm1
 | 
						|
    movq [r2+r3*2], mm5
 | 
						|
    psllq      mm0, 8
 | 
						|
    PALIGNR    mm6, mm0, 7, mm2
 | 
						|
    movq [r4+r3*1], mm6
 | 
						|
    psllq      mm0, 8
 | 
						|
    PALIGNR    mm5, mm0, 7, mm1
 | 
						|
    movq [r4+r3*2], mm5
 | 
						|
    RET
 | 
						|
 | 
						|
%macro PRED8x8L_VERTICAL_RIGHT 0
 | 
						|
cglobal pred8x8l_vertical_right_8, 4,5,7
 | 
						|
    ; manually spill XMM registers for Win64 because
 | 
						|
    ; the code here is initialized with INIT_MMX
 | 
						|
    WIN64_SPILL_XMM 7
 | 
						|
    sub          r0, r3
 | 
						|
    lea          r4, [r0+r3*2]
 | 
						|
    movq        mm0, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm0, [r0+r3*0-8]
 | 
						|
    movq        mm1, [r4+r3*1-8]
 | 
						|
    punpckhbw   mm1, [r0+r3*2-8]
 | 
						|
    mov          r4, r0
 | 
						|
    punpckhwd   mm1, mm0
 | 
						|
    lea          r0, [r0+r3*4]
 | 
						|
    movq        mm2, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm2, [r0+r3*0-8]
 | 
						|
    lea          r0, [r0+r3*2]
 | 
						|
    movq        mm3, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm3, [r0+r3*0-8]
 | 
						|
    punpckhwd   mm3, mm2
 | 
						|
    punpckhdq   mm3, mm1
 | 
						|
    lea          r0, [r0+r3*2]
 | 
						|
    movq        mm0, [r0+r3*0-8]
 | 
						|
    movq        mm1, [r4]
 | 
						|
    mov          r0, r4
 | 
						|
    movq        mm4, mm3
 | 
						|
    movq        mm2, mm3
 | 
						|
    PALIGNR     mm4, mm0, 7, mm0
 | 
						|
    PALIGNR     mm1, mm2, 1, mm2
 | 
						|
    test        r1d, r1d
 | 
						|
    jnz .do_left
 | 
						|
.fix_lt_1:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm4
 | 
						|
    psrlq       mm5, 56
 | 
						|
    psllq       mm5, 48
 | 
						|
    pxor        mm1, mm5
 | 
						|
    jmp .do_left
 | 
						|
.fix_lt_2:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm2
 | 
						|
    psllq       mm5, 56
 | 
						|
    psrlq       mm5, 56
 | 
						|
    pxor        mm2, mm5
 | 
						|
    test        r2d, r2d
 | 
						|
    jnz .do_top
 | 
						|
.fix_tr_1:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm1
 | 
						|
    psrlq       mm5, 56
 | 
						|
    psllq       mm5, 56
 | 
						|
    pxor        mm1, mm5
 | 
						|
    jmp .do_top
 | 
						|
.do_left:
 | 
						|
    movq        mm0, mm4
 | 
						|
    PRED4x4_LOWPASS mm2, mm1, mm4, mm3, mm5
 | 
						|
    movq2dq    xmm0, mm2
 | 
						|
    movq        mm0, [r0-8]
 | 
						|
    movq        mm3, [r0]
 | 
						|
    movq        mm1, [r0+8]
 | 
						|
    movq        mm2, mm3
 | 
						|
    movq        mm4, mm3
 | 
						|
    PALIGNR     mm2, mm0, 7, mm0
 | 
						|
    PALIGNR     mm1, mm4, 1, mm4
 | 
						|
    test        r1d, r1d
 | 
						|
    jz .fix_lt_2
 | 
						|
    test        r2d, r2d
 | 
						|
    jz .fix_tr_1
 | 
						|
.do_top:
 | 
						|
    PRED4x4_LOWPASS mm6, mm2, mm1, mm3, mm5
 | 
						|
    lea           r1, [r0+r3*2]
 | 
						|
    movq2dq     xmm4, mm6
 | 
						|
    pslldq      xmm4, 8
 | 
						|
    por         xmm0, xmm4
 | 
						|
    movdqa      xmm6, [pw_ff00]
 | 
						|
    movdqa      xmm1, xmm0
 | 
						|
    lea           r2, [r1+r3*2]
 | 
						|
    movdqa      xmm2, xmm0
 | 
						|
    movdqa      xmm3, xmm0
 | 
						|
    pslldq      xmm0, 1
 | 
						|
    pslldq      xmm1, 2
 | 
						|
    pavgb       xmm2, xmm0
 | 
						|
INIT_XMM cpuname
 | 
						|
    PRED4x4_LOWPASS xmm4, xmm3, xmm1, xmm0, xmm5
 | 
						|
    pandn       xmm6, xmm4
 | 
						|
    movdqa      xmm5, xmm4
 | 
						|
    psrlw       xmm4, 8
 | 
						|
    packuswb    xmm6, xmm4
 | 
						|
    movhlps     xmm4, xmm6
 | 
						|
    movhps [r0+r3*2], xmm5
 | 
						|
    movhps [r0+r3*1], xmm2
 | 
						|
    psrldq      xmm5, 4
 | 
						|
    movss       xmm5, xmm6
 | 
						|
    psrldq      xmm2, 4
 | 
						|
    movss       xmm2, xmm4
 | 
						|
    lea           r0, [r2+r3*2]
 | 
						|
    psrldq      xmm5, 1
 | 
						|
    psrldq      xmm2, 1
 | 
						|
    movq        [r0+r3*2], xmm5
 | 
						|
    movq        [r0+r3*1], xmm2
 | 
						|
    psrldq      xmm5, 1
 | 
						|
    psrldq      xmm2, 1
 | 
						|
    movq        [r2+r3*2], xmm5
 | 
						|
    movq        [r2+r3*1], xmm2
 | 
						|
    psrldq      xmm5, 1
 | 
						|
    psrldq      xmm2, 1
 | 
						|
    movq        [r1+r3*2], xmm5
 | 
						|
    movq        [r1+r3*1], xmm2
 | 
						|
    RET
 | 
						|
%endmacro
 | 
						|
 | 
						|
INIT_MMX sse2
 | 
						|
PRED8x8L_VERTICAL_RIGHT
 | 
						|
INIT_MMX ssse3
 | 
						|
PRED8x8L_VERTICAL_RIGHT
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred8x8l_vertical_left_8(uint8_t *src, int has_topleft,
 | 
						|
;                                  int has_topright, int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
%macro PRED8x8L_VERTICAL_LEFT 0
 | 
						|
cglobal pred8x8l_vertical_left_8, 4,4
 | 
						|
    sub          r0, r3
 | 
						|
    movq        mm0, [r0-8]
 | 
						|
    movq        mm3, [r0]
 | 
						|
    movq        mm1, [r0+8]
 | 
						|
    movq        mm2, mm3
 | 
						|
    movq        mm4, mm3
 | 
						|
    PALIGNR     mm2, mm0, 7, mm0
 | 
						|
    PALIGNR     mm1, mm4, 1, mm4
 | 
						|
    test        r1d, r1d
 | 
						|
    jz .fix_lt_2
 | 
						|
    test        r2d, r2d
 | 
						|
    jz .fix_tr_1
 | 
						|
    jmp .do_top
 | 
						|
.fix_lt_2:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm2
 | 
						|
    psllq       mm5, 56
 | 
						|
    psrlq       mm5, 56
 | 
						|
    pxor        mm2, mm5
 | 
						|
    test        r2d, r2d
 | 
						|
    jnz .do_top
 | 
						|
.fix_tr_1:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm1
 | 
						|
    psrlq       mm5, 56
 | 
						|
    psllq       mm5, 56
 | 
						|
    pxor        mm1, mm5
 | 
						|
    jmp .do_top
 | 
						|
.fix_tr_2:
 | 
						|
    punpckhbw   mm3, mm3
 | 
						|
    pshufw      mm1, mm3, 0xFF
 | 
						|
    jmp .do_topright
 | 
						|
.do_top:
 | 
						|
    PRED4x4_LOWPASS mm4, mm2, mm1, mm3, mm5
 | 
						|
    movq2dq    xmm4, mm4
 | 
						|
    test        r2d, r2d
 | 
						|
    jz .fix_tr_2
 | 
						|
    movq        mm0, [r0+8]
 | 
						|
    movq        mm5, mm0
 | 
						|
    movq        mm2, mm0
 | 
						|
    movq        mm4, mm0
 | 
						|
    psrlq       mm5, 56
 | 
						|
    PALIGNR     mm2, mm3, 7, mm3
 | 
						|
    PALIGNR     mm5, mm4, 1, mm4
 | 
						|
    PRED4x4_LOWPASS mm1, mm2, mm5, mm0, mm4
 | 
						|
.do_topright:
 | 
						|
    movq2dq   xmm3, mm1
 | 
						|
    lea         r1, [r0+r3*2]
 | 
						|
    pslldq    xmm3, 8
 | 
						|
    por       xmm4, xmm3
 | 
						|
    movdqa    xmm2, xmm4
 | 
						|
    movdqa    xmm1, xmm4
 | 
						|
    movdqa    xmm3, xmm4
 | 
						|
    psrldq    xmm2, 1
 | 
						|
    pslldq    xmm1, 1
 | 
						|
    pavgb     xmm3, xmm2
 | 
						|
    lea         r2, [r1+r3*2]
 | 
						|
INIT_XMM cpuname
 | 
						|
    PRED4x4_LOWPASS xmm0, xmm1, xmm2, xmm4, xmm5
 | 
						|
    psrldq    xmm0, 1
 | 
						|
    movq [r0+r3*1], xmm3
 | 
						|
    movq [r0+r3*2], xmm0
 | 
						|
    lea         r0, [r2+r3*2]
 | 
						|
    psrldq    xmm3, 1
 | 
						|
    psrldq    xmm0, 1
 | 
						|
    movq [r1+r3*1], xmm3
 | 
						|
    movq [r1+r3*2], xmm0
 | 
						|
    psrldq    xmm3, 1
 | 
						|
    psrldq    xmm0, 1
 | 
						|
    movq [r2+r3*1], xmm3
 | 
						|
    movq [r2+r3*2], xmm0
 | 
						|
    psrldq    xmm3, 1
 | 
						|
    psrldq    xmm0, 1
 | 
						|
    movq [r0+r3*1], xmm3
 | 
						|
    movq [r0+r3*2], xmm0
 | 
						|
    RET
 | 
						|
%endmacro
 | 
						|
 | 
						|
INIT_MMX sse2
 | 
						|
PRED8x8L_VERTICAL_LEFT
 | 
						|
INIT_MMX ssse3
 | 
						|
PRED8x8L_VERTICAL_LEFT
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred8x8l_horizontal_up_8(uint8_t *src, int has_topleft,
 | 
						|
;                                  int has_topright, int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
%macro PRED8x8L_HORIZONTAL_UP 0
 | 
						|
cglobal pred8x8l_horizontal_up_8, 4,4
 | 
						|
    sub          r0, r3
 | 
						|
    lea          r2, [r0+r3*2]
 | 
						|
    movq        mm0, [r0+r3*1-8]
 | 
						|
    test        r1d, r1d
 | 
						|
    lea          r1, [r0+r3]
 | 
						|
    cmovnz       r1, r0
 | 
						|
    punpckhbw   mm0, [r1+r3*0-8]
 | 
						|
    movq        mm1, [r2+r3*1-8]
 | 
						|
    punpckhbw   mm1, [r0+r3*2-8]
 | 
						|
    mov          r2, r0
 | 
						|
    punpckhwd   mm1, mm0
 | 
						|
    lea          r0, [r0+r3*4]
 | 
						|
    movq        mm2, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm2, [r0+r3*0-8]
 | 
						|
    lea          r0, [r0+r3*2]
 | 
						|
    movq        mm3, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm3, [r0+r3*0-8]
 | 
						|
    punpckhwd   mm3, mm2
 | 
						|
    punpckhdq   mm3, mm1
 | 
						|
    lea          r0, [r0+r3*2]
 | 
						|
    movq        mm0, [r0+r3*0-8]
 | 
						|
    movq        mm1, [r1+r3*0-8]
 | 
						|
    mov          r0, r2
 | 
						|
    movq        mm4, mm3
 | 
						|
    movq        mm2, mm3
 | 
						|
    PALIGNR     mm4, mm0, 7, mm0
 | 
						|
    PALIGNR     mm1, mm2, 1, mm2
 | 
						|
    movq       mm0, mm4
 | 
						|
    PRED4x4_LOWPASS mm2, mm1, mm4, mm3, mm5
 | 
						|
    movq       mm4, mm0
 | 
						|
    movq       mm7, mm2
 | 
						|
    PRED4x4_LOWPASS mm1, mm3, mm0, mm4, mm5
 | 
						|
    psllq      mm1, 56
 | 
						|
    PALIGNR    mm7, mm1, 7, mm3
 | 
						|
    lea         r1, [r0+r3*2]
 | 
						|
    pshufw     mm0, mm7, 00011011b ; l6 l7 l4 l5 l2 l3 l0 l1
 | 
						|
    psllq      mm7, 56             ; l7 .. .. .. .. .. .. ..
 | 
						|
    movq       mm2, mm0
 | 
						|
    psllw      mm0, 8
 | 
						|
    psrlw      mm2, 8
 | 
						|
    por        mm2, mm0            ; l7 l6 l5 l4 l3 l2 l1 l0
 | 
						|
    movq       mm3, mm2
 | 
						|
    movq       mm4, mm2
 | 
						|
    movq       mm5, mm2
 | 
						|
    psrlq      mm2, 8
 | 
						|
    psrlq      mm3, 16
 | 
						|
    lea         r2, [r1+r3*2]
 | 
						|
    por        mm2, mm7            ; l7 l7 l6 l5 l4 l3 l2 l1
 | 
						|
    punpckhbw  mm7, mm7
 | 
						|
    por        mm3, mm7            ; l7 l7 l7 l6 l5 l4 l3 l2
 | 
						|
    pavgb      mm4, mm2
 | 
						|
    PRED4x4_LOWPASS mm1, mm3, mm5, mm2, mm6
 | 
						|
    movq       mm5, mm4
 | 
						|
    punpcklbw  mm4, mm1            ; p4 p3 p2 p1
 | 
						|
    punpckhbw  mm5, mm1            ; p8 p7 p6 p5
 | 
						|
    movq       mm6, mm5
 | 
						|
    movq       mm7, mm5
 | 
						|
    movq       mm0, mm5
 | 
						|
    PALIGNR    mm5, mm4, 2, mm1
 | 
						|
    pshufw     mm1, mm6, 11111001b
 | 
						|
    PALIGNR    mm6, mm4, 4, mm2
 | 
						|
    pshufw     mm2, mm7, 11111110b
 | 
						|
    PALIGNR    mm7, mm4, 6, mm3
 | 
						|
    pshufw     mm3, mm0, 11111111b
 | 
						|
    movq [r0+r3*1], mm4
 | 
						|
    movq [r0+r3*2], mm5
 | 
						|
    lea         r0, [r2+r3*2]
 | 
						|
    movq [r1+r3*1], mm6
 | 
						|
    movq [r1+r3*2], mm7
 | 
						|
    movq [r2+r3*1], mm0
 | 
						|
    movq [r2+r3*2], mm1
 | 
						|
    movq [r0+r3*1], mm2
 | 
						|
    movq [r0+r3*2], mm3
 | 
						|
    RET
 | 
						|
%endmacro
 | 
						|
 | 
						|
INIT_MMX mmxext
 | 
						|
PRED8x8L_HORIZONTAL_UP
 | 
						|
INIT_MMX ssse3
 | 
						|
PRED8x8L_HORIZONTAL_UP
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred8x8l_horizontal_down_8(uint8_t *src, int has_topleft,
 | 
						|
;                                    int has_topright, int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
INIT_MMX mmxext
 | 
						|
cglobal pred8x8l_horizontal_down_8, 4,5
 | 
						|
    sub          r0, r3
 | 
						|
    lea          r4, [r0+r3*2]
 | 
						|
    movq        mm0, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm0, [r0+r3*0-8]
 | 
						|
    movq        mm1, [r4+r3*1-8]
 | 
						|
    punpckhbw   mm1, [r0+r3*2-8]
 | 
						|
    mov          r4, r0
 | 
						|
    punpckhwd   mm1, mm0
 | 
						|
    lea          r0, [r0+r3*4]
 | 
						|
    movq        mm2, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm2, [r0+r3*0-8]
 | 
						|
    lea          r0, [r0+r3*2]
 | 
						|
    movq        mm3, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm3, [r0+r3*0-8]
 | 
						|
    punpckhwd   mm3, mm2
 | 
						|
    punpckhdq   mm3, mm1
 | 
						|
    lea          r0, [r0+r3*2]
 | 
						|
    movq        mm0, [r0+r3*0-8]
 | 
						|
    movq        mm1, [r4]
 | 
						|
    mov          r0, r4
 | 
						|
    movq        mm4, mm3
 | 
						|
    movq        mm2, mm3
 | 
						|
    PALIGNR     mm4, mm0, 7, mm0
 | 
						|
    PALIGNR     mm1, mm2, 1, mm2
 | 
						|
    test        r1d, r1d
 | 
						|
    jnz .do_left
 | 
						|
.fix_lt_1:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm4
 | 
						|
    psrlq       mm5, 56
 | 
						|
    psllq       mm5, 48
 | 
						|
    pxor        mm1, mm5
 | 
						|
    jmp .do_left
 | 
						|
.fix_lt_2:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm2
 | 
						|
    psllq       mm5, 56
 | 
						|
    psrlq       mm5, 56
 | 
						|
    pxor        mm2, mm5
 | 
						|
    test        r2d, r2d
 | 
						|
    jnz .do_top
 | 
						|
.fix_tr_1:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm1
 | 
						|
    psrlq       mm5, 56
 | 
						|
    psllq       mm5, 56
 | 
						|
    pxor        mm1, mm5
 | 
						|
    jmp .do_top
 | 
						|
.do_left:
 | 
						|
    movq        mm0, mm4
 | 
						|
    PRED4x4_LOWPASS mm2, mm1, mm4, mm3, mm5
 | 
						|
    movq        mm4, mm0
 | 
						|
    movq        mm7, mm2
 | 
						|
    movq        mm6, mm2
 | 
						|
    PRED4x4_LOWPASS mm1, mm3, mm0, mm4, mm5
 | 
						|
    psllq       mm1, 56
 | 
						|
    PALIGNR     mm7, mm1, 7, mm3
 | 
						|
    movq        mm0, [r0-8]
 | 
						|
    movq        mm3, [r0]
 | 
						|
    movq        mm1, [r0+8]
 | 
						|
    movq        mm2, mm3
 | 
						|
    movq        mm4, mm3
 | 
						|
    PALIGNR     mm2, mm0, 7, mm0
 | 
						|
    PALIGNR     mm1, mm4, 1, mm4
 | 
						|
    test        r1d, r1d
 | 
						|
    jz .fix_lt_2
 | 
						|
    test        r2d, r2d
 | 
						|
    jz .fix_tr_1
 | 
						|
.do_top:
 | 
						|
    PRED4x4_LOWPASS mm4, mm2, mm1, mm3, mm5
 | 
						|
    movq       mm5, mm4
 | 
						|
    lea         r1, [r0+r3*2]
 | 
						|
    psllq      mm7, 56
 | 
						|
    movq       mm2, mm5
 | 
						|
    movq       mm3, mm6
 | 
						|
    movq       mm4, mm2
 | 
						|
    PALIGNR    mm2, mm6, 7, mm5
 | 
						|
    PALIGNR    mm6, mm7, 7, mm0
 | 
						|
    lea         r2, [r1+r3*2]
 | 
						|
    PALIGNR    mm4, mm3, 1, mm7
 | 
						|
    movq       mm5, mm3
 | 
						|
    pavgb      mm3, mm6
 | 
						|
    PRED4x4_LOWPASS mm0, mm4, mm6, mm5, mm7
 | 
						|
    movq       mm4, mm2
 | 
						|
    movq       mm1, mm2
 | 
						|
    lea         r4, [r2+r3*2]
 | 
						|
    psrlq      mm4, 16
 | 
						|
    psrlq      mm1, 8
 | 
						|
    PRED4x4_LOWPASS mm6, mm4, mm2, mm1, mm5
 | 
						|
    movq       mm7, mm3
 | 
						|
    punpcklbw  mm3, mm0
 | 
						|
    punpckhbw  mm7, mm0
 | 
						|
    movq       mm1, mm7
 | 
						|
    movq       mm0, mm7
 | 
						|
    movq       mm4, mm7
 | 
						|
    movq [r4+r3*2], mm3
 | 
						|
    PALIGNR    mm7, mm3, 2, mm5
 | 
						|
    movq [r4+r3*1], mm7
 | 
						|
    PALIGNR    mm1, mm3, 4, mm5
 | 
						|
    movq [r2+r3*2], mm1
 | 
						|
    PALIGNR    mm0, mm3, 6, mm3
 | 
						|
    movq [r2+r3*1], mm0
 | 
						|
    movq       mm2, mm6
 | 
						|
    movq       mm3, mm6
 | 
						|
    movq [r1+r3*2], mm4
 | 
						|
    PALIGNR    mm6, mm4, 2, mm5
 | 
						|
    movq [r1+r3*1], mm6
 | 
						|
    PALIGNR    mm2, mm4, 4, mm5
 | 
						|
    movq [r0+r3*2], mm2
 | 
						|
    PALIGNR    mm3, mm4, 6, mm4
 | 
						|
    movq [r0+r3*1], mm3
 | 
						|
    RET
 | 
						|
 | 
						|
%macro PRED8x8L_HORIZONTAL_DOWN 0
 | 
						|
cglobal pred8x8l_horizontal_down_8, 4,5
 | 
						|
    sub          r0, r3
 | 
						|
    lea          r4, [r0+r3*2]
 | 
						|
    movq        mm0, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm0, [r0+r3*0-8]
 | 
						|
    movq        mm1, [r4+r3*1-8]
 | 
						|
    punpckhbw   mm1, [r0+r3*2-8]
 | 
						|
    mov          r4, r0
 | 
						|
    punpckhwd   mm1, mm0
 | 
						|
    lea          r0, [r0+r3*4]
 | 
						|
    movq        mm2, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm2, [r0+r3*0-8]
 | 
						|
    lea          r0, [r0+r3*2]
 | 
						|
    movq        mm3, [r0+r3*1-8]
 | 
						|
    punpckhbw   mm3, [r0+r3*0-8]
 | 
						|
    punpckhwd   mm3, mm2
 | 
						|
    punpckhdq   mm3, mm1
 | 
						|
    lea          r0, [r0+r3*2]
 | 
						|
    movq        mm0, [r0+r3*0-8]
 | 
						|
    movq        mm1, [r4]
 | 
						|
    mov          r0, r4
 | 
						|
    movq        mm4, mm3
 | 
						|
    movq        mm2, mm3
 | 
						|
    PALIGNR     mm4, mm0, 7, mm0
 | 
						|
    PALIGNR     mm1, mm2, 1, mm2
 | 
						|
    test        r1d, r1d
 | 
						|
    jnz .do_left
 | 
						|
.fix_lt_1:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm4
 | 
						|
    psrlq       mm5, 56
 | 
						|
    psllq       mm5, 48
 | 
						|
    pxor        mm1, mm5
 | 
						|
    jmp .do_left
 | 
						|
.fix_lt_2:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm2
 | 
						|
    psllq       mm5, 56
 | 
						|
    psrlq       mm5, 56
 | 
						|
    pxor        mm2, mm5
 | 
						|
    test        r2d, r2d
 | 
						|
    jnz .do_top
 | 
						|
.fix_tr_1:
 | 
						|
    movq        mm5, mm3
 | 
						|
    pxor        mm5, mm1
 | 
						|
    psrlq       mm5, 56
 | 
						|
    psllq       mm5, 56
 | 
						|
    pxor        mm1, mm5
 | 
						|
    jmp .do_top
 | 
						|
.fix_tr_2:
 | 
						|
    punpckhbw   mm3, mm3
 | 
						|
    pshufw      mm1, mm3, 0xFF
 | 
						|
    jmp .do_topright
 | 
						|
.do_left:
 | 
						|
    movq        mm0, mm4
 | 
						|
    PRED4x4_LOWPASS mm2, mm1, mm4, mm3, mm5
 | 
						|
    movq2dq    xmm0, mm2
 | 
						|
    pslldq     xmm0, 8
 | 
						|
    movq        mm4, mm0
 | 
						|
    PRED4x4_LOWPASS mm1, mm3, mm0, mm4, mm5
 | 
						|
    movq2dq    xmm2, mm1
 | 
						|
    pslldq     xmm2, 15
 | 
						|
    psrldq     xmm2, 8
 | 
						|
    por        xmm0, xmm2
 | 
						|
    movq        mm0, [r0-8]
 | 
						|
    movq        mm3, [r0]
 | 
						|
    movq        mm1, [r0+8]
 | 
						|
    movq        mm2, mm3
 | 
						|
    movq        mm4, mm3
 | 
						|
    PALIGNR     mm2, mm0, 7, mm0
 | 
						|
    PALIGNR     mm1, mm4, 1, mm4
 | 
						|
    test        r1d, r1d
 | 
						|
    jz .fix_lt_2
 | 
						|
    test        r2d, r2d
 | 
						|
    jz .fix_tr_1
 | 
						|
.do_top:
 | 
						|
    PRED4x4_LOWPASS mm4, mm2, mm1, mm3, mm5
 | 
						|
    movq2dq    xmm1, mm4
 | 
						|
    test        r2d, r2d
 | 
						|
    jz .fix_tr_2
 | 
						|
    movq        mm0, [r0+8]
 | 
						|
    movq        mm5, mm0
 | 
						|
    movq        mm2, mm0
 | 
						|
    movq        mm4, mm0
 | 
						|
    psrlq       mm5, 56
 | 
						|
    PALIGNR     mm2, mm3, 7, mm3
 | 
						|
    PALIGNR     mm5, mm4, 1, mm4
 | 
						|
    PRED4x4_LOWPASS mm1, mm2, mm5, mm0, mm4
 | 
						|
.do_topright:
 | 
						|
    movq2dq    xmm5, mm1
 | 
						|
    pslldq     xmm5, 8
 | 
						|
    por        xmm1, xmm5
 | 
						|
INIT_XMM cpuname
 | 
						|
    lea         r2, [r4+r3*2]
 | 
						|
    movdqa    xmm2, xmm1
 | 
						|
    movdqa    xmm3, xmm1
 | 
						|
    PALIGNR   xmm1, xmm0, 7, xmm4
 | 
						|
    PALIGNR   xmm2, xmm0, 9, xmm5
 | 
						|
    lea         r1, [r2+r3*2]
 | 
						|
    PALIGNR   xmm3, xmm0, 8, xmm0
 | 
						|
    movdqa    xmm4, xmm1
 | 
						|
    pavgb     xmm4, xmm3
 | 
						|
    lea         r0, [r1+r3*2]
 | 
						|
    PRED4x4_LOWPASS xmm0, xmm1, xmm2, xmm3, xmm5
 | 
						|
    punpcklbw xmm4, xmm0
 | 
						|
    movhlps   xmm0, xmm4
 | 
						|
    movq   [r0+r3*2], xmm4
 | 
						|
    movq   [r2+r3*2], xmm0
 | 
						|
    psrldq xmm4, 2
 | 
						|
    psrldq xmm0, 2
 | 
						|
    movq   [r0+r3*1], xmm4
 | 
						|
    movq   [r2+r3*1], xmm0
 | 
						|
    psrldq xmm4, 2
 | 
						|
    psrldq xmm0, 2
 | 
						|
    movq   [r1+r3*2], xmm4
 | 
						|
    movq   [r4+r3*2], xmm0
 | 
						|
    psrldq xmm4, 2
 | 
						|
    psrldq xmm0, 2
 | 
						|
    movq   [r1+r3*1], xmm4
 | 
						|
    movq   [r4+r3*1], xmm0
 | 
						|
    RET
 | 
						|
%endmacro
 | 
						|
 | 
						|
INIT_MMX sse2
 | 
						|
PRED8x8L_HORIZONTAL_DOWN
 | 
						|
INIT_MMX ssse3
 | 
						|
PRED8x8L_HORIZONTAL_DOWN
 | 
						|
 | 
						|
;-------------------------------------------------------------------------------
 | 
						|
; void ff_pred4x4_dc_8_mmxext(uint8_t *src, const uint8_t *topright, int stride)
 | 
						|
;-------------------------------------------------------------------------------
 | 
						|
 | 
						|
INIT_MMX mmxext
 | 
						|
cglobal pred4x4_dc_8, 3,5
 | 
						|
    pxor   mm7, mm7
 | 
						|
    mov     r4, r0
 | 
						|
    sub     r0, r2
 | 
						|
    movd   mm0, [r0]
 | 
						|
    psadbw mm0, mm7
 | 
						|
    movzx  r1d, byte [r0+r2*1-1]
 | 
						|
    movd   r3d, mm0
 | 
						|
    add    r3d, r1d
 | 
						|
    movzx  r1d, byte [r0+r2*2-1]
 | 
						|
    lea     r0, [r0+r2*2]
 | 
						|
    add    r3d, r1d
 | 
						|
    movzx  r1d, byte [r0+r2*1-1]
 | 
						|
    add    r3d, r1d
 | 
						|
    movzx  r1d, byte [r0+r2*2-1]
 | 
						|
    add    r3d, r1d
 | 
						|
    add    r3d, 4
 | 
						|
    shr    r3d, 3
 | 
						|
    imul   r3d, 0x01010101
 | 
						|
    mov   [r4+r2*0], r3d
 | 
						|
    mov   [r0+r2*0], r3d
 | 
						|
    mov   [r0+r2*1], r3d
 | 
						|
    mov   [r0+r2*2], r3d
 | 
						|
    RET
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred4x4_tm_vp8_8_mmxext(uint8_t *src, const uint8_t *topright,
 | 
						|
;                                 int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
%macro PRED4x4_TM 0
 | 
						|
cglobal pred4x4_tm_vp8_8, 3,6
 | 
						|
    sub        r0, r2
 | 
						|
    pxor      mm7, mm7
 | 
						|
    movd      mm0, [r0]
 | 
						|
    punpcklbw mm0, mm7
 | 
						|
    movzx     r4d, byte [r0-1]
 | 
						|
    mov       r5d, 2
 | 
						|
.loop:
 | 
						|
    movzx     r1d, byte [r0+r2*1-1]
 | 
						|
    movzx     r3d, byte [r0+r2*2-1]
 | 
						|
    sub       r1d, r4d
 | 
						|
    sub       r3d, r4d
 | 
						|
    movd      mm2, r1d
 | 
						|
    movd      mm4, r3d
 | 
						|
%if cpuflag(mmxext)
 | 
						|
    pshufw    mm2, mm2, 0
 | 
						|
    pshufw    mm4, mm4, 0
 | 
						|
%else
 | 
						|
    punpcklwd mm2, mm2
 | 
						|
    punpcklwd mm4, mm4
 | 
						|
    punpckldq mm2, mm2
 | 
						|
    punpckldq mm4, mm4
 | 
						|
%endif
 | 
						|
    paddw     mm2, mm0
 | 
						|
    paddw     mm4, mm0
 | 
						|
    packuswb  mm2, mm2
 | 
						|
    packuswb  mm4, mm4
 | 
						|
    movd [r0+r2*1], mm2
 | 
						|
    movd [r0+r2*2], mm4
 | 
						|
    lea        r0, [r0+r2*2]
 | 
						|
    dec       r5d
 | 
						|
    jg .loop
 | 
						|
    REP_RET
 | 
						|
%endmacro
 | 
						|
 | 
						|
INIT_MMX mmx
 | 
						|
PRED4x4_TM
 | 
						|
INIT_MMX mmxext
 | 
						|
PRED4x4_TM
 | 
						|
 | 
						|
INIT_XMM ssse3
 | 
						|
cglobal pred4x4_tm_vp8_8, 3,3
 | 
						|
    sub         r0, r2
 | 
						|
    movq       mm6, [tm_shuf]
 | 
						|
    pxor       mm1, mm1
 | 
						|
    movd       mm0, [r0]
 | 
						|
    punpcklbw  mm0, mm1
 | 
						|
    movd       mm7, [r0-4]
 | 
						|
    pshufb     mm7, mm6
 | 
						|
    lea         r1, [r0+r2*2]
 | 
						|
    movd       mm2, [r0+r2*1-4]
 | 
						|
    movd       mm3, [r0+r2*2-4]
 | 
						|
    movd       mm4, [r1+r2*1-4]
 | 
						|
    movd       mm5, [r1+r2*2-4]
 | 
						|
    pshufb     mm2, mm6
 | 
						|
    pshufb     mm3, mm6
 | 
						|
    pshufb     mm4, mm6
 | 
						|
    pshufb     mm5, mm6
 | 
						|
    psubw      mm0, mm7
 | 
						|
    paddw      mm2, mm0
 | 
						|
    paddw      mm3, mm0
 | 
						|
    paddw      mm4, mm0
 | 
						|
    paddw      mm5, mm0
 | 
						|
    packuswb   mm2, mm2
 | 
						|
    packuswb   mm3, mm3
 | 
						|
    packuswb   mm4, mm4
 | 
						|
    packuswb   mm5, mm5
 | 
						|
    movd [r0+r2*1], mm2
 | 
						|
    movd [r0+r2*2], mm3
 | 
						|
    movd [r1+r2*1], mm4
 | 
						|
    movd [r1+r2*2], mm5
 | 
						|
    RET
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred4x4_vertical_vp8_8_mmxext(uint8_t *src, const uint8_t *topright,
 | 
						|
;                                       int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
INIT_MMX mmxext
 | 
						|
cglobal pred4x4_vertical_vp8_8, 3,3
 | 
						|
    sub       r0, r2
 | 
						|
    movd      m1, [r0-1]
 | 
						|
    movd      m0, [r0]
 | 
						|
    mova      m2, m0   ;t0 t1 t2 t3
 | 
						|
    punpckldq m0, [r1] ;t0 t1 t2 t3 t4 t5 t6 t7
 | 
						|
    lea       r1, [r0+r2*2]
 | 
						|
    psrlq     m0, 8    ;t1 t2 t3 t4
 | 
						|
    PRED4x4_LOWPASS m3, m1, m0, m2, m4
 | 
						|
    movd [r0+r2*1], m3
 | 
						|
    movd [r0+r2*2], m3
 | 
						|
    movd [r1+r2*1], m3
 | 
						|
    movd [r1+r2*2], m3
 | 
						|
    RET
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred4x4_down_left_8_mmxext(uint8_t *src, const uint8_t *topright,
 | 
						|
;                                    int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
INIT_MMX mmxext
 | 
						|
cglobal pred4x4_down_left_8, 3,3
 | 
						|
    sub       r0, r2
 | 
						|
    movq      m1, [r0]
 | 
						|
    punpckldq m1, [r1]
 | 
						|
    movq      m2, m1
 | 
						|
    movq      m3, m1
 | 
						|
    psllq     m1, 8
 | 
						|
    pxor      m2, m1
 | 
						|
    psrlq     m2, 8
 | 
						|
    pxor      m2, m3
 | 
						|
    PRED4x4_LOWPASS m0, m1, m2, m3, m4
 | 
						|
    lea       r1, [r0+r2*2]
 | 
						|
    psrlq     m0, 8
 | 
						|
    movd      [r0+r2*1], m0
 | 
						|
    psrlq     m0, 8
 | 
						|
    movd      [r0+r2*2], m0
 | 
						|
    psrlq     m0, 8
 | 
						|
    movd      [r1+r2*1], m0
 | 
						|
    psrlq     m0, 8
 | 
						|
    movd      [r1+r2*2], m0
 | 
						|
    RET
 | 
						|
 | 
						|
;------------------------------------------------------------------------------
 | 
						|
; void ff_pred4x4_vertical_left_8_mmxext(uint8_t *src, const uint8_t *topright,
 | 
						|
;                                        int stride)
 | 
						|
;------------------------------------------------------------------------------
 | 
						|
 | 
						|
INIT_MMX mmxext
 | 
						|
cglobal pred4x4_vertical_left_8, 3,3
 | 
						|
    sub       r0, r2
 | 
						|
    movq      m1, [r0]
 | 
						|
    punpckldq m1, [r1]
 | 
						|
    movq      m3, m1
 | 
						|
    movq      m2, m1
 | 
						|
    psrlq     m3, 8
 | 
						|
    psrlq     m2, 16
 | 
						|
    movq      m4, m3
 | 
						|
    pavgb     m4, m1
 | 
						|
    PRED4x4_LOWPASS m0, m1, m2, m3, m5
 | 
						|
    lea       r1, [r0+r2*2]
 | 
						|
    movh      [r0+r2*1], m4
 | 
						|
    movh      [r0+r2*2], m0
 | 
						|
    psrlq     m4, 8
 | 
						|
    psrlq     m0, 8
 | 
						|
    movh      [r1+r2*1], m4
 | 
						|
    movh      [r1+r2*2], m0
 | 
						|
    RET
 | 
						|
 | 
						|
;------------------------------------------------------------------------------
 | 
						|
; void ff_pred4x4_horizontal_up_8_mmxext(uint8_t *src, const uint8_t *topright,
 | 
						|
;                                        int stride)
 | 
						|
;------------------------------------------------------------------------------
 | 
						|
 | 
						|
INIT_MMX mmxext
 | 
						|
cglobal pred4x4_horizontal_up_8, 3,3
 | 
						|
    sub       r0, r2
 | 
						|
    lea       r1, [r0+r2*2]
 | 
						|
    movd      m0, [r0+r2*1-4]
 | 
						|
    punpcklbw m0, [r0+r2*2-4]
 | 
						|
    movd      m1, [r1+r2*1-4]
 | 
						|
    punpcklbw m1, [r1+r2*2-4]
 | 
						|
    punpckhwd m0, m1
 | 
						|
    movq      m1, m0
 | 
						|
    punpckhbw m1, m1
 | 
						|
    pshufw    m1, m1, 0xFF
 | 
						|
    punpckhdq m0, m1
 | 
						|
    movq      m2, m0
 | 
						|
    movq      m3, m0
 | 
						|
    movq      m7, m0
 | 
						|
    psrlq     m2, 16
 | 
						|
    psrlq     m3, 8
 | 
						|
    pavgb     m7, m3
 | 
						|
    PRED4x4_LOWPASS m4, m0, m2, m3, m5
 | 
						|
    punpcklbw m7, m4
 | 
						|
    movd    [r0+r2*1], m7
 | 
						|
    psrlq    m7, 16
 | 
						|
    movd    [r0+r2*2], m7
 | 
						|
    psrlq    m7, 16
 | 
						|
    movd    [r1+r2*1], m7
 | 
						|
    movd    [r1+r2*2], m1
 | 
						|
    RET
 | 
						|
 | 
						|
;------------------------------------------------------------------------------
 | 
						|
; void ff_pred4x4_horizontal_down_8_mmxext(uint8_t *src,
 | 
						|
;                                          const uint8_t *topright, int stride)
 | 
						|
;------------------------------------------------------------------------------
 | 
						|
 | 
						|
INIT_MMX mmxext
 | 
						|
cglobal pred4x4_horizontal_down_8, 3,3
 | 
						|
    sub       r0, r2
 | 
						|
    lea       r1, [r0+r2*2]
 | 
						|
    movh      m0, [r0-4]      ; lt ..
 | 
						|
    punpckldq m0, [r0]        ; t3 t2 t1 t0 lt .. .. ..
 | 
						|
    psllq     m0, 8           ; t2 t1 t0 lt .. .. .. ..
 | 
						|
    movd      m1, [r1+r2*2-4] ; l3
 | 
						|
    punpcklbw m1, [r1+r2*1-4] ; l2 l3
 | 
						|
    movd      m2, [r0+r2*2-4] ; l1
 | 
						|
    punpcklbw m2, [r0+r2*1-4] ; l0 l1
 | 
						|
    punpckhwd m1, m2          ; l0 l1 l2 l3
 | 
						|
    punpckhdq m1, m0          ; t2 t1 t0 lt l0 l1 l2 l3
 | 
						|
    movq      m0, m1
 | 
						|
    movq      m2, m1
 | 
						|
    movq      m5, m1
 | 
						|
    psrlq     m0, 16          ; .. .. t2 t1 t0 lt l0 l1
 | 
						|
    psrlq     m2, 8           ; .. t2 t1 t0 lt l0 l1 l2
 | 
						|
    pavgb     m5, m2
 | 
						|
    PRED4x4_LOWPASS m3, m1, m0, m2, m4
 | 
						|
    punpcklbw m5, m3
 | 
						|
    psrlq     m3, 32
 | 
						|
    PALIGNR   m3, m5, 6, m4
 | 
						|
    movh      [r1+r2*2], m5
 | 
						|
    psrlq     m5, 16
 | 
						|
    movh      [r1+r2*1], m5
 | 
						|
    psrlq     m5, 16
 | 
						|
    movh      [r0+r2*2], m5
 | 
						|
    movh      [r0+r2*1], m3
 | 
						|
    RET
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred4x4_vertical_right_8_mmxext(uint8_t *src,
 | 
						|
;                                         const uint8_t *topright, int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
INIT_MMX mmxext
 | 
						|
cglobal pred4x4_vertical_right_8, 3,3
 | 
						|
    sub     r0, r2
 | 
						|
    lea     r1, [r0+r2*2]
 | 
						|
    movh    m0, [r0]                    ; ........t3t2t1t0
 | 
						|
    movq    m5, m0
 | 
						|
    PALIGNR m0, [r0-8], 7, m1           ; ......t3t2t1t0lt
 | 
						|
    pavgb   m5, m0
 | 
						|
    PALIGNR m0, [r0+r2*1-8], 7, m1      ; ....t3t2t1t0ltl0
 | 
						|
    movq    m1, m0
 | 
						|
    PALIGNR m0, [r0+r2*2-8], 7, m2      ; ..t3t2t1t0ltl0l1
 | 
						|
    movq    m2, m0
 | 
						|
    PALIGNR m0, [r1+r2*1-8], 7, m3      ; t3t2t1t0ltl0l1l2
 | 
						|
    PRED4x4_LOWPASS m3, m1, m0, m2, m4
 | 
						|
    movq    m1, m3
 | 
						|
    psrlq   m3, 16
 | 
						|
    psllq   m1, 48
 | 
						|
    movh    [r0+r2*1], m5
 | 
						|
    movh    [r0+r2*2], m3
 | 
						|
    PALIGNR m5, m1, 7, m2
 | 
						|
    psllq   m1, 8
 | 
						|
    movh    [r1+r2*1], m5
 | 
						|
    PALIGNR m3, m1, 7, m1
 | 
						|
    movh    [r1+r2*2], m3
 | 
						|
    RET
 | 
						|
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
; void ff_pred4x4_down_right_8_mmxext(uint8_t *src, const uint8_t *topright,
 | 
						|
;                                     int stride)
 | 
						|
;-----------------------------------------------------------------------------
 | 
						|
 | 
						|
INIT_MMX mmxext
 | 
						|
cglobal pred4x4_down_right_8, 3,3
 | 
						|
    sub       r0, r2
 | 
						|
    lea       r1, [r0+r2*2]
 | 
						|
    movq      m1, [r1-8]
 | 
						|
    movq      m2, [r0+r2*1-8]
 | 
						|
    punpckhbw m2, [r0-8]
 | 
						|
    movh      m3, [r0]
 | 
						|
    punpckhwd m1, m2
 | 
						|
    PALIGNR   m3, m1, 5, m1
 | 
						|
    movq      m1, m3
 | 
						|
    PALIGNR   m3, [r1+r2*1-8], 7, m4
 | 
						|
    movq      m2, m3
 | 
						|
    PALIGNR   m3, [r1+r2*2-8], 7, m4
 | 
						|
    PRED4x4_LOWPASS m0, m3, m1, m2, m4
 | 
						|
    movh      [r1+r2*2], m0
 | 
						|
    psrlq     m0, 8
 | 
						|
    movh      [r1+r2*1], m0
 | 
						|
    psrlq     m0, 8
 | 
						|
    movh      [r0+r2*2], m0
 | 
						|
    psrlq     m0, 8
 | 
						|
    movh      [r0+r2*1], m0
 | 
						|
    RET
 |