316 lines
8.2 KiB
NASM
316 lines
8.2 KiB
NASM
;******************************************************************************
|
|
;* x86 optimized Format Conversion Utils
|
|
;* Copyright (c) 2008 Loren Merritt
|
|
;* Copyright (c) 2012 Justin Ruggles <justin.ruggles@gmail.com>
|
|
;*
|
|
;* This file is part of Libav.
|
|
;*
|
|
;* Libav is free software; you can redistribute it and/or
|
|
;* modify it under the terms of the GNU Lesser General Public
|
|
;* License as published by the Free Software Foundation; either
|
|
;* version 2.1 of the License, or (at your option) any later version.
|
|
;*
|
|
;* Libav is distributed in the hope that it will be useful,
|
|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
;* Lesser General Public License for more details.
|
|
;*
|
|
;* You should have received a copy of the GNU Lesser General Public
|
|
;* License along with Libav; if not, write to the Free Software
|
|
;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
;******************************************************************************
|
|
|
|
%include "x86inc.asm"
|
|
%include "x86util.asm"
|
|
%include "util.asm"
|
|
|
|
SECTION_RODATA 32
|
|
|
|
pf_s32_inv_scale: times 8 dd 0x30000000
|
|
pf_s32_scale: times 8 dd 0x4f000000
|
|
pf_s16_inv_scale: times 4 dd 0x38000000
|
|
pf_s16_scale: times 4 dd 0x47000000
|
|
|
|
SECTION_TEXT
|
|
|
|
;------------------------------------------------------------------------------
|
|
; void ff_conv_s16_to_s32(int32_t *dst, const int16_t *src, int len);
|
|
;------------------------------------------------------------------------------
|
|
|
|
INIT_XMM sse2
|
|
cglobal conv_s16_to_s32, 3,3,3, dst, src, len
|
|
lea lenq, [2*lend]
|
|
lea dstq, [dstq+2*lenq]
|
|
add srcq, lenq
|
|
neg lenq
|
|
.loop:
|
|
mova m2, [srcq+lenq]
|
|
pxor m0, m0
|
|
pxor m1, m1
|
|
punpcklwd m0, m2
|
|
punpckhwd m1, m2
|
|
mova [dstq+2*lenq ], m0
|
|
mova [dstq+2*lenq+mmsize], m1
|
|
add lenq, mmsize
|
|
jl .loop
|
|
REP_RET
|
|
|
|
;------------------------------------------------------------------------------
|
|
; void ff_conv_s16_to_flt(float *dst, const int16_t *src, int len);
|
|
;------------------------------------------------------------------------------
|
|
|
|
%macro CONV_S16_TO_FLT 0
|
|
cglobal conv_s16_to_flt, 3,3,3, dst, src, len
|
|
lea lenq, [2*lend]
|
|
add srcq, lenq
|
|
lea dstq, [dstq + 2*lenq]
|
|
neg lenq
|
|
mova m2, [pf_s16_inv_scale]
|
|
ALIGN 16
|
|
.loop:
|
|
mova m0, [srcq+lenq]
|
|
S16_TO_S32_SX 0, 1
|
|
cvtdq2ps m0, m0
|
|
cvtdq2ps m1, m1
|
|
mulps m0, m2
|
|
mulps m1, m2
|
|
mova [dstq+2*lenq ], m0
|
|
mova [dstq+2*lenq+mmsize], m1
|
|
add lenq, mmsize
|
|
jl .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_XMM sse2
|
|
CONV_S16_TO_FLT
|
|
INIT_XMM sse4
|
|
CONV_S16_TO_FLT
|
|
|
|
;------------------------------------------------------------------------------
|
|
; void ff_conv_s32_to_s16(int16_t *dst, const int32_t *src, int len);
|
|
;------------------------------------------------------------------------------
|
|
|
|
%macro CONV_S32_TO_S16 0
|
|
cglobal conv_s32_to_s16, 3,3,4, dst, src, len
|
|
lea lenq, [2*lend]
|
|
lea srcq, [srcq+2*lenq]
|
|
add dstq, lenq
|
|
neg lenq
|
|
.loop:
|
|
mova m0, [srcq+2*lenq ]
|
|
mova m1, [srcq+2*lenq+ mmsize]
|
|
mova m2, [srcq+2*lenq+2*mmsize]
|
|
mova m3, [srcq+2*lenq+3*mmsize]
|
|
psrad m0, 16
|
|
psrad m1, 16
|
|
psrad m2, 16
|
|
psrad m3, 16
|
|
packssdw m0, m1
|
|
packssdw m2, m3
|
|
mova [dstq+lenq ], m0
|
|
mova [dstq+lenq+mmsize], m2
|
|
add lenq, mmsize*2
|
|
jl .loop
|
|
%if mmsize == 8
|
|
emms
|
|
RET
|
|
%else
|
|
REP_RET
|
|
%endif
|
|
%endmacro
|
|
|
|
INIT_MMX mmx
|
|
CONV_S32_TO_S16
|
|
INIT_XMM sse2
|
|
CONV_S32_TO_S16
|
|
|
|
;------------------------------------------------------------------------------
|
|
; void ff_conv_s32_to_flt(float *dst, const int32_t *src, int len);
|
|
;------------------------------------------------------------------------------
|
|
|
|
%macro CONV_S32_TO_FLT 0
|
|
cglobal conv_s32_to_flt, 3,3,3, dst, src, len
|
|
lea lenq, [4*lend]
|
|
add srcq, lenq
|
|
add dstq, lenq
|
|
neg lenq
|
|
mova m0, [pf_s32_inv_scale]
|
|
ALIGN 16
|
|
.loop:
|
|
cvtdq2ps m1, [srcq+lenq ]
|
|
cvtdq2ps m2, [srcq+lenq+mmsize]
|
|
mulps m1, m1, m0
|
|
mulps m2, m2, m0
|
|
mova [dstq+lenq ], m1
|
|
mova [dstq+lenq+mmsize], m2
|
|
add lenq, mmsize*2
|
|
jl .loop
|
|
%if mmsize == 32
|
|
vzeroupper
|
|
RET
|
|
%else
|
|
REP_RET
|
|
%endif
|
|
%endmacro
|
|
|
|
INIT_XMM sse2
|
|
CONV_S32_TO_FLT
|
|
%if HAVE_AVX
|
|
INIT_YMM avx
|
|
CONV_S32_TO_FLT
|
|
%endif
|
|
|
|
;------------------------------------------------------------------------------
|
|
; void ff_conv_flt_to_s16(int16_t *dst, const float *src, int len);
|
|
;------------------------------------------------------------------------------
|
|
|
|
INIT_XMM sse2
|
|
cglobal conv_flt_to_s16, 3,3,5, dst, src, len
|
|
lea lenq, [2*lend]
|
|
lea srcq, [srcq+2*lenq]
|
|
add dstq, lenq
|
|
neg lenq
|
|
mova m4, [pf_s16_scale]
|
|
.loop:
|
|
mova m0, [srcq+2*lenq ]
|
|
mova m1, [srcq+2*lenq+1*mmsize]
|
|
mova m2, [srcq+2*lenq+2*mmsize]
|
|
mova m3, [srcq+2*lenq+3*mmsize]
|
|
mulps m0, m4
|
|
mulps m1, m4
|
|
mulps m2, m4
|
|
mulps m3, m4
|
|
cvtps2dq m0, m0
|
|
cvtps2dq m1, m1
|
|
cvtps2dq m2, m2
|
|
cvtps2dq m3, m3
|
|
packssdw m0, m1
|
|
packssdw m2, m3
|
|
mova [dstq+lenq ], m0
|
|
mova [dstq+lenq+mmsize], m2
|
|
add lenq, mmsize*2
|
|
jl .loop
|
|
REP_RET
|
|
|
|
;------------------------------------------------------------------------------
|
|
; void ff_conv_flt_to_s32(int32_t *dst, const float *src, int len);
|
|
;------------------------------------------------------------------------------
|
|
|
|
%macro CONV_FLT_TO_S32 0
|
|
cglobal conv_flt_to_s32, 3,3,5, dst, src, len
|
|
lea lenq, [lend*4]
|
|
add srcq, lenq
|
|
add dstq, lenq
|
|
neg lenq
|
|
mova m4, [pf_s32_scale]
|
|
.loop:
|
|
mulps m0, m4, [srcq+lenq ]
|
|
mulps m1, m4, [srcq+lenq+1*mmsize]
|
|
mulps m2, m4, [srcq+lenq+2*mmsize]
|
|
mulps m3, m4, [srcq+lenq+3*mmsize]
|
|
cvtps2dq m0, m0
|
|
cvtps2dq m1, m1
|
|
cvtps2dq m2, m2
|
|
cvtps2dq m3, m3
|
|
mova [dstq+lenq ], m0
|
|
mova [dstq+lenq+1*mmsize], m1
|
|
mova [dstq+lenq+2*mmsize], m2
|
|
mova [dstq+lenq+3*mmsize], m3
|
|
add lenq, mmsize*4
|
|
jl .loop
|
|
%if mmsize == 32
|
|
vzeroupper
|
|
RET
|
|
%else
|
|
REP_RET
|
|
%endif
|
|
%endmacro
|
|
|
|
INIT_XMM sse2
|
|
CONV_FLT_TO_S32
|
|
%if HAVE_AVX
|
|
INIT_YMM avx
|
|
CONV_FLT_TO_S32
|
|
%endif
|
|
|
|
;-----------------------------------------------------------------------------
|
|
; void ff_conv_fltp_to_flt_6ch(float *dst, float *const *src, int len,
|
|
; int channels);
|
|
;-----------------------------------------------------------------------------
|
|
|
|
%macro CONV_FLTP_TO_FLT_6CH 0
|
|
cglobal conv_fltp_to_flt_6ch, 2,8,7, dst, src, src1, src2, src3, src4, src5, len
|
|
%if ARCH_X86_64
|
|
mov lend, r2d
|
|
%else
|
|
%define lend dword r2m
|
|
%endif
|
|
mov src1q, [srcq+1*gprsize]
|
|
mov src2q, [srcq+2*gprsize]
|
|
mov src3q, [srcq+3*gprsize]
|
|
mov src4q, [srcq+4*gprsize]
|
|
mov src5q, [srcq+5*gprsize]
|
|
mov srcq, [srcq]
|
|
sub src1q, srcq
|
|
sub src2q, srcq
|
|
sub src3q, srcq
|
|
sub src4q, srcq
|
|
sub src5q, srcq
|
|
.loop:
|
|
mova m0, [srcq ]
|
|
mova m1, [srcq+src1q]
|
|
mova m2, [srcq+src2q]
|
|
mova m3, [srcq+src3q]
|
|
mova m4, [srcq+src4q]
|
|
mova m5, [srcq+src5q]
|
|
%if cpuflag(sse4)
|
|
SBUTTERFLYPS 0, 1, 6
|
|
SBUTTERFLYPS 2, 3, 6
|
|
SBUTTERFLYPS 4, 5, 6
|
|
|
|
blendps m6, m4, m0, 1100b
|
|
movlhps m0, m2
|
|
movhlps m4, m2
|
|
blendps m2, m5, m1, 1100b
|
|
movlhps m1, m3
|
|
movhlps m5, m3
|
|
|
|
movaps [dstq ], m0
|
|
movaps [dstq+16], m6
|
|
movaps [dstq+32], m4
|
|
movaps [dstq+48], m1
|
|
movaps [dstq+64], m2
|
|
movaps [dstq+80], m5
|
|
%else ; mmx
|
|
SBUTTERFLY dq, 0, 1, 6
|
|
SBUTTERFLY dq, 2, 3, 6
|
|
SBUTTERFLY dq, 4, 5, 6
|
|
|
|
movq [dstq ], m0
|
|
movq [dstq+ 8], m2
|
|
movq [dstq+16], m4
|
|
movq [dstq+24], m1
|
|
movq [dstq+32], m3
|
|
movq [dstq+40], m5
|
|
%endif
|
|
add srcq, mmsize
|
|
add dstq, mmsize*6
|
|
sub lend, mmsize/4
|
|
jg .loop
|
|
%if mmsize == 8
|
|
emms
|
|
RET
|
|
%else
|
|
REP_RET
|
|
%endif
|
|
%endmacro
|
|
|
|
INIT_MMX mmx
|
|
CONV_FLTP_TO_FLT_6CH
|
|
INIT_XMM sse4
|
|
CONV_FLTP_TO_FLT_6CH
|
|
%if HAVE_AVX
|
|
INIT_XMM avx
|
|
CONV_FLTP_TO_FLT_6CH
|
|
%endif
|