c2140b8af1
Changes 'The VP8 project' to 'The WebM project', for consistency with other webmproject.org repositories. Fixes issue #97. Change-Id: I37c13ed5fbdb9d334ceef71c6350e9febed9bbba
598 lines
16 KiB
C
598 lines
16 KiB
C
/*
|
|
* Copyright (c) 2010 The WebM project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
|
|
#include "variance.h"
|
|
#include "pragmas.h"
|
|
#include "vpx_ports/mem.h"
|
|
|
|
extern void filter_block1d_h6_mmx
|
|
(
|
|
unsigned char *src_ptr,
|
|
unsigned short *output_ptr,
|
|
unsigned int src_pixels_per_line,
|
|
unsigned int pixel_step,
|
|
unsigned int output_height,
|
|
unsigned int output_width,
|
|
short *vp7_filter
|
|
);
|
|
extern void filter_block1d_v6_mmx
|
|
(
|
|
short *src_ptr,
|
|
unsigned char *output_ptr,
|
|
unsigned int pixels_per_line,
|
|
unsigned int pixel_step,
|
|
unsigned int output_height,
|
|
unsigned int output_width,
|
|
short *vp7_filter
|
|
);
|
|
|
|
extern unsigned int vp8_get_mb_ss_mmx(short *src_ptr);
|
|
extern unsigned int vp8_get8x8var_mmx
|
|
(
|
|
unsigned char *src_ptr,
|
|
int source_stride,
|
|
unsigned char *ref_ptr,
|
|
int recon_stride,
|
|
unsigned int *SSE,
|
|
int *Sum
|
|
);
|
|
extern unsigned int vp8_get4x4var_mmx
|
|
(
|
|
unsigned char *src_ptr,
|
|
int source_stride,
|
|
unsigned char *ref_ptr,
|
|
int recon_stride,
|
|
unsigned int *SSE,
|
|
int *Sum
|
|
);
|
|
extern unsigned int vp8_get4x4sse_cs_mmx
|
|
(
|
|
unsigned char *src_ptr,
|
|
int source_stride,
|
|
unsigned char *ref_ptr,
|
|
int recon_stride
|
|
);
|
|
extern void vp8_filter_block2d_bil4x4_var_mmx
|
|
(
|
|
unsigned char *ref_ptr,
|
|
int ref_pixels_per_line,
|
|
unsigned char *src_ptr,
|
|
int src_pixels_per_line,
|
|
const short *HFilter,
|
|
const short *VFilter,
|
|
int *sum,
|
|
unsigned int *sumsquared
|
|
);
|
|
extern void vp8_filter_block2d_bil_var_mmx
|
|
(
|
|
unsigned char *ref_ptr,
|
|
int ref_pixels_per_line,
|
|
unsigned char *src_ptr,
|
|
int src_pixels_per_line,
|
|
unsigned int Height,
|
|
const short *HFilter,
|
|
const short *VFilter,
|
|
int *sum,
|
|
unsigned int *sumsquared
|
|
);
|
|
extern unsigned int vp8_get16x16pred_error_mmx
|
|
(
|
|
unsigned char *src_ptr,
|
|
int src_stride,
|
|
unsigned char *ref_ptr,
|
|
int ref_stride
|
|
);
|
|
|
|
|
|
void vp8_test_get_mb_ss(void)
|
|
{
|
|
short zz[] =
|
|
{
|
|
-4, -4, -4, -4, 4, 4, 4, 4, -4, -4, -4, -4, 4, 4, 4, 4,
|
|
-2, -2, -2, -2, 2, 2, 2, 2, -2, -2, -2, -2, 2, 2, 2, 2,
|
|
-4, -4, -4, -4, 4, 4, 4, 4, -4, -4, -4, -4, 4, 4, 4, 4,
|
|
-4, -4, -4, -4, 4, 4, 4, 4, -4, -4, -4, -4, 4, 4, 4, 4,
|
|
-4, -4, -4, -4, 4, 4, 4, 4, -4, -4, -4, -4, 4, 4, 4, 4,
|
|
-4, -4, -4, -4, 4, 4, 4, 4, -4, -4, -4, -4, 4, 4, 4, 4,
|
|
-4, -4, -4, -4, 4, 4, 4, 4, -4, -4, -4, -4, 4, 4, 4, 4,
|
|
-4, -4, -4, -4, 4, 4, 4, 4, -4, -4, -4, -4, 4, 4, 4, 4,
|
|
-4, -4, -4, -4, 4, 4, 4, 4, -4, -4, -4, -4, 4, 4, 4, 4,
|
|
-4, -4, -4, -4, 4, 4, 4, 4, -4, -4, -4, -4, 4, 4, 4, 4,
|
|
-4, -4, -4, -4, 4, 4, 4, 4, -4, -4, -4, -4, 4, 4, 4, 4,
|
|
-4, -4, -4, -4, 4, 4, 4, 4, -4, -4, -4, -4, 4, 4, 4, 4,
|
|
-4, -4, -4, -4, 4, 4, 4, 4, -4, -4, -4, -4, 4, 4, 4, 4,
|
|
-3, -3, -3, -3, 3, 3, 3, 3, -3, -3, -3, -3, 3, 3, 3, 3,
|
|
-4, -4, -4, -4, 4, 4, 4, 4, -4, -4, -4, -4, 4, 4, 4, 4,
|
|
-4, -4, -4, -4, 4, 4, 4, 4, -4, -4, -4, -4, 4, 4, 4, 4,
|
|
};
|
|
int s = 0, x = vp8_get_mb_ss_mmx(zz);
|
|
{
|
|
int y;
|
|
|
|
for (y = 0; y < 256; y++)
|
|
s += (zz[y] * zz[y]);
|
|
}
|
|
|
|
x += 0;
|
|
}
|
|
|
|
|
|
unsigned int vp8_get16x16var_mmx(
|
|
unsigned char *src_ptr,
|
|
int source_stride,
|
|
unsigned char *ref_ptr,
|
|
int recon_stride,
|
|
unsigned *SSE,
|
|
unsigned *SUM
|
|
)
|
|
{
|
|
unsigned int sse0, sse1, sse2, sse3, var;
|
|
int sum0, sum1, sum2, sum3, avg;
|
|
|
|
|
|
vp8_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0) ;
|
|
vp8_get8x8var_mmx(src_ptr + 8, source_stride, ref_ptr + 8, recon_stride, &sse1, &sum1);
|
|
vp8_get8x8var_mmx(src_ptr + 8 * source_stride, source_stride, ref_ptr + 8 * recon_stride, recon_stride, &sse2, &sum2) ;
|
|
vp8_get8x8var_mmx(src_ptr + 8 * source_stride + 8, source_stride, ref_ptr + 8 * recon_stride + 8, recon_stride, &sse3, &sum3);
|
|
|
|
var = sse0 + sse1 + sse2 + sse3;
|
|
avg = sum0 + sum1 + sum2 + sum3;
|
|
|
|
*SSE = var;
|
|
*SUM = avg;
|
|
return (var - ((avg * avg) >> 8));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
unsigned int vp8_variance4x4_mmx(
|
|
unsigned char *src_ptr,
|
|
int source_stride,
|
|
unsigned char *ref_ptr,
|
|
int recon_stride,
|
|
unsigned int *sse)
|
|
{
|
|
unsigned int var;
|
|
int avg;
|
|
|
|
vp8_get4x4var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg) ;
|
|
*sse = var;
|
|
return (var - ((avg * avg) >> 4));
|
|
|
|
}
|
|
|
|
unsigned int vp8_variance8x8_mmx(
|
|
unsigned char *src_ptr,
|
|
int source_stride,
|
|
unsigned char *ref_ptr,
|
|
int recon_stride,
|
|
unsigned int *sse)
|
|
{
|
|
unsigned int var;
|
|
int avg;
|
|
|
|
vp8_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg) ;
|
|
*sse = var;
|
|
|
|
return (var - ((avg * avg) >> 6));
|
|
|
|
}
|
|
|
|
unsigned int vp8_mse16x16_mmx(
|
|
unsigned char *src_ptr,
|
|
int source_stride,
|
|
unsigned char *ref_ptr,
|
|
int recon_stride,
|
|
unsigned int *sse)
|
|
{
|
|
unsigned int sse0, sse1, sse2, sse3, var;
|
|
int sum0, sum1, sum2, sum3;
|
|
|
|
|
|
vp8_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0) ;
|
|
vp8_get8x8var_mmx(src_ptr + 8, source_stride, ref_ptr + 8, recon_stride, &sse1, &sum1);
|
|
vp8_get8x8var_mmx(src_ptr + 8 * source_stride, source_stride, ref_ptr + 8 * recon_stride, recon_stride, &sse2, &sum2) ;
|
|
vp8_get8x8var_mmx(src_ptr + 8 * source_stride + 8, source_stride, ref_ptr + 8 * recon_stride + 8, recon_stride, &sse3, &sum3);
|
|
|
|
var = sse0 + sse1 + sse2 + sse3;
|
|
*sse = var;
|
|
return var;
|
|
}
|
|
|
|
|
|
unsigned int vp8_variance16x16_mmx(
|
|
unsigned char *src_ptr,
|
|
int source_stride,
|
|
unsigned char *ref_ptr,
|
|
int recon_stride,
|
|
int *sse)
|
|
{
|
|
unsigned int sse0, sse1, sse2, sse3, var;
|
|
int sum0, sum1, sum2, sum3, avg;
|
|
|
|
|
|
vp8_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0) ;
|
|
vp8_get8x8var_mmx(src_ptr + 8, source_stride, ref_ptr + 8, recon_stride, &sse1, &sum1);
|
|
vp8_get8x8var_mmx(src_ptr + 8 * source_stride, source_stride, ref_ptr + 8 * recon_stride, recon_stride, &sse2, &sum2) ;
|
|
vp8_get8x8var_mmx(src_ptr + 8 * source_stride + 8, source_stride, ref_ptr + 8 * recon_stride + 8, recon_stride, &sse3, &sum3);
|
|
|
|
var = sse0 + sse1 + sse2 + sse3;
|
|
avg = sum0 + sum1 + sum2 + sum3;
|
|
*sse = var;
|
|
return (var - ((avg * avg) >> 8));
|
|
}
|
|
|
|
unsigned int vp8_variance16x8_mmx(
|
|
unsigned char *src_ptr,
|
|
int source_stride,
|
|
unsigned char *ref_ptr,
|
|
int recon_stride,
|
|
unsigned int *sse)
|
|
{
|
|
unsigned int sse0, sse1, var;
|
|
int sum0, sum1, avg;
|
|
|
|
vp8_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0) ;
|
|
vp8_get8x8var_mmx(src_ptr + 8, source_stride, ref_ptr + 8, recon_stride, &sse1, &sum1);
|
|
|
|
var = sse0 + sse1;
|
|
avg = sum0 + sum1;
|
|
*sse = var;
|
|
return (var - ((avg * avg) >> 7));
|
|
|
|
}
|
|
|
|
|
|
unsigned int vp8_variance8x16_mmx(
|
|
unsigned char *src_ptr,
|
|
int source_stride,
|
|
unsigned char *ref_ptr,
|
|
int recon_stride,
|
|
unsigned int *sse)
|
|
{
|
|
unsigned int sse0, sse1, var;
|
|
int sum0, sum1, avg;
|
|
|
|
vp8_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0) ;
|
|
vp8_get8x8var_mmx(src_ptr + 8 * source_stride, source_stride, ref_ptr + 8 * recon_stride, recon_stride, &sse1, &sum1) ;
|
|
|
|
var = sse0 + sse1;
|
|
avg = sum0 + sum1;
|
|
*sse = var;
|
|
|
|
return (var - ((avg * avg) >> 7));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
///////////////////////////////////////////////////////////////////////////
|
|
// the mmx function that does the bilinear filtering and var calculation //
|
|
// int one pass //
|
|
///////////////////////////////////////////////////////////////////////////
|
|
DECLARE_ALIGNED(16, const short, vp8_vp7_bilinear_filters_mmx[8][8]) =
|
|
{
|
|
{ 128, 128, 128, 128, 0, 0, 0, 0 },
|
|
{ 112, 112, 112, 112, 16, 16, 16, 16 },
|
|
{ 96, 96, 96, 96, 32, 32, 32, 32 },
|
|
{ 80, 80, 80, 80, 48, 48, 48, 48 },
|
|
{ 64, 64, 64, 64, 64, 64, 64, 64 },
|
|
{ 48, 48, 48, 48, 80, 80, 80, 80 },
|
|
{ 32, 32, 32, 32, 96, 96, 96, 96 },
|
|
{ 16, 16, 16, 16, 112, 112, 112, 112 }
|
|
};
|
|
|
|
unsigned int vp8_sub_pixel_variance4x4_mmx
|
|
(
|
|
unsigned char *src_ptr,
|
|
int src_pixels_per_line,
|
|
int xoffset,
|
|
int yoffset,
|
|
unsigned char *dst_ptr,
|
|
int dst_pixels_per_line,
|
|
unsigned int *sse)
|
|
|
|
{
|
|
int xsum;
|
|
unsigned int xxsum;
|
|
vp8_filter_block2d_bil4x4_var_mmx(
|
|
src_ptr, src_pixels_per_line,
|
|
dst_ptr, dst_pixels_per_line,
|
|
vp8_vp7_bilinear_filters_mmx[xoffset], vp8_vp7_bilinear_filters_mmx[yoffset],
|
|
&xsum, &xxsum
|
|
);
|
|
*sse = xxsum;
|
|
return (xxsum - ((xsum * xsum) >> 4));
|
|
}
|
|
|
|
|
|
unsigned int vp8_sub_pixel_variance8x8_mmx
|
|
(
|
|
unsigned char *src_ptr,
|
|
int src_pixels_per_line,
|
|
int xoffset,
|
|
int yoffset,
|
|
unsigned char *dst_ptr,
|
|
int dst_pixels_per_line,
|
|
unsigned int *sse
|
|
)
|
|
{
|
|
|
|
int xsum;
|
|
unsigned int xxsum;
|
|
vp8_filter_block2d_bil_var_mmx(
|
|
src_ptr, src_pixels_per_line,
|
|
dst_ptr, dst_pixels_per_line, 8,
|
|
vp8_vp7_bilinear_filters_mmx[xoffset], vp8_vp7_bilinear_filters_mmx[yoffset],
|
|
&xsum, &xxsum
|
|
);
|
|
*sse = xxsum;
|
|
return (xxsum - ((xsum * xsum) >> 6));
|
|
}
|
|
|
|
unsigned int vp8_sub_pixel_variance16x16_mmx
|
|
(
|
|
unsigned char *src_ptr,
|
|
int src_pixels_per_line,
|
|
int xoffset,
|
|
int yoffset,
|
|
unsigned char *dst_ptr,
|
|
int dst_pixels_per_line,
|
|
unsigned int *sse
|
|
)
|
|
{
|
|
|
|
int xsum0, xsum1;
|
|
unsigned int xxsum0, xxsum1;
|
|
|
|
|
|
vp8_filter_block2d_bil_var_mmx(
|
|
src_ptr, src_pixels_per_line,
|
|
dst_ptr, dst_pixels_per_line, 16,
|
|
vp8_vp7_bilinear_filters_mmx[xoffset], vp8_vp7_bilinear_filters_mmx[yoffset],
|
|
&xsum0, &xxsum0
|
|
);
|
|
|
|
|
|
vp8_filter_block2d_bil_var_mmx(
|
|
src_ptr + 8, src_pixels_per_line,
|
|
dst_ptr + 8, dst_pixels_per_line, 16,
|
|
vp8_vp7_bilinear_filters_mmx[xoffset], vp8_vp7_bilinear_filters_mmx[yoffset],
|
|
&xsum1, &xxsum1
|
|
);
|
|
|
|
xsum0 += xsum1;
|
|
xxsum0 += xxsum1;
|
|
|
|
*sse = xxsum0;
|
|
return (xxsum0 - ((xsum0 * xsum0) >> 8));
|
|
|
|
|
|
}
|
|
|
|
unsigned int vp8_sub_pixel_mse16x16_mmx(
|
|
unsigned char *src_ptr,
|
|
int src_pixels_per_line,
|
|
int xoffset,
|
|
int yoffset,
|
|
unsigned char *dst_ptr,
|
|
int dst_pixels_per_line,
|
|
unsigned int *sse
|
|
)
|
|
{
|
|
vp8_sub_pixel_variance16x16_mmx(src_ptr, src_pixels_per_line, xoffset, yoffset, dst_ptr, dst_pixels_per_line, sse);
|
|
return *sse;
|
|
}
|
|
|
|
unsigned int vp8_sub_pixel_variance16x8_mmx
|
|
(
|
|
unsigned char *src_ptr,
|
|
int src_pixels_per_line,
|
|
int xoffset,
|
|
int yoffset,
|
|
unsigned char *dst_ptr,
|
|
int dst_pixels_per_line,
|
|
unsigned int *sse
|
|
)
|
|
{
|
|
int xsum0, xsum1;
|
|
unsigned int xxsum0, xxsum1;
|
|
|
|
|
|
vp8_filter_block2d_bil_var_mmx(
|
|
src_ptr, src_pixels_per_line,
|
|
dst_ptr, dst_pixels_per_line, 8,
|
|
vp8_vp7_bilinear_filters_mmx[xoffset], vp8_vp7_bilinear_filters_mmx[yoffset],
|
|
&xsum0, &xxsum0
|
|
);
|
|
|
|
|
|
vp8_filter_block2d_bil_var_mmx(
|
|
src_ptr + 8, src_pixels_per_line,
|
|
dst_ptr + 8, dst_pixels_per_line, 8,
|
|
vp8_vp7_bilinear_filters_mmx[xoffset], vp8_vp7_bilinear_filters_mmx[yoffset],
|
|
&xsum1, &xxsum1
|
|
);
|
|
|
|
xsum0 += xsum1;
|
|
xxsum0 += xxsum1;
|
|
|
|
*sse = xxsum0;
|
|
return (xxsum0 - ((xsum0 * xsum0) >> 7));
|
|
}
|
|
|
|
unsigned int vp8_sub_pixel_variance8x16_mmx
|
|
(
|
|
unsigned char *src_ptr,
|
|
int src_pixels_per_line,
|
|
int xoffset,
|
|
int yoffset,
|
|
unsigned char *dst_ptr,
|
|
int dst_pixels_per_line,
|
|
int *sse
|
|
)
|
|
{
|
|
int xsum;
|
|
unsigned int xxsum;
|
|
vp8_filter_block2d_bil_var_mmx(
|
|
src_ptr, src_pixels_per_line,
|
|
dst_ptr, dst_pixels_per_line, 16,
|
|
vp8_vp7_bilinear_filters_mmx[xoffset], vp8_vp7_bilinear_filters_mmx[yoffset],
|
|
&xsum, &xxsum
|
|
);
|
|
*sse = xxsum;
|
|
return (xxsum - ((xsum * xsum) >> 7));
|
|
}
|
|
|
|
unsigned int vp8_i_variance16x16_mmx(
|
|
unsigned char *src_ptr,
|
|
int source_stride,
|
|
unsigned char *ref_ptr,
|
|
int recon_stride,
|
|
unsigned int *sse)
|
|
{
|
|
unsigned int sse0, sse1, sse2, sse3, var;
|
|
int sum0, sum1, sum2, sum3, avg;
|
|
|
|
|
|
vp8_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0) ;
|
|
vp8_get8x8var_mmx(src_ptr + 8, source_stride, ref_ptr + 8, recon_stride, &sse1, &sum1);
|
|
vp8_get8x8var_mmx(src_ptr + (source_stride >> 1), source_stride, ref_ptr + (recon_stride >> 1), recon_stride, &sse2, &sum2) ;
|
|
vp8_get8x8var_mmx(src_ptr + (source_stride >> 1) + 8, source_stride, ref_ptr + (recon_stride >> 1) + 8, recon_stride, &sse3, &sum3);
|
|
|
|
var = sse0 + sse1 + sse2 + sse3;
|
|
avg = sum0 + sum1 + sum2 + sum3;
|
|
*sse = var;
|
|
return (var - ((avg * avg) >> 8));
|
|
|
|
}
|
|
|
|
unsigned int vp8_i_variance8x16_mmx(
|
|
unsigned char *src_ptr,
|
|
int source_stride,
|
|
unsigned char *ref_ptr,
|
|
int recon_stride,
|
|
unsigned int *sse)
|
|
{
|
|
unsigned int sse0, sse1, var;
|
|
int sum0, sum1, avg;
|
|
vp8_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0) ;
|
|
vp8_get8x8var_mmx(src_ptr + (source_stride >> 1), source_stride, ref_ptr + (recon_stride >> 1), recon_stride, &sse1, &sum1) ;
|
|
|
|
var = sse0 + sse1;
|
|
avg = sum0 + sum1;
|
|
|
|
*sse = var;
|
|
return (var - ((avg * avg) >> 7));
|
|
|
|
}
|
|
|
|
unsigned int vp8_i_sub_pixel_variance16x16_mmx
|
|
(
|
|
unsigned char *src_ptr,
|
|
int src_pixels_per_line,
|
|
int xoffset,
|
|
int yoffset,
|
|
unsigned char *dst_ptr,
|
|
int dst_pixels_per_line,
|
|
unsigned int *sse
|
|
)
|
|
{
|
|
int xsum0, xsum1;
|
|
unsigned int xxsum0, xxsum1;
|
|
int f2soffset = (src_pixels_per_line >> 1);
|
|
int f2doffset = (dst_pixels_per_line >> 1);
|
|
|
|
|
|
vp8_filter_block2d_bil_var_mmx(
|
|
src_ptr, src_pixels_per_line,
|
|
dst_ptr, dst_pixels_per_line, 8,
|
|
vp8_vp7_bilinear_filters_mmx[xoffset], vp8_vp7_bilinear_filters_mmx[yoffset],
|
|
&xsum0, &xxsum0
|
|
);
|
|
|
|
|
|
vp8_filter_block2d_bil_var_mmx(
|
|
src_ptr + 8, src_pixels_per_line,
|
|
dst_ptr + 8, dst_pixels_per_line, 8,
|
|
vp8_vp7_bilinear_filters_mmx[xoffset], vp8_vp7_bilinear_filters_mmx[yoffset],
|
|
&xsum1, &xxsum1
|
|
);
|
|
|
|
xsum0 += xsum1;
|
|
xxsum0 += xxsum1;
|
|
|
|
vp8_filter_block2d_bil_var_mmx(
|
|
src_ptr + f2soffset, src_pixels_per_line,
|
|
dst_ptr + f2doffset, dst_pixels_per_line, 8,
|
|
vp8_vp7_bilinear_filters_mmx[xoffset], vp8_vp7_bilinear_filters_mmx[yoffset],
|
|
&xsum1, &xxsum1
|
|
);
|
|
|
|
xsum0 += xsum1;
|
|
xxsum0 += xxsum1;
|
|
|
|
vp8_filter_block2d_bil_var_mmx(
|
|
src_ptr + f2soffset + 8, src_pixels_per_line,
|
|
dst_ptr + f2doffset + 8, dst_pixels_per_line, 8,
|
|
vp8_vp7_bilinear_filters_mmx[xoffset], vp8_vp7_bilinear_filters_mmx[yoffset],
|
|
&xsum1, &xxsum1
|
|
);
|
|
|
|
xsum0 += xsum1;
|
|
xxsum0 += xxsum1;
|
|
*sse = xxsum0;
|
|
return (xxsum0 - ((xsum0 * xsum0) >> 8));
|
|
}
|
|
|
|
|
|
unsigned int vp8_i_sub_pixel_variance8x16_mmx
|
|
(
|
|
unsigned char *src_ptr,
|
|
int src_pixels_per_line,
|
|
int xoffset,
|
|
int yoffset,
|
|
unsigned char *dst_ptr,
|
|
int dst_pixels_per_line,
|
|
unsigned int *sse
|
|
)
|
|
{
|
|
int xsum0, xsum1;
|
|
unsigned int xxsum0, xxsum1;
|
|
int f2soffset = (src_pixels_per_line >> 1);
|
|
int f2doffset = (dst_pixels_per_line >> 1);
|
|
|
|
|
|
vp8_filter_block2d_bil_var_mmx(
|
|
src_ptr, src_pixels_per_line,
|
|
dst_ptr, dst_pixels_per_line, 8,
|
|
vp8_vp7_bilinear_filters_mmx[xoffset], vp8_vp7_bilinear_filters_mmx[yoffset],
|
|
&xsum0, &xxsum0
|
|
);
|
|
|
|
|
|
vp8_filter_block2d_bil_var_mmx(
|
|
src_ptr + f2soffset, src_pixels_per_line,
|
|
dst_ptr + f2doffset, dst_pixels_per_line, 8,
|
|
vp8_vp7_bilinear_filters_mmx[xoffset], vp8_vp7_bilinear_filters_mmx[yoffset],
|
|
&xsum1, &xxsum1
|
|
);
|
|
|
|
xsum0 += xsum1;
|
|
xxsum0 += xxsum1;
|
|
*sse = xxsum0;
|
|
return (xxsum0 - ((xsum0 * xsum0) >> 7));
|
|
}
|