Refactored Tegra related macro usage
This commit is contained in:
parent
48ea65e69c
commit
bdfd4c3062
@ -1943,6 +1943,52 @@ typedef void (*ReduceFunc)( const Mat& src, Mat& dst );
|
|||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#define reduceSumR8u32s reduceR_<uchar, int, OpAdd<int> >
|
||||||
|
#define reduceSumR8u32f reduceR_<uchar, float, OpAdd<int> >
|
||||||
|
#define reduceSumR8u64f reduceR_<uchar, double,OpAdd<int> >
|
||||||
|
#define reduceSumR16u32f reduceR_<ushort,float, OpAdd<float> >
|
||||||
|
#define reduceSumR16u64f reduceR_<ushort,double,OpAdd<double> >
|
||||||
|
#define reduceSumR16s32f reduceR_<short, float, OpAdd<float> >
|
||||||
|
#define reduceSumR16s64f reduceR_<short, double,OpAdd<double> >
|
||||||
|
#define reduceSumR32f32f reduceR_<float, float, OpAdd<float> >
|
||||||
|
#define reduceSumR32f64f reduceR_<float, double,OpAdd<double> >
|
||||||
|
#define reduceSumR64f64f reduceR_<double,double,OpAdd<double> >
|
||||||
|
|
||||||
|
#define reduceMaxR8u reduceR_<uchar, uchar, OpMax<uchar> >
|
||||||
|
#define reduceMaxR16u reduceR_<ushort,ushort,OpMax<ushort> >
|
||||||
|
#define reduceMaxR16s reduceR_<short, short, OpMax<short> >
|
||||||
|
#define reduceMaxR32f reduceR_<float, float, OpMax<float> >
|
||||||
|
#define reduceMaxR64f reduceR_<double,double,OpMax<double> >
|
||||||
|
|
||||||
|
#define reduceMinR8u reduceR_<uchar, uchar, OpMin<uchar> >
|
||||||
|
#define reduceMinR16u reduceR_<ushort,ushort,OpMin<ushort> >
|
||||||
|
#define reduceMinR16s reduceR_<short, short, OpMin<short> >
|
||||||
|
#define reduceMinR32f reduceR_<float, float, OpMin<float> >
|
||||||
|
#define reduceMinR64f reduceR_<double,double,OpMin<double> >
|
||||||
|
|
||||||
|
#define reduceSumC8u32s reduceC_<uchar, int, OpAdd<int> >
|
||||||
|
#define reduceSumC8u32f reduceC_<uchar, float, OpAdd<int> >
|
||||||
|
#define reduceSumC8u64f reduceC_<uchar, double,OpAdd<int> >
|
||||||
|
#define reduceSumC16u32f reduceC_<ushort,float, OpAdd<float> >
|
||||||
|
#define reduceSumC16u64f reduceC_<ushort,double,OpAdd<double> >
|
||||||
|
#define reduceSumC16s32f reduceC_<short, float, OpAdd<float> >
|
||||||
|
#define reduceSumC16s64f reduceC_<short, double,OpAdd<double> >
|
||||||
|
#define reduceSumC32f32f reduceC_<float, float, OpAdd<float> >
|
||||||
|
#define reduceSumC32f64f reduceC_<float, double,OpAdd<double> >
|
||||||
|
#define reduceSumC64f64f reduceC_<double,double,OpAdd<double> >
|
||||||
|
|
||||||
|
#define reduceMaxC8u reduceC_<uchar, uchar, OpMax<uchar> >
|
||||||
|
#define reduceMaxC16u reduceC_<ushort,ushort,OpMax<ushort> >
|
||||||
|
#define reduceMaxC16s reduceC_<short, short, OpMax<short> >
|
||||||
|
#define reduceMaxC32f reduceC_<float, float, OpMax<float> >
|
||||||
|
#define reduceMaxC64f reduceC_<double,double,OpMax<double> >
|
||||||
|
|
||||||
|
#define reduceMinC8u reduceC_<uchar, uchar, OpMin<uchar> >
|
||||||
|
#define reduceMinC16u reduceC_<ushort,ushort,OpMin<ushort> >
|
||||||
|
#define reduceMinC16s reduceC_<short, short, OpMin<short> >
|
||||||
|
#define reduceMinC32f reduceC_<float, float, OpMin<float> >
|
||||||
|
#define reduceMinC64f reduceC_<double,double,OpMin<double> >
|
||||||
|
|
||||||
void cv::reduce(InputArray _src, OutputArray _dst, int dim, int op, int dtype)
|
void cv::reduce(InputArray _src, OutputArray _dst, int dim, int op, int dtype)
|
||||||
{
|
{
|
||||||
Mat src = _src.getMat();
|
Mat src = _src.getMat();
|
||||||
@ -1977,75 +2023,51 @@ void cv::reduce(InputArray _src, OutputArray _dst, int dim, int op, int dtype)
|
|||||||
if( op == CV_REDUCE_SUM )
|
if( op == CV_REDUCE_SUM )
|
||||||
{
|
{
|
||||||
if(sdepth == CV_8U && ddepth == CV_32S)
|
if(sdepth == CV_8U && ddepth == CV_32S)
|
||||||
#ifdef HAVE_TEGRA_OPTIMIZATION
|
func = GET_OPTIMIZED(reduceSumR8u32s);
|
||||||
func = tegra::getTegraOptimized_reduceR8uAdd(tegra::reduceR8uAdd);
|
|
||||||
#else
|
|
||||||
func = reduceR_<uchar,int,OpAdd<int> >;
|
|
||||||
#endif
|
|
||||||
else if(sdepth == CV_8U && ddepth == CV_32F)
|
else if(sdepth == CV_8U && ddepth == CV_32F)
|
||||||
func = reduceR_<uchar,float,OpAdd<int> >;
|
func = reduceSumR8u32f;
|
||||||
else if(sdepth == CV_8U && ddepth == CV_64F)
|
else if(sdepth == CV_8U && ddepth == CV_64F)
|
||||||
func = reduceR_<uchar,double,OpAdd<int> >;
|
func = reduceSumR8u64f;
|
||||||
else if(sdepth == CV_16U && ddepth == CV_32F)
|
else if(sdepth == CV_16U && ddepth == CV_32F)
|
||||||
func = reduceR_<ushort,float,OpAdd<float> >;
|
func = reduceSumR16u32f;
|
||||||
else if(sdepth == CV_16U && ddepth == CV_64F)
|
else if(sdepth == CV_16U && ddepth == CV_64F)
|
||||||
func = reduceR_<ushort,double,OpAdd<double> >;
|
func = reduceSumR16u64f;
|
||||||
else if(sdepth == CV_16S && ddepth == CV_32F)
|
else if(sdepth == CV_16S && ddepth == CV_32F)
|
||||||
func = reduceR_<short,float,OpAdd<float> >;
|
func = reduceSumR16s32f;
|
||||||
else if(sdepth == CV_16S && ddepth == CV_64F)
|
else if(sdepth == CV_16S && ddepth == CV_64F)
|
||||||
func = reduceR_<short,double,OpAdd<double> >;
|
func = reduceSumR16s64f;
|
||||||
else if(sdepth == CV_32F && ddepth == CV_32F)
|
else if(sdepth == CV_32F && ddepth == CV_32F)
|
||||||
#ifdef HAVE_TEGRA_OPTIMIZATION
|
func = GET_OPTIMIZED(reduceSumR32f32f);
|
||||||
func = tegra::getTegraOptimized_reduceR32fAdd(tegra::reduceR32fAdd);
|
|
||||||
#else
|
|
||||||
func = reduceR_<float,float,OpAdd<float> >;
|
|
||||||
#endif
|
|
||||||
else if(sdepth == CV_32F && ddepth == CV_64F)
|
else if(sdepth == CV_32F && ddepth == CV_64F)
|
||||||
func = reduceR_<float,double,OpAdd<double> >;
|
func = reduceSumR32f64f;
|
||||||
else if(sdepth == CV_64F && ddepth == CV_64F)
|
else if(sdepth == CV_64F && ddepth == CV_64F)
|
||||||
func = reduceR_<double,double,OpAdd<double> >;
|
func = reduceSumR64f64f;
|
||||||
}
|
}
|
||||||
else if(op == CV_REDUCE_MAX)
|
else if(op == CV_REDUCE_MAX)
|
||||||
{
|
{
|
||||||
if(sdepth == CV_8U && ddepth == CV_8U)
|
if(sdepth == CV_8U && ddepth == CV_8U)
|
||||||
#ifdef HAVE_TEGRA_OPTIMIZATION
|
func = GET_OPTIMIZED(reduceMaxR8u);
|
||||||
func = tegra::getTegraOptimized_reduceR8uMax(tegra::reduceR8uMax);
|
|
||||||
#else
|
|
||||||
func = reduceR_<uchar, uchar, OpMax<uchar> >;
|
|
||||||
#endif
|
|
||||||
else if(sdepth == CV_16U && ddepth == CV_16U)
|
else if(sdepth == CV_16U && ddepth == CV_16U)
|
||||||
func = reduceR_<ushort, ushort, OpMax<ushort> >;
|
func = reduceMaxR16u;
|
||||||
else if(sdepth == CV_16S && ddepth == CV_16S)
|
else if(sdepth == CV_16S && ddepth == CV_16S)
|
||||||
func = reduceR_<short, short, OpMax<short> >;
|
func = reduceMaxR16s;
|
||||||
else if(sdepth == CV_32F && ddepth == CV_32F)
|
else if(sdepth == CV_32F && ddepth == CV_32F)
|
||||||
#ifdef HAVE_TEGRA_OPTIMIZATION
|
func = GET_OPTIMIZED(reduceMaxR32f);
|
||||||
func = tegra::getTegraOptimized_reduceR32fMax(tegra::reduceR32fMax);
|
|
||||||
#else
|
|
||||||
func = reduceR_<float, float, OpMax<float> >;
|
|
||||||
#endif
|
|
||||||
else if(sdepth == CV_64F && ddepth == CV_64F)
|
else if(sdepth == CV_64F && ddepth == CV_64F)
|
||||||
func = reduceR_<double, double, OpMax<double> >;
|
func = reduceMaxR64f;
|
||||||
}
|
}
|
||||||
else if(op == CV_REDUCE_MIN)
|
else if(op == CV_REDUCE_MIN)
|
||||||
{
|
{
|
||||||
if(sdepth == CV_8U && ddepth == CV_8U)
|
if(sdepth == CV_8U && ddepth == CV_8U)
|
||||||
#ifdef HAVE_TEGRA_OPTIMIZATION
|
func = GET_OPTIMIZED(reduceMinR8u);
|
||||||
func = tegra::getTegraOptimized_reduceR8uMin(tegra::reduceR8uMin);
|
|
||||||
#else
|
|
||||||
func = reduceR_<uchar, uchar, OpMin<uchar> >;
|
|
||||||
#endif
|
|
||||||
else if(sdepth == CV_16U && ddepth == CV_16U)
|
else if(sdepth == CV_16U && ddepth == CV_16U)
|
||||||
func = reduceR_<ushort, ushort, OpMin<ushort> >;
|
func = reduceMinR16u;
|
||||||
else if(sdepth == CV_16S && ddepth == CV_16S)
|
else if(sdepth == CV_16S && ddepth == CV_16S)
|
||||||
func = reduceR_<short, short, OpMin<short> >;
|
func = reduceMinR16s;
|
||||||
else if(sdepth == CV_32F && ddepth == CV_32F)
|
else if(sdepth == CV_32F && ddepth == CV_32F)
|
||||||
#ifdef HAVE_TEGRA_OPTIMIZATION
|
func = GET_OPTIMIZED(reduceMinR32f);
|
||||||
func = tegra::getTegraOptimized_reduceR32fMin(tegra::reduceR32fMin);
|
|
||||||
#else
|
|
||||||
func = reduceR_<float, float, OpMin<float> >;
|
|
||||||
#endif
|
|
||||||
else if(sdepth == CV_64F && ddepth == CV_64F)
|
else if(sdepth == CV_64F && ddepth == CV_64F)
|
||||||
func = reduceR_<double, double, OpMin<double> >;
|
func = reduceMinR64f;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
else
|
else
|
||||||
@ -2053,55 +2075,51 @@ void cv::reduce(InputArray _src, OutputArray _dst, int dim, int op, int dtype)
|
|||||||
if(op == CV_REDUCE_SUM)
|
if(op == CV_REDUCE_SUM)
|
||||||
{
|
{
|
||||||
if(sdepth == CV_8U && ddepth == CV_32S)
|
if(sdepth == CV_8U && ddepth == CV_32S)
|
||||||
#ifdef HAVE_TEGRA_OPTIMIZATION
|
func = GET_OPTIMIZED(reduceSumC8u32s);
|
||||||
func = tegra::getTegraOptimized_reduceC8uAdd(tegra::reduceC8uAdd);
|
|
||||||
#else
|
|
||||||
func = reduceC_<uchar,int,OpAdd<int> >;
|
|
||||||
#endif
|
|
||||||
else if(sdepth == CV_8U && ddepth == CV_32F)
|
else if(sdepth == CV_8U && ddepth == CV_32F)
|
||||||
func = reduceC_<uchar,float,OpAdd<int> >;
|
func = reduceSumC8u32f;
|
||||||
else if(sdepth == CV_8U && ddepth == CV_64F)
|
else if(sdepth == CV_8U && ddepth == CV_64F)
|
||||||
func = reduceC_<uchar,double,OpAdd<int> >;
|
func = reduceSumC8u64f;
|
||||||
else if(sdepth == CV_16U && ddepth == CV_32F)
|
else if(sdepth == CV_16U && ddepth == CV_32F)
|
||||||
func = reduceC_<ushort,float,OpAdd<float> >;
|
func = reduceSumC16u32f;
|
||||||
else if(sdepth == CV_16U && ddepth == CV_64F)
|
else if(sdepth == CV_16U && ddepth == CV_64F)
|
||||||
func = reduceC_<ushort,double,OpAdd<double> >;
|
func = reduceSumC16u64f;
|
||||||
else if(sdepth == CV_16S && ddepth == CV_32F)
|
else if(sdepth == CV_16S && ddepth == CV_32F)
|
||||||
func = reduceC_<short,float,OpAdd<float> >;
|
func = reduceSumC16s32f;
|
||||||
else if(sdepth == CV_16S && ddepth == CV_64F)
|
else if(sdepth == CV_16S && ddepth == CV_64F)
|
||||||
func = reduceC_<short,double,OpAdd<double> >;
|
func = reduceSumC16s64f;
|
||||||
else if(sdepth == CV_32F && ddepth == CV_32F)
|
else if(sdepth == CV_32F && ddepth == CV_32F)
|
||||||
func = reduceC_<float,float,OpAdd<float> >;
|
func = GET_OPTIMIZED(reduceSumC32f32f);
|
||||||
else if(sdepth == CV_32F && ddepth == CV_64F)
|
else if(sdepth == CV_32F && ddepth == CV_64F)
|
||||||
func = reduceC_<float,double,OpAdd<double> >;
|
func = reduceSumC32f64f;
|
||||||
else if(sdepth == CV_64F && ddepth == CV_64F)
|
else if(sdepth == CV_64F && ddepth == CV_64F)
|
||||||
func = reduceC_<double,double,OpAdd<double> >;
|
func = reduceSumC64f64f;
|
||||||
}
|
}
|
||||||
else if(op == CV_REDUCE_MAX)
|
else if(op == CV_REDUCE_MAX)
|
||||||
{
|
{
|
||||||
if(sdepth == CV_8U && ddepth == CV_8U)
|
if(sdepth == CV_8U && ddepth == CV_8U)
|
||||||
func = reduceC_<uchar, uchar, OpMax<uchar> >;
|
func = GET_OPTIMIZED(reduceMaxC8u);
|
||||||
else if(sdepth == CV_16U && ddepth == CV_16U)
|
else if(sdepth == CV_16U && ddepth == CV_16U)
|
||||||
func = reduceC_<ushort, ushort, OpMax<ushort> >;
|
func = reduceMaxC16u;
|
||||||
else if(sdepth == CV_16S && ddepth == CV_16S)
|
else if(sdepth == CV_16S && ddepth == CV_16S)
|
||||||
func = reduceC_<short, short, OpMax<short> >;
|
func = reduceMaxC16s;
|
||||||
else if(sdepth == CV_32F && ddepth == CV_32F)
|
else if(sdepth == CV_32F && ddepth == CV_32F)
|
||||||
func = reduceC_<float, float, OpMax<float> >;
|
func = GET_OPTIMIZED(reduceMaxC32f);
|
||||||
else if(sdepth == CV_64F && ddepth == CV_64F)
|
else if(sdepth == CV_64F && ddepth == CV_64F)
|
||||||
func = reduceC_<double, double, OpMax<double> >;
|
func = reduceMaxC64f;
|
||||||
}
|
}
|
||||||
else if(op == CV_REDUCE_MIN)
|
else if(op == CV_REDUCE_MIN)
|
||||||
{
|
{
|
||||||
if(sdepth == CV_8U && ddepth == CV_8U)
|
if(sdepth == CV_8U && ddepth == CV_8U)
|
||||||
func = reduceC_<uchar, uchar, OpMin<uchar> >;
|
func = GET_OPTIMIZED(reduceMinC8u);
|
||||||
else if(sdepth == CV_16U && ddepth == CV_16U)
|
else if(sdepth == CV_16U && ddepth == CV_16U)
|
||||||
func = reduceC_<ushort, ushort, OpMin<ushort> >;
|
func = reduceMinC16u;
|
||||||
else if(sdepth == CV_16S && ddepth == CV_16S)
|
else if(sdepth == CV_16S && ddepth == CV_16S)
|
||||||
func = reduceC_<short, short, OpMin<short> >;
|
func = reduceMinC16s;
|
||||||
else if(sdepth == CV_32F && ddepth == CV_32F)
|
else if(sdepth == CV_32F && ddepth == CV_32F)
|
||||||
func = reduceC_<float, float, OpMin<float> >;
|
func = GET_OPTIMIZED(reduceMinC32f);
|
||||||
else if(sdepth == CV_64F && ddepth == CV_64F)
|
else if(sdepth == CV_64F && ddepth == CV_64F)
|
||||||
func = reduceC_<double, double, OpMin<double> >;
|
func = reduceMinC64f;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -2113,7 +2131,7 @@ void cv::reduce(InputArray _src, OutputArray _dst, int dim, int op, int dtype)
|
|||||||
|
|
||||||
if( op0 == CV_REDUCE_AVG )
|
if( op0 == CV_REDUCE_AVG )
|
||||||
temp.convertTo(dst, dst.type(), 1./(dim == 0 ? src.rows : src.cols));
|
temp.convertTo(dst, dst.type(), 1./(dim == 0 ? src.rows : src.cols));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
//////////////////////////////////////// sort ///////////////////////////////////////////
|
//////////////////////////////////////// sort ///////////////////////////////////////////
|
||||||
|
@ -1140,16 +1140,16 @@ normDiffL2_(const T* src1, const T* src2, const uchar* mask, ST* _result, int le
|
|||||||
|
|
||||||
|
|
||||||
#define CV_DEF_NORM_FUNC(L, suffix, type, ntype) \
|
#define CV_DEF_NORM_FUNC(L, suffix, type, ntype) \
|
||||||
static int norm##L##_##suffix(const type* src, const uchar* mask, ntype* r, int len, int cn) \
|
static int norm##L##_##suffix(const type* src, const uchar* mask, ntype* r, int len, int cn) \
|
||||||
{ return norm##L##_(src, mask, r, len, cn); } \
|
{ return norm##L##_(src, mask, r, len, cn); } \
|
||||||
static int normDiff##L##_##suffix(const type* src1, const type* src2, \
|
static int normDiff##L##_##suffix(const type* src1, const type* src2, \
|
||||||
const uchar* mask, ntype* r, int len, int cn) \
|
const uchar* mask, ntype* r, int len, int cn) \
|
||||||
{ return normDiff##L##_(src1, src2, mask, r, (int)len, cn); }
|
{ return normDiff##L##_(src1, src2, mask, r, (int)len, cn); }
|
||||||
|
|
||||||
#define CV_DEF_NORM_ALL(suffix, type, inftype, l1type, l2type) \
|
#define CV_DEF_NORM_ALL(suffix, type, inftype, l1type, l2type) \
|
||||||
CV_DEF_NORM_FUNC(Inf, suffix, type, inftype) \
|
CV_DEF_NORM_FUNC(Inf, suffix, type, inftype) \
|
||||||
CV_DEF_NORM_FUNC(L1, suffix, type, l1type) \
|
CV_DEF_NORM_FUNC(L1, suffix, type, l1type) \
|
||||||
CV_DEF_NORM_FUNC(L2, suffix, type, l2type)
|
CV_DEF_NORM_FUNC(L2, suffix, type, l2type)
|
||||||
|
|
||||||
CV_DEF_NORM_ALL(8u, uchar, int, int, int)
|
CV_DEF_NORM_ALL(8u, uchar, int, int, int)
|
||||||
CV_DEF_NORM_ALL(8s, schar, int, int, int)
|
CV_DEF_NORM_ALL(8s, schar, int, int, int)
|
||||||
|
Loading…
x
Reference in New Issue
Block a user