diff --git a/modules/gpu/src/cuda/color.cu b/modules/gpu/src/cuda/color.cu index a16c82cdc..71f84e33d 100644 --- a/modules/gpu/src/cuda/color.cu +++ b/modules/gpu/src/cuda/color.cu @@ -53,6 +53,173 @@ namespace cv { namespace gpu { namespace device enum { smart_shift = 4 }; }; + template <> struct TransformFunctorTraits : DefaultTransformFunctorTraits + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + template <> struct TransformFunctorTraits : DefaultTransformFunctorTraits + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + template <> struct TransformFunctorTraits : DefaultTransformFunctorTraits + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + template <> struct TransformFunctorTraits : DefaultTransformFunctorTraits + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + + template <> struct TransformFunctorTraits : DefaultTransformFunctorTraits + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + template <> struct TransformFunctorTraits : DefaultTransformFunctorTraits + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + template <> struct TransformFunctorTraits : DefaultTransformFunctorTraits + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + template <> struct TransformFunctorTraits : DefaultTransformFunctorTraits + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + + template <> struct TransformFunctorTraits::functor_type> : DefaultTransformFunctorTraits::functor_type> + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + + template <> struct TransformFunctorTraits : DefaultTransformFunctorTraits + { + enum { smart_shift = 4 }; + }; + template <> struct TransformFunctorTraits : DefaultTransformFunctorTraits + { + enum { smart_shift = 4 }; + }; + + template <> struct TransformFunctorTraits::functor_type> : DefaultTransformFunctorTraits::functor_type> + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + template <> struct TransformFunctorTraits::functor_type> : DefaultTransformFunctorTraits::functor_type> + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + + template <> struct TransformFunctorTraits::functor_type> : DefaultTransformFunctorTraits::functor_type> + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + template <> struct TransformFunctorTraits::functor_type> : DefaultTransformFunctorTraits::functor_type> + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + + template <> struct TransformFunctorTraits::functor_type> : DefaultTransformFunctorTraits::functor_type> + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + template <> struct TransformFunctorTraits::functor_type> : DefaultTransformFunctorTraits::functor_type> + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + + template <> struct TransformFunctorTraits::functor_type> : DefaultTransformFunctorTraits::functor_type> + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + template <> struct TransformFunctorTraits::functor_type> : DefaultTransformFunctorTraits::functor_type> + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + + template <> struct TransformFunctorTraits::functor_type> : DefaultTransformFunctorTraits::functor_type> + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + template <> struct TransformFunctorTraits::functor_type> : DefaultTransformFunctorTraits::functor_type> + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + + template <> struct TransformFunctorTraits::functor_type> : DefaultTransformFunctorTraits::functor_type> + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + template <> struct TransformFunctorTraits::functor_type> : DefaultTransformFunctorTraits::functor_type> + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + + template <> struct TransformFunctorTraits::functor_type> : DefaultTransformFunctorTraits::functor_type> + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + template <> struct TransformFunctorTraits::functor_type> : DefaultTransformFunctorTraits::functor_type> + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + + template <> struct TransformFunctorTraits::functor_type> : DefaultTransformFunctorTraits::functor_type> + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + template <> struct TransformFunctorTraits::functor_type> : DefaultTransformFunctorTraits::functor_type> + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + + template <> struct TransformFunctorTraits::functor_type> : DefaultTransformFunctorTraits::functor_type> + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + template <> struct TransformFunctorTraits::functor_type> : DefaultTransformFunctorTraits::functor_type> + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + + template <> struct TransformFunctorTraits::functor_type> : DefaultTransformFunctorTraits::functor_type> + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + template <> struct TransformFunctorTraits::functor_type> : DefaultTransformFunctorTraits::functor_type> + { + enum { smart_block_dim_y = 8 }; + enum { smart_shift = 4 }; + }; + #define OPENCV_GPU_IMPLEMENT_CVTCOLOR(name, traits) \ void name(const DevMem2D& src, const DevMem2D& dst, cudaStream_t stream) \ { \ diff --git a/modules/gpu/src/opencv2/gpu/device/detail/color.hpp b/modules/gpu/src/opencv2/gpu/device/detail/color.hpp index 467a0ca09..ad5e81551 100644 --- a/modules/gpu/src/opencv2/gpu/device/detail/color.hpp +++ b/modules/gpu/src/opencv2/gpu/device/detail/color.hpp @@ -121,10 +121,10 @@ namespace cv { namespace gpu { namespace device { uint dst = 0; - dst |= (0xff & (src >> 16)); - dst |= (0xff & (src >> 8)) << 8; - dst |= (0xff & (src)) << 16; - dst |= (0xff & (src >> 24)) << 24; + dst |= (0xffu & (src >> 16)); + dst |= (0xffu & (src >> 8)) << 8; + dst |= (0xffu & (src)) << 16; + dst |= (0xffu & (src >> 24)) << 24; return dst; } @@ -145,13 +145,20 @@ namespace cv { namespace gpu { namespace device namespace detail { - template struct RGB2RGB5x5Converter; + template struct RGB2RGB5x5Converter; template struct RGB2RGB5x5Converter<6, bidx> { - template static __device__ __forceinline__ ushort cvt(const T& src) + static __device__ __forceinline__ ushort cvt(const uchar3& src) { return (ushort)(((&src.x)[bidx] >> 3) | ((src.y & ~3) << 3) | (((&src.x)[bidx^2] & ~7) << 8)); } + static __device__ __forceinline__ ushort cvt(uint src) + { + uint b = 0xffu & (src >> (bidx * 8)); + uint g = 0xffu & (src >> 8); + uint r = 0xffu & (src >> ((bidx ^ 2) * 8)); + return (ushort)((b >> 3) | ((g & ~3) << 3) | ((r & ~7) << 8)); + } }; template struct RGB2RGB5x5Converter<5, bidx> { @@ -159,17 +166,29 @@ namespace cv { namespace gpu { namespace device { return (ushort)(((&src.x)[bidx] >> 3) | ((src.y & ~7) << 2) | (((&src.x)[bidx^2] & ~7) << 7)); } - static __device__ __forceinline__ ushort cvt(const uchar4& src) + static __device__ __forceinline__ ushort cvt(uint src) { - return (ushort)(((&src.x)[bidx] >> 3) | ((src.y & ~7) << 2) | (((&src.x)[bidx^2] & ~7) << 7) | (src.w ? 0x8000 : 0)); + uint b = 0xffu & (src >> (bidx * 8)); + uint g = 0xffu & (src >> 8); + uint r = 0xffu & (src >> ((bidx ^ 2) * 8)); + uint a = 0xffu & (src >> 24); + return (ushort)((b >> 3) | ((g & ~7) << 2) | ((r & ~7) << 7) | (a * 0x8000)); } }; - template struct RGB2RGB5x5 : public unary_function + template struct RGB2RGB5x5; + template struct RGB2RGB5x5<3, bidx,green_bits> : unary_function { - __device__ __forceinline__ ushort operator()(const T& src) const + __device__ __forceinline__ ushort operator()(const uchar3& src) const { - return RGB2RGB5x5Converter::cvt(src); + return RGB2RGB5x5Converter::cvt(src); + } + }; + template struct RGB2RGB5x5<4, bidx,green_bits> : unary_function + { + __device__ __forceinline__ ushort operator()(uint src) const + { + return RGB2RGB5x5Converter::cvt(src); } }; } @@ -177,44 +196,68 @@ namespace cv { namespace gpu { namespace device #define OPENCV_GPU_IMPLEMENT_RGB2RGB5x5_TRAITS(name, scn, bidx, green_bits) \ struct name ## _traits \ { \ - typedef detail::RGB2RGB5x5 functor_type; \ + typedef detail::RGB2RGB5x5 functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::RGB2RGB5x5(); \ + return functor_type(); \ } \ }; namespace detail { - template struct RGB5x52RGBConverter; + template struct RGB5x52RGBConverter; template struct RGB5x52RGBConverter<5, bidx> { - template static __device__ void cvt(uint src, D& dst) + static __device__ __forceinline__ void cvt(uint src, uchar3& dst) { - (&dst.x)[bidx] = (uchar)(src << 3); - dst.y = (uchar)((src >> 2) & ~7); - (&dst.x)[bidx ^ 2] = (uchar)((src >> 7) & ~7); - setAlpha(dst, (uchar)(src & 0x8000 ? 255 : 0)); + (&dst.x)[bidx] = src << 3; + dst.y = (src >> 2) & ~7; + (&dst.x)[bidx ^ 2] = (src >> 7) & ~7; + } + static __device__ __forceinline__ void cvt(uint src, uint& dst) + { + dst = 0; + + dst |= (0xffu & (src << 3)) << (bidx * 8); + dst |= (0xffu & ((src >> 2) & ~7)) << 8; + dst |= (0xffu & ((src >> 7) & ~7)) << ((bidx ^ 2) * 8); + dst |= ((src & 0x8000) * 0xffu) << 24; } }; template struct RGB5x52RGBConverter<6, bidx> { - template - static __device__ void cvt(uint src, D& dst) + static __device__ __forceinline__ void cvt(uint src, uchar3& dst) { - (&dst.x)[bidx] = (uchar)(src << 3); - dst.y = (uchar)((src >> 3) & ~3); - (&dst.x)[bidx ^ 2] = (uchar)((src >> 8) & ~7); - setAlpha(dst, (uchar)(255)); + (&dst.x)[bidx] = src << 3; + dst.y = (src >> 3) & ~3; + (&dst.x)[bidx ^ 2] = (src >> 8) & ~7; + } + static __device__ __forceinline__ void cvt(uint src, uint& dst) + { + dst = 0xffu << 24; + + dst |= (0xffu & (src << 3)) << (bidx * 8); + dst |= (0xffu &((src >> 3) & ~3)) << 8; + dst |= (0xffu & ((src >> 8) & ~7)) << ((bidx ^ 2) * 8); } }; - template struct RGB5x52RGB : public unary_function + template struct RGB5x52RGB; + template struct RGB5x52RGB<3, bidx, green_bits> : unary_function { - __device__ __forceinline__ D operator()(ushort src) const + __device__ __forceinline__ uchar3 operator()(ushort src) const { - D dst; - RGB5x52RGBConverter::cvt((uint)src, dst); + uchar3 dst; + RGB5x52RGBConverter::cvt(src, dst); + return dst; + } + }; + template struct RGB5x52RGB<4, bidx, green_bits> : unary_function + { + __device__ __forceinline__ uint operator()(ushort src) const + { + uint dst; + RGB5x52RGBConverter::cvt(src, dst); return dst; } }; @@ -223,10 +266,10 @@ namespace cv { namespace gpu { namespace device #define OPENCV_GPU_IMPLEMENT_RGB5x52RGB_TRAITS(name, dcn, bidx, green_bits) \ struct name ## _traits \ { \ - typedef detail::RGB5x52RGB functor_type; \ + typedef detail::RGB5x52RGB functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::RGB5x52RGB(); \ + return functor_type(); \ } \ }; @@ -234,13 +277,28 @@ namespace cv { namespace gpu { namespace device namespace detail { - template struct Gray2RGB : public unary_function + template struct Gray2RGB : unary_function::vec_type> { - __device__ __forceinline__ D operator()(const T& src) const + __device__ __forceinline__ typename TypeVec::vec_type operator()(T src) const { - D dst; + typename TypeVec::vec_type dst; + dst.z = dst.y = dst.x = src; - setAlpha(dst, ColorChannel::elem_type>::max()); + setAlpha(dst, ColorChannel::max()); + + return dst; + } + }; + template <> struct Gray2RGB : unary_function + { + __device__ __forceinline__ uint operator()(uint src) const + { + uint dst = 0xffu << 24; + + dst |= src; + dst |= src << 8; + dst |= src << 16; + return dst; } }; @@ -249,16 +307,16 @@ namespace cv { namespace gpu { namespace device #define OPENCV_GPU_IMPLEMENT_GRAY2RGB_TRAITS(name, dcn) \ template struct name ## _traits \ { \ - typedef detail::Gray2RGB::vec_type> functor_type; \ + typedef detail::Gray2RGB functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::Gray2RGB::vec_type>(); \ + return functor_type(); \ } \ }; namespace detail { - template struct Gray2RGB5x5Converter; + template struct Gray2RGB5x5Converter; template<> struct Gray2RGB5x5Converter<6> { static __device__ __forceinline__ ushort cvt(uint t) @@ -275,11 +333,11 @@ namespace cv { namespace gpu { namespace device } }; - template struct Gray2RGB5x5 : public unary_function + template struct Gray2RGB5x5 : unary_function { - __device__ __forceinline__ ushort operator()(uchar src) const + __device__ __forceinline__ ushort operator()(uint src) const { - return Gray2RGB5x5Converter::cvt((uint)src); + return Gray2RGB5x5Converter::cvt(src); } }; } @@ -290,7 +348,7 @@ namespace cv { namespace gpu { namespace device typedef detail::Gray2RGB5x5 functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::Gray2RGB5x5(); \ + return functor_type(); \ } \ }; @@ -298,7 +356,7 @@ namespace cv { namespace gpu { namespace device namespace detail { - template struct RGB5x52GrayConverter; + template struct RGB5x52GrayConverter; template <> struct RGB5x52GrayConverter<6> { static __device__ __forceinline__ uchar cvt(uint t) @@ -314,11 +372,11 @@ namespace cv { namespace gpu { namespace device } }; - template struct RGB5x52Gray : public unary_function + template struct RGB5x52Gray : unary_function { - __device__ __forceinline__ uchar operator()(ushort src) const + __device__ __forceinline__ uchar operator()(uint src) const { - return RGB5x52GrayConverter::cvt((uint)src); + return RGB5x52GrayConverter::cvt(src); } }; } @@ -329,7 +387,7 @@ namespace cv { namespace gpu { namespace device typedef detail::RGB5x52Gray functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::RGB5x52Gray(); \ + return functor_type(); \ } \ }; @@ -339,27 +397,41 @@ namespace cv { namespace gpu { namespace device { return (T)CV_DESCALE((unsigned)(src[bidx] * B2Y + src[1] * G2Y + src[bidx^2] * R2Y), yuv_shift); } + template static __device__ __forceinline__ uchar RGB2GrayConvert(uint src) + { + uint b = 0xffu & (src >> (bidx * 8)); + uint g = 0xffu & (src >> 8); + uint r = 0xffu & (src >> ((bidx ^ 2) * 8)); + return CV_DESCALE((uint)(b * B2Y + g * G2Y + r * R2Y), yuv_shift); + } template static __device__ __forceinline__ float RGB2GrayConvert(const float* src) { return src[bidx] * 0.114f + src[1] * 0.587f + src[bidx^2] * 0.299f; } - template struct RGB2Gray : public unary_function + template struct RGB2Gray : unary_function::vec_type, T> { - __device__ __forceinline__ D operator()(const T& src) const + __device__ __forceinline__ T operator()(const typename TypeVec::vec_type& src) const { return RGB2GrayConvert(&src.x); } }; + template struct RGB2Gray : unary_function + { + __device__ __forceinline__ uchar operator()(uint src) const + { + return RGB2GrayConvert(src); + } + }; } #define OPENCV_GPU_IMPLEMENT_RGB2GRAY_TRAITS(name, scn, bidx) \ template struct name ## _traits \ { \ - typedef detail::RGB2Gray::vec_type, T, bidx> functor_type; \ + typedef detail::RGB2Gray functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::RGB2Gray::vec_type, T, bidx>(); \ + return functor_type(); \ } \ }; @@ -382,6 +454,22 @@ namespace cv { namespace gpu { namespace device dst.y = saturate_cast(Cr); dst.z = saturate_cast(Cb); } + template static __device__ uint RGB2YUVConvert(uint src) + { + const uint delta = ColorChannel::half() * (1 << yuv_shift); + + const uint Y = CV_DESCALE((0xffu & src) * c_RGB2YUVCoeffs_i[bidx^2] + (0xffu & (src >> 8)) * c_RGB2YUVCoeffs_i[1] + (0xffu & (src >> 16)) * c_RGB2YUVCoeffs_i[bidx], yuv_shift); + const uint Cr = CV_DESCALE(((0xffu & (src >> ((bidx ^ 2) * 8))) - Y) * c_RGB2YUVCoeffs_i[3] + delta, yuv_shift); + const uint Cb = CV_DESCALE(((0xffu & (src >> (bidx * 8))) - Y) * c_RGB2YUVCoeffs_i[4] + delta, yuv_shift); + + uint dst = 0; + + dst |= saturate_cast(Y); + dst |= saturate_cast(Cr) << 8; + dst |= saturate_cast(Cb) << 16; + + return dst; + } template static __device__ __forceinline__ void RGB2YUVConvert(const float* src, D& dst) { dst.x = src[0] * c_RGB2YUVCoeffs_f[bidx^2] + src[1] * c_RGB2YUVCoeffs_f[1] + src[2] * c_RGB2YUVCoeffs_f[bidx]; @@ -389,24 +477,31 @@ namespace cv { namespace gpu { namespace device dst.z = (src[bidx] - dst.x) * c_RGB2YUVCoeffs_f[4] + ColorChannel::half(); } - template struct RGB2YUV : public unary_function + template struct RGB2YUV : unary_function::vec_type, typename TypeVec::vec_type> { - __device__ __forceinline__ D operator ()(const T& src) const + __device__ __forceinline__ typename TypeVec::vec_type operator ()(const typename TypeVec::vec_type& src) const { - D dst; + typename TypeVec::vec_type dst; RGB2YUVConvert(&src.x, dst); return dst; } }; + template struct RGB2YUV : unary_function + { + __device__ __forceinline__ uint operator ()(uint src) const + { + return RGB2YUVConvert(src); + } + }; } #define OPENCV_GPU_IMPLEMENT_RGB2YUV_TRAITS(name, scn, dcn, bidx) \ template struct name ## _traits \ { \ - typedef detail::RGB2YUV::vec_type, typename TypeVec::vec_type, bidx> functor_type; \ + typedef detail::RGB2YUV functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::RGB2YUV::vec_type, typename TypeVec::vec_type, bidx>(); \ + return functor_type(); \ } \ }; @@ -425,6 +520,24 @@ namespace cv { namespace gpu { namespace device dst[1] = saturate_cast(g); dst[bidx^2] = saturate_cast(r); } + template static __device__ uint YUV2RGBConvert(uint src) + { + const int x = 0xff & (src); + const int y = 0xff & (src >> 8); + const int z = 0xff & (src >> 16); + + const uint b = x + CV_DESCALE((z - ColorChannel::half()) * c_YUV2RGBCoeffs_i[3], yuv_shift); + const uint g = x + CV_DESCALE((z - ColorChannel::half()) * c_YUV2RGBCoeffs_i[2] + (y - ColorChannel::half()) * c_YUV2RGBCoeffs_i[1], yuv_shift); + const uint r = x + CV_DESCALE((y - ColorChannel::half()) * c_YUV2RGBCoeffs_i[0], yuv_shift); + + uint dst = 0xffu << 24; + + dst |= saturate_cast(b) << (bidx * 8); + dst |= saturate_cast(g) << 8; + dst |= saturate_cast(r) << ((bidx ^ 2) * 8); + + return dst; + } template static __device__ __forceinline__ void YUV2RGBConvert(const T& src, float* dst) { dst[bidx] = src.x + (src.z - ColorChannel::half()) * c_YUV2RGBCoeffs_f[3]; @@ -432,27 +545,34 @@ namespace cv { namespace gpu { namespace device dst[bidx^2] = src.x + (src.y - ColorChannel::half()) * c_YUV2RGBCoeffs_f[0]; } - template struct YUV2RGB : public unary_function + template struct YUV2RGB : unary_function::vec_type, typename TypeVec::vec_type> { - __device__ __forceinline__ D operator ()(const T& src) const + __device__ __forceinline__ typename TypeVec::vec_type operator ()(const typename TypeVec::vec_type& src) const { - D dst; + typename TypeVec::vec_type dst; YUV2RGBConvert(src, &dst.x); - setAlpha(dst, ColorChannel::elem_type>::max()); + setAlpha(dst, ColorChannel::max()); return dst; } }; + template struct YUV2RGB : unary_function + { + __device__ __forceinline__ uint operator ()(uint src) const + { + return YUV2RGBConvert(src); + } + }; } #define OPENCV_GPU_IMPLEMENT_YUV2RGB_TRAITS(name, scn, dcn, bidx) \ template struct name ## _traits \ { \ - typedef detail::YUV2RGB::vec_type, typename TypeVec::vec_type, bidx> functor_type; \ + typedef detail::YUV2RGB functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::YUV2RGB::vec_type, typename TypeVec::vec_type, bidx>(); \ + return functor_type(); \ } \ }; @@ -462,7 +582,7 @@ namespace cv { namespace gpu { namespace device { __constant__ float c_RGB2YCrCbCoeffs_f[5] = {0.299f, 0.587f, 0.114f, 0.713f, 0.564f}; __constant__ int c_RGB2YCrCbCoeffs_i[5] = {R2Y, G2Y, B2Y, 11682, 9241}; - + template static __device__ void RGB2YCrCbConvert(const T* src, D& dst) { const int delta = ColorChannel::half() * (1 << yuv_shift); @@ -475,6 +595,22 @@ namespace cv { namespace gpu { namespace device dst.y = saturate_cast(Cr); dst.z = saturate_cast(Cb); } + template static __device__ uint RGB2YCrCbConvert(uint src) + { + const int delta = ColorChannel::half() * (1 << yuv_shift); + + const uint Y = CV_DESCALE((0xffu & src) * c_RGB2YCrCbCoeffs_i[bidx^2] + (0xffu & (src >> 8)) * c_RGB2YCrCbCoeffs_i[1] + (0xffu & (src >> 16)) * c_RGB2YCrCbCoeffs_i[bidx], yuv_shift); + const uint Cr = CV_DESCALE(((0xffu & (src >> ((bidx ^ 2) * 8))) - Y) * c_RGB2YCrCbCoeffs_i[3] + delta, yuv_shift); + const uint Cb = CV_DESCALE(((0xffu & (src >> (bidx * 8))) - Y) * c_RGB2YCrCbCoeffs_i[4] + delta, yuv_shift); + + uint dst = 0; + + dst |= saturate_cast(Y); + dst |= saturate_cast(Cr) << 8; + dst |= saturate_cast(Cb) << 16; + + return dst; + } template static __device__ __forceinline__ void RGB2YCrCbConvert(const float* src, D& dst) { dst.x = src[0] * c_RGB2YCrCbCoeffs_f[bidx^2] + src[1] * c_RGB2YCrCbCoeffs_f[1] + src[2] * c_RGB2YCrCbCoeffs_f[bidx]; @@ -482,24 +618,31 @@ namespace cv { namespace gpu { namespace device dst.z = (src[bidx] - dst.x) * c_RGB2YCrCbCoeffs_f[4] + ColorChannel::half(); } - template struct RGB2YCrCb : public unary_function + template struct RGB2YCrCb : unary_function::vec_type, typename TypeVec::vec_type> { - __device__ __forceinline__ D operator ()(const T& src) const + __device__ __forceinline__ typename TypeVec::vec_type operator ()(const typename TypeVec::vec_type& src) const { - D dst; + typename TypeVec::vec_type dst; RGB2YCrCbConvert(&src.x, dst); return dst; } }; + template struct RGB2YCrCb : unary_function + { + __device__ __forceinline__ uint operator ()(uint src) const + { + return RGB2YCrCbConvert(src); + } + }; } #define OPENCV_GPU_IMPLEMENT_RGB2YCrCb_TRAITS(name, scn, dcn, bidx) \ template struct name ## _traits \ { \ - typedef detail::RGB2YCrCb::vec_type, typename TypeVec::vec_type, bidx> functor_type; \ + typedef detail::RGB2YCrCb functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::RGB2YCrCb::vec_type, typename TypeVec::vec_type, bidx>(); \ + return functor_type(); \ } \ }; @@ -518,6 +661,24 @@ namespace cv { namespace gpu { namespace device dst[1] = saturate_cast(g); dst[bidx^2] = saturate_cast(r); } + template static __device__ uint YCrCb2RGBConvert(uint src) + { + const int x = 0xff & (src); + const int y = 0xff & (src >> 8); + const int z = 0xff & (src >> 16); + + const uint b = x + CV_DESCALE((z - ColorChannel::half()) * c_YCrCb2RGBCoeffs_i[3], yuv_shift); + const uint g = x + CV_DESCALE((z - ColorChannel::half()) * c_YCrCb2RGBCoeffs_i[2] + (y - ColorChannel::half()) * c_YCrCb2RGBCoeffs_i[1], yuv_shift); + const uint r = x + CV_DESCALE((y - ColorChannel::half()) * c_YCrCb2RGBCoeffs_i[0], yuv_shift); + + uint dst = 0xffu << 24; + + dst |= saturate_cast(b) << (bidx * 8); + dst |= saturate_cast(g) << 8; + dst |= saturate_cast(r) << ((bidx ^ 2) * 8); + + return dst; + } template __device__ __forceinline__ void YCrCb2RGBConvert(const T& src, float* dst) { dst[bidx] = src.x + (src.z - ColorChannel::half()) * c_YCrCb2RGBCoeffs_f[3]; @@ -525,27 +686,34 @@ namespace cv { namespace gpu { namespace device dst[bidx^2] = src.x + (src.y - ColorChannel::half()) * c_YCrCb2RGBCoeffs_f[0]; } - template struct YCrCb2RGB : public unary_function + template struct YCrCb2RGB : unary_function::vec_type, typename TypeVec::vec_type> { - __device__ __forceinline__ D operator ()(const T& src) const + __device__ __forceinline__ typename TypeVec::vec_type operator ()(const typename TypeVec::vec_type& src) const { - D dst; + typename TypeVec::vec_type dst; YCrCb2RGBConvert(src, &dst.x); - setAlpha(dst, ColorChannel::elem_type>::max()); + setAlpha(dst, ColorChannel::max()); return dst; } }; + template struct YCrCb2RGB : unary_function + { + __device__ __forceinline__ uint operator ()(uint src) const + { + return YCrCb2RGBConvert(src); + } + }; } #define OPENCV_GPU_IMPLEMENT_YCrCb2RGB_TRAITS(name, scn, dcn, bidx) \ template struct name ## _traits \ { \ - typedef detail::YCrCb2RGB::vec_type, typename TypeVec::vec_type, bidx> functor_type; \ + typedef detail::YCrCb2RGB functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::YCrCb2RGB::vec_type, typename TypeVec::vec_type, bidx>(); \ + return functor_type(); \ } \ }; @@ -562,6 +730,24 @@ namespace cv { namespace gpu { namespace device dst.y = saturate_cast(CV_DESCALE(src[bidx^2] * c_RGB2XYZ_D65i[3] + src[1] * c_RGB2XYZ_D65i[4] + src[bidx] * c_RGB2XYZ_D65i[5], xyz_shift)); dst.z = saturate_cast(CV_DESCALE(src[bidx^2] * c_RGB2XYZ_D65i[6] + src[1] * c_RGB2XYZ_D65i[7] + src[bidx] * c_RGB2XYZ_D65i[8], xyz_shift)); } + template static __device__ __forceinline__ uint RGB2XYZConvert(uint src) + { + const uint b = 0xffu & (src >> (bidx * 8)); + const uint g = 0xffu & (src >> 8); + const uint r = 0xffu & (src >> ((bidx ^ 2) * 8)); + + const uint x = saturate_cast(CV_DESCALE(r * c_RGB2XYZ_D65i[0] + g * c_RGB2XYZ_D65i[1] + b * c_RGB2XYZ_D65i[2], xyz_shift)); + const uint y = saturate_cast(CV_DESCALE(r * c_RGB2XYZ_D65i[3] + g * c_RGB2XYZ_D65i[4] + b * c_RGB2XYZ_D65i[5], xyz_shift)); + const uint z = saturate_cast(CV_DESCALE(r * c_RGB2XYZ_D65i[6] + g * c_RGB2XYZ_D65i[7] + b * c_RGB2XYZ_D65i[8], xyz_shift)); + + uint dst = 0; + + dst |= x; + dst |= y << 8; + dst |= z << 16; + + return dst; + } template static __device__ __forceinline__ void RGB2XYZConvert(const float* src, D& dst) { dst.x = src[bidx^2] * c_RGB2XYZ_D65f[0] + src[1] * c_RGB2XYZ_D65f[1] + src[bidx] * c_RGB2XYZ_D65f[2]; @@ -569,24 +755,33 @@ namespace cv { namespace gpu { namespace device dst.z = src[bidx^2] * c_RGB2XYZ_D65f[6] + src[1] * c_RGB2XYZ_D65f[7] + src[bidx] * c_RGB2XYZ_D65f[8]; } - template struct RGB2XYZ : public unary_function + template struct RGB2XYZ : unary_function::vec_type, typename TypeVec::vec_type> { - __device__ __forceinline__ D operator()(const T& src) const + __device__ __forceinline__ typename TypeVec::vec_type operator()(const typename TypeVec::vec_type& src) const { - D dst; + typename TypeVec::vec_type dst; + RGB2XYZConvert(&src.x, dst); + return dst; } }; + template struct RGB2XYZ : unary_function + { + __device__ __forceinline__ uint operator()(uint src) const + { + return RGB2XYZConvert(src); + } + }; } #define OPENCV_GPU_IMPLEMENT_RGB2XYZ_TRAITS(name, scn, dcn, bidx) \ template struct name ## _traits \ { \ - typedef detail::RGB2XYZ::vec_type, typename TypeVec::vec_type, bidx> functor_type; \ + typedef detail::RGB2XYZ functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::RGB2XYZ::vec_type, typename TypeVec::vec_type, bidx>(); \ + return functor_type(); \ } \ }; @@ -601,6 +796,24 @@ namespace cv { namespace gpu { namespace device dst[1] = saturate_cast(CV_DESCALE(src.x * c_XYZ2sRGB_D65i[3] + src.y * c_XYZ2sRGB_D65i[4] + src.z * c_XYZ2sRGB_D65i[5], xyz_shift)); dst[bidx] = saturate_cast(CV_DESCALE(src.x * c_XYZ2sRGB_D65i[6] + src.y * c_XYZ2sRGB_D65i[7] + src.z * c_XYZ2sRGB_D65i[8], xyz_shift)); } + template static __device__ __forceinline__ uint XYZ2RGBConvert(uint src) + { + const int x = 0xff & src; + const int y = 0xff & (src >> 8); + const int z = 0xff & (src >> 16); + + const uint r = saturate_cast(CV_DESCALE(x * c_XYZ2sRGB_D65i[0] + y * c_XYZ2sRGB_D65i[1] + z * c_XYZ2sRGB_D65i[2], xyz_shift)); + const uint g = saturate_cast(CV_DESCALE(x * c_XYZ2sRGB_D65i[3] + y * c_XYZ2sRGB_D65i[4] + z * c_XYZ2sRGB_D65i[5], xyz_shift)); + const uint b = saturate_cast(CV_DESCALE(x * c_XYZ2sRGB_D65i[6] + y * c_XYZ2sRGB_D65i[7] + z * c_XYZ2sRGB_D65i[8], xyz_shift)); + + uint dst = 0xffu << 24; + + dst |= b << (bidx * 8); + dst |= g << 8; + dst |= r << ((bidx ^ 2) * 8); + + return dst; + } template static __device__ __forceinline__ void XYZ2RGBConvert(const T& src, float* dst) { dst[bidx^2] = src.x * c_XYZ2sRGB_D65f[0] + src.y * c_XYZ2sRGB_D65f[1] + src.z * c_XYZ2sRGB_D65f[2]; @@ -608,25 +821,34 @@ namespace cv { namespace gpu { namespace device dst[bidx] = src.x * c_XYZ2sRGB_D65f[6] + src.y * c_XYZ2sRGB_D65f[7] + src.z * c_XYZ2sRGB_D65f[8]; } - template struct XYZ2RGB : public unary_function + template struct XYZ2RGB : unary_function::vec_type, typename TypeVec::vec_type> { - __device__ __forceinline__ D operator()(const T& src) const + __device__ __forceinline__ typename TypeVec::vec_type operator()(const typename TypeVec::vec_type& src) const { - D dst; + typename TypeVec::vec_type dst; + XYZ2RGBConvert(src, &dst.x); - setAlpha(dst, ColorChannel::elem_type>::max()); + setAlpha(dst, ColorChannel::max()); + return dst; } }; + template struct XYZ2RGB : unary_function + { + __device__ __forceinline__ uint operator()(uint src) const + { + return XYZ2RGBConvert(src); + } + }; } #define OPENCV_GPU_IMPLEMENT_XYZ2RGB_TRAITS(name, scn, dcn, bidx) \ template struct name ## _traits \ { \ - typedef detail::XYZ2RGB::vec_type, typename TypeVec::vec_type, bidx> functor_type; \ + typedef detail::XYZ2RGB functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::XYZ2RGB::vec_type, typename TypeVec::vec_type, bidx>(); \ + return functor_type(); \ } \ }; @@ -637,7 +859,7 @@ namespace cv { namespace gpu { namespace device __constant__ int c_HsvDivTable [256] = {0, 1044480, 522240, 348160, 261120, 208896, 174080, 149211, 130560, 116053, 104448, 94953, 87040, 80345, 74606, 69632, 65280, 61440, 58027, 54973, 52224, 49737, 47476, 45412, 43520, 41779, 40172, 38684, 37303, 36017, 34816, 33693, 32640, 31651, 30720, 29842, 29013, 28229, 27486, 26782, 26112, 25475, 24869, 24290, 23738, 23211, 22706, 22223, 21760, 21316, 20890, 20480, 20086, 19707, 19342, 18991, 18651, 18324, 18008, 17703, 17408, 17123, 16846, 16579, 16320, 16069, 15825, 15589, 15360, 15137, 14921, 14711, 14507, 14308, 14115, 13926, 13743, 13565, 13391, 13221, 13056, 12895, 12738, 12584, 12434, 12288, 12145, 12006, 11869, 11736, 11605, 11478, 11353, 11231, 11111, 10995, 10880, 10768, 10658, 10550, 10445, 10341, 10240, 10141, 10043, 9947, 9854, 9761, 9671, 9582, 9495, 9410, 9326, 9243, 9162, 9082, 9004, 8927, 8852, 8777, 8704, 8632, 8561, 8492, 8423, 8356, 8290, 8224, 8160, 8097, 8034, 7973, 7913, 7853, 7795, 7737, 7680, 7624, 7569, 7514, 7461, 7408, 7355, 7304, 7253, 7203, 7154, 7105, 7057, 7010, 6963, 6917, 6872, 6827, 6782, 6739, 6695, 6653, 6611, 6569, 6528, 6487, 6447, 6408, 6369, 6330, 6292, 6254, 6217, 6180, 6144, 6108, 6073, 6037, 6003, 5968, 5935, 5901, 5868, 5835, 5803, 5771, 5739, 5708, 5677, 5646, 5615, 5585, 5556, 5526, 5497, 5468, 5440, 5412, 5384, 5356, 5329, 5302, 5275, 5249, 5222, 5196, 5171, 5145, 5120, 5095, 5070, 5046, 5022, 4998, 4974, 4950, 4927, 4904, 4881, 4858, 4836, 4813, 4791, 4769, 4748, 4726, 4705, 4684, 4663, 4642, 4622, 4601, 4581, 4561, 4541, 4522, 4502, 4483, 4464, 4445, 4426, 4407, 4389, 4370, 4352, 4334, 4316, 4298, 4281, 4263, 4246, 4229, 4212, 4195, 4178, 4161, 4145, 4128, 4112, 4096}; __constant__ int c_HsvDivTable180[256] = {0, 122880, 61440, 40960, 30720, 24576, 20480, 17554, 15360, 13653, 12288, 11171, 10240, 9452, 8777, 8192, 7680, 7228, 6827, 6467, 6144, 5851, 5585, 5343, 5120, 4915, 4726, 4551, 4389, 4237, 4096, 3964, 3840, 3724, 3614, 3511, 3413, 3321, 3234, 3151, 3072, 2997, 2926, 2858, 2793, 2731, 2671, 2614, 2560, 2508, 2458, 2409, 2363, 2318, 2276, 2234, 2194, 2156, 2119, 2083, 2048, 2014, 1982, 1950, 1920, 1890, 1862, 1834, 1807, 1781, 1755, 1731, 1707, 1683, 1661, 1638, 1617, 1596, 1575, 1555, 1536, 1517, 1499, 1480, 1463, 1446, 1429, 1412, 1396, 1381, 1365, 1350, 1336, 1321, 1307, 1293, 1280, 1267, 1254, 1241, 1229, 1217, 1205, 1193, 1182, 1170, 1159, 1148, 1138, 1127, 1117, 1107, 1097, 1087, 1078, 1069, 1059, 1050, 1041, 1033, 1024, 1016, 1007, 999, 991, 983, 975, 968, 960, 953, 945, 938, 931, 924, 917, 910, 904, 897, 890, 884, 878, 871, 865, 859, 853, 847, 842, 836, 830, 825, 819, 814, 808, 803, 798, 793, 788, 783, 778, 773, 768, 763, 759, 754, 749, 745, 740, 736, 731, 727, 723, 719, 714, 710, 706, 702, 698, 694, 690, 686, 683, 679, 675, 671, 668, 664, 661, 657, 654, 650, 647, 643, 640, 637, 633, 630, 627, 624, 621, 617, 614, 611, 608, 605, 602, 599, 597, 594, 591, 588, 585, 582, 580, 577, 574, 572, 569, 566, 564, 561, 559, 556, 554, 551, 549, 546, 544, 541, 539, 537, 534, 532, 530, 527, 525, 523, 521, 518, 516, 514, 512, 510, 508, 506, 504, 502, 500, 497, 495, 493, 492, 490, 488, 486, 484, 482}; __constant__ int c_HsvDivTable256[256] = {0, 174763, 87381, 58254, 43691, 34953, 29127, 24966, 21845, 19418, 17476, 15888, 14564, 13443, 12483, 11651, 10923, 10280, 9709, 9198, 8738, 8322, 7944, 7598, 7282, 6991, 6722, 6473, 6242, 6026, 5825, 5638, 5461, 5296, 5140, 4993, 4855, 4723, 4599, 4481, 4369, 4263, 4161, 4064, 3972, 3884, 3799, 3718, 3641, 3567, 3495, 3427, 3361, 3297, 3236, 3178, 3121, 3066, 3013, 2962, 2913, 2865, 2819, 2774, 2731, 2689, 2648, 2608, 2570, 2533, 2497, 2461, 2427, 2394, 2362, 2330, 2300, 2270, 2241, 2212, 2185, 2158, 2131, 2106, 2081, 2056, 2032, 2009, 1986, 1964, 1942, 1920, 1900, 1879, 1859, 1840, 1820, 1802, 1783, 1765, 1748, 1730, 1713, 1697, 1680, 1664, 1649, 1633, 1618, 1603, 1589, 1574, 1560, 1547, 1533, 1520, 1507, 1494, 1481, 1469, 1456, 1444, 1432, 1421, 1409, 1398, 1387, 1376, 1365, 1355, 1344, 1334, 1324, 1314, 1304, 1295, 1285, 1276, 1266, 1257, 1248, 1239, 1231, 1222, 1214, 1205, 1197, 1189, 1181, 1173, 1165, 1157, 1150, 1142, 1135, 1128, 1120, 1113, 1106, 1099, 1092, 1085, 1079, 1072, 1066, 1059, 1053, 1046, 1040, 1034, 1028, 1022, 1016, 1010, 1004, 999, 993, 987, 982, 976, 971, 966, 960, 955, 950, 945, 940, 935, 930, 925, 920, 915, 910, 906, 901, 896, 892, 887, 883, 878, 874, 869, 865, 861, 857, 853, 848, 844, 840, 836, 832, 828, 824, 820, 817, 813, 809, 805, 802, 798, 794, 791, 787, 784, 780, 777, 773, 770, 767, 763, 760, 757, 753, 750, 747, 744, 741, 737, 734, 731, 728, 725, 722, 719, 716, 713, 710, 708, 705, 702, 699, 696, 694, 691, 688, 685}; - + template static __device__ void RGB2HSVConvert(const uchar* src, D& dst) { const int hsv_shift = 12; @@ -654,18 +876,53 @@ namespace cv { namespace gpu { namespace device vmin = ::min(vmin, r); diff = v - vmin; - vr = v == r ? -1 : 0; - vg = v == g ? -1 : 0; + vr = (v == r) * -1; + vg = (v == g) * -1; s = (diff * c_HsvDivTable[v] + (1 << (hsv_shift-1))) >> hsv_shift; h = (vr & (g - b)) + (~vr & ((vg & (b - r + 2 * diff)) + ((~vg) & (r - g + 4 * diff)))); h = (h * hdiv_table[diff] + (1 << (hsv_shift-1))) >> hsv_shift; - h += h < 0 ? hr : 0; + h += (h < 0) * hr; dst.x = saturate_cast(h); dst.y = (uchar)s; dst.z = (uchar)v; } + template static __device__ uint RGB2HSVConvert(uint src) + { + const int hsv_shift = 12; + const int* hdiv_table = hr == 180 ? c_HsvDivTable180 : c_HsvDivTable256; + + const int b = 0xff & (src >> (bidx * 8)); + const int g = 0xff & (src >> 8); + const int r = 0xff & (src >> ((bidx ^ 2) * 8)); + + int h, s, v = b; + int vmin = b, diff; + int vr, vg; + + v = ::max(v, g); + v = ::max(v, r); + vmin = ::min(vmin, g); + vmin = ::min(vmin, r); + + diff = v - vmin; + vr = (v == r) * -1; + vg = (v == g) * -1; + + s = (diff * c_HsvDivTable[v] + (1 << (hsv_shift-1))) >> hsv_shift; + h = (vr & (g - b)) + (~vr & ((vg & (b - r + 2 * diff)) + ((~vg) & (r - g + 4 * diff)))); + h = (h * hdiv_table[diff] + (1 << (hsv_shift-1))) >> hsv_shift; + h += (h < 0) * hr; + + uint dst = 0; + + dst |= saturate_cast(h); + dst |= (0xffu & s) << 8; + dst |= (0xffu & v) << 16; + + return dst; + } template static __device__ void RGB2HSVConvert(const float* src, D& dst) { const float hscale = hr * (1.f / 360.f); @@ -685,62 +942,67 @@ namespace cv { namespace gpu { namespace device s = diff / (float)(::fabs(v) + numeric_limits::epsilon()); diff = (float)(60. / (diff + numeric_limits::epsilon())); - if (v == r) - h = (g - b) * diff; - else if (v == g) - h = (b - r) * diff + 120.f; - else - h = (r - g) * diff + 240.f; - - if (h < 0) h += 360.f; + h = (v == r) * (g - b) * diff; + h += (v != r && v == g) * ((b - r) * diff + 120.f); + h += (v != r && v != g) * ((r - g) * diff + 240.f); + h += (h < 0) * 360.f; dst.x = h * hscale; dst.y = s; dst.z = v; } - template struct RGB2HSV : public unary_function + template struct RGB2HSV : unary_function::vec_type, typename TypeVec::vec_type> { - __device__ __forceinline__ D operator()(const T& src) const + __device__ __forceinline__ typename TypeVec::vec_type operator()(const typename TypeVec::vec_type& src) const { - D dst; + typename TypeVec::vec_type dst; + RGB2HSVConvert(&src.x, dst); + return dst; } }; + template struct RGB2HSV : unary_function + { + __device__ __forceinline__ uint operator()(uint src) const + { + return RGB2HSVConvert(src); + } + }; } #define OPENCV_GPU_IMPLEMENT_RGB2HSV_TRAITS(name, scn, dcn, bidx) \ template struct name ## _traits \ { \ - typedef detail::RGB2HSV::vec_type, typename TypeVec::vec_type, bidx, 180> functor_type; \ + typedef detail::RGB2HSV functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::RGB2HSV::vec_type, typename TypeVec::vec_type, bidx, 180>(); \ + return functor_type(); \ } \ }; \ template struct name ## _full_traits \ { \ - typedef detail::RGB2HSV::vec_type, typename TypeVec::vec_type, bidx, 256> functor_type; \ + typedef detail::RGB2HSV functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::RGB2HSV::vec_type, typename TypeVec::vec_type, bidx, 256>(); \ + return functor_type(); \ } \ }; \ template <> struct name ## _traits \ { \ - typedef detail::RGB2HSV::vec_type, typename TypeVec::vec_type, bidx, 360> functor_type; \ + typedef detail::RGB2HSV functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::RGB2HSV::vec_type, typename TypeVec::vec_type, bidx, 360>(); \ + return functor_type(); \ } \ }; \ template <> struct name ## _full_traits \ { \ - typedef detail::RGB2HSV::vec_type, typename TypeVec::vec_type, bidx, 360> functor_type; \ + typedef detail::RGB2HSV functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::RGB2HSV::vec_type, typename TypeVec::vec_type, bidx, 360>(); \ + return functor_type(); \ } \ }; @@ -748,31 +1010,30 @@ namespace cv { namespace gpu { namespace device { __constant__ int c_HsvSectorData[6][3] = { {1,3,0}, {1,0,2}, {3,0,1}, {0,2,1}, {0,1,3}, {2,1,0} }; - template static __device__ void HSV2RGBConvert(const T& src, float* dst) + template static __device__ void HSV2RGBConvert(const T& src, float* dst) { - const float hscale = 6.f / HR; + const float hscale = 6.f / hr; float h = src.x, s = src.y, v = src.z; - float b, g, r; + float b = v, g = v, r = v; - if( s == 0 ) - b = g = r = v; - else + if (s != 0) { - float tab[4]; - int sector; h *= hscale; + if( h < 0 ) do h += 6; while( h < 0 ); else if( h >= 6 ) do h -= 6; while( h >= 6 ); - sector = __float2int_rd(h); + + int sector = __float2int_rd(h); h -= sector; + float tab[4]; tab[0] = v; - tab[1] = v*(1.f - s); - tab[2] = v*(1.f - s*h); - tab[3] = v*(1.f - s*(1.f - h)); + tab[1] = v * (1.f - s); + tab[2] = v * (1.f - s * h); + tab[3] = v * (1.f - s * (1.f - h)); b = tab[c_HsvSectorData[sector][0]]; g = tab[c_HsvSectorData[sector][1]]; @@ -788,8 +1049,8 @@ namespace cv { namespace gpu { namespace device float3 buf; buf.x = src.x; - buf.y = src.y * (1.f/255.f); - buf.z = src.z * (1.f/255.f); + buf.y = src.y * (1.f / 255.f); + buf.z = src.z * (1.f / 255.f); HSV2RGBConvert(buf, &buf.x); @@ -797,50 +1058,77 @@ namespace cv { namespace gpu { namespace device dst[1] = saturate_cast(buf.y * 255.f); dst[2] = saturate_cast(buf.z * 255.f); } - - template struct HSV2RGB : public unary_function + template static __device__ uint HSV2RGBConvert(uint src) { - __device__ __forceinline__ D operator()(const T& src) const + float3 buf; + + buf.x = src & 0xff; + buf.y = ((src >> 8) & 0xff) * (1.f/255.f); + buf.z = ((src >> 16) & 0xff) * (1.f/255.f); + + HSV2RGBConvert(buf, &buf.x); + + uint dst = 0xffu << 24; + + dst |= saturate_cast(buf.x * 255.f); + dst |= saturate_cast(buf.y * 255.f) << 8; + dst |= saturate_cast(buf.z * 255.f) << 16; + + return dst; + } + + template struct HSV2RGB : unary_function::vec_type, typename TypeVec::vec_type> + { + __device__ __forceinline__ typename TypeVec::vec_type operator()(const typename TypeVec::vec_type& src) const { - D dst; - HSV2RGBConvert(src, &dst.x); - setAlpha(dst, ColorChannel::elem_type>::max()); + typename TypeVec::vec_type dst; + + HSV2RGBConvert(src, &dst.x); + setAlpha(dst, ColorChannel::max()); + return dst; } }; + template struct HSV2RGB : unary_function + { + __device__ __forceinline__ uint operator()(uint src) const + { + return HSV2RGBConvert(src); + } + }; } #define OPENCV_GPU_IMPLEMENT_HSV2RGB_TRAITS(name, scn, dcn, bidx) \ template struct name ## _traits \ { \ - typedef detail::HSV2RGB::vec_type, typename TypeVec::vec_type, bidx, 180> functor_type; \ + typedef detail::HSV2RGB functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::HSV2RGB::vec_type, typename TypeVec::vec_type, bidx, 180>(); \ + return functor_type(); \ } \ }; \ template struct name ## _full_traits \ { \ - typedef detail::HSV2RGB::vec_type, typename TypeVec::vec_type, bidx, 255> functor_type; \ + typedef detail::HSV2RGB functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::HSV2RGB::vec_type, typename TypeVec::vec_type, bidx, 255>(); \ + return functor_type(); \ } \ }; \ template <> struct name ## _traits \ { \ - typedef detail::HSV2RGB::vec_type, typename TypeVec::vec_type, bidx, 360> functor_type; \ + typedef detail::HSV2RGB functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::HSV2RGB::vec_type, typename TypeVec::vec_type, bidx, 360>(); \ + return functor_type(); \ } \ }; \ template <> struct name ## _full_traits \ { \ - typedef detail::HSV2RGB::vec_type, typename TypeVec::vec_type, bidx, 360> functor_type; \ + typedef detail::HSV2RGB functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::HSV2RGB::vec_type, typename TypeVec::vec_type, bidx, 360>(); \ + return functor_type(); \ } \ }; @@ -867,17 +1155,15 @@ namespace cv { namespace gpu { namespace device if (diff > numeric_limits::epsilon()) { - s = l < 0.5f ? diff / (vmax + vmin) : diff / (2.0f - vmax - vmin); + s = (l < 0.5f) * diff / (vmax + vmin); + s += (l >= 0.5f) * diff / (2.0f - vmax - vmin); + diff = 60.f / diff; - if (vmax == r) - h = (g - b)*diff; - else if (vmax == g) - h = (b - r)*diff + 120.f; - else - h = (r - g)*diff + 240.f; - - if (h < 0.f) h += 360.f; + h = (vmax == r) * (g - b) * diff; + h += (vmax != r && vmax == g) * ((b - r) * diff + 120.f); + h += (vmax != r && vmax != g) * ((r - g) * diff + 240.f); + h += (h < 0.f) * 360.f; } dst.x = h * hscale; @@ -888,9 +1174,9 @@ namespace cv { namespace gpu { namespace device { float3 buf; - buf.x = src[0]*(1.f/255.f); - buf.y = src[1]*(1.f/255.f); - buf.z = src[2]*(1.f/255.f); + buf.x = src[0] * (1.f / 255.f); + buf.y = src[1] * (1.f / 255.f); + buf.z = src[2] * (1.f / 255.f); RGB2HLSConvert(&buf.x, buf); @@ -898,49 +1184,76 @@ namespace cv { namespace gpu { namespace device dst.y = saturate_cast(buf.y*255.f); dst.z = saturate_cast(buf.z*255.f); } - - template struct RGB2HLS : public unary_function + template static __device__ uint RGB2HLSConvert(uint src) { - __device__ __forceinline__ D operator()(const T& src) const + float3 buf; + + buf.x = (0xff & src) * (1.f / 255.f); + buf.y = (0xff & (src >> 8)) * (1.f / 255.f); + buf.z = (0xff & (src >> 16)) * (1.f / 255.f); + + RGB2HLSConvert(&buf.x, buf); + + uint dst = 0xffu << 24; + + dst |= saturate_cast(buf.x); + dst |= saturate_cast(buf.y * 255.f) << 8; + dst |= saturate_cast(buf.z * 255.f) << 16; + + return dst; + } + + template struct RGB2HLS : unary_function::vec_type, typename TypeVec::vec_type> + { + __device__ __forceinline__ typename TypeVec::vec_type operator()(const typename TypeVec::vec_type& src) const { - D dst; + typename TypeVec::vec_type dst; + RGB2HLSConvert(&src.x, dst); + return dst; } }; + template struct RGB2HLS : unary_function + { + __device__ __forceinline__ uint operator()(uint src) const + { + return RGB2HLSConvert(src); + } + }; } #define OPENCV_GPU_IMPLEMENT_RGB2HLS_TRAITS(name, scn, dcn, bidx) \ template struct name ## _traits \ { \ - typedef detail::RGB2HLS::vec_type, typename TypeVec::vec_type, bidx, 180> functor_type; \ + typedef detail::RGB2HLS functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::RGB2HLS::vec_type, typename TypeVec::vec_type, bidx, 180>(); \ + return functor_type(); \ } \ }; \ template struct name ## _full_traits \ { \ - typedef detail::RGB2HLS::vec_type, typename TypeVec::vec_type, bidx, 256> functor_type; \ + typedef detail::RGB2HLS functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::RGB2HLS::vec_type, typename TypeVec::vec_type, bidx, 256>(); \ + return functor_type(); \ } \ }; \ template <> struct name ## _traits \ { \ - typedef detail::RGB2HLS::vec_type, typename TypeVec::vec_type, bidx, 360> functor_type; \ + typedef detail::RGB2HLS functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::RGB2HLS::vec_type, typename TypeVec::vec_type, bidx, 360>(); \ + return functor_type(); \ } \ }; \ template <> struct name ## _full_traits \ { \ - typedef detail::RGB2HLS::vec_type, typename TypeVec::vec_type, bidx, 360> functor_type; \ + typedef detail::RGB2HLS functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::RGB2HLS::vec_type, typename TypeVec::vec_type, bidx, 360>(); \ + return functor_type(); \ } \ }; @@ -948,21 +1261,17 @@ namespace cv { namespace gpu { namespace device { __constant__ int c_HlsSectorData[6][3] = { {1,3,0}, {1,0,2}, {3,0,1}, {0,2,1}, {0,1,3}, {2,1,0} }; - template static __device__ void HLS2RGBConvert(const T& src, float* dst) + template static __device__ void HLS2RGBConvert(const T& src, float* dst) { - const float hscale = 6.0f / HR; + const float hscale = 6.0f / hr; float h = src.x, l = src.y, s = src.z; - float b, g, r; + float b = l, g = l, r = l; - if (s == 0) - b = g = r = l; - else + if (s != 0) { - float tab[4]; - int sector; - - float p2 = l <= 0.5f ? l * (1 + s) : l + s - l * s; + float p2 = (l <= 0.5f) * l * (1 + s); + p2 += (l > 0.5f) * (l + s - l * s); float p1 = 2 * l - p2; h *= hscale; @@ -972,9 +1281,12 @@ namespace cv { namespace gpu { namespace device else if( h >= 6 ) do h -= 6; while( h >= 6 ); + int sector; sector = __float2int_rd(h); + h -= sector; + float tab[4]; tab[0] = p2; tab[1] = p1; tab[2] = p1 + (p2 - p1) * (1 - h); @@ -989,64 +1301,91 @@ namespace cv { namespace gpu { namespace device dst[1] = g; dst[bidx^2] = r; } - template static __device__ void HLS2RGBConvert(const T& src, uchar* dst) + template static __device__ void HLS2RGBConvert(const T& src, uchar* dst) { float3 buf; buf.x = src.x; - buf.y = src.y*(1.f/255.f); - buf.z = src.z*(1.f/255.f); + buf.y = src.y * (1.f / 255.f); + buf.z = src.z * (1.f / 255.f); - HLS2RGBConvert(buf, &buf.x); + HLS2RGBConvert(buf, &buf.x); - dst[0] = saturate_cast(buf.x*255.f); - dst[1] = saturate_cast(buf.y*255.f); - dst[2] = saturate_cast(buf.z*255.f); + dst[0] = saturate_cast(buf.x * 255.f); + dst[1] = saturate_cast(buf.y * 255.f); + dst[2] = saturate_cast(buf.z * 255.f); + } + template static __device__ uint HLS2RGBConvert(uint src) + { + float3 buf; + + buf.x = 0xff & src; + buf.y = (0xff & (src >> 8)) * (1.f / 255.f); + buf.z = (0xff & (src >> 16)) * (1.f / 255.f); + + HLS2RGBConvert(buf, &buf.x); + + uint dst = 0xffu << 24; + + dst |= saturate_cast(buf.x * 255.f); + dst |= saturate_cast(buf.y * 255.f) << 8; + dst |= saturate_cast(buf.z * 255.f) << 16; + + return dst; } - template struct HLS2RGB : public unary_function + template struct HLS2RGB : unary_function::vec_type, typename TypeVec::vec_type> { - __device__ __forceinline__ D operator()(const T& src) const + __device__ __forceinline__ typename TypeVec::vec_type operator()(const typename TypeVec::vec_type& src) const { - D dst; - HLS2RGBConvert(src, &dst.x); - setAlpha(dst, ColorChannel::elem_type>::max()); + typename TypeVec::vec_type dst; + + HLS2RGBConvert(src, &dst.x); + setAlpha(dst, ColorChannel::max()); + return dst; } }; + template struct HLS2RGB : unary_function + { + __device__ __forceinline__ uint operator()(uint src) const + { + return HLS2RGBConvert(src); + } + }; } #define OPENCV_GPU_IMPLEMENT_HLS2RGB_TRAITS(name, scn, dcn, bidx) \ template struct name ## _traits \ { \ - typedef detail::HLS2RGB::vec_type, typename TypeVec::vec_type, bidx, 180> functor_type; \ + typedef detail::HLS2RGB functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::HLS2RGB::vec_type, typename TypeVec::vec_type, bidx, 180>(); \ + return functor_type(); \ } \ }; \ template struct name ## _full_traits \ { \ - typedef detail::HLS2RGB::vec_type, typename TypeVec::vec_type, bidx, 255> functor_type; \ + typedef detail::HLS2RGB functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::HLS2RGB::vec_type, typename TypeVec::vec_type, bidx, 255>(); \ + return functor_type(); \ } \ }; \ template <> struct name ## _traits \ { \ - typedef detail::HLS2RGB::vec_type, typename TypeVec::vec_type, bidx, 360> functor_type; \ + typedef detail::HLS2RGB functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::HLS2RGB::vec_type, typename TypeVec::vec_type, bidx, 360>(); \ + return functor_type(); \ } \ }; \ template <> struct name ## _full_traits \ { \ - typedef detail::HLS2RGB::vec_type, typename TypeVec::vec_type, bidx, 360> functor_type; \ + typedef detail::HLS2RGB functor_type; \ static __host__ __device__ __forceinline__ functor_type create_functor() \ { \ - return detail::HLS2RGB::vec_type, typename TypeVec::vec_type, bidx, 360>(); \ + return functor_type(); \ } \ }; }}} diff --git a/modules/gpu/test/test_imgproc.cpp b/modules/gpu/test/test_imgproc.cpp index 9f337834e..cd9bf8410 100644 --- a/modules/gpu/test/test_imgproc.cpp +++ b/modules/gpu/test/test_imgproc.cpp @@ -1352,6 +1352,63 @@ TEST_P(CvtColor, RGB2XYZ) EXPECT_MAT_NEAR(dst_gold, dst, 1e-5); } +TEST_P(CvtColor, BGR2XYZ4) +{ + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src = img; + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_BGR2XYZ); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_BGR2XYZ, 4); + + gpuRes.download(dst); + ); + + ASSERT_EQ(4, dst.channels()); + + cv::Mat channels[4]; + cv::split(dst, channels); + cv::merge(channels, 3, dst); + + EXPECT_MAT_NEAR(dst_gold, dst, 1e-5); +} + +TEST_P(CvtColor, BGRA2XYZ4) +{ + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2BGRA); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_BGR2XYZ); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_BGR2XYZ, 4); + + gpuRes.download(dst); + ); + + ASSERT_EQ(4, dst.channels()); + + cv::Mat channels[4]; + cv::split(dst, channels); + cv::merge(channels, 3, dst); + + EXPECT_MAT_NEAR(dst_gold, dst, 1e-5); +} + TEST_P(CvtColor, XYZ2BGR) { PRINT_PARAM(devInfo); @@ -1398,6 +1455,62 @@ TEST_P(CvtColor, XYZ2RGB) EXPECT_MAT_NEAR(dst_gold, dst, 1e-5); } +TEST_P(CvtColor, XYZ42BGR) +{ + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2XYZ); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_XYZ2BGR); + + cv::Mat channels[4]; + cv::split(src, channels); + channels[3] = cv::Mat(src.size(), type, cv::Scalar::all(0)); + cv::merge(channels, 4, src); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_XYZ2BGR); + + gpuRes.download(dst); + ); + + EXPECT_MAT_NEAR(dst_gold, dst, 1e-5); +} + +TEST_P(CvtColor, XYZ42BGRA) +{ + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2XYZ); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_XYZ2BGR, 4); + + cv::Mat channels[4]; + cv::split(src, channels); + channels[3] = cv::Mat(src.size(), type, cv::Scalar::all(0)); + cv::merge(channels, 4, src); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_XYZ2BGR, 4); + + gpuRes.download(dst); + ); + + EXPECT_MAT_NEAR(dst_gold, dst, 1e-5); +} + TEST_P(CvtColor, BGR2YCrCb) { PRINT_PARAM(devInfo); @@ -1443,6 +1556,63 @@ TEST_P(CvtColor, RGB2YCrCb) EXPECT_MAT_NEAR(dst_gold, dst, 1e-5); } +TEST_P(CvtColor, BGR2YCrCb4) +{ + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src = img; + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_BGR2YCrCb); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_BGR2YCrCb, 4); + + gpuRes.download(dst); + ); + + ASSERT_EQ(4, dst.channels()); + + cv::Mat channels[4]; + cv::split(dst, channels); + cv::merge(channels, 3, dst); + + EXPECT_MAT_NEAR(dst_gold, dst, 1e-5); +} + +TEST_P(CvtColor, RGBA2YCrCb4) +{ + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2RGBA); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_BGR2YCrCb); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_BGR2YCrCb, 4); + + gpuRes.download(dst); + ); + + ASSERT_EQ(4, dst.channels()); + + cv::Mat channels[4]; + cv::split(dst, channels); + cv::merge(channels, 3, dst); + + EXPECT_MAT_NEAR(dst_gold, dst, 1e-5); +} + TEST_P(CvtColor, YCrCb2BGR) { PRINT_PARAM(devInfo); @@ -1489,6 +1659,62 @@ TEST_P(CvtColor, YCrCb2RGB) EXPECT_MAT_NEAR(dst_gold, dst, 1e-5); } +TEST_P(CvtColor, YCrCb42RGB) +{ + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2YCrCb); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_YCrCb2RGB); + + cv::Mat channels[4]; + cv::split(src, channels); + channels[3] = cv::Mat(src.size(), type, cv::Scalar::all(0)); + cv::merge(channels, 4, src); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_YCrCb2RGB); + + gpuRes.download(dst); + ); + + EXPECT_MAT_NEAR(dst_gold, dst, 1e-5); +} + +TEST_P(CvtColor, YCrCb42RGBA) +{ + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2YCrCb); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_YCrCb2RGB, 4); + + cv::Mat channels[4]; + cv::split(src, channels); + channels[3] = cv::Mat(src.size(), type, cv::Scalar::all(0)); + cv::merge(channels, 4, src); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_YCrCb2RGB, 4); + + gpuRes.download(dst); + ); + + EXPECT_MAT_NEAR(dst_gold, dst, 1e-5); +} + TEST_P(CvtColor, BGR2HSV) { if (type == CV_16U) @@ -1540,6 +1766,70 @@ TEST_P(CvtColor, RGB2HSV) EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); } +TEST_P(CvtColor, RGB2HSV4) +{ + if (type == CV_16U) + return; + + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2RGB); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_RGB2HSV); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_RGB2HSV, 4); + + gpuRes.download(dst); + ); + + ASSERT_EQ(4, dst.channels()); + + cv::Mat channels[4]; + cv::split(dst, channels); + cv::merge(channels, 3, dst); + + EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); +} + +TEST_P(CvtColor, RGBA2HSV4) +{ + if (type == CV_16U) + return; + + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2RGBA); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_RGB2HSV); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_RGB2HSV, 4); + + gpuRes.download(dst); + ); + + ASSERT_EQ(4, dst.channels()); + + cv::Mat channels[4]; + cv::split(dst, channels); + cv::merge(channels, 3, dst); + + EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); +} + TEST_P(CvtColor, BGR2HLS) { if (type == CV_16U) @@ -1591,6 +1881,70 @@ TEST_P(CvtColor, RGB2HLS) EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); } +TEST_P(CvtColor, RGB2HLS4) +{ + if (type == CV_16U) + return; + + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2RGB); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_RGB2HLS); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_RGB2HLS, 4); + + gpuRes.download(dst); + ); + + ASSERT_EQ(4, dst.channels()); + + cv::Mat channels[4]; + cv::split(dst, channels); + cv::merge(channels, 3, dst); + + EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); +} + +TEST_P(CvtColor, RGBA2HLS4) +{ + if (type == CV_16U) + return; + + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2RGBA); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_RGB2HLS); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_RGB2HLS, 4); + + gpuRes.download(dst); + ); + + ASSERT_EQ(4, dst.channels()); + + cv::Mat channels[4]; + cv::split(dst, channels); + cv::merge(channels, 3, dst); + + EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); +} + TEST_P(CvtColor, HSV2BGR) { if (type == CV_16U) @@ -1643,6 +1997,68 @@ TEST_P(CvtColor, HSV2RGB) EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); } +TEST_P(CvtColor, HSV42BGR) +{ + if (type == CV_16U) + return; + + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2HSV); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_HSV2BGR); + + cv::Mat channels[4]; + cv::split(src, channels); + channels[3] = cv::Mat(src.size(), type, cv::Scalar::all(0)); + cv::merge(channels, 4, src); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_HSV2BGR); + + gpuRes.download(dst); + ); + + EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); +} + +TEST_P(CvtColor, HSV42BGRA) +{ + if (type == CV_16U) + return; + + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2HSV); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_HSV2BGR, 4); + + cv::Mat channels[4]; + cv::split(src, channels); + channels[3] = cv::Mat(src.size(), type, cv::Scalar::all(0)); + cv::merge(channels, 4, src); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_HSV2BGR, 4); + + gpuRes.download(dst); + ); + + EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); +} + TEST_P(CvtColor, HLS2BGR) { if (type == CV_16U) @@ -1695,6 +2111,68 @@ TEST_P(CvtColor, HLS2RGB) EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); } +TEST_P(CvtColor, HLS42RGB) +{ + if (type == CV_16U) + return; + + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2HLS); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_HLS2RGB); + + cv::Mat channels[4]; + cv::split(src, channels); + channels[3] = cv::Mat(src.size(), type, cv::Scalar::all(0)); + cv::merge(channels, 4, src); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_HLS2RGB); + + gpuRes.download(dst); + ); + + EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); +} + +TEST_P(CvtColor, HLS42RGBA) +{ + if (type == CV_16U) + return; + + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2HLS); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_HLS2RGB, 4); + + cv::Mat channels[4]; + cv::split(src, channels); + channels[3] = cv::Mat(src.size(), type, cv::Scalar::all(0)); + cv::merge(channels, 4, src); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_HLS2RGB, 4); + + gpuRes.download(dst); + ); + + EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); +} + TEST_P(CvtColor, BGR2HSV_FULL) { if (type == CV_16U) @@ -1746,6 +2224,70 @@ TEST_P(CvtColor, RGB2HSV_FULL) EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); } +TEST_P(CvtColor, RGB2HSV4_FULL) +{ + if (type == CV_16U) + return; + + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2RGB); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_RGB2HSV_FULL); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_RGB2HSV_FULL, 4); + + gpuRes.download(dst); + ); + + ASSERT_EQ(4, dst.channels()); + + cv::Mat channels[4]; + cv::split(dst, channels); + cv::merge(channels, 3, dst); + + EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); +} + +TEST_P(CvtColor, RGBA2HSV4_FULL) +{ + if (type == CV_16U) + return; + + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2RGBA); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_RGB2HSV_FULL); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_RGB2HSV_FULL, 4); + + gpuRes.download(dst); + ); + + ASSERT_EQ(4, dst.channels()); + + cv::Mat channels[4]; + cv::split(dst, channels); + cv::merge(channels, 3, dst); + + EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); +} + TEST_P(CvtColor, BGR2HLS_FULL) { if (type == CV_16U) @@ -1797,6 +2339,70 @@ TEST_P(CvtColor, RGB2HLS_FULL) EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); } +TEST_P(CvtColor, RGB2HLS4_FULL) +{ + if (type == CV_16U) + return; + + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2RGB); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_RGB2HLS_FULL); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_RGB2HLS_FULL, 4); + + gpuRes.download(dst); + ); + + ASSERT_EQ(4, dst.channels()); + + cv::Mat channels[4]; + cv::split(dst, channels); + cv::merge(channels, 3, dst); + + EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); +} + +TEST_P(CvtColor, RGBA2HLS4_FULL) +{ + if (type == CV_16U) + return; + + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2RGBA); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_RGB2HLS_FULL); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_RGB2HLS_FULL, 4); + + gpuRes.download(dst); + ); + + ASSERT_EQ(4, dst.channels()); + + cv::Mat channels[4]; + cv::split(dst, channels); + cv::merge(channels, 3, dst); + + EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); +} + TEST_P(CvtColor, HSV2BGR_FULL) { if (type == CV_16U) @@ -1849,6 +2455,68 @@ TEST_P(CvtColor, HSV2RGB_FULL) EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); } +TEST_P(CvtColor, HSV42RGB_FULL) +{ + if (type == CV_16U) + return; + + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2HSV_FULL); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_HSV2RGB_FULL); + + cv::Mat channels[4]; + cv::split(src, channels); + channels[3] = cv::Mat(src.size(), type, cv::Scalar::all(0)); + cv::merge(channels, 4, src); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_HSV2RGB_FULL); + + gpuRes.download(dst); + ); + + EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); +} + +TEST_P(CvtColor, HSV42RGBA_FULL) +{ + if (type == CV_16U) + return; + + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2HSV_FULL); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_HSV2RGB_FULL, 4); + + cv::Mat channels[4]; + cv::split(src, channels); + channels[3] = cv::Mat(src.size(), type, cv::Scalar::all(0)); + cv::merge(channels, 4, src); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_HSV2RGB_FULL, 4); + + gpuRes.download(dst); + ); + + EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); +} + TEST_P(CvtColor, HLS2BGR_FULL) { if (type == CV_16U) @@ -1901,6 +2569,68 @@ TEST_P(CvtColor, HLS2RGB_FULL) EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); } +TEST_P(CvtColor, HLS42RGB_FULL) +{ + if (type == CV_16U) + return; + + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2HLS_FULL); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_HLS2RGB_FULL); + + cv::Mat channels[4]; + cv::split(src, channels); + channels[3] = cv::Mat(src.size(), type, cv::Scalar::all(0)); + cv::merge(channels, 4, src); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_HLS2RGB_FULL); + + gpuRes.download(dst); + ); + + EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); +} + +TEST_P(CvtColor, HLS42RGBA_FULL) +{ + if (type == CV_16U) + return; + + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2HLS_FULL); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_HLS2RGB_FULL, 4); + + cv::Mat channels[4]; + cv::split(src, channels); + channels[3] = cv::Mat(src.size(), type, cv::Scalar::all(0)); + cv::merge(channels, 4, src); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_HLS2RGB_FULL, 4); + + gpuRes.download(dst); + ); + + EXPECT_MAT_NEAR(dst_gold, dst, type == CV_32F ? 1e-2 : 1); +} + TEST_P(CvtColor, BGR2YUV) { PRINT_PARAM(devInfo); @@ -1969,6 +2699,62 @@ TEST_P(CvtColor, YUV2BGR) EXPECT_MAT_NEAR(dst_gold, dst, 1e-5); } +TEST_P(CvtColor, YUV42BGR) +{ + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2YUV); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_YUV2BGR); + + cv::Mat channels[4]; + cv::split(src, channels); + channels[3] = cv::Mat(src.size(), type, cv::Scalar::all(0)); + cv::merge(channels, 4, src); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_YUV2BGR); + + gpuRes.download(dst); + ); + + EXPECT_MAT_NEAR(dst_gold, dst, 1e-5); +} + +TEST_P(CvtColor, YUV42BGRA) +{ + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2YUV); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_YUV2BGR, 4); + + cv::Mat channels[4]; + cv::split(src, channels); + channels[3] = cv::Mat(src.size(), type, cv::Scalar::all(0)); + cv::merge(channels, 4, src); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_YUV2BGR, 4); + + gpuRes.download(dst); + ); + + EXPECT_MAT_NEAR(dst_gold, dst, 1e-5); +} + TEST_P(CvtColor, YUV2RGB) { PRINT_PARAM(devInfo); @@ -1992,6 +2778,63 @@ TEST_P(CvtColor, YUV2RGB) EXPECT_MAT_NEAR(dst_gold, dst, 1e-5); } +TEST_P(CvtColor, BGR2YUV4) +{ + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src = img; + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_BGR2YUV); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_BGR2YUV, 4); + + gpuRes.download(dst); + ); + + ASSERT_EQ(4, dst.channels()); + + cv::Mat channels[4]; + cv::split(dst, channels); + cv::merge(channels, 3, dst); + + EXPECT_MAT_NEAR(dst_gold, dst, 1e-5); +} + +TEST_P(CvtColor, RGBA2YUV4) +{ + PRINT_PARAM(devInfo); + PRINT_TYPE(type); + + cv::Mat src; + cv::cvtColor(img, src, CV_BGR2RGBA); + cv::Mat dst_gold; + cv::cvtColor(src, dst_gold, CV_RGB2YUV); + + cv::Mat dst; + + ASSERT_NO_THROW( + cv::gpu::GpuMat gpuRes; + + cv::gpu::cvtColor(cv::gpu::GpuMat(src), gpuRes, CV_RGB2YUV, 4); + + gpuRes.download(dst); + ); + + ASSERT_EQ(4, dst.channels()); + + cv::Mat channels[4]; + cv::split(dst, channels); + cv::merge(channels, 3, dst); + + EXPECT_MAT_NEAR(dst_gold, dst, 1e-5); +} + INSTANTIATE_TEST_CASE_P(ImgProc, CvtColor, testing::Combine( testing::ValuesIn(devices()), testing::Values(CV_8U, CV_16U, CV_32F)));