[*] Fixed #1464
[~] NCVPyramid uses tr1 and thus can be compiled with CL, commented out on linux [+] Moved reduction functors to NCVAlg [*] Warnings in NCV
This commit is contained in:
parent
325e0b1ab8
commit
2cb9192604
@ -565,14 +565,7 @@ __global__ void applyHaarClassifierClassifierParallel(Ncv32u *d_IImg, Ncv32u IIm
|
|||||||
curRootNodeOffset += NUM_THREADS_CLASSIFIERPARALLEL;
|
curRootNodeOffset += NUM_THREADS_CLASSIFIERPARALLEL;
|
||||||
}
|
}
|
||||||
|
|
||||||
struct functorAddValues
|
Ncv32f finalStageSum = subReduce<Ncv32f, functorAddValues<Ncv32f>, NUM_THREADS_CLASSIFIERPARALLEL>(curStageSum);
|
||||||
{
|
|
||||||
__device__ void reduce(Ncv32f &in1out, Ncv32f &in2)
|
|
||||||
{
|
|
||||||
in1out += in2;
|
|
||||||
}
|
|
||||||
};
|
|
||||||
Ncv32f finalStageSum = subReduce<Ncv32f, functorAddValues, NUM_THREADS_CLASSIFIERPARALLEL>(curStageSum);
|
|
||||||
|
|
||||||
if (finalStageSum < stageThreshold)
|
if (finalStageSum < stageThreshold)
|
||||||
{
|
{
|
||||||
|
@ -278,6 +278,7 @@ NCVMemStackAllocator::NCVMemStackAllocator(NCVMemoryType memT, size_t capacity,
|
|||||||
{
|
{
|
||||||
NcvBool bProperAlignment = (alignment & (alignment-1)) == 0;
|
NcvBool bProperAlignment = (alignment & (alignment-1)) == 0;
|
||||||
ncvAssertPrintCheck(bProperAlignment, "NCVMemStackAllocator ctor:: _alignment not power of 2");
|
ncvAssertPrintCheck(bProperAlignment, "NCVMemStackAllocator ctor:: _alignment not power of 2");
|
||||||
|
ncvAssertPrintCheck(memT != NCVMemoryTypeNone, "NCVMemStackAllocator ctor:: Incorrect allocator type");
|
||||||
|
|
||||||
allocBegin = NULL;
|
allocBegin = NULL;
|
||||||
|
|
||||||
@ -295,6 +296,7 @@ NCVMemStackAllocator::NCVMemStackAllocator(NCVMemoryType memT, size_t capacity,
|
|||||||
case NCVMemoryTypeHostPageable:
|
case NCVMemoryTypeHostPageable:
|
||||||
allocBegin = (Ncv8u *)malloc(capacity);
|
allocBegin = (Ncv8u *)malloc(capacity);
|
||||||
break;
|
break;
|
||||||
|
default:;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
else
|
else
|
||||||
@ -335,6 +337,7 @@ NCVMemStackAllocator::~NCVMemStackAllocator()
|
|||||||
case NCVMemoryTypeHostPageable:
|
case NCVMemoryTypeHostPageable:
|
||||||
free(allocBegin);
|
free(allocBegin);
|
||||||
break;
|
break;
|
||||||
|
default:;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -455,6 +458,7 @@ NCVStatus NCVMemNativeAllocator::alloc(NCVMemSegment &seg, size_t size)
|
|||||||
case NCVMemoryTypeHostPageable:
|
case NCVMemoryTypeHostPageable:
|
||||||
seg.begin.ptr = (Ncv8u *)malloc(size);
|
seg.begin.ptr = (Ncv8u *)malloc(size);
|
||||||
break;
|
break;
|
||||||
|
default:;
|
||||||
}
|
}
|
||||||
|
|
||||||
this->currentSize += alignUp(size, this->_alignment);
|
this->currentSize += alignUp(size, this->_alignment);
|
||||||
@ -487,6 +491,7 @@ NCVStatus NCVMemNativeAllocator::dealloc(NCVMemSegment &seg)
|
|||||||
case NCVMemoryTypeHostPageable:
|
case NCVMemoryTypeHostPageable:
|
||||||
free(seg.begin.ptr);
|
free(seg.begin.ptr);
|
||||||
break;
|
break;
|
||||||
|
default:;
|
||||||
}
|
}
|
||||||
|
|
||||||
seg.clear();
|
seg.clear();
|
||||||
|
@ -52,6 +52,36 @@ static T divUp(T a, T b)
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
template<typename T>
|
||||||
|
struct functorAddValues
|
||||||
|
{
|
||||||
|
static __device__ __inline__ void reduce(T &in1out, T &in2)
|
||||||
|
{
|
||||||
|
in1out += in2;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
|
||||||
|
template<typename T>
|
||||||
|
struct functorMinValues
|
||||||
|
{
|
||||||
|
static __device__ __inline__ void reduce(T &in1out, T &in2)
|
||||||
|
{
|
||||||
|
in1out = in1out > in2 ? in2 : in1out;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
|
||||||
|
template<typename T>
|
||||||
|
struct functorMaxValues
|
||||||
|
{
|
||||||
|
static __device__ __inline__ void reduce(T &in1out, T &in2)
|
||||||
|
{
|
||||||
|
in1out = in1out > in2 ? in1out : in2;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
|
||||||
template<typename Tdata, class Tfunc, Ncv32u nThreads>
|
template<typename Tdata, class Tfunc, Ncv32u nThreads>
|
||||||
static __device__ Tdata subReduce(Tdata threadElem)
|
static __device__ Tdata subReduce(Tdata threadElem)
|
||||||
{
|
{
|
||||||
|
@ -46,6 +46,7 @@
|
|||||||
#include "NCVPyramid.hpp"
|
#include "NCVPyramid.hpp"
|
||||||
#include "NCVPixelOperations.hpp"
|
#include "NCVPixelOperations.hpp"
|
||||||
|
|
||||||
|
#ifdef _WIN32
|
||||||
|
|
||||||
template<typename T, Ncv32u CN> struct __average4_CN {static T _average4_CN(const T &p00, const T &p01, const T &p10, const T &p11);};
|
template<typename T, Ncv32u CN> struct __average4_CN {static T _average4_CN(const T &p00, const T &p01, const T &p10, const T &p11);};
|
||||||
|
|
||||||
@ -395,3 +396,5 @@ template class NCVImagePyramid<uint4>;
|
|||||||
template class NCVImagePyramid<float1>;
|
template class NCVImagePyramid<float1>;
|
||||||
template class NCVImagePyramid<float3>;
|
template class NCVImagePyramid<float3>;
|
||||||
template class NCVImagePyramid<float4>;
|
template class NCVImagePyramid<float4>;
|
||||||
|
|
||||||
|
#endif //_WIN32
|
||||||
|
@ -47,6 +47,7 @@
|
|||||||
#include <vector>
|
#include <vector>
|
||||||
#include "NCV.hpp"
|
#include "NCV.hpp"
|
||||||
|
|
||||||
|
#ifdef _WIN32
|
||||||
|
|
||||||
template <class T>
|
template <class T>
|
||||||
class NCV_EXPORTS NCVMatrixStack
|
class NCV_EXPORTS NCVMatrixStack
|
||||||
@ -93,5 +94,6 @@ private:
|
|||||||
Ncv32u nLayers;
|
Ncv32u nLayers;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
#endif //_WIN32
|
||||||
|
|
||||||
#endif //_ncvpyramid_hpp_
|
#endif //_ncvpyramid_hpp_
|
||||||
|
Loading…
x
Reference in New Issue
Block a user