Merged the branch /branches/opencv_pthread_framework/opencv into the trunk.

The branch contains changes, concerning adding ThreadingFramework -- temporary lightweight decision for ARM instead of Intel TBB.
Also some changes in Android highgui camera class were made.
This commit is contained in:
Leonid Beynenson 2011-04-29 14:20:24 +00:00
parent e202b13069
commit 2806db93d8
6 changed files with 127 additions and 9 deletions

View File

@ -683,6 +683,14 @@ if (WITH_TBB)
endif() endif()
endif(WITH_TBB) endif(WITH_TBB)
#Threading Framework -- temporary decision for ARM-s instead of TBB
if (NOT HAVE_TBB)
file(GLOB THREADING_FRAMEWORK_HEADER "${CMAKE_SOURCE_DIR}/modules/core/include/opencv2/core/threading_framework.hpp")
file(GLOB THREADING_FRAMEWORK_SOURCE "${CMAKE_SOURCE_DIR}/modules/core/src/threading_framework.cpp")
if(THREADING_FRAMEWORK_HEADER AND THREADING_FRAMEWORK_SOURCE)
set(HAVE_THREADING_FRAMEWORK 1)
endif()
endif()
############################ Intel IPP ############################# ############################ Intel IPP #############################
set(IPP_FOUND) set(IPP_FOUND)
@ -1453,6 +1461,11 @@ if(HAVE_TBB)
message(STATUS " Use TBB: YES") message(STATUS " Use TBB: YES")
else() else()
message(STATUS " Use TBB: NO") message(STATUS " Use TBB: NO")
if(HAVE_THREADING_FRAMEWORK)
message(STATUS " Use ThreadingFramework: YES")
else()
message(STATUS " Use ThreadingFramework: NO")
endif()
endif() endif()
if (HAVE_CUDA) if (HAVE_CUDA)

View File

@ -157,6 +157,9 @@
/* Intel Threading Building Blocks */ /* Intel Threading Building Blocks */
#cmakedefine HAVE_TBB #cmakedefine HAVE_TBB
/* Threading Framework --- temporary decision for ARM instead of Intel TBB */
#cmakedefine HAVE_THREADING_FRAMEWORK
/* Eigen2 Matrix & Linear Algebra Library */ /* Eigen2 Matrix & Linear Algebra Library */
#cmakedefine HAVE_EIGEN2 #cmakedefine HAVE_EIGEN2

View File

@ -185,11 +185,24 @@ CV_INLINE IppiSize ippiSize(int width, int height)
int _begin, _end, _grainsize; int _begin, _end, _grainsize;
}; };
#ifdef HAVE_THREADING_FRAMEWORK
#include "threading_framework.hpp"
template<typename Body>
static void parallel_for( const BlockedRange& range, const Body& body )
{
tf::parallel_for<Body>(range, body);
}
typedef tf::ConcurrentVector<Rect> ConcurrentRectVector;
#else
template<typename Body> static inline template<typename Body> static inline
void parallel_for( const BlockedRange& range, const Body& body ) void parallel_for( const BlockedRange& range, const Body& body )
{ {
body(range); body(range);
} }
typedef std::vector<Rect> ConcurrentRectVector;
#endif
template<typename Iterator, typename Body> static inline template<typename Iterator, typename Body> static inline
void parallel_do( Iterator first, Iterator last, const Body& body ) void parallel_do( Iterator first, Iterator last, const Body& body )
@ -206,7 +219,6 @@ CV_INLINE IppiSize ippiSize(int width, int height)
body(range); body(range);
} }
typedef std::vector<Rect> ConcurrentRectVector;
} }
#endif #endif
#endif #endif

View File

@ -396,7 +396,8 @@ enum
enum enum
{ {
CV_CAP_ANDROID_COLOR_FRAME = 0, //BGR CV_CAP_ANDROID_COLOR_FRAME = 0, //BGR
CV_CAP_ANDROID_GREY_FRAME = 1 //Y CV_CAP_ANDROID_GREY_FRAME = 1, //Y
CV_CAP_ANDROID_COLOR_FRAME_RGB = 2 //RGB
}; };
/* retrieve or set capture properties */ /* retrieve or set capture properties */

View File

@ -110,6 +110,7 @@ private:
void prepareCacheForYUV420i(int width, int height); void prepareCacheForYUV420i(int width, int height);
static bool convertYUV420i2Grey(int width, int height, const unsigned char* yuv, cv::Mat& resmat); static bool convertYUV420i2Grey(int width, int height, const unsigned char* yuv, cv::Mat& resmat);
static bool convertYUV420i2BGR888(int width, int height, const unsigned char* yuv, cv::Mat& resmat); static bool convertYUV420i2BGR888(int width, int height, const unsigned char* yuv, cv::Mat& resmat);
static bool convertYUV420i2RGB888(int width, int height, const unsigned char* yuv, cv::Mat& resmat);
friend class HighguiAndroidCameraActivity; friend class HighguiAndroidCameraActivity;
@ -209,13 +210,18 @@ CvCapture_Android::~CvCapture_Android()
((HighguiAndroidCameraActivity*)m_activity)->LogFramesRate(); ((HighguiAndroidCameraActivity*)m_activity)->LogFramesRate();
//m_activity->disconnect() will be automatically called inside destructor; //m_activity->disconnect() will be automatically called inside destructor;
delete m_activity;
delete m_frameYUV420i; delete m_frameYUV420i;
delete m_frameYUV420inext; delete m_frameYUV420inext;
m_activity = 0;
m_frameYUV420i = 0; m_frameYUV420i = 0;
m_frameYUV420inext = 0; m_frameYUV420inext = 0;
pthread_mutex_lock(&m_nextFrameMutex);
pthread_cond_broadcast(&m_nextFrameCond);
pthread_mutex_unlock(&m_nextFrameMutex);
delete m_activity;
m_activity = 0;
pthread_mutex_destroy(&m_nextFrameMutex); pthread_mutex_destroy(&m_nextFrameMutex);
pthread_cond_destroy(&m_nextFrameCond); pthread_cond_destroy(&m_nextFrameCond);
} }
@ -223,7 +229,7 @@ CvCapture_Android::~CvCapture_Android()
double CvCapture_Android::getProperty( int propIdx ) double CvCapture_Android::getProperty( int propIdx )
{ {
switch ( propIdx ) switch ( propIdx )
{ {
case CV_CAP_PROP_FRAME_WIDTH: case CV_CAP_PROP_FRAME_WIDTH:
return (double)m_activity->getFrameWidth(); return (double)m_activity->getFrameWidth();
@ -272,7 +278,11 @@ bool CvCapture_Android::grabFrame()
} }
m_waitingNextFrame = true; m_waitingNextFrame = true;
pthread_cond_wait(&m_nextFrameCond, &m_nextFrameMutex); pthread_cond_wait(&m_nextFrameCond, &m_nextFrameMutex);
pthread_mutex_unlock(&m_nextFrameMutex); int res=pthread_mutex_unlock(&m_nextFrameMutex);
if (res) {
LOGE("Error in CvCapture_Android::grabFrame: pthread_mutex_unlock returned %d --- probably, this object has been destroyed", res);
return false;
}
m_framesGrabbed++; m_framesGrabbed++;
return true; return true;
@ -297,6 +307,12 @@ IplImage* CvCapture_Android::retrieveFrame( int outputType )
return 0; return 0;
image = m_frameGray.getIplImagePtr(); image = m_frameGray.getIplImagePtr();
break; break;
case CV_CAP_ANDROID_COLOR_FRAME_RGB:
if (!m_hasColor)
if (!(m_hasColor = convertYUV420i2RGB888(m_width, m_height, m_frameYUV420i, m_frameColor.mat)))
return 0;
image = m_frameColor.getIplImagePtr();
break;
default: default:
LOGE("Unsupported frame output format: %d", outputType); LOGE("Unsupported frame output format: %d", outputType);
CV_Error( CV_StsOutOfRange, "Output frame format is not supported." ); CV_Error( CV_StsOutOfRange, "Output frame format is not supported." );
@ -365,7 +381,7 @@ bool CvCapture_Android::convertYUV420i2Grey(int width, int height, const unsigne
{ {
if (yuv == 0) return false; if (yuv == 0) return false;
resmat.create(width, height, CV_8UC1); resmat.create(height, width, CV_8UC1);
unsigned char* matBuff = resmat.ptr<unsigned char> (0); unsigned char* matBuff = resmat.ptr<unsigned char> (0);
memcpy(matBuff, yuv, width * height); memcpy(matBuff, yuv, width * height);
return !resmat.empty(); return !resmat.empty();
@ -444,6 +460,79 @@ bool CvCapture_Android::convertYUV420i2BGR888(int width, int height, const unsig
return !resmat.empty(); return !resmat.empty();
} }
bool CvCapture_Android::convertYUV420i2RGB888(int width, int height, const unsigned char* yuv, cv::Mat& resmat) ///TODO: FIXME: copiued from BGR variant
{
if (yuv == 0) return false;
CV_Assert(width % 2 == 0 && height % 2 == 0);
resmat.create(height, width, CV_8UC3);
unsigned char* y1 = (unsigned char*)yuv;
unsigned char* uv = y1 + width * height;
//B = 1.164(Y - 16) + 2.018(U - 128)
//G = 1.164(Y - 16) - 0.813(V - 128) - 0.391(U - 128)
//R = 1.164(Y - 16) + 1.596(V - 128)
for (int j = 0; j < height; j+=2, y1+=width*2, uv+=width)
{
unsigned char* row1 = resmat.ptr<unsigned char>(j);
unsigned char* row2 = resmat.ptr<unsigned char>(j+1);
unsigned char* y2 = y1 + width;
for(int i = 0; i < width; i+=2,row1+=6,row2+=6)
{
// unsigned char cr = uv[i];
// unsigned char cb = uv[i+1];
// row1[0] = y1[i];
// row1[1] = cr;
// row1[2] = cb;
// row1[3] = y1[i+1];
// row1[4] = cr;
// row1[5] = cb;
// row2[0] = y2[i];
// row2[1] = cr;
// row2[2] = cb;
// row2[3] = y2[i+1];
// row2[4] = cr;
// row2[5] = cb;
int cr = uv[i] - 128;
int cb = uv[i+1] - 128;
int buv = 409 * cr + 128;
int guv = 128 - 100 * cb - 208 * cr;
int ruv = 516 * cb + 128;
int y00 = (y1[i] - 16) * 298;
row1[0] = clamp((y00 + buv) >> 8);
row1[1] = clamp((y00 + guv) >> 8);
row1[2] = clamp((y00 + ruv) >> 8);
int y01 = (y1[i+1] - 16) * 298;
row1[3] = clamp((y01 + buv) >> 8);
row1[4] = clamp((y01 + guv) >> 8);
row1[5] = clamp((y01 + ruv) >> 8);
int y10 = (y2[i] - 16) * 298;
row2[0] = clamp((y10 + buv) >> 8);
row2[1] = clamp((y10 + guv) >> 8);
row2[2] = clamp((y10 + ruv) >> 8);
int y11 = (y2[i+1] - 16) * 298;
row2[3] = clamp((y11 + buv) >> 8);
row2[4] = clamp((y11 + guv) >> 8);
row2[5] = clamp((y11 + ruv) >> 8);
}
}
return !resmat.empty();
}
CvCapture* cvCreateCameraCapture_Android( int cameraId ) CvCapture* cvCreateCameraCapture_Android( int cameraId )
{ {

View File

@ -1195,7 +1195,7 @@ void CascadeClassifier::detectMultiScale( const Mat& image, vector<Rect>& object
int yStep = factor > 2. ? 1 : 2; int yStep = factor > 2. ? 1 : 2;
int stripCount, stripSize; int stripCount, stripSize;
#ifdef HAVE_TBB #if defined(HAVE_TBB) || defined(HAVE_THREADING_FRAMEWORK)
const int PTS_PER_THREAD = 1000; const int PTS_PER_THREAD = 1000;
stripCount = ((processingRectSize.width/yStep)*(processingRectSize.height + yStep-1)/yStep + PTS_PER_THREAD/2)/PTS_PER_THREAD; stripCount = ((processingRectSize.width/yStep)*(processingRectSize.height + yStep-1)/yStep + PTS_PER_THREAD/2)/PTS_PER_THREAD;
stripCount = std::min(std::max(stripCount, 1), 100); stripCount = std::min(std::max(stripCount, 1), 100);