Android native camera: added BGR output format; added methods to configure output frame size.
This commit is contained in:
parent
1965b297d0
commit
31e77a3bd9
@ -392,9 +392,8 @@ enum
|
||||
//supported by Android camera output formats
|
||||
enum
|
||||
{
|
||||
CV_CAP_ANDROID_COLOR_FRAME = 1, //TODO: check RGB or BGR?
|
||||
CV_CAP_ANDROID_GREY_FRAME = 0,
|
||||
CV_CAP_ANDROID_YUV_FRAME = 2
|
||||
CV_CAP_ANDROID_COLOR_FRAME = 0, //BGR
|
||||
CV_CAP_ANDROID_GREY_FRAME = 1 //Y
|
||||
};
|
||||
|
||||
/* retrieve or set capture properties */
|
||||
|
@ -1,4 +1,4 @@
|
||||
/*M///////////////////////////////////////////////////////////////////////////////////////
|
||||
/*M///////////////////////////////////////////////////////////////////////////////////////
|
||||
//
|
||||
// IMPORTANT: READ BEFORE DOWNLOADING, COPYING, INSTALLING OR USING.
|
||||
//
|
||||
@ -58,7 +58,7 @@ class HighguiAndroidCameraActivity;
|
||||
class CvCapture_Android : public CvCapture
|
||||
{
|
||||
public:
|
||||
CvCapture_Android();
|
||||
CvCapture_Android(int);
|
||||
virtual ~CvCapture_Android();
|
||||
|
||||
virtual double getProperty(int propIdx);
|
||||
@ -81,66 +81,74 @@ protected:
|
||||
|
||||
CameraActivity* m_activity;
|
||||
|
||||
//raw from camera
|
||||
int m_width;
|
||||
int m_height;
|
||||
unsigned char *m_frameYUV420i;
|
||||
unsigned char *m_frameYUV420inext;
|
||||
|
||||
void setFrame(const void* buffer, int bufferSize);
|
||||
|
||||
private:
|
||||
bool m_isOpened;
|
||||
bool m_CameraParamsChanged;
|
||||
|
||||
OutputMap *m_frameYUV;
|
||||
OutputMap *m_frameYUVnext;
|
||||
//frames counter for statistics
|
||||
int m_framesGrabbed;
|
||||
|
||||
//cached converted frames
|
||||
OutputMap m_frameGray;
|
||||
OutputMap m_frameColor;
|
||||
bool m_hasGray;
|
||||
bool m_hasColor;
|
||||
|
||||
//synchronization
|
||||
pthread_mutex_t m_nextFrameMutex;
|
||||
pthread_cond_t m_nextFrameCond;
|
||||
volatile bool m_waitingNextFrame;
|
||||
|
||||
int m_framesGrabbed;
|
||||
void prepareCacheForYUV420i(int width, int height);
|
||||
static bool convertYUV420i2Grey(int width, int height, const unsigned char* yuv, cv::Mat& resmat);
|
||||
static bool convertYUV420i2BGR888(int width, int height, const unsigned char* yuv, cv::Mat& resmat);
|
||||
|
||||
|
||||
friend class HighguiAndroidCameraActivity;
|
||||
|
||||
void onFrame(const void* buffer, int bufferSize);
|
||||
|
||||
void convertBufferToYUV(const void* buffer, int size, int width, int height);
|
||||
static bool convertYUVToGrey(const cv::Mat& yuv, cv::Mat& resmat);
|
||||
static bool convertYUVToColor(const cv::Mat& yuv, cv::Mat& resmat);
|
||||
};
|
||||
|
||||
|
||||
class HighguiAndroidCameraActivity : public CameraActivity
|
||||
{
|
||||
public:
|
||||
public:
|
||||
HighguiAndroidCameraActivity(CvCapture_Android* capture)
|
||||
{
|
||||
m_capture = capture;
|
||||
m_framesReceived = 0;
|
||||
m_capture = capture;
|
||||
m_framesReceived = 0;
|
||||
}
|
||||
|
||||
virtual bool onFrameBuffer(void* buffer, int bufferSize)
|
||||
{
|
||||
LOGD("buffer addr:%p size:%d",buffer, bufferSize);
|
||||
if(isConnected() && buffer != 0 && bufferSize > 0)
|
||||
{
|
||||
m_framesReceived++;
|
||||
if (m_capture->m_waitingNextFrame)
|
||||
{
|
||||
m_capture->onFrame(buffer, bufferSize);
|
||||
pthread_mutex_lock(&m_capture->m_nextFrameMutex);
|
||||
m_capture->m_waitingNextFrame = false;//set flag that no more frames required at this moment
|
||||
pthread_cond_broadcast(&m_capture->m_nextFrameCond);
|
||||
pthread_mutex_unlock(&m_capture->m_nextFrameMutex);
|
||||
}
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
if(isConnected() && buffer != 0 && bufferSize > 0)
|
||||
{
|
||||
m_framesReceived++;
|
||||
if (m_capture->m_waitingNextFrame)
|
||||
{
|
||||
m_capture->setFrame(buffer, bufferSize);
|
||||
pthread_mutex_lock(&m_capture->m_nextFrameMutex);
|
||||
m_capture->m_waitingNextFrame = false;//set flag that no more frames required at this moment
|
||||
pthread_cond_broadcast(&m_capture->m_nextFrameCond);
|
||||
pthread_mutex_unlock(&m_capture->m_nextFrameMutex);
|
||||
}
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
}
|
||||
|
||||
void LogFramesRate()
|
||||
{
|
||||
LOGI("FRAMES received: %d grabbed: %d", m_framesReceived, m_capture->m_framesGrabbed);
|
||||
LOGI("FRAMES received: %d grabbed: %d", m_framesReceived, m_capture->m_framesGrabbed);
|
||||
}
|
||||
|
||||
private:
|
||||
private:
|
||||
CvCapture_Android* m_capture;
|
||||
int m_framesReceived;
|
||||
};
|
||||
@ -154,195 +162,292 @@ IplImage* CvCapture_Android::OutputMap::getIplImagePtr()
|
||||
return &iplHeader;
|
||||
}
|
||||
|
||||
CvCapture_Android::CvCapture_Android(int cameraId)
|
||||
{
|
||||
//defaults
|
||||
m_width = 0;
|
||||
m_height = 0;
|
||||
m_activity = 0;
|
||||
m_isOpened = false;
|
||||
m_frameYUV420i = 0;
|
||||
m_frameYUV420inext = 0;
|
||||
m_hasGray = false;
|
||||
m_hasColor = false;
|
||||
m_waitingNextFrame = false;
|
||||
m_framesGrabbed = 0;
|
||||
m_CameraParamsChanged = false;
|
||||
|
||||
//try connect to camera
|
||||
m_activity = new HighguiAndroidCameraActivity(this);
|
||||
|
||||
if (m_activity == 0) return;
|
||||
|
||||
pthread_mutex_init(&m_nextFrameMutex, NULL);
|
||||
pthread_cond_init (&m_nextFrameCond, NULL);
|
||||
|
||||
CameraActivity::ErrorCode errcode = m_activity->connect(cameraId);
|
||||
|
||||
if(errcode == CameraActivity::NO_ERROR)
|
||||
m_isOpened = true;
|
||||
else
|
||||
{
|
||||
LOGE("Native_camera returned opening error: %d", errcode);
|
||||
delete m_activity;
|
||||
m_activity = 0;
|
||||
}
|
||||
}
|
||||
|
||||
bool CvCapture_Android::isOpened() const
|
||||
{
|
||||
return m_isOpened;
|
||||
}
|
||||
|
||||
CvCapture_Android::CvCapture_Android()
|
||||
{
|
||||
//defaults
|
||||
m_activity = 0;
|
||||
m_isOpened = false;
|
||||
m_frameYUV = 0;
|
||||
m_frameYUVnext = 0;
|
||||
m_hasGray = false;
|
||||
m_hasColor = false;
|
||||
m_waitingNextFrame = false;
|
||||
m_framesGrabbed = 0;
|
||||
|
||||
//try connect to camera
|
||||
m_activity = new HighguiAndroidCameraActivity(this);
|
||||
|
||||
if (m_activity == 0) return;
|
||||
pthread_mutex_init(&m_nextFrameMutex, NULL);
|
||||
pthread_cond_init (&m_nextFrameCond, NULL);
|
||||
|
||||
CameraActivity::ErrorCode errcode = m_activity->connect();
|
||||
if(errcode == CameraActivity::NO_ERROR)
|
||||
{
|
||||
m_isOpened = true;
|
||||
m_frameYUV = new OutputMap();
|
||||
m_frameYUVnext = new OutputMap();
|
||||
}
|
||||
else
|
||||
{
|
||||
LOGE("Native_camera returned opening error: %d", errcode);
|
||||
delete m_activity;
|
||||
m_activity = 0;
|
||||
}
|
||||
}
|
||||
|
||||
CvCapture_Android::~CvCapture_Android()
|
||||
{
|
||||
if (m_activity)
|
||||
{
|
||||
((HighguiAndroidCameraActivity*)m_activity)->LogFramesRate();
|
||||
if (m_activity)
|
||||
{
|
||||
((HighguiAndroidCameraActivity*)m_activity)->LogFramesRate();
|
||||
|
||||
//m_activity->disconnect() will be automatically called inside destructor;
|
||||
delete m_activity;
|
||||
delete m_frameYUV;
|
||||
delete m_frameYUVnext;
|
||||
m_activity = 0;
|
||||
m_frameYUV = 0;
|
||||
m_frameYUVnext = 0;
|
||||
|
||||
pthread_mutex_destroy(&m_nextFrameMutex);
|
||||
pthread_cond_destroy(&m_nextFrameCond);
|
||||
}
|
||||
//m_activity->disconnect() will be automatically called inside destructor;
|
||||
delete m_activity;
|
||||
delete m_frameYUV420i;
|
||||
delete m_frameYUV420inext;
|
||||
m_activity = 0;
|
||||
m_frameYUV420i = 0;
|
||||
m_frameYUV420inext = 0;
|
||||
|
||||
pthread_mutex_destroy(&m_nextFrameMutex);
|
||||
pthread_cond_destroy(&m_nextFrameCond);
|
||||
}
|
||||
}
|
||||
|
||||
double CvCapture_Android::getProperty( int propIdx )
|
||||
{
|
||||
switch ( propIdx )
|
||||
{
|
||||
switch ( propIdx )
|
||||
{
|
||||
case CV_CAP_PROP_FRAME_WIDTH:
|
||||
return (double)CameraActivity::getFrameWidth();
|
||||
return (double)m_activity->getFrameWidth();
|
||||
case CV_CAP_PROP_FRAME_HEIGHT:
|
||||
return (double)CameraActivity::getFrameHeight();
|
||||
return (double)m_activity->getFrameHeight();
|
||||
default:
|
||||
CV_Error( CV_StsError, "Failed attempt to GET unsupported camera property." );
|
||||
break;
|
||||
}
|
||||
return -1.0;
|
||||
CV_Error( CV_StsOutOfRange, "Failed attempt to GET unsupported camera property." );
|
||||
break;
|
||||
}
|
||||
return -1.0;
|
||||
}
|
||||
|
||||
bool CvCapture_Android::setProperty( int propIdx, double propValue )
|
||||
{
|
||||
bool res = false;
|
||||
if( isOpened() )
|
||||
{
|
||||
switch ( propIdx )
|
||||
bool res = false;
|
||||
if( isOpened() )
|
||||
{
|
||||
default:
|
||||
CV_Error( CV_StsError, "Failed attempt to SET unsupported camera property." );
|
||||
break;
|
||||
switch ( propIdx )
|
||||
{
|
||||
case CV_CAP_PROP_FRAME_WIDTH:
|
||||
m_activity->setProperty(ANDROID_CAMERA_PROPERTY_FRAMEWIDTH, propValue);
|
||||
break;
|
||||
case CV_CAP_PROP_FRAME_HEIGHT:
|
||||
m_activity->setProperty(ANDROID_CAMERA_PROPERTY_FRAMEHEIGHT, propValue);
|
||||
break;
|
||||
default:
|
||||
CV_Error( CV_StsOutOfRange, "Failed attempt to SET unsupported camera property." );
|
||||
break;
|
||||
}
|
||||
m_CameraParamsChanged = true;
|
||||
}
|
||||
}
|
||||
|
||||
return res;
|
||||
return res;
|
||||
}
|
||||
|
||||
bool CvCapture_Android::grabFrame()
|
||||
{
|
||||
if( !isOpened() )
|
||||
return false;
|
||||
if( !isOpened() )
|
||||
return false;
|
||||
|
||||
pthread_mutex_lock(&m_nextFrameMutex);
|
||||
m_waitingNextFrame = true;
|
||||
pthread_cond_wait(&m_nextFrameCond, &m_nextFrameMutex);
|
||||
pthread_mutex_unlock(&m_nextFrameMutex);
|
||||
m_framesGrabbed++;
|
||||
return true;
|
||||
}
|
||||
pthread_mutex_lock(&m_nextFrameMutex);
|
||||
if (m_CameraParamsChanged)
|
||||
{
|
||||
m_activity->applyProperties();
|
||||
m_CameraParamsChanged = false;
|
||||
}
|
||||
m_waitingNextFrame = true;
|
||||
pthread_cond_wait(&m_nextFrameCond, &m_nextFrameMutex);
|
||||
pthread_mutex_unlock(&m_nextFrameMutex);
|
||||
|
||||
void CvCapture_Android::onFrame(const void* buffer, int bufferSize)
|
||||
{
|
||||
LOGD("Buffer available: %p + %d", buffer, bufferSize);
|
||||
|
||||
convertBufferToYUV(buffer, bufferSize, CameraActivity::getFrameWidth(), CameraActivity::getFrameHeight());
|
||||
|
||||
//swap current and new frames
|
||||
OutputMap* tmp = m_frameYUV;
|
||||
m_frameYUV = m_frameYUVnext;
|
||||
m_frameYUVnext = tmp;
|
||||
|
||||
//discard cached frames
|
||||
m_hasGray = false;
|
||||
m_hasColor = false;
|
||||
m_framesGrabbed++;
|
||||
return true;
|
||||
}
|
||||
|
||||
IplImage* CvCapture_Android::retrieveFrame( int outputType )
|
||||
{
|
||||
IplImage* image = 0;
|
||||
if (0 != m_frameYUV && !m_frameYUV->mat.empty())
|
||||
{
|
||||
switch(outputType)
|
||||
IplImage* image = 0;
|
||||
if (0 != m_frameYUV420i)
|
||||
{
|
||||
case CV_CAP_ANDROID_YUV_FRAME:
|
||||
image = m_frameYUV->getIplImagePtr();
|
||||
break;
|
||||
case CV_CAP_ANDROID_GREY_FRAME:
|
||||
if (!m_hasGray)
|
||||
if (!(m_hasGray = convertYUVToGrey(m_frameYUV->mat, m_frameGray.mat)))
|
||||
image = 0;
|
||||
image = m_frameGray.getIplImagePtr();
|
||||
break;
|
||||
case CV_CAP_ANDROID_COLOR_FRAME:
|
||||
if (!m_hasColor)
|
||||
if (!(m_hasColor = convertYUVToColor(m_frameYUV->mat, m_frameColor.mat)))
|
||||
image = 0;
|
||||
image = m_frameColor.getIplImagePtr();
|
||||
break;
|
||||
default:
|
||||
LOGE("Unsupported frame output format: %d", outputType);
|
||||
image = 0;
|
||||
break;
|
||||
switch(outputType)
|
||||
{
|
||||
case CV_CAP_ANDROID_COLOR_FRAME:
|
||||
if (!m_hasColor)
|
||||
if (!(m_hasColor = convertYUV420i2BGR888(m_width, m_height, m_frameYUV420i, m_frameColor.mat)))
|
||||
return 0;
|
||||
image = m_frameColor.getIplImagePtr();
|
||||
break;
|
||||
case CV_CAP_ANDROID_GREY_FRAME:
|
||||
if (!m_hasGray)
|
||||
if (!(m_hasGray = convertYUV420i2Grey(m_width, m_height, m_frameYUV420i, m_frameGray.mat)))
|
||||
return 0;
|
||||
image = m_frameGray.getIplImagePtr();
|
||||
break;
|
||||
default:
|
||||
LOGE("Unsupported frame output format: %d", outputType);
|
||||
CV_Error( CV_StsOutOfRange, "Output frame format is not supported." );
|
||||
image = 0;
|
||||
break;
|
||||
}
|
||||
}
|
||||
}
|
||||
return image;
|
||||
return image;
|
||||
}
|
||||
|
||||
void CvCapture_Android::setFrame(const void* buffer, int bufferSize)
|
||||
{
|
||||
int width = m_activity->getFrameWidth();
|
||||
int height = m_activity->getFrameHeight();
|
||||
int expectedSize = (width * height * 3) >> 1;
|
||||
|
||||
if ( expectedSize != bufferSize)
|
||||
{
|
||||
LOGE("ERROR reading YUV420i buffer: width=%d, height=%d, size=%d, receivedSize=%d", width, height, expectedSize, bufferSize);
|
||||
return;
|
||||
}
|
||||
|
||||
//allocate memery if needed
|
||||
prepareCacheForYUV420i(width, height);
|
||||
|
||||
//copy data
|
||||
memcpy(m_frameYUV420inext, buffer, bufferSize);
|
||||
|
||||
//swap current and new frames
|
||||
unsigned char* tmp = m_frameYUV420i;
|
||||
m_frameYUV420i = m_frameYUV420inext;
|
||||
m_frameYUV420inext = tmp;
|
||||
|
||||
//discard cached frames
|
||||
m_hasGray = false;
|
||||
m_hasColor = false;
|
||||
}
|
||||
|
||||
void CvCapture_Android::prepareCacheForYUV420i(int width, int height)
|
||||
{
|
||||
if (width != m_width || height != m_height)
|
||||
{
|
||||
m_width = width;
|
||||
m_height = height;
|
||||
unsigned char *tmp = m_frameYUV420inext;
|
||||
m_frameYUV420inext = new unsigned char [width * height * 3 / 2];
|
||||
delete[] tmp;
|
||||
|
||||
tmp = m_frameYUV420i;
|
||||
m_frameYUV420i = new unsigned char [width * height * 3 / 2];
|
||||
delete[] tmp;
|
||||
}
|
||||
}
|
||||
|
||||
inline unsigned char clamp(int value)
|
||||
{
|
||||
if (value <= 0)
|
||||
return 0;
|
||||
if (value >= 255)
|
||||
return (unsigned char)255;
|
||||
return (unsigned char)value;
|
||||
}
|
||||
|
||||
|
||||
void CvCapture_Android::convertBufferToYUV(const void* buffer, int size, int width, int height)
|
||||
bool CvCapture_Android::convertYUV420i2Grey(int width, int height, const unsigned char* yuv, cv::Mat& resmat)
|
||||
{
|
||||
cv::Size buffSize(width, height + (height / 2));
|
||||
if (buffSize.area() != size)
|
||||
{
|
||||
LOGE("ERROR convertBufferToYuv_Mat: width=%d, height=%d, buffSize=%d x %d, buffSize.area()=%d, size=%d",
|
||||
width, height, buffSize.width, buffSize.height, buffSize.area(), size);
|
||||
|
||||
return;
|
||||
}
|
||||
if (yuv == 0) return false;
|
||||
|
||||
m_frameYUVnext->mat.create(buffSize, CV_8UC1);
|
||||
uchar* matBuff = m_frameYUVnext->mat.ptr<uchar> (0);
|
||||
memcpy(matBuff, buffer, size);
|
||||
resmat.create(width, height, CV_8UC1);
|
||||
unsigned char* matBuff = resmat.ptr<unsigned char> (0);
|
||||
memcpy(matBuff, yuv, width * height);
|
||||
return !resmat.empty();
|
||||
}
|
||||
|
||||
bool CvCapture_Android::convertYUVToGrey(const cv::Mat& yuv, cv::Mat& resmat)
|
||||
bool CvCapture_Android::convertYUV420i2BGR888(int width, int height, const unsigned char* yuv, cv::Mat& resmat)
|
||||
{
|
||||
if (yuv.empty())
|
||||
return false;
|
||||
if (yuv == 0) return false;
|
||||
CV_Assert(width % 2 == 0 && height % 2 == 0);
|
||||
|
||||
resmat = yuv(cv::Range(0, yuv.rows * (2.0f / 3)), cv::Range::all());
|
||||
resmat.create(height, width, CV_8UC3);
|
||||
|
||||
return !resmat.empty();
|
||||
}
|
||||
unsigned char* y1 = (unsigned char*)yuv;
|
||||
unsigned char* uv = y1 + width * height;
|
||||
|
||||
bool CvCapture_Android::convertYUVToColor(const cv::Mat& yuv, cv::Mat& resmat)
|
||||
{
|
||||
if (yuv.empty())
|
||||
return false;
|
||||
//B = 1.164(Y - 16) + 2.018(U - 128)
|
||||
//G = 1.164(Y - 16) - 0.813(V - 128) - 0.391(U - 128)
|
||||
//R = 1.164(Y - 16) + 1.596(V - 128)
|
||||
|
||||
cv::cvtColor(yuv, resmat, CV_YUV2RGB);
|
||||
return !resmat.empty();
|
||||
for (int j = 0; j < height; j+=2, y1+=width*2, uv+=width)
|
||||
{
|
||||
unsigned char* row1 = resmat.ptr<unsigned char>(j);
|
||||
unsigned char* row2 = resmat.ptr<unsigned char>(j+1);
|
||||
unsigned char* y2 = y1 + width;
|
||||
|
||||
for(int i = 0; i < width; i+=2,row1+=6,row2+=6)
|
||||
{
|
||||
// unsigned char cr = uv[i];
|
||||
// unsigned char cb = uv[i+1];
|
||||
|
||||
// row1[0] = y1[i];
|
||||
// row1[1] = cr;
|
||||
// row1[2] = cb;
|
||||
|
||||
// row1[3] = y1[i+1];
|
||||
// row1[4] = cr;
|
||||
// row1[5] = cb;
|
||||
|
||||
// row2[0] = y2[i];
|
||||
// row2[1] = cr;
|
||||
// row2[2] = cb;
|
||||
|
||||
// row2[3] = y2[i+1];
|
||||
// row2[4] = cr;
|
||||
// row2[5] = cb;
|
||||
|
||||
int cr = uv[i] - 128;
|
||||
int cb = uv[i+1] - 128;
|
||||
|
||||
int ruv = 409 * cr + 128;
|
||||
int guv = 128 - 100 * cb - 208 * cr;
|
||||
int buv = 516 * cb + 128;
|
||||
|
||||
int y00 = (y1[i] - 16) * 298;
|
||||
row1[0] = clamp((y00 + buv) >> 8);
|
||||
row1[1] = clamp((y00 + guv) >> 8);
|
||||
row1[2] = clamp((y00 + ruv) >> 8);
|
||||
|
||||
int y01 = (y1[i+1] - 16) * 298;
|
||||
row1[3] = clamp((y01 + buv) >> 8);
|
||||
row1[4] = clamp((y01 + guv) >> 8);
|
||||
row1[5] = clamp((y01 + ruv) >> 8);
|
||||
|
||||
int y10 = (y2[i] - 16) * 298;
|
||||
row2[0] = clamp((y10 + buv) >> 8);
|
||||
row2[1] = clamp((y10 + guv) >> 8);
|
||||
row2[2] = clamp((y10 + ruv) >> 8);
|
||||
|
||||
int y11 = (y2[i+1] - 16) * 298;
|
||||
row2[3] = clamp((y11 + buv) >> 8);
|
||||
row2[4] = clamp((y11 + guv) >> 8);
|
||||
row2[5] = clamp((y11 + ruv) >> 8);
|
||||
}
|
||||
}
|
||||
|
||||
return !resmat.empty();
|
||||
}
|
||||
|
||||
|
||||
CvCapture* cvCreateCameraCapture_Android( int /*index*/ )
|
||||
CvCapture* cvCreateCameraCapture_Android( int cameraId )
|
||||
{
|
||||
CvCapture_Android* capture = new CvCapture_Android();
|
||||
CvCapture_Android* capture = new CvCapture_Android(cameraId);
|
||||
|
||||
if( capture->isOpened() )
|
||||
return capture;
|
||||
|
Loading…
x
Reference in New Issue
Block a user