webrtc/talk/media/base/videoframe.cc

385 lines
14 KiB
C++
Raw Normal View History

/*
* libjingle
* Copyright 2011 Google Inc.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are met:
*
* 1. Redistributions of source code must retain the above copyright notice,
* this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright notice,
* this list of conditions and the following disclaimer in the documentation
* and/or other materials provided with the distribution.
* 3. The name of the author may not be used to endorse or promote products
* derived from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR IMPLIED
* WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO
* EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
* PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS;
* OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
* WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR
* OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF
* ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#include "talk/media/base/videoframe.h"
#include <string.h>
#if !defined(DISABLE_YUV)
#include "libyuv/compare.h"
#include "libyuv/planar_functions.h"
#include "libyuv/scale.h"
#endif
#include "talk/media/base/videocommon.h"
#include "webrtc/base/logging.h"
namespace cricket {
// Round to 2 pixels because Chroma channels are half size.
#define ROUNDTO2(v) (v & ~1)
rtc::StreamResult VideoFrame::Write(rtc::StreamInterface* stream,
int* error) {
rtc::StreamResult result = rtc::SR_SUCCESS;
const uint8* src_y = GetYPlane();
const uint8* src_u = GetUPlane();
const uint8* src_v = GetVPlane();
if (!src_y || !src_u || !src_v) {
return result; // Nothing to write.
}
const int32 y_pitch = GetYPitch();
const int32 u_pitch = GetUPitch();
const int32 v_pitch = GetVPitch();
const size_t width = GetWidth();
const size_t height = GetHeight();
const size_t half_width = (width + 1) >> 1;
const size_t half_height = (height + 1) >> 1;
// Write Y.
for (size_t row = 0; row < height; ++row) {
result = stream->Write(src_y + row * y_pitch, width, NULL, error);
if (result != rtc::SR_SUCCESS) {
return result;
}
}
// Write U.
for (size_t row = 0; row < half_height; ++row) {
result = stream->Write(src_u + row * u_pitch, half_width, NULL, error);
if (result != rtc::SR_SUCCESS) {
return result;
}
}
// Write V.
for (size_t row = 0; row < half_height; ++row) {
result = stream->Write(src_v + row * v_pitch, half_width, NULL, error);
if (result != rtc::SR_SUCCESS) {
return result;
}
}
return result;
}
bool VideoFrame::CopyToPlanes(
uint8* dst_y, uint8* dst_u, uint8* dst_v,
int32 dst_pitch_y, int32 dst_pitch_u, int32 dst_pitch_v) const {
#if !defined(DISABLE_YUV)
int32 src_width = static_cast<int>(GetWidth());
int32 src_height = static_cast<int>(GetHeight());
return libyuv::I420Copy(GetYPlane(), GetYPitch(),
GetUPlane(), GetUPitch(),
GetVPlane(), GetVPitch(),
dst_y, dst_pitch_y,
dst_u, dst_pitch_u,
dst_v, dst_pitch_v,
src_width, src_height) == 0;
#else
int uv_size = GetUPitch() * GetChromaHeight();
memcpy(dst_y, GetYPlane(), GetWidth() * GetHeight());
memcpy(dst_u, GetUPlane(), uv_size);
memcpy(dst_v, GetVPlane(), uv_size);
return true;
#endif
}
void VideoFrame::CopyToFrame(VideoFrame* dst) const {
if (!dst) {
LOG(LS_ERROR) << "NULL dst pointer.";
return;
}
CopyToPlanes(dst->GetYPlane(), dst->GetUPlane(), dst->GetVPlane(),
dst->GetYPitch(), dst->GetUPitch(), dst->GetVPitch());
}
// TODO(fbarchard): Handle odd width/height with rounding.
void VideoFrame::StretchToPlanes(
uint8* dst_y, uint8* dst_u, uint8* dst_v,
int32 dst_pitch_y, int32 dst_pitch_u, int32 dst_pitch_v,
size_t width, size_t height, bool interpolate, bool vert_crop) const {
if (!GetYPlane() || !GetUPlane() || !GetVPlane()) {
LOG(LS_ERROR) << "NULL plane pointer.";
return;
}
size_t src_width = GetWidth();
size_t src_height = GetHeight();
if (width == src_width && height == src_height) {
CopyToPlanes(dst_y, dst_u, dst_v, dst_pitch_y, dst_pitch_u, dst_pitch_v);
return;
}
const uint8* src_y = GetYPlane();
const uint8* src_u = GetUPlane();
const uint8* src_v = GetVPlane();
if (vert_crop) {
// Adjust the input width:height ratio to be the same as the output ratio.
if (src_width * height > src_height * width) {
// Reduce the input width, but keep size/position aligned for YuvScaler
src_width = ROUNDTO2(src_height * width / height);
int32 iwidth_offset = ROUNDTO2((GetWidth() - src_width) / 2);
src_y += iwidth_offset;
src_u += iwidth_offset / 2;
src_v += iwidth_offset / 2;
} else if (src_width * height < src_height * width) {
// Reduce the input height.
src_height = src_width * height / width;
int32 iheight_offset = static_cast<int32>(
(GetHeight() - src_height) >> 2);
iheight_offset <<= 1; // Ensure that iheight_offset is even.
src_y += iheight_offset * GetYPitch();
src_u += iheight_offset / 2 * GetUPitch();
src_v += iheight_offset / 2 * GetVPitch();
}
}
// TODO(fbarchard): Implement a simple scale for non-libyuv.
#if !defined(DISABLE_YUV)
// Scale to the output I420 frame.
libyuv::Scale(src_y, src_u, src_v,
GetYPitch(), GetUPitch(), GetVPitch(),
static_cast<int>(src_width), static_cast<int>(src_height),
dst_y, dst_u, dst_v, dst_pitch_y, dst_pitch_u, dst_pitch_v,
static_cast<int>(width), static_cast<int>(height), interpolate);
#endif
}
size_t VideoFrame::StretchToBuffer(size_t dst_width, size_t dst_height,
uint8* dst_buffer, size_t size,
bool interpolate, bool vert_crop) const {
if (!dst_buffer) {
LOG(LS_ERROR) << "NULL dst_buffer pointer.";
return 0;
}
size_t needed = SizeOf(dst_width, dst_height);
if (needed <= size) {
uint8* dst_y = dst_buffer;
uint8* dst_u = dst_y + dst_width * dst_height;
uint8* dst_v = dst_u + ((dst_width + 1) >> 1) * ((dst_height + 1) >> 1);
StretchToPlanes(dst_y, dst_u, dst_v,
static_cast<int32>(dst_width),
static_cast<int32>((dst_width + 1) >> 1),
static_cast<int32>((dst_width + 1) >> 1),
dst_width, dst_height, interpolate, vert_crop);
}
return needed;
}
void VideoFrame::StretchToFrame(VideoFrame* dst,
bool interpolate, bool vert_crop) const {
if (!dst) {
LOG(LS_ERROR) << "NULL dst pointer.";
return;
}
StretchToPlanes(dst->GetYPlane(), dst->GetUPlane(), dst->GetVPlane(),
dst->GetYPitch(), dst->GetUPitch(), dst->GetVPitch(),
dst->GetWidth(), dst->GetHeight(),
interpolate, vert_crop);
dst->SetElapsedTime(GetElapsedTime());
dst->SetTimeStamp(GetTimeStamp());
}
VideoFrame* VideoFrame::Stretch(size_t dst_width, size_t dst_height,
bool interpolate, bool vert_crop) const {
VideoFrame* dest = CreateEmptyFrame(static_cast<int>(dst_width),
static_cast<int>(dst_height),
GetPixelWidth(), GetPixelHeight(),
GetElapsedTime(), GetTimeStamp());
if (dest) {
StretchToFrame(dest, interpolate, vert_crop);
}
return dest;
}
bool VideoFrame::SetToBlack() {
#if !defined(DISABLE_YUV)
return libyuv::I420Rect(GetYPlane(), GetYPitch(),
GetUPlane(), GetUPitch(),
GetVPlane(), GetVPitch(),
0, 0,
static_cast<int>(GetWidth()),
static_cast<int>(GetHeight()),
16, 128, 128) == 0;
#else
int uv_size = GetUPitch() * GetChromaHeight();
memset(GetYPlane(), 16, GetWidth() * GetHeight());
memset(GetUPlane(), 128, uv_size);
memset(GetVPlane(), 128, uv_size);
return true;
#endif
}
static const size_t kMaxSampleSize = 1000000000u;
// Returns whether a sample is valid.
bool VideoFrame::Validate(uint32 fourcc, int w, int h,
const uint8 *sample, size_t sample_size) {
if (h < 0) {
h = -h;
}
// 16384 is maximum resolution for VP8 codec.
if (w < 1 || w > 16384 || h < 1 || h > 16384) {
LOG(LS_ERROR) << "Invalid dimensions: " << w << "x" << h;
return false;
}
uint32 format = CanonicalFourCC(fourcc);
int expected_bpp = 8;
switch (format) {
case FOURCC_I400:
case FOURCC_RGGB:
case FOURCC_BGGR:
case FOURCC_GRBG:
case FOURCC_GBRG:
expected_bpp = 8;
break;
case FOURCC_I420:
case FOURCC_I411:
case FOURCC_YU12:
case FOURCC_YV12:
case FOURCC_M420:
case FOURCC_Q420:
case FOURCC_NV21:
case FOURCC_NV12:
expected_bpp = 12;
break;
case FOURCC_I422:
case FOURCC_YV16:
case FOURCC_YUY2:
case FOURCC_UYVY:
case FOURCC_RGBP:
case FOURCC_RGBO:
case FOURCC_R444:
expected_bpp = 16;
break;
case FOURCC_I444:
case FOURCC_YV24:
case FOURCC_24BG:
case FOURCC_RAW:
expected_bpp = 24;
break;
case FOURCC_ABGR:
case FOURCC_BGRA:
case FOURCC_ARGB:
expected_bpp = 32;
break;
case FOURCC_MJPG:
case FOURCC_H264:
expected_bpp = 0;
break;
default:
expected_bpp = 8; // Expect format is at least 8 bits per pixel.
break;
}
size_t expected_size = (w * expected_bpp + 7) / 8 * h;
// For compressed formats, expect 4 bits per 16 x 16 macro. I420 would be
// 6 bits, but grey can be 4 bits.
if (expected_bpp == 0) {
expected_size = ((w + 15) / 16) * ((h + 15) / 16) * 4 / 8;
}
if (sample == NULL) {
LOG(LS_ERROR) << "NULL sample pointer."
<< " format: " << GetFourccName(format)
<< " bpp: " << expected_bpp
<< " size: " << w << "x" << h
<< " expected: " << expected_size
<< " " << sample_size;
return false;
}
// TODO(fbarchard): Make function to dump information about frames.
uint8 four_samples[4] = { 0, 0, 0, 0 };
for (size_t i = 0; i < ARRAY_SIZE(four_samples) && i < sample_size; ++i) {
four_samples[i] = sample[i];
}
if (sample_size < expected_size) {
LOG(LS_ERROR) << "Size field is too small."
<< " format: " << GetFourccName(format)
<< " bpp: " << expected_bpp
<< " size: " << w << "x" << h
<< " " << sample_size
<< " expected: " << expected_size
<< " sample[0..3]: " << static_cast<int>(four_samples[0])
<< ", " << static_cast<int>(four_samples[1])
<< ", " << static_cast<int>(four_samples[2])
<< ", " << static_cast<int>(four_samples[3]);
return false;
}
if (sample_size > kMaxSampleSize) {
LOG(LS_WARNING) << "Size field is invalid."
<< " format: " << GetFourccName(format)
<< " bpp: " << expected_bpp
<< " size: " << w << "x" << h
<< " " << sample_size
<< " expected: " << 2 * expected_size
<< " sample[0..3]: " << static_cast<int>(four_samples[0])
<< ", " << static_cast<int>(four_samples[1])
<< ", " << static_cast<int>(four_samples[2])
<< ", " << static_cast<int>(four_samples[3]);
return false;
}
// Show large size warning once every 100 frames.
// TODO(fbarchard): Make frame counter atomic for thread safety.
static int large_warn100 = 0;
size_t large_expected_size = expected_size * 2;
if (expected_bpp >= 8 &&
(sample_size > large_expected_size || sample_size > kMaxSampleSize) &&
large_warn100 % 100 == 0) {
++large_warn100;
LOG(LS_WARNING) << "Size field is too large."
<< " format: " << GetFourccName(format)
<< " bpp: " << expected_bpp
<< " size: " << w << "x" << h
<< " bytes: " << sample_size
<< " expected: " << large_expected_size
<< " sample[0..3]: " << static_cast<int>(four_samples[0])
<< ", " << static_cast<int>(four_samples[1])
<< ", " << static_cast<int>(four_samples[2])
<< ", " << static_cast<int>(four_samples[3]);
}
// TODO(fbarchard): Add duplicate pixel check.
// TODO(fbarchard): Use frame counter atomic for thread safety.
static bool valid_once = true;
if (valid_once) {
valid_once = false;
LOG(LS_INFO) << "Validate frame passed."
<< " format: " << GetFourccName(format)
<< " bpp: " << expected_bpp
<< " size: " << w << "x" << h
<< " bytes: " << sample_size
<< " expected: " << expected_size
<< " sample[0..3]: " << static_cast<int>(four_samples[0])
<< ", " << static_cast<int>(four_samples[1])
<< ", " << static_cast<int>(four_samples[2])
<< ", " << static_cast<int>(four_samples[3]);
}
return true;
}
} // namespace cricket