/* * libjingle * Copyright 2011 Google Inc. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are met: * * 1. Redistributions of source code must retain the above copyright notice, * this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright notice, * this list of conditions and the following disclaimer in the documentation * and/or other materials provided with the distribution. * 3. The name of the author may not be used to endorse or promote products * derived from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR IMPLIED * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF * MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO * EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; * OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR * OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF * ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #include "talk/media/base/videoframe.h" #include #if !defined(DISABLE_YUV) #include "libyuv/compare.h" #include "libyuv/planar_functions.h" #include "libyuv/scale.h" #endif #include "talk/media/base/videocommon.h" #include "webrtc/base/logging.h" namespace cricket { // Round to 2 pixels because Chroma channels are half size. #define ROUNDTO2(v) (v & ~1) rtc::StreamResult VideoFrame::Write(rtc::StreamInterface* stream, int* error) { rtc::StreamResult result = rtc::SR_SUCCESS; const uint8* src_y = GetYPlane(); const uint8* src_u = GetUPlane(); const uint8* src_v = GetVPlane(); if (!src_y || !src_u || !src_v) { return result; // Nothing to write. } const int32 y_pitch = GetYPitch(); const int32 u_pitch = GetUPitch(); const int32 v_pitch = GetVPitch(); const size_t width = GetWidth(); const size_t height = GetHeight(); const size_t half_width = (width + 1) >> 1; const size_t half_height = (height + 1) >> 1; // Write Y. for (size_t row = 0; row < height; ++row) { result = stream->Write(src_y + row * y_pitch, width, NULL, error); if (result != rtc::SR_SUCCESS) { return result; } } // Write U. for (size_t row = 0; row < half_height; ++row) { result = stream->Write(src_u + row * u_pitch, half_width, NULL, error); if (result != rtc::SR_SUCCESS) { return result; } } // Write V. for (size_t row = 0; row < half_height; ++row) { result = stream->Write(src_v + row * v_pitch, half_width, NULL, error); if (result != rtc::SR_SUCCESS) { return result; } } return result; } bool VideoFrame::CopyToPlanes( uint8* dst_y, uint8* dst_u, uint8* dst_v, int32 dst_pitch_y, int32 dst_pitch_u, int32 dst_pitch_v) const { #if !defined(DISABLE_YUV) int32 src_width = static_cast(GetWidth()); int32 src_height = static_cast(GetHeight()); return libyuv::I420Copy(GetYPlane(), GetYPitch(), GetUPlane(), GetUPitch(), GetVPlane(), GetVPitch(), dst_y, dst_pitch_y, dst_u, dst_pitch_u, dst_v, dst_pitch_v, src_width, src_height) == 0; #else int uv_size = GetUPitch() * GetChromaHeight(); memcpy(dst_y, GetYPlane(), GetWidth() * GetHeight()); memcpy(dst_u, GetUPlane(), uv_size); memcpy(dst_v, GetVPlane(), uv_size); return true; #endif } void VideoFrame::CopyToFrame(VideoFrame* dst) const { if (!dst) { LOG(LS_ERROR) << "NULL dst pointer."; return; } CopyToPlanes(dst->GetYPlane(), dst->GetUPlane(), dst->GetVPlane(), dst->GetYPitch(), dst->GetUPitch(), dst->GetVPitch()); } size_t VideoFrame::ConvertToRgbBuffer(uint32 to_fourcc, uint8* buffer, size_t size, int stride_rgb) const { const size_t needed = std::abs(stride_rgb) * GetHeight(); if (size < needed) { LOG(LS_WARNING) << "RGB buffer is not large enough"; return needed; } if (libyuv::ConvertFromI420(GetYPlane(), GetYPitch(), GetUPlane(), GetUPitch(), GetVPlane(), GetVPitch(), buffer, stride_rgb, static_cast(GetWidth()), static_cast(GetHeight()), to_fourcc)) { LOG(LS_ERROR) << "RGB type not supported: " << to_fourcc; return 0; // 0 indicates error } return needed; } // TODO(fbarchard): Handle odd width/height with rounding. void VideoFrame::StretchToPlanes( uint8* dst_y, uint8* dst_u, uint8* dst_v, int32 dst_pitch_y, int32 dst_pitch_u, int32 dst_pitch_v, size_t width, size_t height, bool interpolate, bool vert_crop) const { if (!GetYPlane() || !GetUPlane() || !GetVPlane()) { LOG(LS_ERROR) << "NULL plane pointer."; return; } size_t src_width = GetWidth(); size_t src_height = GetHeight(); if (width == src_width && height == src_height) { CopyToPlanes(dst_y, dst_u, dst_v, dst_pitch_y, dst_pitch_u, dst_pitch_v); return; } const uint8* src_y = GetYPlane(); const uint8* src_u = GetUPlane(); const uint8* src_v = GetVPlane(); if (vert_crop) { // Adjust the input width:height ratio to be the same as the output ratio. if (src_width * height > src_height * width) { // Reduce the input width, but keep size/position aligned for YuvScaler src_width = ROUNDTO2(src_height * width / height); int32 iwidth_offset = ROUNDTO2((GetWidth() - src_width) / 2); src_y += iwidth_offset; src_u += iwidth_offset / 2; src_v += iwidth_offset / 2; } else if (src_width * height < src_height * width) { // Reduce the input height. src_height = src_width * height / width; int32 iheight_offset = static_cast( (GetHeight() - src_height) >> 2); iheight_offset <<= 1; // Ensure that iheight_offset is even. src_y += iheight_offset * GetYPitch(); src_u += iheight_offset / 2 * GetUPitch(); src_v += iheight_offset / 2 * GetVPitch(); } } // TODO(fbarchard): Implement a simple scale for non-libyuv. #if !defined(DISABLE_YUV) // Scale to the output I420 frame. libyuv::Scale(src_y, src_u, src_v, GetYPitch(), GetUPitch(), GetVPitch(), static_cast(src_width), static_cast(src_height), dst_y, dst_u, dst_v, dst_pitch_y, dst_pitch_u, dst_pitch_v, static_cast(width), static_cast(height), interpolate); #endif } size_t VideoFrame::StretchToBuffer(size_t dst_width, size_t dst_height, uint8* dst_buffer, size_t size, bool interpolate, bool vert_crop) const { if (!dst_buffer) { LOG(LS_ERROR) << "NULL dst_buffer pointer."; return 0; } size_t needed = SizeOf(dst_width, dst_height); if (needed <= size) { uint8* dst_y = dst_buffer; uint8* dst_u = dst_y + dst_width * dst_height; uint8* dst_v = dst_u + ((dst_width + 1) >> 1) * ((dst_height + 1) >> 1); StretchToPlanes(dst_y, dst_u, dst_v, static_cast(dst_width), static_cast((dst_width + 1) >> 1), static_cast((dst_width + 1) >> 1), dst_width, dst_height, interpolate, vert_crop); } return needed; } void VideoFrame::StretchToFrame(VideoFrame* dst, bool interpolate, bool vert_crop) const { if (!dst) { LOG(LS_ERROR) << "NULL dst pointer."; return; } StretchToPlanes(dst->GetYPlane(), dst->GetUPlane(), dst->GetVPlane(), dst->GetYPitch(), dst->GetUPitch(), dst->GetVPitch(), dst->GetWidth(), dst->GetHeight(), interpolate, vert_crop); dst->SetElapsedTime(GetElapsedTime()); dst->SetTimeStamp(GetTimeStamp()); } VideoFrame* VideoFrame::Stretch(size_t dst_width, size_t dst_height, bool interpolate, bool vert_crop) const { VideoFrame* dest = CreateEmptyFrame(static_cast(dst_width), static_cast(dst_height), GetPixelWidth(), GetPixelHeight(), GetElapsedTime(), GetTimeStamp()); if (dest) { StretchToFrame(dest, interpolate, vert_crop); } return dest; } bool VideoFrame::SetToBlack() { #if !defined(DISABLE_YUV) return libyuv::I420Rect(GetYPlane(), GetYPitch(), GetUPlane(), GetUPitch(), GetVPlane(), GetVPitch(), 0, 0, static_cast(GetWidth()), static_cast(GetHeight()), 16, 128, 128) == 0; #else int uv_size = GetUPitch() * GetChromaHeight(); memset(GetYPlane(), 16, GetWidth() * GetHeight()); memset(GetUPlane(), 128, uv_size); memset(GetVPlane(), 128, uv_size); return true; #endif } static const size_t kMaxSampleSize = 1000000000u; // Returns whether a sample is valid. bool VideoFrame::Validate(uint32 fourcc, int w, int h, const uint8 *sample, size_t sample_size) { if (h < 0) { h = -h; } // 16384 is maximum resolution for VP8 codec. if (w < 1 || w > 16384 || h < 1 || h > 16384) { LOG(LS_ERROR) << "Invalid dimensions: " << w << "x" << h; return false; } uint32 format = CanonicalFourCC(fourcc); int expected_bpp = 8; switch (format) { case FOURCC_I400: case FOURCC_RGGB: case FOURCC_BGGR: case FOURCC_GRBG: case FOURCC_GBRG: expected_bpp = 8; break; case FOURCC_I420: case FOURCC_I411: case FOURCC_YU12: case FOURCC_YV12: case FOURCC_M420: case FOURCC_Q420: case FOURCC_NV21: case FOURCC_NV12: expected_bpp = 12; break; case FOURCC_I422: case FOURCC_YV16: case FOURCC_YUY2: case FOURCC_UYVY: case FOURCC_RGBP: case FOURCC_RGBO: case FOURCC_R444: expected_bpp = 16; break; case FOURCC_I444: case FOURCC_YV24: case FOURCC_24BG: case FOURCC_RAW: expected_bpp = 24; break; case FOURCC_ABGR: case FOURCC_BGRA: case FOURCC_ARGB: expected_bpp = 32; break; case FOURCC_MJPG: case FOURCC_H264: expected_bpp = 0; break; default: expected_bpp = 8; // Expect format is at least 8 bits per pixel. break; } size_t expected_size = (w * expected_bpp + 7) / 8 * h; // For compressed formats, expect 4 bits per 16 x 16 macro. I420 would be // 6 bits, but grey can be 4 bits. if (expected_bpp == 0) { expected_size = ((w + 15) / 16) * ((h + 15) / 16) * 4 / 8; } if (sample == NULL) { LOG(LS_ERROR) << "NULL sample pointer." << " format: " << GetFourccName(format) << " bpp: " << expected_bpp << " size: " << w << "x" << h << " expected: " << expected_size << " " << sample_size; return false; } // TODO(fbarchard): Make function to dump information about frames. uint8 four_samples[4] = { 0, 0, 0, 0 }; for (size_t i = 0; i < ARRAY_SIZE(four_samples) && i < sample_size; ++i) { four_samples[i] = sample[i]; } if (sample_size < expected_size) { LOG(LS_ERROR) << "Size field is too small." << " format: " << GetFourccName(format) << " bpp: " << expected_bpp << " size: " << w << "x" << h << " " << sample_size << " expected: " << expected_size << " sample[0..3]: " << static_cast(four_samples[0]) << ", " << static_cast(four_samples[1]) << ", " << static_cast(four_samples[2]) << ", " << static_cast(four_samples[3]); return false; } if (sample_size > kMaxSampleSize) { LOG(LS_WARNING) << "Size field is invalid." << " format: " << GetFourccName(format) << " bpp: " << expected_bpp << " size: " << w << "x" << h << " " << sample_size << " expected: " << 2 * expected_size << " sample[0..3]: " << static_cast(four_samples[0]) << ", " << static_cast(four_samples[1]) << ", " << static_cast(four_samples[2]) << ", " << static_cast(four_samples[3]); return false; } // Show large size warning once every 100 frames. // TODO(fbarchard): Make frame counter atomic for thread safety. static int large_warn100 = 0; size_t large_expected_size = expected_size * 2; if (expected_bpp >= 8 && (sample_size > large_expected_size || sample_size > kMaxSampleSize) && large_warn100 % 100 == 0) { ++large_warn100; LOG(LS_WARNING) << "Size field is too large." << " format: " << GetFourccName(format) << " bpp: " << expected_bpp << " size: " << w << "x" << h << " bytes: " << sample_size << " expected: " << large_expected_size << " sample[0..3]: " << static_cast(four_samples[0]) << ", " << static_cast(four_samples[1]) << ", " << static_cast(four_samples[2]) << ", " << static_cast(four_samples[3]); } // TODO(fbarchard): Add duplicate pixel check. // TODO(fbarchard): Use frame counter atomic for thread safety. static bool valid_once = true; if (valid_once) { valid_once = false; LOG(LS_INFO) << "Validate frame passed." << " format: " << GetFourccName(format) << " bpp: " << expected_bpp << " size: " << w << "x" << h << " bytes: " << sample_size << " expected: " << expected_size << " sample[0..3]: " << static_cast(four_samples[0]) << ", " << static_cast(four_samples[1]) << ", " << static_cast(four_samples[2]) << ", " << static_cast(four_samples[3]); } return true; } } // namespace cricket