Refactor receiver.h/.cc.

TEST=video_coding_unittests, vie_auto_test --automated

Review URL: https://webrtc-codereview.appspot.com/994008

git-svn-id: http://webrtc.googlecode.com/svn/trunk@3336 4adac7df-926f-26a2-2b94-8c16560cd09d
This commit is contained in:
stefan@webrtc.org 2013-01-07 08:49:41 +00:00
parent 1926d33344
commit 1ea4b502ef
3 changed files with 436 additions and 519 deletions

View File

@ -8,488 +8,416 @@
* be found in the AUTHORS file in the root of the source tree. * be found in the AUTHORS file in the root of the source tree.
*/ */
#include "modules/video_coding/main/source/receiver.h" #include "webrtc/modules/video_coding/main/source/receiver.h"
#include <assert.h> #include <assert.h>
#include "modules/video_coding/main/interface/video_coding.h" #include "webrtc/modules/video_coding/main/interface/video_coding.h"
#include "modules/video_coding/main/source/encoded_frame.h" #include "webrtc/modules/video_coding/main/source/encoded_frame.h"
#include "modules/video_coding/main/source/internal_defines.h" #include "webrtc/modules/video_coding/main/source/internal_defines.h"
#include "modules/video_coding/main/source/media_opt_util.h" #include "webrtc/modules/video_coding/main/source/media_opt_util.h"
#include "modules/video_coding/main/source/tick_time_base.h" #include "webrtc/modules/video_coding/main/source/tick_time_base.h"
#include "system_wrappers/interface/trace.h" #include "webrtc/system_wrappers/interface/trace.h"
namespace webrtc { namespace webrtc {
VCMReceiver::VCMReceiver(VCMTiming& timing, VCMReceiver::VCMReceiver(VCMTiming* timing,
TickTimeBase* clock, TickTimeBase* clock,
WebRtc_Word32 vcmId, int32_t vcm_id,
WebRtc_Word32 receiverId, int32_t receiver_id,
bool master) bool master)
: _critSect(CriticalSectionWrapper::CreateCriticalSection()), : crit_sect_(CriticalSectionWrapper::CreateCriticalSection()),
_vcmId(vcmId), vcm_id_(vcm_id),
_clock(clock), clock_(clock),
_receiverId(receiverId), receiver_id_(receiver_id),
_master(master), master_(master),
_jitterBuffer(_clock, vcmId, receiverId, master), jitter_buffer_(clock_, vcm_id, receiver_id, master),
_timing(timing), timing_(timing),
_renderWaitEvent(*new VCMEvent()), render_wait_event_(),
_state(kPassive) {} state_(kPassive) {}
VCMReceiver::~VCMReceiver() VCMReceiver::~VCMReceiver() {
{ render_wait_event_.Set();
_renderWaitEvent.Set(); delete crit_sect_;
delete &_renderWaitEvent;
delete _critSect;
} }
void void VCMReceiver::Reset() {
VCMReceiver::Reset() CriticalSectionScoped cs(crit_sect_);
{ if (!jitter_buffer_.Running()) {
CriticalSectionScoped cs(_critSect); jitter_buffer_.Start();
if (!_jitterBuffer.Running()) } else {
{ jitter_buffer_.Flush();
_jitterBuffer.Start(); }
} render_wait_event_.Reset();
else if (master_) {
{ state_ = kReceiving;
_jitterBuffer.Flush(); } else {
} state_ = kPassive;
_renderWaitEvent.Reset(); }
if (_master)
{
_state = kReceiving;
}
else
{
_state = kPassive;
}
} }
WebRtc_Word32 int32_t VCMReceiver::Initialize() {
VCMReceiver::Initialize() CriticalSectionScoped cs(crit_sect_);
{ Reset();
CriticalSectionScoped cs(_critSect); if (!master_) {
Reset(); SetNackMode(kNoNack);
if (!_master) }
{ return VCM_OK;
SetNackMode(kNoNack); }
}
void VCMReceiver::UpdateRtt(uint32_t rtt) {
jitter_buffer_.UpdateRtt(rtt);
}
int32_t VCMReceiver::InsertPacket(const VCMPacket& packet, uint16_t frame_width,
uint16_t frame_height) {
// Find an empty frame.
VCMEncodedFrame* buffer = NULL;
const int32_t error = jitter_buffer_.GetFrame(packet, buffer);
if (error == VCM_OLD_PACKET_ERROR) {
return VCM_OK; return VCM_OK;
} else if (error != VCM_OK) {
return error;
}
assert(buffer);
{
CriticalSectionScoped cs(crit_sect_);
if (frame_width && frame_height) {
buffer->SetEncodedSize(static_cast<uint32_t>(frame_width),
static_cast<uint32_t>(frame_height));
}
if (master_) {
// Only trace the primary receiver to make it possible to parse and plot
// the trace file.
WEBRTC_TRACE(webrtc::kTraceDebug, webrtc::kTraceVideoCoding,
VCMId(vcm_id_, receiver_id_),
"Packet seq_no %u of frame %u at %u",
packet.seqNum, packet.timestamp,
MaskWord64ToUWord32(clock_->MillisecondTimestamp()));
}
const int64_t now_ms = clock_->MillisecondTimestamp();
int64_t render_time_ms = timing_->RenderTimeMs(packet.timestamp, now_ms);
if (render_time_ms < 0) {
// Render time error. Assume that this is due to some change in the
// incoming video stream and reset the JB and the timing.
jitter_buffer_.Flush();
timing_->Reset(clock_->MillisecondTimestamp());
return VCM_FLUSH_INDICATOR;
} else if (render_time_ms < now_ms - kMaxVideoDelayMs) {
WEBRTC_TRACE(webrtc::kTraceWarning, webrtc::kTraceVideoCoding,
VCMId(vcm_id_, receiver_id_),
"This frame should have been rendered more than %u ms ago."
"Flushing jitter buffer and resetting timing.",
kMaxVideoDelayMs);
jitter_buffer_.Flush();
timing_->Reset(clock_->MillisecondTimestamp());
return VCM_FLUSH_INDICATOR;
} else if (timing_->TargetVideoDelay() > kMaxVideoDelayMs) {
WEBRTC_TRACE(webrtc::kTraceWarning, webrtc::kTraceVideoCoding,
VCMId(vcm_id_, receiver_id_),
"More than %u ms target delay. Flushing jitter buffer and"
"resetting timing.", kMaxVideoDelayMs);
jitter_buffer_.Flush();
timing_->Reset(clock_->MillisecondTimestamp());
return VCM_FLUSH_INDICATOR;
}
// First packet received belonging to this frame.
if (buffer->Length() == 0) {
const int64_t now_ms = clock_->MillisecondTimestamp();
if (master_) {
// Only trace the primary receiver to make it possible to parse and plot
// the trace file.
WEBRTC_TRACE(webrtc::kTraceDebug, webrtc::kTraceVideoCoding,
VCMId(vcm_id_, receiver_id_),
"First packet of frame %u at %u", packet.timestamp,
MaskWord64ToUWord32(now_ms));
}
render_time_ms = timing_->RenderTimeMs(packet.timestamp, now_ms);
if (render_time_ms >= 0) {
buffer->SetRenderTime(render_time_ms);
} else {
buffer->SetRenderTime(now_ms);
}
}
// Insert packet into the jitter buffer both media and empty packets.
const VCMFrameBufferEnum
ret = jitter_buffer_.InsertPacket(buffer, packet);
if (ret == kFlushIndicator) {
return VCM_FLUSH_INDICATOR;
} else if (ret < 0) {
WEBRTC_TRACE(webrtc::kTraceError, webrtc::kTraceVideoCoding,
VCMId(vcm_id_, receiver_id_),
"Error inserting packet seq_no=%u, time_stamp=%u",
packet.seqNum, packet.timestamp);
return VCM_JITTER_BUFFER_ERROR;
}
}
return VCM_OK;
} }
void VCMReceiver::UpdateRtt(WebRtc_UWord32 rtt) VCMEncodedFrame* VCMReceiver::FrameForDecoding(
{ uint16_t max_wait_time_ms,
_jitterBuffer.UpdateRtt(rtt); int64_t& next_render_time_ms,
bool render_timing,
VCMReceiver* dual_receiver) {
// No need to enter the critical section here since the jitter buffer
// is thread-safe.
FrameType incoming_frame_type = kVideoFrameDelta;
next_render_time_ms = -1;
const int64_t start_time_ms = clock_->MillisecondTimestamp();
int64_t ret = jitter_buffer_.NextTimestamp(max_wait_time_ms,
&incoming_frame_type,
&next_render_time_ms);
if (ret < 0) {
// No timestamp in jitter buffer at the moment.
return NULL;
}
const uint32_t time_stamp = static_cast<uint32_t>(ret);
// Update the timing.
timing_->SetRequiredDelay(jitter_buffer_.EstimatedJitterMs());
timing_->UpdateCurrentDelay(time_stamp);
const int32_t temp_wait_time = max_wait_time_ms -
static_cast<int32_t>(clock_->MillisecondTimestamp() - start_time_ms);
uint16_t new_max_wait_time = static_cast<uint16_t>(VCM_MAX(temp_wait_time,
0));
VCMEncodedFrame* frame = NULL;
if (render_timing) {
frame = FrameForDecoding(new_max_wait_time, next_render_time_ms,
dual_receiver);
} else {
frame = FrameForRendering(new_max_wait_time, next_render_time_ms,
dual_receiver);
}
if (frame != NULL) {
bool retransmitted = false;
const int64_t last_packet_time_ms =
jitter_buffer_.LastPacketTime(frame, &retransmitted);
if (last_packet_time_ms >= 0 && !retransmitted) {
// We don't want to include timestamps which have suffered from
// retransmission here, since we compensate with extra retransmission
// delay within the jitter estimate.
timing_->IncomingTimestamp(time_stamp, last_packet_time_ms);
}
if (dual_receiver != NULL) {
dual_receiver->UpdateState(*frame);
}
}
return frame;
} }
WebRtc_Word32 VCMEncodedFrame* VCMReceiver::FrameForDecoding(
VCMReceiver::InsertPacket(const VCMPacket& packet, uint16_t max_wait_time_ms,
WebRtc_UWord16 frameWidth, int64_t next_render_time_ms,
WebRtc_UWord16 frameHeight) VCMReceiver* dual_receiver) {
{ // How long can we wait until we must decode the next frame.
// Find an empty frame uint32_t wait_time_ms = timing_->MaxWaitingTime(
VCMEncodedFrame *buffer = NULL; next_render_time_ms, clock_->MillisecondTimestamp());
const WebRtc_Word32 error = _jitterBuffer.GetFrame(packet, buffer);
if (error == VCM_OLD_PACKET_ERROR) // Try to get a complete frame from the jitter buffer.
{ VCMEncodedFrame* frame = jitter_buffer_.GetCompleteFrameForDecoding(0);
return VCM_OK;
if (frame == NULL && max_wait_time_ms == 0 && wait_time_ms > 0) {
// If we're not allowed to wait for frames to get complete we must
// calculate if it's time to decode, and if it's not we will just return
// for now.
return NULL;
}
if (frame == NULL && VCM_MIN(wait_time_ms, max_wait_time_ms) == 0) {
// No time to wait for a complete frame, check if we have an incomplete.
const bool dual_receiver_enabled_and_passive = (dual_receiver != NULL &&
dual_receiver->State() == kPassive &&
dual_receiver->NackMode() == kNackInfinite);
if (dual_receiver_enabled_and_passive &&
!jitter_buffer_.CompleteSequenceWithNextFrame()) {
// Jitter buffer state might get corrupt with this frame.
dual_receiver->CopyJitterBufferStateFromReceiver(*this);
frame = jitter_buffer_.GetFrameForDecoding();
assert(frame);
} else {
frame = jitter_buffer_.GetFrameForDecoding();
} }
else if (error != VCM_OK) }
{ if (frame == NULL) {
return error; // Wait for a complete frame.
frame = jitter_buffer_.GetCompleteFrameForDecoding(max_wait_time_ms);
}
if (frame == NULL) {
// Get an incomplete frame.
if (timing_->MaxWaitingTime(next_render_time_ms,
clock_->MillisecondTimestamp()) > 0) {
// Still time to wait for a complete frame.
return NULL;
} }
assert(buffer);
{
CriticalSectionScoped cs(_critSect);
if (frameWidth && frameHeight) // No time left to wait, we must decode this frame now.
{ const bool dual_receiver_enabled_and_passive = (dual_receiver != NULL &&
buffer->SetEncodedSize(static_cast<WebRtc_UWord32>(frameWidth), dual_receiver->State() == kPassive &&
static_cast<WebRtc_UWord32>(frameHeight)); dual_receiver->NackMode() == kNackInfinite);
} if (dual_receiver_enabled_and_passive &&
!jitter_buffer_.CompleteSequenceWithNextFrame()) {
if (_master) // Jitter buffer state might get corrupt with this frame.
{ dual_receiver->CopyJitterBufferStateFromReceiver(*this);
// Only trace the primary receiver to make it possible
// to parse and plot the trace file.
WEBRTC_TRACE(webrtc::kTraceDebug, webrtc::kTraceVideoCoding,
VCMId(_vcmId, _receiverId),
"Packet seqNo %u of frame %u at %u",
packet.seqNum, packet.timestamp,
MaskWord64ToUWord32(_clock->MillisecondTimestamp()));
}
const WebRtc_Word64 nowMs = _clock->MillisecondTimestamp();
WebRtc_Word64 renderTimeMs = _timing.RenderTimeMs(packet.timestamp, nowMs);
if (renderTimeMs < 0)
{
// Render time error. Assume that this is due to some change in
// the incoming video stream and reset the JB and the timing.
_jitterBuffer.Flush();
_timing.Reset(_clock->MillisecondTimestamp());
return VCM_FLUSH_INDICATOR;
}
else if (renderTimeMs < nowMs - kMaxVideoDelayMs)
{
WEBRTC_TRACE(webrtc::kTraceWarning, webrtc::kTraceVideoCoding, VCMId(_vcmId, _receiverId),
"This frame should have been rendered more than %u ms ago."
"Flushing jitter buffer and resetting timing.", kMaxVideoDelayMs);
_jitterBuffer.Flush();
_timing.Reset(_clock->MillisecondTimestamp());
return VCM_FLUSH_INDICATOR;
}
else if (_timing.TargetVideoDelay() > kMaxVideoDelayMs)
{
WEBRTC_TRACE(webrtc::kTraceWarning, webrtc::kTraceVideoCoding, VCMId(_vcmId, _receiverId),
"More than %u ms target delay. Flushing jitter buffer and resetting timing.",
kMaxVideoDelayMs);
_jitterBuffer.Flush();
_timing.Reset(_clock->MillisecondTimestamp());
return VCM_FLUSH_INDICATOR;
}
// First packet received belonging to this frame.
if (buffer->Length() == 0)
{
const WebRtc_Word64 nowMs = _clock->MillisecondTimestamp();
if (_master)
{
// Only trace the primary receiver to make it possible to parse and plot the trace file.
WEBRTC_TRACE(webrtc::kTraceDebug, webrtc::kTraceVideoCoding, VCMId(_vcmId, _receiverId),
"First packet of frame %u at %u", packet.timestamp,
MaskWord64ToUWord32(nowMs));
}
renderTimeMs = _timing.RenderTimeMs(packet.timestamp, nowMs);
if (renderTimeMs >= 0)
{
buffer->SetRenderTime(renderTimeMs);
}
else
{
buffer->SetRenderTime(nowMs);
}
}
// Insert packet into the jitter buffer
// both media and empty packets
const VCMFrameBufferEnum
ret = _jitterBuffer.InsertPacket(buffer, packet);
if (ret == kFlushIndicator) {
return VCM_FLUSH_INDICATOR;
} else if (ret < 0) {
WEBRTC_TRACE(webrtc::kTraceError, webrtc::kTraceVideoCoding,
VCMId(_vcmId, _receiverId),
"Error inserting packet seqNo=%u, timeStamp=%u",
packet.seqNum, packet.timestamp);
return VCM_JITTER_BUFFER_ERROR;
}
} }
return VCM_OK;
frame = jitter_buffer_.GetFrameForDecoding();
}
return frame;
} }
VCMEncodedFrame* VCMReceiver::FrameForDecoding(WebRtc_UWord16 maxWaitTimeMs, VCMEncodedFrame* VCMReceiver::FrameForRendering(uint16_t max_wait_time_ms,
WebRtc_Word64& nextRenderTimeMs, int64_t next_render_time_ms,
bool renderTiming, VCMReceiver* dual_receiver) {
VCMReceiver* dualReceiver) // How long MUST we wait until we must decode the next frame. This is
{ // different for the case where we have a renderer which can render at a
// No need to enter the critical section here since the jitter buffer // specified time. Here we must wait as long as possible before giving the
// is thread-safe. // frame to the decoder, which will render the frame as soon as it has been
FrameType incomingFrameType = kVideoFrameDelta; // decoded.
nextRenderTimeMs = -1; uint32_t wait_time_ms = timing_->MaxWaitingTime(
const WebRtc_Word64 startTimeMs = _clock->MillisecondTimestamp(); next_render_time_ms, clock_->MillisecondTimestamp());
WebRtc_Word64 ret = _jitterBuffer.NextTimestamp(maxWaitTimeMs, if (max_wait_time_ms < wait_time_ms) {
&incomingFrameType, // If we're not allowed to wait until the frame is supposed to be rendered
&nextRenderTimeMs); // we will have to return NULL for now.
if (ret < 0) return NULL;
{ }
// No timestamp in jitter buffer at the moment // Wait until it's time to render.
return NULL; render_wait_event_.Wait(wait_time_ms);
}
const WebRtc_UWord32 timeStamp = static_cast<WebRtc_UWord32>(ret);
// Update the timing // Get a complete frame if possible.
_timing.SetRequiredDelay(_jitterBuffer.EstimatedJitterMs()); VCMEncodedFrame* frame = jitter_buffer_.GetCompleteFrameForDecoding(0);
_timing.UpdateCurrentDelay(timeStamp);
const WebRtc_Word32 tempWaitTime = maxWaitTimeMs - if (frame == NULL) {
static_cast<WebRtc_Word32>(_clock->MillisecondTimestamp() - startTimeMs); // Get an incomplete frame.
WebRtc_UWord16 newMaxWaitTime = static_cast<WebRtc_UWord16>(VCM_MAX(tempWaitTime, 0)); const bool dual_receiver_enabled_and_passive = (dual_receiver != NULL &&
dual_receiver->State() == kPassive &&
VCMEncodedFrame* frame = NULL; dual_receiver->NackMode() == kNackInfinite);
if (dual_receiver_enabled_and_passive &&
if (renderTiming) !jitter_buffer_.CompleteSequenceWithNextFrame()) {
{ // Jitter buffer state might get corrupt with this frame.
frame = FrameForDecoding(newMaxWaitTime, nextRenderTimeMs, dualReceiver); dual_receiver->CopyJitterBufferStateFromReceiver(*this);
}
else
{
frame = FrameForRendering(newMaxWaitTime, nextRenderTimeMs, dualReceiver);
} }
if (frame != NULL) frame = jitter_buffer_.GetFrameForDecoding();
{ }
bool retransmitted = false; return frame;
const WebRtc_Word64 lastPacketTimeMs =
_jitterBuffer.LastPacketTime(frame, &retransmitted);
if (lastPacketTimeMs >= 0 && !retransmitted)
{
// We don't want to include timestamps which have suffered from retransmission
// here, since we compensate with extra retransmission delay within
// the jitter estimate.
_timing.IncomingTimestamp(timeStamp, lastPacketTimeMs);
}
if (dualReceiver != NULL)
{
dualReceiver->UpdateState(*frame);
}
}
return frame;
} }
VCMEncodedFrame* void VCMReceiver::ReleaseFrame(VCMEncodedFrame* frame) {
VCMReceiver::FrameForDecoding(WebRtc_UWord16 maxWaitTimeMs, jitter_buffer_.ReleaseFrame(frame);
WebRtc_Word64 nextRenderTimeMs,
VCMReceiver* dualReceiver)
{
// How long can we wait until we must decode the next frame
WebRtc_UWord32 waitTimeMs = _timing.MaxWaitingTime(nextRenderTimeMs,
_clock->MillisecondTimestamp());
// Try to get a complete frame from the jitter buffer
VCMEncodedFrame* frame = _jitterBuffer.GetCompleteFrameForDecoding(0);
if (frame == NULL && maxWaitTimeMs == 0 && waitTimeMs > 0)
{
// If we're not allowed to wait for frames to get complete we must
// calculate if it's time to decode, and if it's not we will just return
// for now.
return NULL;
}
if (frame == NULL && VCM_MIN(waitTimeMs, maxWaitTimeMs) == 0)
{
// No time to wait for a complete frame,
// check if we have an incomplete
const bool dualReceiverEnabledAndPassive = (dualReceiver != NULL &&
dualReceiver->State() == kPassive &&
dualReceiver->NackMode() == kNackInfinite);
if (dualReceiverEnabledAndPassive &&
!_jitterBuffer.CompleteSequenceWithNextFrame())
{
// Jitter buffer state might get corrupt with this frame.
dualReceiver->CopyJitterBufferStateFromReceiver(*this);
frame = _jitterBuffer.GetFrameForDecoding();
assert(frame);
} else {
frame = _jitterBuffer.GetFrameForDecoding();
}
}
if (frame == NULL)
{
// Wait for a complete frame
frame = _jitterBuffer.GetCompleteFrameForDecoding(maxWaitTimeMs);
}
if (frame == NULL)
{
// Get an incomplete frame
if (_timing.MaxWaitingTime(nextRenderTimeMs,
_clock->MillisecondTimestamp()) > 0)
{
// Still time to wait for a complete frame
return NULL;
}
// No time left to wait, we must decode this frame now.
const bool dualReceiverEnabledAndPassive = (dualReceiver != NULL &&
dualReceiver->State() == kPassive &&
dualReceiver->NackMode() == kNackInfinite);
if (dualReceiverEnabledAndPassive &&
!_jitterBuffer.CompleteSequenceWithNextFrame())
{
// Jitter buffer state might get corrupt with this frame.
dualReceiver->CopyJitterBufferStateFromReceiver(*this);
}
frame = _jitterBuffer.GetFrameForDecoding();
}
return frame;
} }
VCMEncodedFrame* void VCMReceiver::ReceiveStatistics(uint32_t* bitrate,
VCMReceiver::FrameForRendering(WebRtc_UWord16 maxWaitTimeMs, uint32_t* framerate) {
WebRtc_Word64 nextRenderTimeMs, assert(bitrate);
VCMReceiver* dualReceiver) assert(framerate);
{ jitter_buffer_.IncomingRateStatistics(framerate, bitrate);
// How long MUST we wait until we must decode the next frame. This is different for the case *bitrate /= 1000; // Should be in kbps.
// where we have a renderer which can render at a specified time. Here we must wait as long
// as possible before giving the frame to the decoder, which will render the frame as soon
// as it has been decoded.
WebRtc_UWord32 waitTimeMs = _timing.MaxWaitingTime(nextRenderTimeMs,
_clock->MillisecondTimestamp());
if (maxWaitTimeMs < waitTimeMs)
{
// If we're not allowed to wait until the frame is supposed to be rendered
// we will have to return NULL for now.
return NULL;
}
// Wait until it's time to render
_renderWaitEvent.Wait(waitTimeMs);
// Get a complete frame if possible
VCMEncodedFrame* frame = _jitterBuffer.GetCompleteFrameForDecoding(0);
if (frame == NULL)
{
// Get an incomplete frame
const bool dualReceiverEnabledAndPassive = dualReceiver != NULL &&
dualReceiver->State() == kPassive &&
dualReceiver->NackMode() == kNackInfinite;
if (dualReceiverEnabledAndPassive && !_jitterBuffer.CompleteSequenceWithNextFrame())
{
// Jitter buffer state might get corrupt with this frame.
dualReceiver->CopyJitterBufferStateFromReceiver(*this);
}
frame = _jitterBuffer.GetFrameForDecoding();
}
return frame;
} }
void void VCMReceiver::ReceivedFrameCount(VCMFrameCount* frame_count) const {
VCMReceiver::ReleaseFrame(VCMEncodedFrame* frame) assert(frame_count);
{ jitter_buffer_.FrameStatistics(&frame_count->numDeltaFrames,
_jitterBuffer.ReleaseFrame(frame); &frame_count->numKeyFrames);
} }
WebRtc_Word32 uint32_t VCMReceiver::DiscardedPackets() const {
VCMReceiver::ReceiveStatistics(WebRtc_UWord32& bitRate, WebRtc_UWord32& frameRate) return jitter_buffer_.num_discarded_packets();
{
_jitterBuffer.IncomingRateStatistics(&frameRate, &bitRate);
bitRate /= 1000; // Should be in kbps
return 0;
} }
WebRtc_Word32 void VCMReceiver::SetNackMode(VCMNackMode nackMode) {
VCMReceiver::ReceivedFrameCount(VCMFrameCount& frameCount) const CriticalSectionScoped cs(crit_sect_);
{ // Default to always having NACK enabled in hybrid mode.
_jitterBuffer.FrameStatistics(&frameCount.numDeltaFrames, jitter_buffer_.SetNackMode(nackMode, kLowRttNackMs, -1);
&frameCount.numKeyFrames); if (!master_) {
return 0; state_ = kPassive; // The dual decoder defaults to passive.
}
} }
WebRtc_UWord32 VCMReceiver::DiscardedPackets() const { VCMNackMode VCMReceiver::NackMode() const {
return _jitterBuffer.num_discarded_packets(); CriticalSectionScoped cs(crit_sect_);
return jitter_buffer_.nack_mode();
} }
void VCMNackStatus VCMReceiver::NackList(uint16_t* nack_list,
VCMReceiver::SetNackMode(VCMNackMode nackMode) uint16_t* size) {
{ bool extended = false;
CriticalSectionScoped cs(_critSect); uint16_t nack_list_size = 0;
// Default to always having NACK enabled in hybrid mode. uint16_t* internal_nack_list = jitter_buffer_.CreateNackList(&nack_list_size,
_jitterBuffer.SetNackMode(nackMode, kLowRttNackMs, -1); &extended);
if (!_master) if (internal_nack_list == NULL && nack_list_size == 0xffff) {
{ // This combination is used to trigger key frame requests.
_state = kPassive; // The dual decoder defaults to passive *size = 0;
} return kNackKeyFrameRequest;
}
if (nack_list_size > *size) {
*size = nack_list_size;
return kNackNeedMoreMemory;
}
if (internal_nack_list != NULL && nack_list_size > 0) {
memcpy(nack_list, internal_nack_list, nack_list_size * sizeof(uint16_t));
}
*size = nack_list_size;
return kNackOk;
} }
VCMNackMode // Decide whether we should change decoder state. This should be done if the
VCMReceiver::NackMode() const // dual decoder has caught up with the decoder decoding with packet losses.
{ bool VCMReceiver::DualDecoderCaughtUp(VCMEncodedFrame* dual_frame,
CriticalSectionScoped cs(_critSect); VCMReceiver& dual_receiver) const {
return _jitterBuffer.nack_mode(); if (dual_frame == NULL) {
}
VCMNackStatus
VCMReceiver::NackList(WebRtc_UWord16* nackList, WebRtc_UWord16& size)
{
bool extended = false;
WebRtc_UWord16 nackListSize = 0;
WebRtc_UWord16* internalNackList = _jitterBuffer.CreateNackList(
&nackListSize, &extended);
if (internalNackList == NULL && nackListSize == 0xffff)
{
// This combination is used to trigger key frame requests.
size = 0;
return kNackKeyFrameRequest;
}
if (nackListSize > size)
{
size = nackListSize;
return kNackNeedMoreMemory;
}
if (internalNackList != NULL && nackListSize > 0) {
memcpy(nackList, internalNackList, nackListSize * sizeof(WebRtc_UWord16));
}
size = nackListSize;
return kNackOk;
}
// Decide whether we should change decoder state. This should be done if the dual decoder
// has caught up with the decoder decoding with packet losses.
bool
VCMReceiver::DualDecoderCaughtUp(VCMEncodedFrame* dualFrame, VCMReceiver& dualReceiver) const
{
if (dualFrame == NULL)
{
return false;
}
if (_jitterBuffer.LastDecodedTimestamp() == dualFrame->TimeStamp())
{
dualReceiver.UpdateState(kWaitForPrimaryDecode);
return true;
}
return false; return false;
}
if (jitter_buffer_.LastDecodedTimestamp() == dual_frame->TimeStamp()) {
dual_receiver.UpdateState(kWaitForPrimaryDecode);
return true;
}
return false;
} }
void void VCMReceiver::CopyJitterBufferStateFromReceiver(
VCMReceiver::CopyJitterBufferStateFromReceiver(const VCMReceiver& receiver) const VCMReceiver& receiver) {
{ jitter_buffer_.CopyFrom(receiver.jitter_buffer_);
_jitterBuffer.CopyFrom(receiver._jitterBuffer);
} }
VCMReceiverState VCMReceiverState VCMReceiver::State() const {
VCMReceiver::State() const CriticalSectionScoped cs(crit_sect_);
{ return state_;
CriticalSectionScoped cs(_critSect);
return _state;
} }
void void VCMReceiver::UpdateState(VCMReceiverState new_state) {
VCMReceiver::UpdateState(VCMReceiverState newState) CriticalSectionScoped cs(crit_sect_);
{ assert(!(state_ == kPassive && new_state == kWaitForPrimaryDecode));
CriticalSectionScoped cs(_critSect); state_ = new_state;
assert(!(_state == kPassive && newState == kWaitForPrimaryDecode));
// assert(!(_state == kReceiving && newState == kPassive));
_state = newState;
}
void
VCMReceiver::UpdateState(VCMEncodedFrame& frame)
{
if (_jitterBuffer.nack_mode() == kNoNack)
{
// Dual decoder mode has not been enabled.
return;
}
// Update the dual receiver state
if (frame.Complete() && frame.FrameType() == kVideoFrameKey)
{
UpdateState(kPassive);
}
if (State() == kWaitForPrimaryDecode &&
frame.Complete() && !frame.MissingFrame())
{
UpdateState(kPassive);
}
if (frame.MissingFrame() || !frame.Complete())
{
// State was corrupted, enable dual receiver.
UpdateState(kReceiving);
}
} }
void VCMReceiver::UpdateState(const VCMEncodedFrame& frame) {
if (jitter_buffer_.nack_mode() == kNoNack) {
// Dual decoder mode has not been enabled.
return;
}
// Update the dual receiver state.
if (frame.Complete() && frame.FrameType() == kVideoFrameKey) {
UpdateState(kPassive);
}
if (State() == kWaitForPrimaryDecode &&
frame.Complete() && !frame.MissingFrame()) {
UpdateState(kPassive);
}
if (frame.MissingFrame() || !frame.Complete()) {
// State was corrupted, enable dual receiver.
UpdateState(kReceiving);
}
} }
} // namespace webrtc

View File

@ -8,94 +8,90 @@
* be found in the AUTHORS file in the root of the source tree. * be found in the AUTHORS file in the root of the source tree.
*/ */
#ifndef WEBRTC_MODULES_VIDEO_CODING_RECEIVER_H_ #ifndef WEBRTC_MODULES_VIDEO_CODING_MAIN_SOURCE_RECEIVER_H_
#define WEBRTC_MODULES_VIDEO_CODING_RECEIVER_H_ #define WEBRTC_MODULES_VIDEO_CODING_MAIN_SOURCE_RECEIVER_H_
#include "critical_section_wrapper.h" #include "webrtc/system_wrappers/interface/critical_section_wrapper.h"
#include "jitter_buffer.h" #include "webrtc/modules/video_coding/main/source/jitter_buffer.h"
#include "modules/video_coding/main/source/tick_time_base.h" #include "webrtc/modules/video_coding/main/source/packet.h"
#include "timing.h" #include "webrtc/modules/video_coding/main/source/tick_time_base.h"
#include "packet.h" #include "webrtc/modules/video_coding/main/source/timing.h"
namespace webrtc namespace webrtc {
{
class VCMEncodedFrame; class VCMEncodedFrame;
enum VCMNackStatus enum VCMNackStatus {
{ kNackOk,
kNackOk, kNackNeedMoreMemory,
kNackNeedMoreMemory, kNackKeyFrameRequest
kNackKeyFrameRequest
}; };
enum VCMReceiverState {
enum VCMReceiverState kReceiving,
{ kPassive,
kReceiving, kWaitForPrimaryDecode
kPassive,
kWaitForPrimaryDecode
}; };
class VCMReceiver class VCMReceiver {
{ public:
public: VCMReceiver(VCMTiming* timing,
VCMReceiver(VCMTiming& timing, TickTimeBase* clock,
TickTimeBase* clock, int32_t vcm_id = -1,
WebRtc_Word32 vcmId = -1, int32_t receiver_id = -1,
WebRtc_Word32 receiverId = -1, bool master = true);
bool master = true); ~VCMReceiver();
~VCMReceiver();
void Reset(); void Reset();
WebRtc_Word32 Initialize(); int32_t Initialize();
void UpdateRtt(WebRtc_UWord32 rtt); void UpdateRtt(uint32_t rtt);
WebRtc_Word32 InsertPacket(const VCMPacket& packet, int32_t InsertPacket(const VCMPacket& packet,
WebRtc_UWord16 frameWidth, uint16_t frame_width,
WebRtc_UWord16 frameHeight); uint16_t frame_height);
VCMEncodedFrame* FrameForDecoding(WebRtc_UWord16 maxWaitTimeMs, VCMEncodedFrame* FrameForDecoding(uint16_t max_wait_time_ms,
WebRtc_Word64& nextRenderTimeMs, int64_t& next_render_time_ms,
bool renderTiming = true, bool render_timing = true,
VCMReceiver* dualReceiver = NULL); VCMReceiver* dual_receiver = NULL);
void ReleaseFrame(VCMEncodedFrame* frame); void ReleaseFrame(VCMEncodedFrame* frame);
WebRtc_Word32 ReceiveStatistics(WebRtc_UWord32& bitRate, WebRtc_UWord32& frameRate); void ReceiveStatistics(uint32_t* bitrate, uint32_t* framerate);
WebRtc_Word32 ReceivedFrameCount(VCMFrameCount& frameCount) const; void ReceivedFrameCount(VCMFrameCount* frame_count) const;
WebRtc_UWord32 DiscardedPackets() const; uint32_t DiscardedPackets() const;
// NACK // NACK.
void SetNackMode(VCMNackMode nackMode); void SetNackMode(VCMNackMode nackMode);
VCMNackMode NackMode() const; VCMNackMode NackMode() const;
VCMNackStatus NackList(WebRtc_UWord16* nackList, WebRtc_UWord16& size); VCMNackStatus NackList(uint16_t* nackList, uint16_t* size);
// Dual decoder // Dual decoder.
bool DualDecoderCaughtUp(VCMEncodedFrame* dualFrame, VCMReceiver& dualReceiver) const; bool DualDecoderCaughtUp(VCMEncodedFrame* dual_frame,
VCMReceiverState State() const; VCMReceiver& dual_receiver) const;
VCMReceiverState State() const;
private: private:
VCMEncodedFrame* FrameForDecoding(WebRtc_UWord16 maxWaitTimeMs, VCMEncodedFrame* FrameForDecoding(uint16_t max_wait_time_ms,
WebRtc_Word64 nextrenderTimeMs, int64_t nextrender_time_ms,
VCMReceiver* dualReceiver); VCMReceiver* dual_receiver);
VCMEncodedFrame* FrameForRendering(WebRtc_UWord16 maxWaitTimeMs, VCMEncodedFrame* FrameForRendering(uint16_t max_wait_time_ms,
WebRtc_Word64 nextrenderTimeMs, int64_t nextrender_time_ms,
VCMReceiver* dualReceiver); VCMReceiver* dual_receiver);
void CopyJitterBufferStateFromReceiver(const VCMReceiver& receiver); void CopyJitterBufferStateFromReceiver(const VCMReceiver& receiver);
void UpdateState(VCMReceiverState newState); void UpdateState(VCMReceiverState new_state);
void UpdateState(VCMEncodedFrame& frame); void UpdateState(const VCMEncodedFrame& frame);
static WebRtc_Word32 GenerateReceiverId(); static int32_t GenerateReceiverId();
CriticalSectionWrapper* _critSect; CriticalSectionWrapper* crit_sect_;
WebRtc_Word32 _vcmId; int32_t vcm_id_;
TickTimeBase* _clock; TickTimeBase* clock_;
WebRtc_Word32 _receiverId; int32_t receiver_id_;
bool _master; bool master_;
VCMJitterBuffer _jitterBuffer; VCMJitterBuffer jitter_buffer_;
VCMTiming& _timing; VCMTiming* timing_;
VCMEvent& _renderWaitEvent; VCMEvent render_wait_event_;
VCMReceiverState _state; VCMReceiverState state_;
static WebRtc_Word32 _receiverIdCounter; static int32_t receiver_id_counter_;
}; };
} // namespace webrtc } // namespace webrtc
#endif // WEBRTC_MODULES_VIDEO_CODING_RECEIVER_H_ #endif // WEBRTC_MODULES_VIDEO_CODING_MAIN_SOURCE_RECEIVER_H_

View File

@ -54,8 +54,8 @@ _receiveCritSect(CriticalSectionWrapper::CreateCriticalSection()),
_receiverInited(false), _receiverInited(false),
_timing(clock_, id, 1), _timing(clock_, id, 1),
_dualTiming(clock_, id, 2, &_timing), _dualTiming(clock_, id, 2, &_timing),
_receiver(_timing, clock_, id, 1), _receiver(&_timing, clock_, id, 1),
_dualReceiver(_dualTiming, clock_, id, 2, false), _dualReceiver(&_dualTiming, clock_, id, 2, false),
_decodedFrameCallback(_timing, clock_), _decodedFrameCallback(_timing, clock_),
_dualDecodedFrameCallback(_dualTiming, clock_), _dualDecodedFrameCallback(_dualTiming, clock_),
_frameTypeCallback(NULL), _frameTypeCallback(NULL),
@ -144,16 +144,8 @@ VideoCodingModuleImpl::Process()
{ {
WebRtc_UWord32 bitRate; WebRtc_UWord32 bitRate;
WebRtc_UWord32 frameRate; WebRtc_UWord32 frameRate;
const WebRtc_Word32 ret = _receiver.ReceiveStatistics(bitRate, _receiver.ReceiveStatistics(&bitRate, &frameRate);
frameRate); _receiveStatsCallback->ReceiveStatistics(bitRate, frameRate);
if (ret == 0)
{
_receiveStatsCallback->ReceiveStatistics(bitRate, frameRate);
}
else if (returnValue == VCM_OK)
{
returnValue = ret;
}
} }
} }
@ -1255,11 +1247,11 @@ VideoCodingModuleImpl::NackList(WebRtc_UWord16* nackList, WebRtc_UWord16& size)
// the dual receiver if the dual receiver is receiving. // the dual receiver if the dual receiver is receiving.
if (_receiver.NackMode() != kNoNack) if (_receiver.NackMode() != kNoNack)
{ {
nackStatus = _receiver.NackList(nackList, size); nackStatus = _receiver.NackList(nackList, &size);
} }
else if (_dualReceiver.State() != kPassive) else if (_dualReceiver.State() != kPassive)
{ {
nackStatus = _dualReceiver.NackList(nackList, size); nackStatus = _dualReceiver.NackList(nackList, &size);
} }
else else
{ {
@ -1294,7 +1286,8 @@ VideoCodingModuleImpl::NackList(WebRtc_UWord16* nackList, WebRtc_UWord16& size)
WebRtc_Word32 WebRtc_Word32
VideoCodingModuleImpl::ReceivedFrameCount(VCMFrameCount& frameCount) const VideoCodingModuleImpl::ReceivedFrameCount(VCMFrameCount& frameCount) const
{ {
return _receiver.ReceivedFrameCount(frameCount); _receiver.ReceivedFrameCount(&frameCount);
return VCM_OK;
} }
WebRtc_UWord32 VideoCodingModuleImpl::DiscardedPackets() const { WebRtc_UWord32 VideoCodingModuleImpl::DiscardedPackets() const {