Fix the chain that propagates the audio frame's rtp and ntp timestamp including:

* In AudioCodingModuleImpl::PlayoutData10Ms, don't reset the timestamp got from GetAudio.
* When there're more than one participant, set AudioFrame's RTP timestamp to 0.
* Copy ntp_time_ms_ in AudioFrame::CopyFrom method.
* In RemixAndResample, pass src frame's timestamp_ and ntp_time_ms_ to the dst frame.
* Fix how |elapsed_time_ms| is computed in channel.cc by adding GetPlayoutFrequency.

Tweaks on ntp_time_ms_:
* Init ntp_time_ms_ to -1 in AudioFrame ctor.
* When there're more than one participant, set AudioFrame's ntp_time_ms_ to an invalid value. I.e. we don't support ntp_time_ms_ in multiple participants case before the mixing is moved to chrome.

Added elapsed_time_ms to AudioFrame and pass it to chrome, where we don't have the information about the rtp timestmp's sample rate, i.e. can't convert rtp timestamp to ms.

BUG=3111
R=henrik.lundin@webrtc.org, turaj@webrtc.org, xians@webrtc.org
TBR=andrew
andrew to take another look on audio_conference_mixer_impl.cc

Review URL: https://webrtc-codereview.appspot.com/14559004

git-svn-id: http://webrtc.googlecode.com/svn/trunk@6346 4adac7df-926f-26a2-2b94-8c16560cd09d
This commit is contained in:
wu@webrtc.org 2014-06-05 20:34:08 +00:00
parent 130fa64d4c
commit 94454b71ad
26 changed files with 168 additions and 100 deletions

View File

@ -729,12 +729,12 @@ void FakeAudioCaptureModule::ReceiveFrameP() {
ResetRecBuffer();
uint32_t nSamplesOut = 0;
#ifdef USE_WEBRTC_DEV_BRANCH
uint32_t rtp_timestamp = 0;
int64_t elapsed_time_ms = 0;
int64_t ntp_time_ms = 0;
if (audio_callback_->NeedMorePlayData(kNumberSamples, kNumberBytesPerSample,
kNumberOfChannels, kSamplesPerSecond,
rec_buffer_, nSamplesOut,
&rtp_timestamp, &ntp_time_ms) != 0) {
&elapsed_time_ms, &ntp_time_ms) != 0) {
ASSERT(false);
}
#else

View File

@ -86,7 +86,7 @@ class FakeAdmTest : public testing::Test,
void* audioSamples,
#ifdef USE_WEBRTC_DEV_BRANCH
uint32_t& nSamplesOut,
uint32_t* rtp_timestamp,
int64_t* elapsed_time_ms,
int64_t* ntp_time_ms) {
#else
uint32_t& nSamplesOut) {
@ -98,7 +98,7 @@ class FakeAdmTest : public testing::Test,
GenerateZeroBuffer(audioSamples, audio_buffer_size);
nSamplesOut = bytes_out / nBytesPerSample;
#ifdef USE_WEBRTC_DEV_BRANCH
*rtp_timestamp = 0;
*elapsed_time_ms = 0;
*ntp_time_ms = 0;
#endif
return 0;

View File

@ -475,10 +475,17 @@ int AcmReceiver::GetAudio(int desired_freq_hz, AudioFrame* audio_frame) {
call_stats_.DecodedByNetEq(audio_frame->speech_type_);
// Computes the RTP timestamp of the first sample in |audio_frame| from
// |PlayoutTimestamp|, which is the timestamp of the last sample of
// |GetPlayoutTimestamp|, which is the timestamp of the last sample of
// |audio_frame|.
audio_frame->timestamp_ =
PlayoutTimestamp() - audio_frame->samples_per_channel_;
uint32_t playout_timestamp = 0;
if (GetPlayoutTimestamp(&playout_timestamp)) {
audio_frame->timestamp_ =
playout_timestamp - audio_frame->samples_per_channel_;
} else {
// Remain 0 until we have a valid |playout_timestamp|.
audio_frame->timestamp_ = 0;
}
return 0;
}
@ -596,13 +603,14 @@ void AcmReceiver::set_id(int id) {
id_ = id;
}
uint32_t AcmReceiver::PlayoutTimestamp() {
bool AcmReceiver::GetPlayoutTimestamp(uint32_t* timestamp) {
if (av_sync_) {
assert(initial_delay_manager_.get());
if (initial_delay_manager_->buffering())
return initial_delay_manager_->playout_timestamp();
if (initial_delay_manager_->buffering()) {
return initial_delay_manager_->GetPlayoutTimestamp(timestamp);
}
}
return neteq_->PlayoutTimestamp();
return neteq_->GetPlayoutTimestamp(timestamp);
}
int AcmReceiver::last_audio_codec_id() const {

View File

@ -242,9 +242,10 @@ class AcmReceiver {
void set_id(int id); // TODO(turajs): can be inline.
//
// Returns the RTP timestamp of the last sample delivered by GetAudio().
// Gets the RTP timestamp of the last sample delivered by GetAudio().
// Returns true if the RTP timestamp is valid, otherwise false.
//
uint32_t PlayoutTimestamp();
bool GetPlayoutTimestamp(uint32_t* timestamp);
//
// Return the index of the codec associated with the last non-CNG/non-DTMF

View File

@ -1776,7 +1776,6 @@ int AudioCodingModuleImpl::PlayoutData10Ms(int desired_freq_hz,
}
audio_frame->id_ = id_;
audio_frame->timestamp_ = 0;
return 0;
}
@ -1917,8 +1916,7 @@ int AudioCodingModuleImpl::ConfigISACBandwidthEstimator(
}
int AudioCodingModuleImpl::PlayoutTimestamp(uint32_t* timestamp) {
*timestamp = receiver_.PlayoutTimestamp();
return 0;
return receiver_.GetPlayoutTimestamp(timestamp) ? 0 : -1;
}
bool AudioCodingModuleImpl::HaveValidEncoder(const char* caller_name) const {

View File

@ -219,6 +219,14 @@ void InitialDelayManager::LatePackets(
return;
}
bool InitialDelayManager::GetPlayoutTimestamp(uint32_t* playout_timestamp) {
if (!buffering_) {
return false;
}
*playout_timestamp = playout_timestamp_;
return true;
}
void InitialDelayManager::DisableBuffering() {
buffering_ = false;
}

View File

@ -65,8 +65,9 @@ class InitialDelayManager {
// sequence of late (or perhaps missing) packets is computed.
void LatePackets(uint32_t timestamp_now, SyncStream* sync_stream);
// Playout timestamp, valid when buffering.
uint32_t playout_timestamp() { return playout_timestamp_; }
// Get playout timestamp.
// Returns true if the timestamp is valid (when buffering), otherwise false.
bool GetPlayoutTimestamp(uint32_t* playout_timestamp);
// True if buffered audio is less than the given initial delay (specified at
// the constructor). Buffering might be disabled by the client of this class.

View File

@ -359,7 +359,9 @@ TEST_F(InitialDelayManagerTest, BufferingAudio) {
EXPECT_TRUE(manager_->buffering());
const uint32_t expected_playout_timestamp = rtp_info_.header.timestamp -
kInitDelayMs * kSamplingRateHz / 1000;
EXPECT_EQ(expected_playout_timestamp, manager_->playout_timestamp());
uint32_t actual_playout_timestamp = 0;
EXPECT_TRUE(manager_->GetPlayoutTimestamp(&actual_playout_timestamp));
EXPECT_EQ(expected_playout_timestamp, actual_playout_timestamp);
NextRtpHeader(&rtp_info_, &rtp_receive_timestamp_);
}

View File

@ -228,8 +228,9 @@ class NetEq {
// Disables post-decode VAD.
virtual void DisableVad() = 0;
// Returns the RTP timestamp for the last sample delivered by GetAudio().
virtual uint32_t PlayoutTimestamp() = 0;
// Gets the RTP timestamp for the last sample delivered by GetAudio().
// Returns true if the RTP timestamp is valid, otherwise false.
virtual bool GetPlayoutTimestamp(uint32_t* timestamp) = 0;
// Not implemented.
virtual int SetTargetNumberOfChannels() = 0;

View File

@ -335,9 +335,15 @@ void NetEqImpl::DisableVad() {
vad_->Disable();
}
uint32_t NetEqImpl::PlayoutTimestamp() {
bool NetEqImpl::GetPlayoutTimestamp(uint32_t* timestamp) {
CriticalSectionScoped lock(crit_sect_.get());
return timestamp_scaler_->ToExternal(playout_timestamp_);
if (first_packet_) {
// We don't have a valid RTP timestamp until we have decoded our first
// RTP packet.
return false;
}
*timestamp = timestamp_scaler_->ToExternal(playout_timestamp_);
return true;
}
int NetEqImpl::LastError() {

View File

@ -166,8 +166,7 @@ class NetEqImpl : public webrtc::NetEq {
// Disables post-decode VAD.
virtual void DisableVad();
// Returns the RTP timestamp for the last sample delivered by GetAudio().
virtual uint32_t PlayoutTimestamp();
virtual bool GetPlayoutTimestamp(uint32_t* timestamp);
virtual int SetTargetNumberOfChannels() { return kNotImplemented; }

View File

@ -477,8 +477,10 @@ TEST_F(NetEqImplTest, VerifyTimestampPropagation) {
// The value of the last of the output samples is the same as the number of
// samples played from the decoded packet. Thus, this number + the RTP
// timestamp should match the playout timestamp.
uint32_t timestamp = 0;
EXPECT_TRUE(neteq_->GetPlayoutTimestamp(&timestamp));
EXPECT_EQ(rtp_header.header.timestamp + output[samples_per_channel - 1],
neteq_->PlayoutTimestamp());
timestamp);
// Check the timestamp for the last value in the sync buffer. This should
// be one full frame length ahead of the RTP timestamp.

View File

@ -228,6 +228,8 @@ class NetEqDecodingTest : public ::testing::Test {
void DuplicateCng();
uint32_t PlayoutTimestamp();
NetEq* neteq_;
FILE* rtp_fp_;
unsigned int sim_clock_;
@ -736,7 +738,7 @@ void NetEqDecodingTest::LongCngWithClockDrift(double drift_factor,
}
EXPECT_EQ(kOutputNormal, type);
int32_t delay_before = timestamp - neteq_->PlayoutTimestamp();
int32_t delay_before = timestamp - PlayoutTimestamp();
// Insert CNG for 1 minute (= 60000 ms).
const int kCngPeriodMs = 100;
@ -829,7 +831,7 @@ void NetEqDecodingTest::LongCngWithClockDrift(double drift_factor,
// Check that the speech starts again within reasonable time.
double time_until_speech_returns_ms = t_ms - speech_restart_time_ms;
EXPECT_LT(time_until_speech_returns_ms, max_time_to_speech_ms);
int32_t delay_after = timestamp - neteq_->PlayoutTimestamp();
int32_t delay_after = timestamp - PlayoutTimestamp();
// Compare delay before and after, and make sure it differs less than 20 ms.
EXPECT_LE(delay_after, delay_before + delay_tolerance_ms * 16);
EXPECT_GE(delay_after, delay_before - delay_tolerance_ms * 16);
@ -1310,7 +1312,7 @@ void NetEqDecodingTest::WrapTest(uint16_t start_seq_no,
ASSERT_EQ(1, num_channels);
// Expect delay (in samples) to be less than 2 packets.
EXPECT_LE(timestamp - neteq_->PlayoutTimestamp(),
EXPECT_LE(timestamp - PlayoutTimestamp(),
static_cast<uint32_t>(kSamples * 2));
}
// Make sure we have actually tested wrap-around.
@ -1391,7 +1393,7 @@ void NetEqDecodingTest::DuplicateCng() {
kMaxBlockSize, out_data_, &out_len, &num_channels, &type));
ASSERT_EQ(kBlockSize16kHz, out_len);
EXPECT_EQ(kOutputCNG, type);
EXPECT_EQ(timestamp - algorithmic_delay_samples, neteq_->PlayoutTimestamp());
EXPECT_EQ(timestamp - algorithmic_delay_samples, PlayoutTimestamp());
// Insert the same CNG packet again. Note that at this point it is old, since
// we have already decoded the first copy of it.
@ -1406,7 +1408,7 @@ void NetEqDecodingTest::DuplicateCng() {
ASSERT_EQ(kBlockSize16kHz, out_len);
EXPECT_EQ(kOutputCNG, type);
EXPECT_EQ(timestamp - algorithmic_delay_samples,
neteq_->PlayoutTimestamp());
PlayoutTimestamp());
}
// Insert speech again.
@ -1422,7 +1424,13 @@ void NetEqDecodingTest::DuplicateCng() {
ASSERT_EQ(kBlockSize16kHz, out_len);
EXPECT_EQ(kOutputNormal, type);
EXPECT_EQ(timestamp + kSamples - algorithmic_delay_samples,
neteq_->PlayoutTimestamp());
PlayoutTimestamp());
}
uint32_t NetEqDecodingTest::PlayoutTimestamp() {
uint32_t playout_timestamp = 0;
EXPECT_TRUE(neteq_->GetPlayoutTimestamp(&playout_timestamp));
return playout_timestamp;
}
TEST_F(NetEqDecodingTest, DiscardDuplicateCng) { DuplicateCng(); }

View File

@ -651,6 +651,11 @@ void AudioConferenceMixerImpl::UpdateToMix(
_audioFramePool->PushMemory(audioFrame);
continue;
}
if (_participantList.size() != 1) {
// TODO(wu): Issue 3390, add support for multiple participants case.
audioFrame->ntp_time_ms_ = -1;
}
// TODO(henrike): this assert triggers in some test cases where SRTP is
// used which prevents NetEQ from making a VAD. Temporarily disable this
// assert until the problem is fixed on a higher level.
@ -950,6 +955,16 @@ int32_t AudioConferenceMixerImpl::MixFromList(
return 0;
}
if (audioFrameList->size() == 1) {
mixedAudio.timestamp_ = audioFrameList->front()->timestamp_;
mixedAudio.elapsed_time_ms_ = audioFrameList->front()->elapsed_time_ms_;
} else {
// TODO(wu): Issue 3390.
// Audio frame timestamp is only supported in one channel case.
mixedAudio.timestamp_ = 0;
mixedAudio.elapsed_time_ms_ = -1;
}
for (AudioFrameList::const_iterator iter = audioFrameList->begin();
iter != audioFrameList->end();
++iter) {

View File

@ -548,15 +548,15 @@ int32_t AudioDeviceBuffer::RequestPlayoutData(uint32_t nSamples)
if (_ptrCbAudioTransport)
{
uint32_t res(0);
uint32_t rtp_timestamp = 0;
int64_t ntp_time_ms = 0;
int64_t elapsed_time_ms = -1;
int64_t ntp_time_ms = -1;
res = _ptrCbAudioTransport->NeedMorePlayData(_playSamples,
playBytesPerSample,
playChannels,
playSampleRate,
&_playBuffer[0],
nSamplesOut,
&rtp_timestamp,
&elapsed_time_ms,
&ntp_time_ms);
if (res != 0)
{

View File

@ -71,7 +71,7 @@ public:
const uint32_t samplesPerSec,
void* audioSamples,
uint32_t& nSamplesOut,
uint32_t* rtp_timestamp,
int64_t* elapsed_time_ms,
int64_t* ntp_time_ms) = 0;
// Method to pass captured data directly and unmixed to network channels.
@ -128,7 +128,7 @@ public:
virtual void PullRenderData(int bits_per_sample, int sample_rate,
int number_of_channels, int number_of_frames,
void* audio_data,
uint32_t* rtp_timestamp,
int64_t* elapsed_time_ms,
int64_t* ntp_time_ms) {}
protected:

View File

@ -117,7 +117,7 @@ class AudioTransportAPI: public AudioTransport {
const uint32_t sampleRate,
void* audioSamples,
uint32_t& nSamplesOut,
uint32_t* rtp_timestamp,
int64_t* elapsed_time_ms,
int64_t* ntp_time_ms) {
play_count_++;
if (play_count_ % 100 == 0) {
@ -152,7 +152,7 @@ class AudioTransportAPI: public AudioTransport {
virtual void PullRenderData(int bits_per_sample, int sample_rate,
int number_of_channels, int number_of_frames,
void* audio_data,
uint32_t* rtp_timestamp,
int64_t* elapsed_time_ms,
int64_t* ntp_time_ms) {}
private:
uint32_t rec_count_;

View File

@ -293,7 +293,7 @@ int32_t AudioTransportImpl::NeedMorePlayData(
const uint32_t samplesPerSec,
void* audioSamples,
uint32_t& nSamplesOut,
uint32_t* rtp_timestamp,
int64_t* elapsed_time_ms,
int64_t* ntp_time_ms)
{
if (_fullDuplex)
@ -554,7 +554,7 @@ void AudioTransportImpl::PullRenderData(int bits_per_sample, int sample_rate,
int number_of_channels,
int number_of_frames,
void* audio_data,
uint32_t* rtp_timestamp,
int64_t* elapsed_time_ms,
int64_t* ntp_time_ms) {}
FuncTestManager::FuncTestManager() :

View File

@ -119,7 +119,7 @@ public:
const uint32_t samplesPerSec,
void* audioSamples,
uint32_t& nSamplesOut,
uint32_t* rtp_timestamp,
int64_t* elapsed_time_ms,
int64_t* ntp_time_ms);
virtual int OnDataAvailable(const int voe_channels[],
@ -141,7 +141,7 @@ public:
virtual void PullRenderData(int bits_per_sample, int sample_rate,
int number_of_channels, int number_of_frames,
void* audio_data,
uint32_t* rtp_timestamp,
int64_t* elapsed_time_ms,
int64_t* ntp_time_ms);
AudioTransportImpl(AudioDeviceModule* audioDevice);

View File

@ -690,6 +690,9 @@ class AudioFrame {
int id_;
// RTP timestamp of the first sample in the AudioFrame.
uint32_t timestamp_;
// Time since the first frame in milliseconds.
// -1 represents an uninitialized value.
int64_t elapsed_time_ms_;
// NTP time of the estimated capture time in local timebase in milliseconds.
// -1 represents an uninitialized value.
int64_t ntp_time_ms_;
@ -720,6 +723,7 @@ inline void AudioFrame::Reset() {
// TODO(wu): Zero is a valid value for |timestamp_|. We should initialize
// to an invalid value, or add a new member to indicate invalidity.
timestamp_ = 0;
elapsed_time_ms_ = -1;
ntp_time_ms_ = -1;
samples_per_channel_ = 0;
sample_rate_hz_ = 0;
@ -759,6 +763,8 @@ inline void AudioFrame::CopyFrom(const AudioFrame& src) {
id_ = src.id_;
timestamp_ = src.timestamp_;
elapsed_time_ms_ = src.elapsed_time_ms_;
ntp_time_ms_ = src.ntp_time_ms_;
samples_per_channel_ = src.samples_per_channel_;
sample_rate_hz_ = src.sample_rate_hz_;
speech_type_ = src.speech_type_;

View File

@ -121,8 +121,8 @@ void FakeAudioDevice::CaptureAudio() {
samples_needed = std::min(kFrequencyHz / time_since_last_playout_ms,
kBufferSizeBytes / 2);
uint32_t samples_out = 0;
uint32_t rtp_timestamp = 0;
int64_t ntp_time_ms = 0;
int64_t elapsed_time_ms = -1;
int64_t ntp_time_ms = -1;
EXPECT_EQ(0,
audio_callback_->NeedMorePlayData(samples_needed,
2,
@ -130,7 +130,7 @@ void FakeAudioDevice::CaptureAudio() {
kFrequencyHz,
playout_buffer_,
samples_out,
&rtp_timestamp,
&elapsed_time_ms,
&ntp_time_ms));
}
}

View File

@ -10,6 +10,7 @@
#include "webrtc/voice_engine/channel.h"
#include "webrtc/base/timeutils.h"
#include "webrtc/common.h"
#include "webrtc/modules/audio_device/include/audio_device.h"
#include "webrtc/modules/audio_processing/include/audio_processing.h"
@ -683,21 +684,30 @@ int32_t Channel::GetAudioFrame(int32_t id, AudioFrame& audioFrame)
// Measure audio level (0-9)
_outputAudioLevel.ComputeLevel(audioFrame);
audioFrame.ntp_time_ms_ = ntp_estimator_->Estimate(audioFrame.timestamp_);
if (!first_frame_arrived_) {
first_frame_arrived_ = true;
if (capture_start_rtp_time_stamp_ < 0 && audioFrame.timestamp_ != 0) {
// The first frame with a valid rtp timestamp.
capture_start_rtp_time_stamp_ = audioFrame.timestamp_;
} else {
}
if (capture_start_rtp_time_stamp_ >= 0) {
// audioFrame.timestamp_ should be valid from now on.
// Compute elapsed time.
int64_t unwrap_timestamp =
rtp_ts_wraparound_handler_->Unwrap(audioFrame.timestamp_);
audioFrame.elapsed_time_ms_ =
(unwrap_timestamp - capture_start_rtp_time_stamp_) /
(GetPlayoutFrequency() / 1000);
// Compute ntp time.
audioFrame.ntp_time_ms_ = ntp_estimator_->Estimate(audioFrame.timestamp_);
// |ntp_time_ms_| won't be valid until at least 2 RTCP SRs are received.
if (audioFrame.ntp_time_ms_ > 0) {
// Compute |capture_start_ntp_time_ms_| so that
// |capture_start_ntp_time_ms_| + |elapsed_time_ms| == |ntp_time_ms_|
// |capture_start_ntp_time_ms_| + |elapsed_time_ms_| == |ntp_time_ms_|
CriticalSectionScoped lock(ts_stats_lock_.get());
uint32_t elapsed_time_ms =
(audioFrame.timestamp_ - capture_start_rtp_time_stamp_) /
(audioFrame.sample_rate_hz_ * 1000);
capture_start_ntp_time_ms_ = audioFrame.ntp_time_ms_ - elapsed_time_ms;
capture_start_ntp_time_ms_ =
audioFrame.ntp_time_ms_ - audioFrame.elapsed_time_ms_;
}
}
@ -875,8 +885,8 @@ Channel::Channel(int32_t channelId,
_numberOfDiscardedPackets(0),
send_sequence_number_(0),
ts_stats_lock_(CriticalSectionWrapper::CreateCriticalSection()),
first_frame_arrived_(false),
capture_start_rtp_time_stamp_(0),
rtp_ts_wraparound_handler_(new rtc::TimestampWrapAroundHandler()),
capture_start_rtp_time_stamp_(-1),
capture_start_ntp_time_ms_(-1),
_engineStatisticsPtr(NULL),
_outputMixerPtr(NULL),
@ -4045,20 +4055,10 @@ void Channel::UpdatePlayoutTimestamp(bool rtcp) {
return;
}
int32_t playout_frequency = audio_coding_->PlayoutFrequency();
CodecInst current_recive_codec;
if (audio_coding_->ReceiveCodec(&current_recive_codec) == 0) {
if (STR_CASE_CMP("G722", current_recive_codec.plname) == 0) {
playout_frequency = 8000;
} else if (STR_CASE_CMP("opus", current_recive_codec.plname) == 0) {
playout_frequency = 48000;
}
}
jitter_buffer_playout_timestamp_ = playout_timestamp;
// Remove the playout delay.
playout_timestamp -= (delay_ms * (playout_frequency / 1000));
playout_timestamp -= (delay_ms * (GetPlayoutFrequency() / 1000));
WEBRTC_TRACE(kTraceStream, kTraceVoice, VoEId(_instanceId,_channelId),
"Channel::UpdatePlayoutTimestamp() => playoutTimestamp = %lu",
@ -4364,30 +4364,11 @@ void Channel::UpdatePacketDelay(uint32_t rtp_timestamp,
rtp_timestamp, sequence_number);
// Get frequency of last received payload
int rtp_receive_frequency = audio_coding_->ReceiveFrequency();
CodecInst current_receive_codec;
if (audio_coding_->ReceiveCodec(&current_receive_codec) != 0) {
return;
}
int rtp_receive_frequency = GetPlayoutFrequency();
// Update the least required delay.
least_required_delay_ms_ = audio_coding_->LeastRequiredDelayMs();
if (STR_CASE_CMP("G722", current_receive_codec.plname) == 0) {
// Even though the actual sampling rate for G.722 audio is
// 16,000 Hz, the RTP clock rate for the G722 payload format is
// 8,000 Hz because that value was erroneously assigned in
// RFC 1890 and must remain unchanged for backward compatibility.
rtp_receive_frequency = 8000;
} else if (STR_CASE_CMP("opus", current_receive_codec.plname) == 0) {
// We are resampling Opus internally to 32,000 Hz until all our
// DSP routines can operate at 48,000 Hz, but the RTP clock
// rate for the Opus payload format is standardized to 48,000 Hz,
// because that is the maximum supported decoding sampling rate.
rtp_receive_frequency = 48000;
}
// |jitter_buffer_playout_timestamp_| updated in UpdatePlayoutTimestamp for
// every incoming packet.
uint32_t timestamp_diff_ms = (rtp_timestamp -
@ -4560,5 +4541,26 @@ int Channel::SetSendRtpHeaderExtension(bool enable, RTPExtensionType type,
return error;
}
int32_t Channel::GetPlayoutFrequency() {
int32_t playout_frequency = audio_coding_->PlayoutFrequency();
CodecInst current_recive_codec;
if (audio_coding_->ReceiveCodec(&current_recive_codec) == 0) {
if (STR_CASE_CMP("G722", current_recive_codec.plname) == 0) {
// Even though the actual sampling rate for G.722 audio is
// 16,000 Hz, the RTP clock rate for the G722 payload format is
// 8,000 Hz because that value was erroneously assigned in
// RFC 1890 and must remain unchanged for backward compatibility.
playout_frequency = 8000;
} else if (STR_CASE_CMP("opus", current_recive_codec.plname) == 0) {
// We are resampling Opus internally to 32,000 Hz until all our
// DSP routines can operate at 48,000 Hz, but the RTP clock
// rate for the Opus payload format is standardized to 48,000 Hz,
// because that is the maximum supported decoding sampling rate.
playout_frequency = 48000;
}
}
return playout_frequency;
}
} // namespace voe
} // namespace webrtc

View File

@ -35,6 +35,11 @@
#include "webrtc/voice_engine/include/voe_dtmf.h"
#endif
namespace rtc {
class TimestampWrapAroundHandler;
}
namespace webrtc {
class AudioDeviceModule;
@ -500,6 +505,8 @@ private:
int SetSendRtpHeaderExtension(bool enable, RTPExtensionType type,
unsigned char id);
int32_t GetPlayoutFrequency();
CriticalSectionWrapper& _fileCritSect;
CriticalSectionWrapper& _callbackCritSect;
CriticalSectionWrapper& volume_settings_critsect_;
@ -553,9 +560,9 @@ private:
scoped_ptr<CriticalSectionWrapper> ts_stats_lock_;
bool first_frame_arrived_;
scoped_ptr<rtc::TimestampWrapAroundHandler> rtp_ts_wraparound_handler_;
// The rtp timestamp of the first played out audio frame.
uint32_t capture_start_rtp_time_stamp_;
int64_t capture_start_rtp_time_stamp_;
// The capture ntp time (in local timebase) of the first played out audio
// frame.
int64_t capture_start_ntp_time_ms_;

View File

@ -65,6 +65,10 @@ void RemixAndResample(const AudioFrame& src_frame,
dst_frame->num_channels_ = 1;
AudioFrameOperations::MonoToStereo(dst_frame);
}
dst_frame->timestamp_ = src_frame.timestamp_;
dst_frame->elapsed_time_ms_ = src_frame.elapsed_time_ms_;
dst_frame->ntp_time_ms_ = src_frame.ntp_time_ms_;
}
void DownConvertToCodecFormat(const int16_t* src_data,

View File

@ -149,7 +149,7 @@ int32_t VoEBaseImpl::NeedMorePlayData(
uint32_t samplesPerSec,
void* audioSamples,
uint32_t& nSamplesOut,
uint32_t* rtp_timestamp,
int64_t* elapsed_time_ms,
int64_t* ntp_time_ms)
{
WEBRTC_TRACE(kTraceStream, kTraceVoice, VoEId(_shared->instance_id(), -1),
@ -160,7 +160,7 @@ int32_t VoEBaseImpl::NeedMorePlayData(
GetPlayoutData(static_cast<int>(samplesPerSec),
static_cast<int>(nChannels),
static_cast<int>(nSamples), true, audioSamples,
rtp_timestamp, ntp_time_ms);
elapsed_time_ms, ntp_time_ms);
nSamplesOut = _audioFrame.samples_per_channel_;
@ -237,13 +237,13 @@ void VoEBaseImpl::PushCaptureData(int voe_channel, const void* audio_data,
void VoEBaseImpl::PullRenderData(int bits_per_sample, int sample_rate,
int number_of_channels, int number_of_frames,
void* audio_data,
uint32_t* rtp_timestamp,
int64_t* elapsed_time_ms,
int64_t* ntp_time_ms) {
assert(bits_per_sample == 16);
assert(number_of_frames == static_cast<int>(sample_rate / 100));
GetPlayoutData(sample_rate, number_of_channels, number_of_frames, false,
audio_data, rtp_timestamp, ntp_time_ms);
audio_data, elapsed_time_ms, ntp_time_ms);
}
int VoEBaseImpl::RegisterVoiceEngineObserver(VoiceEngineObserver& observer)
@ -1087,7 +1087,7 @@ int VoEBaseImpl::ProcessRecordedDataWithAPM(
void VoEBaseImpl::GetPlayoutData(int sample_rate, int number_of_channels,
int number_of_frames, bool feed_data_to_apm,
void* audio_data,
uint32_t* rtp_timestamp,
int64_t* elapsed_time_ms,
int64_t* ntp_time_ms) {
assert(_shared->output_mixer() != NULL);
@ -1110,7 +1110,7 @@ void VoEBaseImpl::GetPlayoutData(int sample_rate, int number_of_channels,
memcpy(audio_data, _audioFrame.data_,
sizeof(int16_t) * number_of_frames * number_of_channels);
*rtp_timestamp = _audioFrame.timestamp_;
*elapsed_time_ms = _audioFrame.elapsed_time_ms_;
*ntp_time_ms = _audioFrame.ntp_time_ms_;
}

View File

@ -80,7 +80,7 @@ public:
uint32_t samplesPerSec,
void* audioSamples,
uint32_t& nSamplesOut,
uint32_t* rtp_timestamp,
int64_t* elapsed_time_ms,
int64_t* ntp_time_ms);
virtual int OnDataAvailable(const int voe_channels[],
@ -105,7 +105,7 @@ public:
virtual void PullRenderData(int bits_per_sample, int sample_rate,
int number_of_channels, int number_of_frames,
void* audio_data,
uint32_t* rtp_timestamp,
int64_t* elapsed_time_ms,
int64_t* ntp_time_ms);
// AudioDeviceObserver
@ -143,7 +143,7 @@ private:
void GetPlayoutData(int sample_rate, int number_of_channels,
int number_of_frames, bool feed_data_to_apm,
void* audio_data,
uint32_t* rtp_timestamp,
int64_t* elapsed_time_ms,
int64_t* ntp_time_ms);
int32_t AddBuildInfo(char* str) const;