
scoped_array is deprecated. This was done using a Chromium clang tool: http://src.chromium.org/viewvc/chrome/trunk/src/tools/clang/rewrite_scoped_ar... except for the few not-built-on-Linux files which were updated manually. TESTED=trybots BUG=2515 R=niklas.enbom@webrtc.org Review URL: https://webrtc-codereview.appspot.com/12429004 git-svn-id: http://webrtc.googlecode.com/svn/trunk@5985 4adac7df-926f-26a2-2b94-8c16560cd09d
1097 lines
39 KiB
C++
1097 lines
39 KiB
C++
/*
|
|
* Copyright (c) 2012 The WebRTC project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#include <math.h>
|
|
#include <stdio.h>
|
|
#include <string.h>
|
|
#ifdef WEBRTC_ANDROID
|
|
#include <sys/stat.h>
|
|
#endif
|
|
|
|
#include <algorithm>
|
|
|
|
#include "webrtc/common.h"
|
|
#include "webrtc/modules/audio_processing/include/audio_processing.h"
|
|
#include "webrtc/modules/audio_processing/test/test_utils.h"
|
|
#include "webrtc/modules/interface/module_common_types.h"
|
|
#include "webrtc/system_wrappers/interface/cpu_features_wrapper.h"
|
|
#include "webrtc/system_wrappers/interface/scoped_ptr.h"
|
|
#include "webrtc/system_wrappers/interface/tick_util.h"
|
|
#include "webrtc/test/testsupport/fileutils.h"
|
|
#include "webrtc/test/testsupport/perf_test.h"
|
|
#ifdef WEBRTC_ANDROID_PLATFORM_BUILD
|
|
#include "gtest/gtest.h"
|
|
#include "external/webrtc/webrtc/modules/audio_processing/debug.pb.h"
|
|
#else
|
|
#include "testing/gtest/include/gtest/gtest.h"
|
|
#include "webrtc/audio_processing/debug.pb.h"
|
|
#endif
|
|
|
|
namespace webrtc {
|
|
|
|
using webrtc::audioproc::Event;
|
|
using webrtc::audioproc::Init;
|
|
using webrtc::audioproc::ReverseStream;
|
|
using webrtc::audioproc::Stream;
|
|
|
|
namespace {
|
|
|
|
void PrintStat(const AudioProcessing::Statistic& stat) {
|
|
printf("%d, %d, %d\n", stat.average,
|
|
stat.maximum,
|
|
stat.minimum);
|
|
}
|
|
|
|
void usage() {
|
|
printf(
|
|
"Usage: process_test [options] [-pb PROTOBUF_FILE]\n"
|
|
" [-ir REVERSE_FILE] [-i PRIMARY_FILE] [-o OUT_FILE]\n");
|
|
printf(
|
|
"process_test is a test application for AudioProcessing.\n\n"
|
|
"When a protobuf debug file is available, specify it with -pb. Alternately,\n"
|
|
"when -ir or -i is used, the specified files will be processed directly in\n"
|
|
"a simulation mode. Otherwise the full set of legacy test files is expected\n"
|
|
"to be present in the working directory. OUT_FILE should be specified\n"
|
|
"without extension to support both int and float output.\n\n");
|
|
printf("Options\n");
|
|
printf("General configuration (only used for the simulation mode):\n");
|
|
printf(" -fs SAMPLE_RATE_HZ\n");
|
|
printf(" -ch CHANNELS_IN CHANNELS_OUT\n");
|
|
printf(" -rch REVERSE_CHANNELS\n");
|
|
printf("\n");
|
|
printf("Component configuration:\n");
|
|
printf(
|
|
"All components are disabled by default. Each block below begins with a\n"
|
|
"flag to enable the component with default settings. The subsequent flags\n"
|
|
"in the block are used to provide configuration settings.\n");
|
|
printf("\n -aec Echo cancellation\n");
|
|
printf(" --drift_compensation\n");
|
|
printf(" --no_drift_compensation\n");
|
|
printf(" --no_echo_metrics\n");
|
|
printf(" --no_delay_logging\n");
|
|
printf(" --aec_suppression_level LEVEL [0 - 2]\n");
|
|
printf(" --extended_filter\n");
|
|
printf("\n -aecm Echo control mobile\n");
|
|
printf(" --aecm_echo_path_in_file FILE\n");
|
|
printf(" --aecm_echo_path_out_file FILE\n");
|
|
printf(" --no_comfort_noise\n");
|
|
printf(" --routing_mode MODE [0 - 4]\n");
|
|
printf("\n -agc Gain control\n");
|
|
printf(" --analog\n");
|
|
printf(" --adaptive_digital\n");
|
|
printf(" --fixed_digital\n");
|
|
printf(" --target_level LEVEL\n");
|
|
printf(" --compression_gain GAIN\n");
|
|
printf(" --limiter\n");
|
|
printf(" --no_limiter\n");
|
|
printf("\n -hpf High pass filter\n");
|
|
printf("\n -ns Noise suppression\n");
|
|
printf(" --ns_low\n");
|
|
printf(" --ns_moderate\n");
|
|
printf(" --ns_high\n");
|
|
printf(" --ns_very_high\n");
|
|
printf(" --ns_prob_file FILE\n");
|
|
printf("\n -vad Voice activity detection\n");
|
|
printf(" --vad_out_file FILE\n");
|
|
printf("\n -expns Experimental noise suppression\n");
|
|
printf("\n Level metrics (enabled by default)\n");
|
|
printf(" --no_level_metrics\n");
|
|
printf("\n");
|
|
printf("Modifiers:\n");
|
|
printf(" --noasm Disable SSE optimization.\n");
|
|
printf(" --add_delay DELAY Add DELAY ms to input value.\n");
|
|
printf(" --delay DELAY Override input delay with DELAY ms.\n");
|
|
printf(" --perf Measure performance.\n");
|
|
printf(" --quiet Suppress text output.\n");
|
|
printf(" --no_progress Suppress progress.\n");
|
|
printf(" --debug_file FILE Dump a debug recording.\n");
|
|
}
|
|
|
|
static float MicLevel2Gain(int level) {
|
|
return pow(10.0f, ((level - 127.0f) / 128.0f * 40.0f) / 20.0f);
|
|
}
|
|
|
|
static void SimulateMic(int mic_level, AudioFrame* frame) {
|
|
mic_level = std::min(std::max(mic_level, 0), 255);
|
|
float mic_gain = MicLevel2Gain(mic_level);
|
|
int num_samples = frame->samples_per_channel_ * frame->num_channels_;
|
|
float v;
|
|
for (int n = 0; n < num_samples; n++) {
|
|
v = floor(frame->data_[n] * mic_gain + 0.5);
|
|
v = std::max(std::min(32767.0f, v), -32768.0f);
|
|
frame->data_[n] = static_cast<int16_t>(v);
|
|
}
|
|
}
|
|
|
|
// void function for gtest.
|
|
void void_main(int argc, char* argv[]) {
|
|
if (argc > 1 && strcmp(argv[1], "--help") == 0) {
|
|
usage();
|
|
return;
|
|
}
|
|
|
|
if (argc < 2) {
|
|
printf("Did you mean to run without arguments?\n");
|
|
printf("Try `process_test --help' for more information.\n\n");
|
|
}
|
|
|
|
scoped_ptr<AudioProcessing> apm(AudioProcessing::Create());
|
|
ASSERT_TRUE(apm.get() != NULL);
|
|
|
|
const char* pb_filename = NULL;
|
|
const char* far_filename = NULL;
|
|
const char* near_filename = NULL;
|
|
std::string out_filename;
|
|
const char* vad_out_filename = NULL;
|
|
const char* ns_prob_filename = NULL;
|
|
const char* aecm_echo_path_in_filename = NULL;
|
|
const char* aecm_echo_path_out_filename = NULL;
|
|
|
|
int32_t sample_rate_hz = 16000;
|
|
|
|
int num_capture_input_channels = 1;
|
|
int num_capture_output_channels = 1;
|
|
int num_render_channels = 1;
|
|
|
|
int samples_per_channel = sample_rate_hz / 100;
|
|
|
|
bool simulating = false;
|
|
bool perf_testing = false;
|
|
bool verbose = true;
|
|
bool progress = true;
|
|
int extra_delay_ms = 0;
|
|
int override_delay_ms = 0;
|
|
|
|
ASSERT_EQ(apm->kNoError, apm->level_estimator()->Enable(true));
|
|
for (int i = 1; i < argc; i++) {
|
|
if (strcmp(argv[i], "-pb") == 0) {
|
|
i++;
|
|
ASSERT_LT(i, argc) << "Specify protobuf filename after -pb";
|
|
pb_filename = argv[i];
|
|
|
|
} else if (strcmp(argv[i], "-ir") == 0) {
|
|
i++;
|
|
ASSERT_LT(i, argc) << "Specify filename after -ir";
|
|
far_filename = argv[i];
|
|
simulating = true;
|
|
|
|
} else if (strcmp(argv[i], "-i") == 0) {
|
|
i++;
|
|
ASSERT_LT(i, argc) << "Specify filename after -i";
|
|
near_filename = argv[i];
|
|
simulating = true;
|
|
|
|
} else if (strcmp(argv[i], "-o") == 0) {
|
|
i++;
|
|
ASSERT_LT(i, argc) << "Specify filename without extension after -o";
|
|
out_filename = argv[i];
|
|
|
|
} else if (strcmp(argv[i], "-fs") == 0) {
|
|
i++;
|
|
ASSERT_LT(i, argc) << "Specify sample rate after -fs";
|
|
ASSERT_EQ(1, sscanf(argv[i], "%d", &sample_rate_hz));
|
|
samples_per_channel = sample_rate_hz / 100;
|
|
|
|
} else if (strcmp(argv[i], "-ch") == 0) {
|
|
i++;
|
|
ASSERT_LT(i + 1, argc) << "Specify number of channels after -ch";
|
|
ASSERT_EQ(1, sscanf(argv[i], "%d", &num_capture_input_channels));
|
|
i++;
|
|
ASSERT_EQ(1, sscanf(argv[i], "%d", &num_capture_output_channels));
|
|
|
|
} else if (strcmp(argv[i], "-rch") == 0) {
|
|
i++;
|
|
ASSERT_LT(i, argc) << "Specify number of channels after -rch";
|
|
ASSERT_EQ(1, sscanf(argv[i], "%d", &num_render_channels));
|
|
|
|
} else if (strcmp(argv[i], "-aec") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->echo_cancellation()->Enable(true));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->echo_cancellation()->enable_metrics(true));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->echo_cancellation()->enable_delay_logging(true));
|
|
|
|
} else if (strcmp(argv[i], "--drift_compensation") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->echo_cancellation()->Enable(true));
|
|
// TODO(ajm): this is enabled in the VQE test app by default. Investigate
|
|
// why it can give better performance despite passing zeros.
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->echo_cancellation()->enable_drift_compensation(true));
|
|
} else if (strcmp(argv[i], "--no_drift_compensation") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->echo_cancellation()->Enable(true));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->echo_cancellation()->enable_drift_compensation(false));
|
|
|
|
} else if (strcmp(argv[i], "--no_echo_metrics") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->echo_cancellation()->Enable(true));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->echo_cancellation()->enable_metrics(false));
|
|
|
|
} else if (strcmp(argv[i], "--no_delay_logging") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->echo_cancellation()->Enable(true));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->echo_cancellation()->enable_delay_logging(false));
|
|
|
|
} else if (strcmp(argv[i], "--no_level_metrics") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->level_estimator()->Enable(false));
|
|
|
|
} else if (strcmp(argv[i], "--aec_suppression_level") == 0) {
|
|
i++;
|
|
ASSERT_LT(i, argc) << "Specify level after --aec_suppression_level";
|
|
int suppression_level;
|
|
ASSERT_EQ(1, sscanf(argv[i], "%d", &suppression_level));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->echo_cancellation()->set_suppression_level(
|
|
static_cast<webrtc::EchoCancellation::SuppressionLevel>(
|
|
suppression_level)));
|
|
|
|
} else if (strcmp(argv[i], "--extended_filter") == 0) {
|
|
Config config;
|
|
config.Set<DelayCorrection>(new DelayCorrection(true));
|
|
apm->SetExtraOptions(config);
|
|
|
|
} else if (strcmp(argv[i], "-aecm") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->echo_control_mobile()->Enable(true));
|
|
|
|
} else if (strcmp(argv[i], "--aecm_echo_path_in_file") == 0) {
|
|
i++;
|
|
ASSERT_LT(i, argc) << "Specify filename after --aecm_echo_path_in_file";
|
|
aecm_echo_path_in_filename = argv[i];
|
|
|
|
} else if (strcmp(argv[i], "--aecm_echo_path_out_file") == 0) {
|
|
i++;
|
|
ASSERT_LT(i, argc) << "Specify filename after --aecm_echo_path_out_file";
|
|
aecm_echo_path_out_filename = argv[i];
|
|
|
|
} else if (strcmp(argv[i], "--no_comfort_noise") == 0) {
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->echo_control_mobile()->enable_comfort_noise(false));
|
|
|
|
} else if (strcmp(argv[i], "--routing_mode") == 0) {
|
|
i++;
|
|
ASSERT_LT(i, argc) << "Specify mode after --routing_mode";
|
|
int routing_mode;
|
|
ASSERT_EQ(1, sscanf(argv[i], "%d", &routing_mode));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->echo_control_mobile()->set_routing_mode(
|
|
static_cast<webrtc::EchoControlMobile::RoutingMode>(
|
|
routing_mode)));
|
|
|
|
} else if (strcmp(argv[i], "-agc") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->gain_control()->Enable(true));
|
|
|
|
} else if (strcmp(argv[i], "--analog") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->gain_control()->Enable(true));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->gain_control()->set_mode(GainControl::kAdaptiveAnalog));
|
|
|
|
} else if (strcmp(argv[i], "--adaptive_digital") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->gain_control()->Enable(true));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->gain_control()->set_mode(GainControl::kAdaptiveDigital));
|
|
|
|
} else if (strcmp(argv[i], "--fixed_digital") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->gain_control()->Enable(true));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->gain_control()->set_mode(GainControl::kFixedDigital));
|
|
|
|
} else if (strcmp(argv[i], "--target_level") == 0) {
|
|
i++;
|
|
int level;
|
|
ASSERT_EQ(1, sscanf(argv[i], "%d", &level));
|
|
|
|
ASSERT_EQ(apm->kNoError, apm->gain_control()->Enable(true));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->gain_control()->set_target_level_dbfs(level));
|
|
|
|
} else if (strcmp(argv[i], "--compression_gain") == 0) {
|
|
i++;
|
|
int gain;
|
|
ASSERT_EQ(1, sscanf(argv[i], "%d", &gain));
|
|
|
|
ASSERT_EQ(apm->kNoError, apm->gain_control()->Enable(true));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->gain_control()->set_compression_gain_db(gain));
|
|
|
|
} else if (strcmp(argv[i], "--limiter") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->gain_control()->Enable(true));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->gain_control()->enable_limiter(true));
|
|
|
|
} else if (strcmp(argv[i], "--no_limiter") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->gain_control()->Enable(true));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->gain_control()->enable_limiter(false));
|
|
|
|
} else if (strcmp(argv[i], "-hpf") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->high_pass_filter()->Enable(true));
|
|
|
|
} else if (strcmp(argv[i], "-ns") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->noise_suppression()->Enable(true));
|
|
|
|
} else if (strcmp(argv[i], "--ns_low") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->noise_suppression()->Enable(true));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->noise_suppression()->set_level(NoiseSuppression::kLow));
|
|
|
|
} else if (strcmp(argv[i], "--ns_moderate") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->noise_suppression()->Enable(true));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->noise_suppression()->set_level(NoiseSuppression::kModerate));
|
|
|
|
} else if (strcmp(argv[i], "--ns_high") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->noise_suppression()->Enable(true));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->noise_suppression()->set_level(NoiseSuppression::kHigh));
|
|
|
|
} else if (strcmp(argv[i], "--ns_very_high") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->noise_suppression()->Enable(true));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->noise_suppression()->set_level(NoiseSuppression::kVeryHigh));
|
|
|
|
} else if (strcmp(argv[i], "--ns_prob_file") == 0) {
|
|
i++;
|
|
ASSERT_LT(i, argc) << "Specify filename after --ns_prob_file";
|
|
ns_prob_filename = argv[i];
|
|
|
|
} else if (strcmp(argv[i], "-vad") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->voice_detection()->Enable(true));
|
|
|
|
} else if (strcmp(argv[i], "--vad_very_low") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->voice_detection()->Enable(true));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->voice_detection()->set_likelihood(
|
|
VoiceDetection::kVeryLowLikelihood));
|
|
|
|
} else if (strcmp(argv[i], "--vad_low") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->voice_detection()->Enable(true));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->voice_detection()->set_likelihood(
|
|
VoiceDetection::kLowLikelihood));
|
|
|
|
} else if (strcmp(argv[i], "--vad_moderate") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->voice_detection()->Enable(true));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->voice_detection()->set_likelihood(
|
|
VoiceDetection::kModerateLikelihood));
|
|
|
|
} else if (strcmp(argv[i], "--vad_high") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->voice_detection()->Enable(true));
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->voice_detection()->set_likelihood(
|
|
VoiceDetection::kHighLikelihood));
|
|
|
|
} else if (strcmp(argv[i], "--vad_out_file") == 0) {
|
|
i++;
|
|
ASSERT_LT(i, argc) << "Specify filename after --vad_out_file";
|
|
vad_out_filename = argv[i];
|
|
|
|
} else if (strcmp(argv[i], "-expns") == 0) {
|
|
ASSERT_EQ(apm->kNoError, apm->EnableExperimentalNs(true));
|
|
|
|
} else if (strcmp(argv[i], "--noasm") == 0) {
|
|
WebRtc_GetCPUInfo = WebRtc_GetCPUInfoNoASM;
|
|
// We need to reinitialize here if components have already been enabled.
|
|
ASSERT_EQ(apm->kNoError, apm->Initialize());
|
|
|
|
} else if (strcmp(argv[i], "--add_delay") == 0) {
|
|
i++;
|
|
ASSERT_EQ(1, sscanf(argv[i], "%d", &extra_delay_ms));
|
|
|
|
} else if (strcmp(argv[i], "--delay") == 0) {
|
|
i++;
|
|
ASSERT_EQ(1, sscanf(argv[i], "%d", &override_delay_ms));
|
|
|
|
} else if (strcmp(argv[i], "--perf") == 0) {
|
|
perf_testing = true;
|
|
|
|
} else if (strcmp(argv[i], "--quiet") == 0) {
|
|
verbose = false;
|
|
progress = false;
|
|
|
|
} else if (strcmp(argv[i], "--no_progress") == 0) {
|
|
progress = false;
|
|
|
|
} else if (strcmp(argv[i], "--debug_file") == 0) {
|
|
i++;
|
|
ASSERT_LT(i, argc) << "Specify filename after --debug_file";
|
|
ASSERT_EQ(apm->kNoError, apm->StartDebugRecording(argv[i]));
|
|
} else {
|
|
FAIL() << "Unrecognized argument " << argv[i];
|
|
}
|
|
}
|
|
// If we're reading a protobuf file, ensure a simulation hasn't also
|
|
// been requested (which makes no sense...)
|
|
ASSERT_FALSE(pb_filename && simulating);
|
|
|
|
if (verbose) {
|
|
printf("Sample rate: %d Hz\n", sample_rate_hz);
|
|
printf("Primary channels: %d (in), %d (out)\n",
|
|
num_capture_input_channels,
|
|
num_capture_output_channels);
|
|
printf("Reverse channels: %d \n", num_render_channels);
|
|
}
|
|
|
|
const std::string out_path = webrtc::test::OutputPath();
|
|
const char far_file_default[] = "apm_far.pcm";
|
|
const char near_file_default[] = "apm_near.pcm";
|
|
const char event_filename[] = "apm_event.dat";
|
|
const char delay_filename[] = "apm_delay.dat";
|
|
const char drift_filename[] = "apm_drift.dat";
|
|
const std::string vad_file_default = out_path + "vad_out.dat";
|
|
const std::string ns_prob_file_default = out_path + "ns_prob.dat";
|
|
|
|
if (!simulating) {
|
|
far_filename = far_file_default;
|
|
near_filename = near_file_default;
|
|
}
|
|
|
|
if (out_filename.size() == 0) {
|
|
out_filename = out_path + "out";
|
|
}
|
|
std::string out_float_filename = out_filename + ".float";
|
|
out_filename += ".pcm";
|
|
|
|
if (!vad_out_filename) {
|
|
vad_out_filename = vad_file_default.c_str();
|
|
}
|
|
|
|
if (!ns_prob_filename) {
|
|
ns_prob_filename = ns_prob_file_default.c_str();
|
|
}
|
|
|
|
FILE* pb_file = NULL;
|
|
FILE* far_file = NULL;
|
|
FILE* near_file = NULL;
|
|
FILE* event_file = NULL;
|
|
FILE* delay_file = NULL;
|
|
FILE* drift_file = NULL;
|
|
FILE* vad_out_file = NULL;
|
|
FILE* ns_prob_file = NULL;
|
|
FILE* aecm_echo_path_in_file = NULL;
|
|
FILE* aecm_echo_path_out_file = NULL;
|
|
|
|
if (pb_filename) {
|
|
pb_file = OpenFile(pb_filename, "rb");
|
|
} else {
|
|
if (far_filename) {
|
|
far_file = OpenFile(far_filename, "rb");
|
|
}
|
|
|
|
near_file = OpenFile(near_filename, "rb");
|
|
if (!simulating) {
|
|
event_file = OpenFile(event_filename, "rb");
|
|
delay_file = OpenFile(delay_filename, "rb");
|
|
drift_file = OpenFile(drift_filename, "rb");
|
|
}
|
|
}
|
|
|
|
int near_size_bytes = 0;
|
|
if (pb_file) {
|
|
struct stat st;
|
|
stat(pb_filename, &st);
|
|
// Crude estimate, but should be good enough.
|
|
near_size_bytes = st.st_size / 3;
|
|
} else {
|
|
struct stat st;
|
|
stat(near_filename, &st);
|
|
near_size_bytes = st.st_size;
|
|
}
|
|
|
|
if (apm->voice_detection()->is_enabled()) {
|
|
vad_out_file = OpenFile(vad_out_filename, "wb");
|
|
}
|
|
|
|
if (apm->noise_suppression()->is_enabled()) {
|
|
ns_prob_file = OpenFile(ns_prob_filename, "wb");
|
|
}
|
|
|
|
if (aecm_echo_path_in_filename != NULL) {
|
|
aecm_echo_path_in_file = OpenFile(aecm_echo_path_in_filename, "rb");
|
|
|
|
const size_t path_size =
|
|
apm->echo_control_mobile()->echo_path_size_bytes();
|
|
scoped_ptr<char[]> echo_path(new char[path_size]);
|
|
ASSERT_EQ(path_size, fread(echo_path.get(),
|
|
sizeof(char),
|
|
path_size,
|
|
aecm_echo_path_in_file));
|
|
EXPECT_EQ(apm->kNoError,
|
|
apm->echo_control_mobile()->SetEchoPath(echo_path.get(),
|
|
path_size));
|
|
fclose(aecm_echo_path_in_file);
|
|
aecm_echo_path_in_file = NULL;
|
|
}
|
|
|
|
if (aecm_echo_path_out_filename != NULL) {
|
|
aecm_echo_path_out_file = OpenFile(aecm_echo_path_out_filename, "wb");
|
|
}
|
|
|
|
size_t read_count = 0;
|
|
int reverse_count = 0;
|
|
int primary_count = 0;
|
|
int near_read_bytes = 0;
|
|
TickInterval acc_ticks;
|
|
|
|
AudioFrame far_frame;
|
|
AudioFrame near_frame;
|
|
|
|
int delay_ms = 0;
|
|
int drift_samples = 0;
|
|
int capture_level = 127;
|
|
int8_t stream_has_voice = 0;
|
|
float ns_speech_prob = 0.0f;
|
|
|
|
TickTime t0 = TickTime::Now();
|
|
TickTime t1 = t0;
|
|
int64_t max_time_us = 0;
|
|
int64_t max_time_reverse_us = 0;
|
|
int64_t min_time_us = 1e6;
|
|
int64_t min_time_reverse_us = 1e6;
|
|
|
|
// TODO(ajm): Ideally we would refactor this block into separate functions,
|
|
// but for now we want to share the variables.
|
|
if (pb_file) {
|
|
Event event_msg;
|
|
scoped_ptr<ChannelBuffer<float> > reverse_cb;
|
|
scoped_ptr<ChannelBuffer<float> > primary_cb;
|
|
int output_sample_rate = 32000;
|
|
AudioProcessing::ChannelLayout output_layout = AudioProcessing::kMono;
|
|
while (ReadMessageFromFile(pb_file, &event_msg)) {
|
|
std::ostringstream trace_stream;
|
|
trace_stream << "Processed frames: " << reverse_count << " (reverse), "
|
|
<< primary_count << " (primary)";
|
|
SCOPED_TRACE(trace_stream.str());
|
|
|
|
if (event_msg.type() == Event::INIT) {
|
|
ASSERT_TRUE(event_msg.has_init());
|
|
const Init msg = event_msg.init();
|
|
|
|
ASSERT_TRUE(msg.has_sample_rate());
|
|
ASSERT_TRUE(msg.has_num_input_channels());
|
|
ASSERT_TRUE(msg.has_num_output_channels());
|
|
ASSERT_TRUE(msg.has_num_reverse_channels());
|
|
int reverse_sample_rate = msg.sample_rate();
|
|
if (msg.has_reverse_sample_rate()) {
|
|
reverse_sample_rate = msg.reverse_sample_rate();
|
|
}
|
|
output_sample_rate = msg.sample_rate();
|
|
if (msg.has_output_sample_rate()) {
|
|
output_sample_rate = msg.output_sample_rate();
|
|
}
|
|
output_layout = LayoutFromChannels(msg.num_output_channels());
|
|
ASSERT_EQ(kNoErr, apm->Initialize(
|
|
msg.sample_rate(),
|
|
output_sample_rate,
|
|
reverse_sample_rate,
|
|
LayoutFromChannels(msg.num_input_channels()),
|
|
output_layout,
|
|
LayoutFromChannels(msg.num_reverse_channels())));
|
|
|
|
samples_per_channel = msg.sample_rate() / 100;
|
|
far_frame.sample_rate_hz_ = msg.sample_rate();
|
|
far_frame.samples_per_channel_ = samples_per_channel;
|
|
far_frame.num_channels_ = msg.num_reverse_channels();
|
|
near_frame.sample_rate_hz_ = msg.sample_rate();
|
|
near_frame.samples_per_channel_ = samples_per_channel;
|
|
near_frame.num_channels_ = msg.num_input_channels();
|
|
reverse_cb.reset(new ChannelBuffer<float>(samples_per_channel,
|
|
msg.num_reverse_channels()));
|
|
primary_cb.reset(new ChannelBuffer<float>(samples_per_channel,
|
|
msg.num_input_channels()));
|
|
|
|
if (verbose) {
|
|
printf("Init at frame: %d (primary), %d (reverse)\n",
|
|
primary_count, reverse_count);
|
|
printf(" Primary rates: %d Hz (in), %d Hz (out)\n",
|
|
msg.sample_rate(), output_sample_rate);
|
|
printf(" Primary channels: %d (in), %d (out)\n",
|
|
msg.num_input_channels(),
|
|
msg.num_output_channels());
|
|
printf(" Reverse rate: %d\n", reverse_sample_rate);
|
|
printf(" Reverse channels: %d\n", msg.num_reverse_channels());
|
|
}
|
|
|
|
} else if (event_msg.type() == Event::REVERSE_STREAM) {
|
|
ASSERT_TRUE(event_msg.has_reverse_stream());
|
|
ReverseStream msg = event_msg.reverse_stream();
|
|
reverse_count++;
|
|
|
|
ASSERT_TRUE(msg.has_data() ^ (msg.channel_size() > 0));
|
|
if (msg.has_data()) {
|
|
ASSERT_EQ(sizeof(int16_t) * samples_per_channel *
|
|
far_frame.num_channels_, msg.data().size());
|
|
memcpy(far_frame.data_, msg.data().data(), msg.data().size());
|
|
} else {
|
|
for (int i = 0; i < msg.channel_size(); ++i) {
|
|
reverse_cb->CopyFrom(msg.channel(i).data(), i);
|
|
}
|
|
}
|
|
|
|
if (perf_testing) {
|
|
t0 = TickTime::Now();
|
|
}
|
|
|
|
if (msg.has_data()) {
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->AnalyzeReverseStream(&far_frame));
|
|
} else {
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->AnalyzeReverseStream(
|
|
reverse_cb->channels(),
|
|
far_frame.samples_per_channel_,
|
|
far_frame.sample_rate_hz_,
|
|
LayoutFromChannels(far_frame.num_channels_)));
|
|
}
|
|
|
|
if (perf_testing) {
|
|
t1 = TickTime::Now();
|
|
TickInterval tick_diff = t1 - t0;
|
|
acc_ticks += tick_diff;
|
|
if (tick_diff.Microseconds() > max_time_reverse_us) {
|
|
max_time_reverse_us = tick_diff.Microseconds();
|
|
}
|
|
if (tick_diff.Microseconds() < min_time_reverse_us) {
|
|
min_time_reverse_us = tick_diff.Microseconds();
|
|
}
|
|
}
|
|
|
|
} else if (event_msg.type() == Event::STREAM) {
|
|
ASSERT_TRUE(event_msg.has_stream());
|
|
const Stream msg = event_msg.stream();
|
|
primary_count++;
|
|
|
|
// ProcessStream could have changed this for the output frame.
|
|
near_frame.num_channels_ = apm->num_input_channels();
|
|
|
|
ASSERT_TRUE(msg.has_input_data() ^ (msg.input_channel_size() > 0));
|
|
if (msg.has_input_data()) {
|
|
ASSERT_EQ(sizeof(int16_t) * samples_per_channel *
|
|
near_frame.num_channels_, msg.input_data().size());
|
|
memcpy(near_frame.data_,
|
|
msg.input_data().data(),
|
|
msg.input_data().size());
|
|
} else {
|
|
for (int i = 0; i < msg.input_channel_size(); ++i) {
|
|
primary_cb->CopyFrom(msg.input_channel(i).data(), i);
|
|
}
|
|
}
|
|
|
|
near_read_bytes += msg.input_data().size();
|
|
if (progress && primary_count % 100 == 0) {
|
|
printf("%.0f%% complete\r",
|
|
(near_read_bytes * 100.0) / near_size_bytes);
|
|
fflush(stdout);
|
|
}
|
|
|
|
if (perf_testing) {
|
|
t0 = TickTime::Now();
|
|
}
|
|
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->gain_control()->set_stream_analog_level(msg.level()));
|
|
delay_ms = msg.delay() + extra_delay_ms;
|
|
if (override_delay_ms) {
|
|
delay_ms = override_delay_ms;
|
|
}
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->set_stream_delay_ms(delay_ms));
|
|
apm->echo_cancellation()->set_stream_drift_samples(msg.drift());
|
|
|
|
if (msg.has_keypress()) {
|
|
apm->set_stream_key_pressed(msg.keypress());
|
|
} else {
|
|
apm->set_stream_key_pressed(true);
|
|
}
|
|
|
|
int err = apm->kNoError;
|
|
if (msg.has_input_data()) {
|
|
err = apm->ProcessStream(&near_frame);
|
|
ASSERT_TRUE(near_frame.num_channels_ == apm->num_output_channels());
|
|
} else {
|
|
err = apm->ProcessStream(
|
|
primary_cb->channels(),
|
|
near_frame.samples_per_channel_,
|
|
near_frame.sample_rate_hz_,
|
|
LayoutFromChannels(near_frame.num_channels_),
|
|
output_sample_rate,
|
|
output_layout,
|
|
primary_cb->channels());
|
|
}
|
|
|
|
if (err == apm->kBadStreamParameterWarning) {
|
|
printf("Bad parameter warning. %s\n", trace_stream.str().c_str());
|
|
}
|
|
ASSERT_TRUE(err == apm->kNoError ||
|
|
err == apm->kBadStreamParameterWarning);
|
|
|
|
stream_has_voice =
|
|
static_cast<int8_t>(apm->voice_detection()->stream_has_voice());
|
|
if (vad_out_file != NULL) {
|
|
ASSERT_EQ(1u, fwrite(&stream_has_voice,
|
|
sizeof(stream_has_voice),
|
|
1,
|
|
vad_out_file));
|
|
}
|
|
|
|
if (ns_prob_file != NULL) {
|
|
ns_speech_prob = apm->noise_suppression()->speech_probability();
|
|
ASSERT_EQ(1u, fwrite(&ns_speech_prob,
|
|
sizeof(ns_speech_prob),
|
|
1,
|
|
ns_prob_file));
|
|
}
|
|
|
|
if (perf_testing) {
|
|
t1 = TickTime::Now();
|
|
TickInterval tick_diff = t1 - t0;
|
|
acc_ticks += tick_diff;
|
|
if (tick_diff.Microseconds() > max_time_us) {
|
|
max_time_us = tick_diff.Microseconds();
|
|
}
|
|
if (tick_diff.Microseconds() < min_time_us) {
|
|
min_time_us = tick_diff.Microseconds();
|
|
}
|
|
}
|
|
|
|
size_t num_samples = samples_per_channel * apm->num_output_channels();
|
|
if (msg.has_input_data()) {
|
|
static FILE* out_file = OpenFile(out_filename, "wb");
|
|
ASSERT_EQ(num_samples, fwrite(near_frame.data_,
|
|
sizeof(*near_frame.data_),
|
|
num_samples,
|
|
out_file));
|
|
} else {
|
|
static FILE* out_float_file = OpenFile(out_float_filename, "wb");
|
|
ASSERT_EQ(num_samples, fwrite(primary_cb->data(),
|
|
sizeof(*primary_cb->data()),
|
|
num_samples,
|
|
out_float_file));
|
|
}
|
|
}
|
|
}
|
|
|
|
ASSERT_TRUE(feof(pb_file));
|
|
|
|
} else {
|
|
enum Events {
|
|
kInitializeEvent,
|
|
kRenderEvent,
|
|
kCaptureEvent,
|
|
kResetEventDeprecated
|
|
};
|
|
int16_t event = 0;
|
|
while (simulating || feof(event_file) == 0) {
|
|
std::ostringstream trace_stream;
|
|
trace_stream << "Processed frames: " << reverse_count << " (reverse), "
|
|
<< primary_count << " (primary)";
|
|
SCOPED_TRACE(trace_stream.str());
|
|
|
|
if (simulating) {
|
|
if (far_file == NULL) {
|
|
event = kCaptureEvent;
|
|
} else {
|
|
if (event == kRenderEvent) {
|
|
event = kCaptureEvent;
|
|
} else {
|
|
event = kRenderEvent;
|
|
}
|
|
}
|
|
} else {
|
|
read_count = fread(&event, sizeof(event), 1, event_file);
|
|
if (read_count != 1) {
|
|
break;
|
|
}
|
|
}
|
|
|
|
far_frame.sample_rate_hz_ = sample_rate_hz;
|
|
far_frame.samples_per_channel_ = samples_per_channel;
|
|
far_frame.num_channels_ = num_render_channels;
|
|
near_frame.sample_rate_hz_ = sample_rate_hz;
|
|
near_frame.samples_per_channel_ = samples_per_channel;
|
|
|
|
if (event == kInitializeEvent || event == kResetEventDeprecated) {
|
|
ASSERT_EQ(1u,
|
|
fread(&sample_rate_hz, sizeof(sample_rate_hz), 1, event_file));
|
|
samples_per_channel = sample_rate_hz / 100;
|
|
|
|
int32_t unused_device_sample_rate_hz;
|
|
ASSERT_EQ(1u,
|
|
fread(&unused_device_sample_rate_hz,
|
|
sizeof(unused_device_sample_rate_hz),
|
|
1,
|
|
event_file));
|
|
|
|
ASSERT_EQ(kNoErr, apm->Initialize(
|
|
sample_rate_hz,
|
|
sample_rate_hz,
|
|
sample_rate_hz,
|
|
LayoutFromChannels(num_capture_input_channels),
|
|
LayoutFromChannels(num_capture_output_channels),
|
|
LayoutFromChannels(num_render_channels)));
|
|
|
|
far_frame.sample_rate_hz_ = sample_rate_hz;
|
|
far_frame.samples_per_channel_ = samples_per_channel;
|
|
far_frame.num_channels_ = num_render_channels;
|
|
near_frame.sample_rate_hz_ = sample_rate_hz;
|
|
near_frame.samples_per_channel_ = samples_per_channel;
|
|
|
|
if (verbose) {
|
|
printf("Init at frame: %d (primary), %d (reverse)\n",
|
|
primary_count, reverse_count);
|
|
printf(" Sample rate: %d Hz\n", sample_rate_hz);
|
|
}
|
|
|
|
} else if (event == kRenderEvent) {
|
|
reverse_count++;
|
|
|
|
size_t size = samples_per_channel * num_render_channels;
|
|
read_count = fread(far_frame.data_,
|
|
sizeof(int16_t),
|
|
size,
|
|
far_file);
|
|
|
|
if (simulating) {
|
|
if (read_count != size) {
|
|
// Read an equal amount from the near file to avoid errors due to
|
|
// not reaching end-of-file.
|
|
EXPECT_EQ(0, fseek(near_file, read_count * sizeof(int16_t),
|
|
SEEK_CUR));
|
|
break; // This is expected.
|
|
}
|
|
} else {
|
|
ASSERT_EQ(size, read_count);
|
|
}
|
|
|
|
if (perf_testing) {
|
|
t0 = TickTime::Now();
|
|
}
|
|
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->AnalyzeReverseStream(&far_frame));
|
|
|
|
if (perf_testing) {
|
|
t1 = TickTime::Now();
|
|
TickInterval tick_diff = t1 - t0;
|
|
acc_ticks += tick_diff;
|
|
if (tick_diff.Microseconds() > max_time_reverse_us) {
|
|
max_time_reverse_us = tick_diff.Microseconds();
|
|
}
|
|
if (tick_diff.Microseconds() < min_time_reverse_us) {
|
|
min_time_reverse_us = tick_diff.Microseconds();
|
|
}
|
|
}
|
|
|
|
} else if (event == kCaptureEvent) {
|
|
primary_count++;
|
|
near_frame.num_channels_ = num_capture_input_channels;
|
|
|
|
size_t size = samples_per_channel * num_capture_input_channels;
|
|
read_count = fread(near_frame.data_,
|
|
sizeof(int16_t),
|
|
size,
|
|
near_file);
|
|
|
|
near_read_bytes += read_count * sizeof(int16_t);
|
|
if (progress && primary_count % 100 == 0) {
|
|
printf("%.0f%% complete\r",
|
|
(near_read_bytes * 100.0) / near_size_bytes);
|
|
fflush(stdout);
|
|
}
|
|
if (simulating) {
|
|
if (read_count != size) {
|
|
break; // This is expected.
|
|
}
|
|
|
|
delay_ms = 0;
|
|
drift_samples = 0;
|
|
} else {
|
|
ASSERT_EQ(size, read_count);
|
|
|
|
// TODO(ajm): sizeof(delay_ms) for current files?
|
|
ASSERT_EQ(1u,
|
|
fread(&delay_ms, 2, 1, delay_file));
|
|
ASSERT_EQ(1u,
|
|
fread(&drift_samples, sizeof(drift_samples), 1, drift_file));
|
|
}
|
|
|
|
if (apm->gain_control()->is_enabled() &&
|
|
apm->gain_control()->mode() == GainControl::kAdaptiveAnalog) {
|
|
SimulateMic(capture_level, &near_frame);
|
|
}
|
|
|
|
if (perf_testing) {
|
|
t0 = TickTime::Now();
|
|
}
|
|
|
|
const int capture_level_in = capture_level;
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->gain_control()->set_stream_analog_level(capture_level));
|
|
delay_ms += extra_delay_ms;
|
|
if (override_delay_ms) {
|
|
delay_ms = override_delay_ms;
|
|
}
|
|
ASSERT_EQ(apm->kNoError,
|
|
apm->set_stream_delay_ms(delay_ms));
|
|
apm->echo_cancellation()->set_stream_drift_samples(drift_samples);
|
|
|
|
apm->set_stream_key_pressed(true);
|
|
|
|
int err = apm->ProcessStream(&near_frame);
|
|
if (err == apm->kBadStreamParameterWarning) {
|
|
printf("Bad parameter warning. %s\n", trace_stream.str().c_str());
|
|
}
|
|
ASSERT_TRUE(err == apm->kNoError ||
|
|
err == apm->kBadStreamParameterWarning);
|
|
ASSERT_TRUE(near_frame.num_channels_ == apm->num_output_channels());
|
|
|
|
capture_level = apm->gain_control()->stream_analog_level();
|
|
|
|
stream_has_voice =
|
|
static_cast<int8_t>(apm->voice_detection()->stream_has_voice());
|
|
if (vad_out_file != NULL) {
|
|
ASSERT_EQ(1u, fwrite(&stream_has_voice,
|
|
sizeof(stream_has_voice),
|
|
1,
|
|
vad_out_file));
|
|
}
|
|
|
|
if (ns_prob_file != NULL) {
|
|
ns_speech_prob = apm->noise_suppression()->speech_probability();
|
|
ASSERT_EQ(1u, fwrite(&ns_speech_prob,
|
|
sizeof(ns_speech_prob),
|
|
1,
|
|
ns_prob_file));
|
|
}
|
|
|
|
if (apm->gain_control()->mode() != GainControl::kAdaptiveAnalog) {
|
|
ASSERT_EQ(capture_level_in, capture_level);
|
|
}
|
|
|
|
if (perf_testing) {
|
|
t1 = TickTime::Now();
|
|
TickInterval tick_diff = t1 - t0;
|
|
acc_ticks += tick_diff;
|
|
if (tick_diff.Microseconds() > max_time_us) {
|
|
max_time_us = tick_diff.Microseconds();
|
|
}
|
|
if (tick_diff.Microseconds() < min_time_us) {
|
|
min_time_us = tick_diff.Microseconds();
|
|
}
|
|
}
|
|
|
|
size = samples_per_channel * near_frame.num_channels_;
|
|
static FILE* out_file = OpenFile(out_filename, "wb");
|
|
ASSERT_EQ(size, fwrite(near_frame.data_,
|
|
sizeof(int16_t),
|
|
size,
|
|
out_file));
|
|
}
|
|
else {
|
|
FAIL() << "Event " << event << " is unrecognized";
|
|
}
|
|
}
|
|
}
|
|
printf("100%% complete\r");
|
|
|
|
if (aecm_echo_path_out_file != NULL) {
|
|
const size_t path_size =
|
|
apm->echo_control_mobile()->echo_path_size_bytes();
|
|
scoped_ptr<char[]> echo_path(new char[path_size]);
|
|
apm->echo_control_mobile()->GetEchoPath(echo_path.get(), path_size);
|
|
ASSERT_EQ(path_size, fwrite(echo_path.get(),
|
|
sizeof(char),
|
|
path_size,
|
|
aecm_echo_path_out_file));
|
|
fclose(aecm_echo_path_out_file);
|
|
aecm_echo_path_out_file = NULL;
|
|
}
|
|
|
|
if (verbose) {
|
|
printf("\nProcessed frames: %d (primary), %d (reverse)\n",
|
|
primary_count, reverse_count);
|
|
|
|
if (apm->level_estimator()->is_enabled()) {
|
|
printf("\n--Level metrics--\n");
|
|
printf("RMS: %d dBFS\n", -apm->level_estimator()->RMS());
|
|
}
|
|
if (apm->echo_cancellation()->are_metrics_enabled()) {
|
|
EchoCancellation::Metrics metrics;
|
|
apm->echo_cancellation()->GetMetrics(&metrics);
|
|
printf("\n--Echo metrics--\n");
|
|
printf("(avg, max, min)\n");
|
|
printf("ERL: ");
|
|
PrintStat(metrics.echo_return_loss);
|
|
printf("ERLE: ");
|
|
PrintStat(metrics.echo_return_loss_enhancement);
|
|
printf("ANLP: ");
|
|
PrintStat(metrics.a_nlp);
|
|
}
|
|
if (apm->echo_cancellation()->is_delay_logging_enabled()) {
|
|
int median = 0;
|
|
int std = 0;
|
|
apm->echo_cancellation()->GetDelayMetrics(&median, &std);
|
|
printf("\n--Delay metrics--\n");
|
|
printf("Median: %3d\n", median);
|
|
printf("Standard deviation: %3d\n", std);
|
|
}
|
|
}
|
|
|
|
if (!pb_file) {
|
|
int8_t temp_int8;
|
|
if (far_file) {
|
|
read_count = fread(&temp_int8, sizeof(temp_int8), 1, far_file);
|
|
EXPECT_NE(0, feof(far_file)) << "Far-end file not fully processed";
|
|
}
|
|
|
|
read_count = fread(&temp_int8, sizeof(temp_int8), 1, near_file);
|
|
EXPECT_NE(0, feof(near_file)) << "Near-end file not fully processed";
|
|
|
|
if (!simulating) {
|
|
read_count = fread(&temp_int8, sizeof(temp_int8), 1, event_file);
|
|
EXPECT_NE(0, feof(event_file)) << "Event file not fully processed";
|
|
read_count = fread(&temp_int8, sizeof(temp_int8), 1, delay_file);
|
|
EXPECT_NE(0, feof(delay_file)) << "Delay file not fully processed";
|
|
read_count = fread(&temp_int8, sizeof(temp_int8), 1, drift_file);
|
|
EXPECT_NE(0, feof(drift_file)) << "Drift file not fully processed";
|
|
}
|
|
}
|
|
|
|
if (perf_testing) {
|
|
if (primary_count > 0) {
|
|
int64_t exec_time = acc_ticks.Milliseconds();
|
|
printf("\nTotal time: %.3f s, file time: %.2f s\n",
|
|
exec_time * 0.001, primary_count * 0.01);
|
|
printf("Time per frame: %.3f ms (average), %.3f ms (max),"
|
|
" %.3f ms (min)\n",
|
|
(exec_time * 1.0) / primary_count,
|
|
(max_time_us + max_time_reverse_us) / 1000.0,
|
|
(min_time_us + min_time_reverse_us) / 1000.0);
|
|
// Record the results with Perf test tools.
|
|
webrtc::test::PrintResult("audioproc", "", "time_per_10ms_frame",
|
|
(exec_time * 1000) / primary_count, "us", false);
|
|
} else {
|
|
printf("Warning: no capture frames\n");
|
|
}
|
|
}
|
|
}
|
|
|
|
} // namespace
|
|
} // namespace webrtc
|
|
|
|
int main(int argc, char* argv[])
|
|
{
|
|
webrtc::void_main(argc, argv);
|
|
|
|
// Optional, but removes memory leak noise from Valgrind.
|
|
google::protobuf::ShutdownProtobufLibrary();
|
|
return 0;
|
|
}
|