Replace scoped_array<T> with scoped_ptr<T[]>.
scoped_array is deprecated. This was done using a Chromium clang tool: http://src.chromium.org/viewvc/chrome/trunk/src/tools/clang/rewrite_scoped_ar... except for the few not-built-on-Linux files which were updated manually. TESTED=trybots BUG=2515 R=niklas.enbom@webrtc.org Review URL: https://webrtc-codereview.appspot.com/12429004 git-svn-id: http://webrtc.googlecode.com/svn/trunk@5985 4adac7df-926f-26a2-2b94-8c16560cd09d
This commit is contained in:
parent
658a94595d
commit
8f69330310
@ -62,7 +62,7 @@ TEST(SincResamplerTest, ChunkedResample) {
|
|||||||
|
|
||||||
static const int kChunks = 2;
|
static const int kChunks = 2;
|
||||||
int max_chunk_size = resampler.ChunkSize() * kChunks;
|
int max_chunk_size = resampler.ChunkSize() * kChunks;
|
||||||
scoped_array<float> resampled_destination(new float[max_chunk_size]);
|
scoped_ptr<float[]> resampled_destination(new float[max_chunk_size]);
|
||||||
|
|
||||||
// Verify requesting ChunkSize() frames causes a single callback.
|
// Verify requesting ChunkSize() frames causes a single callback.
|
||||||
EXPECT_CALL(mock_source, Run(_, _))
|
EXPECT_CALL(mock_source, Run(_, _))
|
||||||
@ -81,7 +81,7 @@ TEST(SincResamplerTest, Flush) {
|
|||||||
MockSource mock_source;
|
MockSource mock_source;
|
||||||
SincResampler resampler(kSampleRateRatio, SincResampler::kDefaultRequestSize,
|
SincResampler resampler(kSampleRateRatio, SincResampler::kDefaultRequestSize,
|
||||||
&mock_source);
|
&mock_source);
|
||||||
scoped_array<float> resampled_destination(new float[resampler.ChunkSize()]);
|
scoped_ptr<float[]> resampled_destination(new float[resampler.ChunkSize()]);
|
||||||
|
|
||||||
// Fill the resampler with junk data.
|
// Fill the resampler with junk data.
|
||||||
EXPECT_CALL(mock_source, Run(_, _))
|
EXPECT_CALL(mock_source, Run(_, _))
|
||||||
@ -266,7 +266,7 @@ TEST_P(SincResamplerTest, Resample) {
|
|||||||
|
|
||||||
// Force an update to the sample rate ratio to ensure dyanmic sample rate
|
// Force an update to the sample rate ratio to ensure dyanmic sample rate
|
||||||
// changes are working correctly.
|
// changes are working correctly.
|
||||||
scoped_array<float> kernel(new float[SincResampler::kKernelStorageSize]);
|
scoped_ptr<float[]> kernel(new float[SincResampler::kKernelStorageSize]);
|
||||||
memcpy(kernel.get(), resampler.get_kernel_for_testing(),
|
memcpy(kernel.get(), resampler.get_kernel_for_testing(),
|
||||||
SincResampler::kKernelStorageSize);
|
SincResampler::kKernelStorageSize);
|
||||||
resampler.SetRatio(M_PI);
|
resampler.SetRatio(M_PI);
|
||||||
@ -278,8 +278,8 @@ TEST_P(SincResamplerTest, Resample) {
|
|||||||
|
|
||||||
// TODO(dalecurtis): If we switch to AVX/SSE optimization, we'll need to
|
// TODO(dalecurtis): If we switch to AVX/SSE optimization, we'll need to
|
||||||
// allocate these on 32-byte boundaries and ensure they're sized % 32 bytes.
|
// allocate these on 32-byte boundaries and ensure they're sized % 32 bytes.
|
||||||
scoped_array<float> resampled_destination(new float[output_samples]);
|
scoped_ptr<float[]> resampled_destination(new float[output_samples]);
|
||||||
scoped_array<float> pure_destination(new float[output_samples]);
|
scoped_ptr<float[]> pure_destination(new float[output_samples]);
|
||||||
|
|
||||||
// Generate resampled signal.
|
// Generate resampled signal.
|
||||||
resampler.Resample(output_samples, resampled_destination.get());
|
resampler.Resample(output_samples, resampled_destination.get());
|
||||||
|
@ -84,7 +84,7 @@ class TestLibYuv : public ::testing::Test {
|
|||||||
|
|
||||||
FILE* source_file_;
|
FILE* source_file_;
|
||||||
I420VideoFrame orig_frame_;
|
I420VideoFrame orig_frame_;
|
||||||
scoped_array<uint8_t> orig_buffer_;
|
scoped_ptr<uint8_t[]> orig_buffer_;
|
||||||
const int width_;
|
const int width_;
|
||||||
const int height_;
|
const int height_;
|
||||||
const int size_y_;
|
const int size_y_;
|
||||||
@ -147,7 +147,7 @@ TEST_F(TestLibYuv, ConvertTest) {
|
|||||||
(width_ + 1) / 2,
|
(width_ + 1) / 2,
|
||||||
(width_ + 1) / 2));
|
(width_ + 1) / 2));
|
||||||
printf("\nConvert #%d I420 <-> I420 \n", j);
|
printf("\nConvert #%d I420 <-> I420 \n", j);
|
||||||
scoped_array<uint8_t> out_i420_buffer(new uint8_t[frame_length_]);
|
scoped_ptr<uint8_t[]> out_i420_buffer(new uint8_t[frame_length_]);
|
||||||
EXPECT_EQ(0, ConvertFromI420(orig_frame_, kI420, 0,
|
EXPECT_EQ(0, ConvertFromI420(orig_frame_, kI420, 0,
|
||||||
out_i420_buffer.get()));
|
out_i420_buffer.get()));
|
||||||
EXPECT_EQ(0, ConvertToI420(kI420, out_i420_buffer.get(), 0, 0,
|
EXPECT_EQ(0, ConvertToI420(kI420, out_i420_buffer.get(), 0, 0,
|
||||||
@ -162,7 +162,7 @@ TEST_F(TestLibYuv, ConvertTest) {
|
|||||||
j++;
|
j++;
|
||||||
|
|
||||||
printf("\nConvert #%d I420 <-> RGB24\n", j);
|
printf("\nConvert #%d I420 <-> RGB24\n", j);
|
||||||
scoped_array<uint8_t> res_rgb_buffer2(new uint8_t[width_ * height_ * 3]);
|
scoped_ptr<uint8_t[]> res_rgb_buffer2(new uint8_t[width_ * height_ * 3]);
|
||||||
// Align the stride values for the output frame.
|
// Align the stride values for the output frame.
|
||||||
int stride_y = 0;
|
int stride_y = 0;
|
||||||
int stride_uv = 0;
|
int stride_uv = 0;
|
||||||
@ -184,7 +184,7 @@ TEST_F(TestLibYuv, ConvertTest) {
|
|||||||
j++;
|
j++;
|
||||||
|
|
||||||
printf("\nConvert #%d I420 <-> UYVY\n", j);
|
printf("\nConvert #%d I420 <-> UYVY\n", j);
|
||||||
scoped_array<uint8_t> out_uyvy_buffer(new uint8_t[width_ * height_ * 2]);
|
scoped_ptr<uint8_t[]> out_uyvy_buffer(new uint8_t[width_ * height_ * 2]);
|
||||||
EXPECT_EQ(0, ConvertFromI420(orig_frame_, kUYVY, 0, out_uyvy_buffer.get()));
|
EXPECT_EQ(0, ConvertFromI420(orig_frame_, kUYVY, 0, out_uyvy_buffer.get()));
|
||||||
EXPECT_EQ(0, ConvertToI420(kUYVY, out_uyvy_buffer.get(), 0, 0, width_,
|
EXPECT_EQ(0, ConvertToI420(kUYVY, out_uyvy_buffer.get(), 0, 0, width_,
|
||||||
height_, 0, kRotateNone, &res_i420_frame));
|
height_, 0, kRotateNone, &res_i420_frame));
|
||||||
@ -196,8 +196,8 @@ TEST_F(TestLibYuv, ConvertTest) {
|
|||||||
j++;
|
j++;
|
||||||
|
|
||||||
printf("\nConvert #%d I420 <-> YV12\n", j);
|
printf("\nConvert #%d I420 <-> YV12\n", j);
|
||||||
scoped_array<uint8_t> outYV120Buffer(new uint8_t[frame_length_]);
|
scoped_ptr<uint8_t[]> outYV120Buffer(new uint8_t[frame_length_]);
|
||||||
scoped_array<uint8_t> res_i420_buffer(new uint8_t[frame_length_]);
|
scoped_ptr<uint8_t[]> res_i420_buffer(new uint8_t[frame_length_]);
|
||||||
I420VideoFrame yv12_frame;
|
I420VideoFrame yv12_frame;
|
||||||
EXPECT_EQ(0, ConvertFromI420(orig_frame_, kYV12, 0, outYV120Buffer.get()));
|
EXPECT_EQ(0, ConvertFromI420(orig_frame_, kYV12, 0, outYV120Buffer.get()));
|
||||||
yv12_frame.CreateFrame(size_y_, outYV120Buffer.get(),
|
yv12_frame.CreateFrame(size_y_, outYV120Buffer.get(),
|
||||||
@ -218,7 +218,7 @@ TEST_F(TestLibYuv, ConvertTest) {
|
|||||||
j++;
|
j++;
|
||||||
|
|
||||||
printf("\nConvert #%d I420 <-> YUY2\n", j);
|
printf("\nConvert #%d I420 <-> YUY2\n", j);
|
||||||
scoped_array<uint8_t> out_yuy2_buffer(new uint8_t[width_ * height_ * 2]);
|
scoped_ptr<uint8_t[]> out_yuy2_buffer(new uint8_t[width_ * height_ * 2]);
|
||||||
EXPECT_EQ(0, ConvertFromI420(orig_frame_, kYUY2, 0, out_yuy2_buffer.get()));
|
EXPECT_EQ(0, ConvertFromI420(orig_frame_, kYUY2, 0, out_yuy2_buffer.get()));
|
||||||
|
|
||||||
EXPECT_EQ(0, ConvertToI420(kYUY2, out_yuy2_buffer.get(), 0, 0, width_,
|
EXPECT_EQ(0, ConvertToI420(kYUY2, out_yuy2_buffer.get(), 0, 0, width_,
|
||||||
@ -231,7 +231,7 @@ TEST_F(TestLibYuv, ConvertTest) {
|
|||||||
psnr = I420PSNR(&orig_frame_, &res_i420_frame);
|
psnr = I420PSNR(&orig_frame_, &res_i420_frame);
|
||||||
EXPECT_EQ(48.0, psnr);
|
EXPECT_EQ(48.0, psnr);
|
||||||
printf("\nConvert #%d I420 <-> RGB565\n", j);
|
printf("\nConvert #%d I420 <-> RGB565\n", j);
|
||||||
scoped_array<uint8_t> out_rgb565_buffer(new uint8_t[width_ * height_ * 2]);
|
scoped_ptr<uint8_t[]> out_rgb565_buffer(new uint8_t[width_ * height_ * 2]);
|
||||||
EXPECT_EQ(0, ConvertFromI420(orig_frame_, kRGB565, 0,
|
EXPECT_EQ(0, ConvertFromI420(orig_frame_, kRGB565, 0,
|
||||||
out_rgb565_buffer.get()));
|
out_rgb565_buffer.get()));
|
||||||
|
|
||||||
@ -250,7 +250,7 @@ TEST_F(TestLibYuv, ConvertTest) {
|
|||||||
EXPECT_GT(ceil(psnr), 40);
|
EXPECT_GT(ceil(psnr), 40);
|
||||||
|
|
||||||
printf("\nConvert #%d I420 <-> ARGB8888\n", j);
|
printf("\nConvert #%d I420 <-> ARGB8888\n", j);
|
||||||
scoped_array<uint8_t> out_argb8888_buffer(new uint8_t[width_ * height_ * 4]);
|
scoped_ptr<uint8_t[]> out_argb8888_buffer(new uint8_t[width_ * height_ * 4]);
|
||||||
EXPECT_EQ(0, ConvertFromI420(orig_frame_, kARGB, 0,
|
EXPECT_EQ(0, ConvertFromI420(orig_frame_, kARGB, 0,
|
||||||
out_argb8888_buffer.get()));
|
out_argb8888_buffer.get()));
|
||||||
|
|
||||||
@ -283,7 +283,7 @@ TEST_F(TestLibYuv, ConvertAlignedFrame) {
|
|||||||
Calc16ByteAlignedStride(width_, &stride_y, &stride_uv);
|
Calc16ByteAlignedStride(width_, &stride_y, &stride_uv);
|
||||||
EXPECT_EQ(0,res_i420_frame.CreateEmptyFrame(width_, height_,
|
EXPECT_EQ(0,res_i420_frame.CreateEmptyFrame(width_, height_,
|
||||||
stride_y, stride_uv, stride_uv));
|
stride_y, stride_uv, stride_uv));
|
||||||
scoped_array<uint8_t> out_i420_buffer(new uint8_t[frame_length_]);
|
scoped_ptr<uint8_t[]> out_i420_buffer(new uint8_t[frame_length_]);
|
||||||
EXPECT_EQ(0, ConvertFromI420(orig_frame_, kI420, 0,
|
EXPECT_EQ(0, ConvertFromI420(orig_frame_, kI420, 0,
|
||||||
out_i420_buffer.get()));
|
out_i420_buffer.get()));
|
||||||
EXPECT_EQ(0, ConvertToI420(kI420, out_i420_buffer.get(), 0, 0,
|
EXPECT_EQ(0, ConvertToI420(kI420, out_i420_buffer.get(), 0, 0,
|
||||||
|
@ -99,7 +99,7 @@ TEST_F(TestScaler, ScaleSendingBufferTooSmall) {
|
|||||||
kI420, kI420,
|
kI420, kI420,
|
||||||
kScalePoint));
|
kScalePoint));
|
||||||
I420VideoFrame test_frame2;
|
I420VideoFrame test_frame2;
|
||||||
scoped_array<uint8_t> orig_buffer(new uint8_t[frame_length_]);
|
scoped_ptr<uint8_t[]> orig_buffer(new uint8_t[frame_length_]);
|
||||||
EXPECT_GT(fread(orig_buffer.get(), 1, frame_length_, source_file_), 0U);
|
EXPECT_GT(fread(orig_buffer.get(), 1, frame_length_, source_file_), 0U);
|
||||||
test_frame_.CreateFrame(size_y_, orig_buffer.get(),
|
test_frame_.CreateFrame(size_y_, orig_buffer.get(),
|
||||||
size_uv_, orig_buffer.get() + size_y_,
|
size_uv_, orig_buffer.get() + size_y_,
|
||||||
@ -442,7 +442,7 @@ void TestScaler::ScaleSequence(ScaleMethod method,
|
|||||||
total_clock = 0;
|
total_clock = 0;
|
||||||
int frame_count = 0;
|
int frame_count = 0;
|
||||||
int src_required_size = CalcBufferSize(kI420, src_width, src_height);
|
int src_required_size = CalcBufferSize(kI420, src_width, src_height);
|
||||||
scoped_array<uint8_t> frame_buffer(new uint8_t[src_required_size]);
|
scoped_ptr<uint8_t[]> frame_buffer(new uint8_t[src_required_size]);
|
||||||
int size_y = src_width * src_height;
|
int size_y = src_width * src_height;
|
||||||
int size_uv = ((src_width + 1) / 2) * ((src_height + 1) / 2);
|
int size_uv = ((src_width + 1) / 2) * ((src_height + 1) / 2);
|
||||||
|
|
||||||
|
@ -22,7 +22,7 @@ FakeAudioDeviceBuffer::FakeAudioDeviceBuffer()
|
|||||||
next_available_buffer_(0),
|
next_available_buffer_(0),
|
||||||
record_channels_(0),
|
record_channels_(0),
|
||||||
play_channels_(0) {
|
play_channels_(0) {
|
||||||
buf_.reset(new scoped_array<int8_t>[kNumBuffers]);
|
buf_.reset(new scoped_ptr<int8_t[]>[kNumBuffers]);
|
||||||
for (int i = 0; i < kNumBuffers; ++i) {
|
for (int i = 0; i < kNumBuffers; ++i) {
|
||||||
buf_[i].reset(new int8_t[buffer_size_bytes()]);
|
buf_[i].reset(new int8_t[buffer_size_bytes()]);
|
||||||
}
|
}
|
||||||
|
@ -55,7 +55,7 @@ class FakeAudioDeviceBuffer : public AudioDeviceBuffer {
|
|||||||
AudioManagerJni audio_manager_;
|
AudioManagerJni audio_manager_;
|
||||||
|
|
||||||
SingleRwFifo fifo_;
|
SingleRwFifo fifo_;
|
||||||
scoped_array<scoped_array<int8_t> > buf_;
|
scoped_ptr<scoped_ptr<int8_t[]>[]> buf_;
|
||||||
int next_available_buffer_;
|
int next_available_buffer_;
|
||||||
|
|
||||||
uint8_t record_channels_;
|
uint8_t record_channels_;
|
||||||
|
@ -398,7 +398,7 @@ TEST(NackTest, ChangeOfListSizeAppliedAndOldElementsRemoved) {
|
|||||||
// Packet lost more than NACK-list size limit.
|
// Packet lost more than NACK-list size limit.
|
||||||
uint16_t num_lost_packets = kNackThreshold + kNackListSize + 5;
|
uint16_t num_lost_packets = kNackThreshold + kNackListSize + 5;
|
||||||
|
|
||||||
scoped_array<uint16_t> seq_num_lost(new uint16_t[num_lost_packets]);
|
scoped_ptr<uint16_t[]> seq_num_lost(new uint16_t[num_lost_packets]);
|
||||||
for (int n = 0; n < num_lost_packets; ++n) {
|
for (int n = 0; n < num_lost_packets; ++n) {
|
||||||
seq_num_lost[n] = ++seq_num;
|
seq_num_lost[n] = ++seq_num;
|
||||||
}
|
}
|
||||||
|
@ -126,7 +126,7 @@ class BackgroundNoise {
|
|||||||
int32_t residual_energy);
|
int32_t residual_energy);
|
||||||
|
|
||||||
size_t num_channels_;
|
size_t num_channels_;
|
||||||
scoped_array<ChannelParameters> channel_parameters_;
|
scoped_ptr<ChannelParameters[]> channel_parameters_;
|
||||||
bool initialized_;
|
bool initialized_;
|
||||||
NetEqBackgroundNoiseMode mode_;
|
NetEqBackgroundNoiseMode mode_;
|
||||||
|
|
||||||
|
@ -167,7 +167,7 @@ class Expand {
|
|||||||
int lag_index_direction_;
|
int lag_index_direction_;
|
||||||
int current_lag_index_;
|
int current_lag_index_;
|
||||||
bool stop_muting_;
|
bool stop_muting_;
|
||||||
scoped_array<ChannelParameters> channel_parameters_;
|
scoped_ptr<ChannelParameters[]> channel_parameters_;
|
||||||
|
|
||||||
DISALLOW_COPY_AND_ASSIGN(Expand);
|
DISALLOW_COPY_AND_ASSIGN(Expand);
|
||||||
};
|
};
|
||||||
|
@ -77,7 +77,6 @@ NetEqImpl::NetEqImpl(int fs,
|
|||||||
accelerate_factory_(accelerate_factory),
|
accelerate_factory_(accelerate_factory),
|
||||||
preemptive_expand_factory_(preemptive_expand_factory),
|
preemptive_expand_factory_(preemptive_expand_factory),
|
||||||
last_mode_(kModeNormal),
|
last_mode_(kModeNormal),
|
||||||
mute_factor_array_(NULL),
|
|
||||||
decoded_buffer_length_(kMaxFrameSize),
|
decoded_buffer_length_(kMaxFrameSize),
|
||||||
decoded_buffer_(new int16_t[decoded_buffer_length_]),
|
decoded_buffer_(new int16_t[decoded_buffer_length_]),
|
||||||
playout_timestamp_(0),
|
playout_timestamp_(0),
|
||||||
|
@ -371,9 +371,9 @@ class NetEqImpl : public webrtc::NetEq {
|
|||||||
int output_size_samples_ GUARDED_BY(crit_sect_);
|
int output_size_samples_ GUARDED_BY(crit_sect_);
|
||||||
int decoder_frame_length_ GUARDED_BY(crit_sect_);
|
int decoder_frame_length_ GUARDED_BY(crit_sect_);
|
||||||
Modes last_mode_ GUARDED_BY(crit_sect_);
|
Modes last_mode_ GUARDED_BY(crit_sect_);
|
||||||
scoped_array<int16_t> mute_factor_array_ GUARDED_BY(crit_sect_);
|
scoped_ptr<int16_t[]> mute_factor_array_ GUARDED_BY(crit_sect_);
|
||||||
size_t decoded_buffer_length_ GUARDED_BY(crit_sect_);
|
size_t decoded_buffer_length_ GUARDED_BY(crit_sect_);
|
||||||
scoped_array<int16_t> decoded_buffer_ GUARDED_BY(crit_sect_);
|
scoped_ptr<int16_t[]> decoded_buffer_ GUARDED_BY(crit_sect_);
|
||||||
uint32_t playout_timestamp_ GUARDED_BY(crit_sect_);
|
uint32_t playout_timestamp_ GUARDED_BY(crit_sect_);
|
||||||
bool new_codec_ GUARDED_BY(crit_sect_);
|
bool new_codec_ GUARDED_BY(crit_sect_);
|
||||||
uint32_t timestamp_ GUARDED_BY(crit_sect_);
|
uint32_t timestamp_ GUARDED_BY(crit_sect_);
|
||||||
|
@ -29,7 +29,7 @@ TimeStretch::ReturnCodes TimeStretch::Process(
|
|||||||
int fs_mult_120 = fs_mult_ * 120; // Corresponds to 15 ms.
|
int fs_mult_120 = fs_mult_ * 120; // Corresponds to 15 ms.
|
||||||
|
|
||||||
const int16_t* signal;
|
const int16_t* signal;
|
||||||
scoped_array<int16_t> signal_array;
|
scoped_ptr<int16_t[]> signal_array;
|
||||||
size_t signal_len;
|
size_t signal_len;
|
||||||
if (num_channels_ == 1) {
|
if (num_channels_ == 1) {
|
||||||
signal = input;
|
signal = input;
|
||||||
|
@ -27,8 +27,7 @@ class AudioLoop {
|
|||||||
AudioLoop()
|
AudioLoop()
|
||||||
: next_index_(0),
|
: next_index_(0),
|
||||||
loop_length_samples_(0),
|
loop_length_samples_(0),
|
||||||
block_length_samples_(0),
|
block_length_samples_(0) {
|
||||||
audio_array_(NULL) {
|
|
||||||
}
|
}
|
||||||
|
|
||||||
virtual ~AudioLoop() {}
|
virtual ~AudioLoop() {}
|
||||||
@ -50,7 +49,7 @@ class AudioLoop {
|
|||||||
size_t next_index_;
|
size_t next_index_;
|
||||||
size_t loop_length_samples_;
|
size_t loop_length_samples_;
|
||||||
size_t block_length_samples_;
|
size_t block_length_samples_;
|
||||||
scoped_array<int16_t> audio_array_;
|
scoped_ptr<int16_t[]> audio_array_;
|
||||||
|
|
||||||
DISALLOW_COPY_AND_ASSIGN(AudioLoop);
|
DISALLOW_COPY_AND_ASSIGN(AudioLoop);
|
||||||
};
|
};
|
||||||
|
@ -56,7 +56,7 @@ class FineAudioBuffer {
|
|||||||
int bytes_per_10_ms_;
|
int bytes_per_10_ms_;
|
||||||
|
|
||||||
// Storage for samples that are not yet asked for.
|
// Storage for samples that are not yet asked for.
|
||||||
scoped_array<int8_t> cache_buffer_;
|
scoped_ptr<int8_t[]> cache_buffer_;
|
||||||
int cached_buffer_start_; // Location of first unread sample.
|
int cached_buffer_start_; // Location of first unread sample.
|
||||||
int cached_bytes_; // Number of bytes stored in cache.
|
int cached_bytes_; // Number of bytes stored in cache.
|
||||||
};
|
};
|
||||||
|
@ -80,7 +80,7 @@ void RunFineBufferTest(int sample_rate, int frame_size_in_samples) {
|
|||||||
FineAudioBuffer fine_buffer(&audio_device_buffer, kFrameSizeBytes,
|
FineAudioBuffer fine_buffer(&audio_device_buffer, kFrameSizeBytes,
|
||||||
sample_rate);
|
sample_rate);
|
||||||
|
|
||||||
scoped_array<int8_t> out_buffer;
|
scoped_ptr<int8_t[]> out_buffer;
|
||||||
out_buffer.reset(
|
out_buffer.reset(
|
||||||
new int8_t[fine_buffer.RequiredBufferSizeBytes()]);
|
new int8_t[fine_buffer.RequiredBufferSizeBytes()]);
|
||||||
for (int i = 0; i < kNumberOfFrames; ++i) {
|
for (int i = 0; i < kNumberOfFrames; ++i) {
|
||||||
|
@ -289,7 +289,7 @@ void OpenSlesInput::AllocateBuffers() {
|
|||||||
fifo_.reset(new SingleRwFifo(num_fifo_buffers_needed_));
|
fifo_.reset(new SingleRwFifo(num_fifo_buffers_needed_));
|
||||||
|
|
||||||
// Allocate the memory area to be used.
|
// Allocate the memory area to be used.
|
||||||
rec_buf_.reset(new scoped_array<int8_t>[TotalBuffersUsed()]);
|
rec_buf_.reset(new scoped_ptr<int8_t[]>[TotalBuffersUsed()]);
|
||||||
for (int i = 0; i < TotalBuffersUsed(); ++i) {
|
for (int i = 0; i < TotalBuffersUsed(); ++i) {
|
||||||
rec_buf_[i].reset(new int8_t[buffer_size_bytes()]);
|
rec_buf_[i].reset(new int8_t[buffer_size_bytes()]);
|
||||||
}
|
}
|
||||||
|
@ -205,7 +205,7 @@ class OpenSlesInput {
|
|||||||
// Audio buffers
|
// Audio buffers
|
||||||
AudioDeviceBuffer* audio_buffer_;
|
AudioDeviceBuffer* audio_buffer_;
|
||||||
// Holds all allocated memory such that it is deallocated properly.
|
// Holds all allocated memory such that it is deallocated properly.
|
||||||
scoped_array<scoped_array<int8_t> > rec_buf_;
|
scoped_ptr<scoped_ptr<int8_t[]>[]> rec_buf_;
|
||||||
// Index in |rec_buf_| pointing to the audio buffer that will be ready the
|
// Index in |rec_buf_| pointing to the audio buffer that will be ready the
|
||||||
// next time RecorderSimpleBufferQueueCallbackHandler is invoked.
|
// next time RecorderSimpleBufferQueueCallbackHandler is invoked.
|
||||||
// Ready means buffer contains audio data from the device.
|
// Ready means buffer contains audio data from the device.
|
||||||
|
@ -340,7 +340,7 @@ void OpenSlesOutput::AllocateBuffers() {
|
|||||||
fifo_.reset(new SingleRwFifo(num_fifo_buffers_needed_));
|
fifo_.reset(new SingleRwFifo(num_fifo_buffers_needed_));
|
||||||
|
|
||||||
// Allocate the memory area to be used.
|
// Allocate the memory area to be used.
|
||||||
play_buf_.reset(new scoped_array<int8_t>[TotalBuffersUsed()]);
|
play_buf_.reset(new scoped_ptr<int8_t[]>[TotalBuffersUsed()]);
|
||||||
int required_buffer_size = fine_buffer_->RequiredBufferSizeBytes();
|
int required_buffer_size = fine_buffer_->RequiredBufferSizeBytes();
|
||||||
for (int i = 0; i < TotalBuffersUsed(); ++i) {
|
for (int i = 0; i < TotalBuffersUsed(); ++i) {
|
||||||
play_buf_[i].reset(new int8_t[required_buffer_size]);
|
play_buf_[i].reset(new int8_t[required_buffer_size]);
|
||||||
|
@ -223,7 +223,7 @@ class OpenSlesOutput : public PlayoutDelayProvider {
|
|||||||
// Audio buffers
|
// Audio buffers
|
||||||
AudioDeviceBuffer* audio_buffer_;
|
AudioDeviceBuffer* audio_buffer_;
|
||||||
scoped_ptr<FineAudioBuffer> fine_buffer_;
|
scoped_ptr<FineAudioBuffer> fine_buffer_;
|
||||||
scoped_array<scoped_array<int8_t> > play_buf_;
|
scoped_ptr<scoped_ptr<int8_t[]>[]> play_buf_;
|
||||||
// Index in |rec_buf_| pointing to the audio buffer that will be ready the
|
// Index in |rec_buf_| pointing to the audio buffer that will be ready the
|
||||||
// next time PlayerSimpleBufferQueueCallbackHandler is invoked.
|
// next time PlayerSimpleBufferQueueCallbackHandler is invoked.
|
||||||
// Ready means buffer is ready to be played out to device.
|
// Ready means buffer is ready to be played out to device.
|
||||||
|
@ -35,7 +35,7 @@ class SingleRwFifo {
|
|||||||
int capacity() const { return capacity_; }
|
int capacity() const { return capacity_; }
|
||||||
|
|
||||||
private:
|
private:
|
||||||
scoped_array<int8_t*> queue_;
|
scoped_ptr<int8_t*[]> queue_;
|
||||||
int capacity_;
|
int capacity_;
|
||||||
|
|
||||||
Atomic32 size_;
|
Atomic32 size_;
|
||||||
|
@ -90,7 +90,7 @@ class SingleRwFifoTest : public testing::Test {
|
|||||||
protected:
|
protected:
|
||||||
SingleRwFifo fifo_;
|
SingleRwFifo fifo_;
|
||||||
// Memory area for proper de-allocation.
|
// Memory area for proper de-allocation.
|
||||||
scoped_array<int8_t> buffer_[kCapacity];
|
scoped_ptr<int8_t[]> buffer_[kCapacity];
|
||||||
std::list<int8_t*> memory_queue_;
|
std::list<int8_t*> memory_queue_;
|
||||||
|
|
||||||
int pushed_;
|
int pushed_;
|
||||||
|
@ -518,7 +518,7 @@ void void_main(int argc, char* argv[]) {
|
|||||||
|
|
||||||
const size_t path_size =
|
const size_t path_size =
|
||||||
apm->echo_control_mobile()->echo_path_size_bytes();
|
apm->echo_control_mobile()->echo_path_size_bytes();
|
||||||
scoped_array<char> echo_path(new char[path_size]);
|
scoped_ptr<char[]> echo_path(new char[path_size]);
|
||||||
ASSERT_EQ(path_size, fread(echo_path.get(),
|
ASSERT_EQ(path_size, fread(echo_path.get(),
|
||||||
sizeof(char),
|
sizeof(char),
|
||||||
path_size,
|
path_size,
|
||||||
@ -1004,7 +1004,7 @@ void void_main(int argc, char* argv[]) {
|
|||||||
if (aecm_echo_path_out_file != NULL) {
|
if (aecm_echo_path_out_file != NULL) {
|
||||||
const size_t path_size =
|
const size_t path_size =
|
||||||
apm->echo_control_mobile()->echo_path_size_bytes();
|
apm->echo_control_mobile()->echo_path_size_bytes();
|
||||||
scoped_array<char> echo_path(new char[path_size]);
|
scoped_ptr<char[]> echo_path(new char[path_size]);
|
||||||
apm->echo_control_mobile()->GetEchoPath(echo_path.get(), path_size);
|
apm->echo_control_mobile()->GetEchoPath(echo_path.get(), path_size);
|
||||||
ASSERT_EQ(path_size, fwrite(echo_path.get(),
|
ASSERT_EQ(path_size, fwrite(echo_path.get(),
|
||||||
sizeof(char),
|
sizeof(char),
|
||||||
|
@ -61,8 +61,8 @@ static void RandomStressTest(int** data_ptr) {
|
|||||||
srand(seed);
|
srand(seed);
|
||||||
for (int i = 0; i < kNumTests; i++) {
|
for (int i = 0; i < kNumTests; i++) {
|
||||||
const int buffer_size = std::max(rand() % kMaxBufferSize, 1);
|
const int buffer_size = std::max(rand() % kMaxBufferSize, 1);
|
||||||
scoped_array<int> write_data(new int[buffer_size]);
|
scoped_ptr<int[]> write_data(new int[buffer_size]);
|
||||||
scoped_array<int> read_data(new int[buffer_size]);
|
scoped_ptr<int[]> read_data(new int[buffer_size]);
|
||||||
scoped_ring_buffer buffer(WebRtc_CreateBuffer(buffer_size, sizeof(int)));
|
scoped_ring_buffer buffer(WebRtc_CreateBuffer(buffer_size, sizeof(int)));
|
||||||
ASSERT_TRUE(buffer.get() != NULL);
|
ASSERT_TRUE(buffer.get() != NULL);
|
||||||
ASSERT_EQ(0, WebRtc_InitBuffer(buffer.get()));
|
ASSERT_EQ(0, WebRtc_InitBuffer(buffer.get()));
|
||||||
|
@ -76,7 +76,7 @@ class Differ {
|
|||||||
int bytes_per_row_;
|
int bytes_per_row_;
|
||||||
|
|
||||||
// Diff information for each block in the image.
|
// Diff information for each block in the image.
|
||||||
scoped_array<DiffInfo> diff_info_;
|
scoped_ptr<DiffInfo[]> diff_info_;
|
||||||
|
|
||||||
// Dimensions and total size of diff info array.
|
// Dimensions and total size of diff info array.
|
||||||
int diff_info_width_;
|
int diff_info_width_;
|
||||||
|
@ -200,8 +200,8 @@ class DifferTest : public testing::Test {
|
|||||||
int buffer_size_;
|
int buffer_size_;
|
||||||
|
|
||||||
// Previous and current screen buffers.
|
// Previous and current screen buffers.
|
||||||
scoped_array<uint8_t> prev_;
|
scoped_ptr<uint8_t[]> prev_;
|
||||||
scoped_array<uint8_t> curr_;
|
scoped_ptr<uint8_t[]> curr_;
|
||||||
|
|
||||||
private:
|
private:
|
||||||
DISALLOW_COPY_AND_ASSIGN(DifferTest);
|
DISALLOW_COPY_AND_ASSIGN(DifferTest);
|
||||||
|
@ -137,7 +137,7 @@ MouseCursor* CreateMouseCursorFromHCursor(HDC dc, HCURSOR cursor) {
|
|||||||
|
|
||||||
int width = bitmap_info.bmWidth;
|
int width = bitmap_info.bmWidth;
|
||||||
int height = bitmap_info.bmHeight;
|
int height = bitmap_info.bmHeight;
|
||||||
scoped_array<uint32_t> mask_data(new uint32_t[width * height]);
|
scoped_ptr<uint32_t[]> mask_data(new uint32_t[width * height]);
|
||||||
|
|
||||||
// Get pixel data from |scoped_mask| converting it to 32bpp along the way.
|
// Get pixel data from |scoped_mask| converting it to 32bpp along the way.
|
||||||
// GetDIBits() sets the alpha component of every pixel to 0.
|
// GetDIBits() sets the alpha component of every pixel to 0.
|
||||||
|
@ -62,7 +62,7 @@ bool ConvertToMouseShapeAndCompare(unsigned left, unsigned right) {
|
|||||||
|
|
||||||
// Get the pixels from |scoped_color|.
|
// Get the pixels from |scoped_color|.
|
||||||
int size = width * height;
|
int size = width * height;
|
||||||
scoped_array<uint32_t> data(new uint32_t[size]);
|
scoped_ptr<uint32_t[]> data(new uint32_t[size]);
|
||||||
EXPECT_TRUE(GetBitmapBits(scoped_color, size * sizeof(uint32_t), data.get()));
|
EXPECT_TRUE(GetBitmapBits(scoped_color, size * sizeof(uint32_t), data.get()));
|
||||||
|
|
||||||
// Compare the 32bpp image in |mouse_shape| with the one loaded from |right|.
|
// Compare the 32bpp image in |mouse_shape| with the one loaded from |right|.
|
||||||
|
@ -34,7 +34,7 @@ class RateStatistics {
|
|||||||
// Counters are kept in buckets (circular buffer), with one bucket
|
// Counters are kept in buckets (circular buffer), with one bucket
|
||||||
// per millisecond.
|
// per millisecond.
|
||||||
const int num_buckets_;
|
const int num_buckets_;
|
||||||
scoped_array<uint32_t> buckets_;
|
scoped_ptr<uint32_t[]> buckets_;
|
||||||
|
|
||||||
// Total count recorded in buckets.
|
// Total count recorded in buckets.
|
||||||
uint32_t accumulated_count_;
|
uint32_t accumulated_count_;
|
||||||
|
@ -198,7 +198,7 @@ class FecPacketMaskMetricsTest : public ::testing::Test {
|
|||||||
int RecoveredMediaPackets(int num_media_packets,
|
int RecoveredMediaPackets(int num_media_packets,
|
||||||
int num_fec_packets,
|
int num_fec_packets,
|
||||||
uint8_t* state) {
|
uint8_t* state) {
|
||||||
scoped_array<uint8_t> state_tmp(
|
scoped_ptr<uint8_t[]> state_tmp(
|
||||||
new uint8_t[num_media_packets + num_fec_packets]);
|
new uint8_t[num_media_packets + num_fec_packets]);
|
||||||
memcpy(state_tmp.get(), state, num_media_packets + num_fec_packets);
|
memcpy(state_tmp.get(), state, num_media_packets + num_fec_packets);
|
||||||
int num_recovered_packets = 0;
|
int num_recovered_packets = 0;
|
||||||
@ -392,7 +392,7 @@ class FecPacketMaskMetricsTest : public ::testing::Test {
|
|||||||
// (which containes the code size parameters/protection length).
|
// (which containes the code size parameters/protection length).
|
||||||
void ComputeMetricsForCode(CodeType code_type,
|
void ComputeMetricsForCode(CodeType code_type,
|
||||||
int code_index) {
|
int code_index) {
|
||||||
scoped_array<double> prob_weight(new double[kNumLossModels]);
|
scoped_ptr<double[]> prob_weight(new double[kNumLossModels]);
|
||||||
memset(prob_weight.get() , 0, sizeof(double) * kNumLossModels);
|
memset(prob_weight.get() , 0, sizeof(double) * kNumLossModels);
|
||||||
MetricsFecCode metrics_code;
|
MetricsFecCode metrics_code;
|
||||||
SetMetricsZero(&metrics_code);
|
SetMetricsZero(&metrics_code);
|
||||||
@ -400,7 +400,7 @@ class FecPacketMaskMetricsTest : public ::testing::Test {
|
|||||||
int num_media_packets = code_params_[code_index].num_media_packets;
|
int num_media_packets = code_params_[code_index].num_media_packets;
|
||||||
int num_fec_packets = code_params_[code_index].num_fec_packets;
|
int num_fec_packets = code_params_[code_index].num_fec_packets;
|
||||||
int tot_num_packets = num_media_packets + num_fec_packets;
|
int tot_num_packets = num_media_packets + num_fec_packets;
|
||||||
scoped_array<uint8_t> state(new uint8_t[tot_num_packets]);
|
scoped_ptr<uint8_t[]> state(new uint8_t[tot_num_packets]);
|
||||||
memset(state.get() , 0, tot_num_packets);
|
memset(state.get() , 0, tot_num_packets);
|
||||||
|
|
||||||
int num_loss_configurations = static_cast<int>(pow(2.0f, tot_num_packets));
|
int num_loss_configurations = static_cast<int>(pow(2.0f, tot_num_packets));
|
||||||
|
@ -473,7 +473,7 @@ TEST_F(VideoCaptureExternalTest, TestExternalCapture) {
|
|||||||
unsigned int length = webrtc::CalcBufferSize(webrtc::kI420,
|
unsigned int length = webrtc::CalcBufferSize(webrtc::kI420,
|
||||||
test_frame_.width(),
|
test_frame_.width(),
|
||||||
test_frame_.height());
|
test_frame_.height());
|
||||||
webrtc::scoped_array<uint8_t> test_buffer(new uint8_t[length]);
|
webrtc::scoped_ptr<uint8_t[]> test_buffer(new uint8_t[length]);
|
||||||
webrtc::ExtractBuffer(test_frame_, length, test_buffer.get());
|
webrtc::ExtractBuffer(test_frame_, length, test_buffer.get());
|
||||||
EXPECT_EQ(0, capture_input_interface_->IncomingFrame(test_buffer.get(),
|
EXPECT_EQ(0, capture_input_interface_->IncomingFrame(test_buffer.get(),
|
||||||
length, capture_callback_.capability(), 0));
|
length, capture_callback_.capability(), 0));
|
||||||
@ -558,7 +558,7 @@ TEST_F(VideoCaptureExternalTest , FrameRate) {
|
|||||||
unsigned int length = webrtc::CalcBufferSize(webrtc::kI420,
|
unsigned int length = webrtc::CalcBufferSize(webrtc::kI420,
|
||||||
test_frame_.width(),
|
test_frame_.width(),
|
||||||
test_frame_.height());
|
test_frame_.height());
|
||||||
webrtc::scoped_array<uint8_t> test_buffer(new uint8_t[length]);
|
webrtc::scoped_ptr<uint8_t[]> test_buffer(new uint8_t[length]);
|
||||||
webrtc::ExtractBuffer(test_frame_, length, test_buffer.get());
|
webrtc::ExtractBuffer(test_frame_, length, test_buffer.get());
|
||||||
EXPECT_EQ(0, capture_input_interface_->IncomingFrame(test_buffer.get(),
|
EXPECT_EQ(0, capture_input_interface_->IncomingFrame(test_buffer.get(),
|
||||||
length, capture_callback_.capability(), 0));
|
length, capture_callback_.capability(), 0));
|
||||||
@ -574,7 +574,7 @@ TEST_F(VideoCaptureExternalTest , FrameRate) {
|
|||||||
unsigned int length = webrtc::CalcBufferSize(webrtc::kI420,
|
unsigned int length = webrtc::CalcBufferSize(webrtc::kI420,
|
||||||
test_frame_.width(),
|
test_frame_.width(),
|
||||||
test_frame_.height());
|
test_frame_.height());
|
||||||
webrtc::scoped_array<uint8_t> test_buffer(new uint8_t[length]);
|
webrtc::scoped_ptr<uint8_t[]> test_buffer(new uint8_t[length]);
|
||||||
webrtc::ExtractBuffer(test_frame_, length, test_buffer.get());
|
webrtc::ExtractBuffer(test_frame_, length, test_buffer.get());
|
||||||
EXPECT_EQ(0, capture_input_interface_->IncomingFrame(test_buffer.get(),
|
EXPECT_EQ(0, capture_input_interface_->IncomingFrame(test_buffer.get(),
|
||||||
length, capture_callback_.capability(), 0));
|
length, capture_callback_.capability(), 0));
|
||||||
@ -592,7 +592,7 @@ TEST_F(VideoCaptureExternalTest, Rotation) {
|
|||||||
unsigned int length = webrtc::CalcBufferSize(webrtc::kI420,
|
unsigned int length = webrtc::CalcBufferSize(webrtc::kI420,
|
||||||
test_frame_.width(),
|
test_frame_.width(),
|
||||||
test_frame_.height());
|
test_frame_.height());
|
||||||
webrtc::scoped_array<uint8_t> test_buffer(new uint8_t[length]);
|
webrtc::scoped_ptr<uint8_t[]> test_buffer(new uint8_t[length]);
|
||||||
webrtc::ExtractBuffer(test_frame_, length, test_buffer.get());
|
webrtc::ExtractBuffer(test_frame_, length, test_buffer.get());
|
||||||
EXPECT_EQ(0, capture_input_interface_->IncomingFrame(test_buffer.get(),
|
EXPECT_EQ(0, capture_input_interface_->IncomingFrame(test_buffer.get(),
|
||||||
length, capture_callback_.capability(), 0));
|
length, capture_callback_.capability(), 0));
|
||||||
|
@ -331,7 +331,7 @@ void VideoProcessorImpl::FrameDecoded(const I420VideoFrame& image) {
|
|||||||
}
|
}
|
||||||
// TODO(mikhal): Extracting the buffer for now - need to update test.
|
// TODO(mikhal): Extracting the buffer for now - need to update test.
|
||||||
int length = CalcBufferSize(kI420, up_image.width(), up_image.height());
|
int length = CalcBufferSize(kI420, up_image.width(), up_image.height());
|
||||||
scoped_array<uint8_t> image_buffer(new uint8_t[length]);
|
scoped_ptr<uint8_t[]> image_buffer(new uint8_t[length]);
|
||||||
length = ExtractBuffer(up_image, length, image_buffer.get());
|
length = ExtractBuffer(up_image, length, image_buffer.get());
|
||||||
// Update our copy of the last successful frame:
|
// Update our copy of the last successful frame:
|
||||||
memcpy(last_successful_frame_buffer_, image_buffer.get(), length);
|
memcpy(last_successful_frame_buffer_, image_buffer.get(), length);
|
||||||
@ -344,7 +344,7 @@ void VideoProcessorImpl::FrameDecoded(const I420VideoFrame& image) {
|
|||||||
// Update our copy of the last successful frame:
|
// Update our copy of the last successful frame:
|
||||||
// TODO(mikhal): Add as a member function, so won't be allocated per frame.
|
// TODO(mikhal): Add as a member function, so won't be allocated per frame.
|
||||||
int length = CalcBufferSize(kI420, image.width(), image.height());
|
int length = CalcBufferSize(kI420, image.width(), image.height());
|
||||||
scoped_array<uint8_t> image_buffer(new uint8_t[length]);
|
scoped_ptr<uint8_t[]> image_buffer(new uint8_t[length]);
|
||||||
length = ExtractBuffer(image, length, image_buffer.get());
|
length = ExtractBuffer(image, length, image_buffer.get());
|
||||||
assert(length > 0);
|
assert(length > 0);
|
||||||
memcpy(last_successful_frame_buffer_, image_buffer.get(), length);
|
memcpy(last_successful_frame_buffer_, image_buffer.get(), length);
|
||||||
|
@ -565,7 +565,7 @@ UnitTest::Perform()
|
|||||||
frameLength = WaitForDecodedFrame();
|
frameLength = WaitForDecodedFrame();
|
||||||
}
|
}
|
||||||
unsigned int length = CalcBufferSize(kI420, width, height);
|
unsigned int length = CalcBufferSize(kI420, width, height);
|
||||||
scoped_array<uint8_t> decoded_buffer(new uint8_t[length]);
|
scoped_ptr<uint8_t[]> decoded_buffer(new uint8_t[length]);
|
||||||
ExtractBuffer(_decodedVideoBuffer, _lengthSourceFrame,
|
ExtractBuffer(_decodedVideoBuffer, _lengthSourceFrame,
|
||||||
decoded_buffer.get());
|
decoded_buffer.get());
|
||||||
EXPECT_TRUE(CheckIfBitExact(decoded_buffer.get(), frameLength, _refDecFrame,
|
EXPECT_TRUE(CheckIfBitExact(decoded_buffer.get(), frameLength, _refDecFrame,
|
||||||
@ -645,7 +645,7 @@ UnitTest::Perform()
|
|||||||
|
|
||||||
// check that decoded frame matches with reference
|
// check that decoded frame matches with reference
|
||||||
unsigned int length = CalcBufferSize(kI420, width, height);
|
unsigned int length = CalcBufferSize(kI420, width, height);
|
||||||
scoped_array<uint8_t> decoded_buffer(new uint8_t[length]);
|
scoped_ptr<uint8_t[]> decoded_buffer(new uint8_t[length]);
|
||||||
ExtractBuffer(_decodedVideoBuffer, length, decoded_buffer.get());
|
ExtractBuffer(_decodedVideoBuffer, length, decoded_buffer.get());
|
||||||
EXPECT_TRUE(CheckIfBitExact(decoded_buffer.get(), length,
|
EXPECT_TRUE(CheckIfBitExact(decoded_buffer.get(), length,
|
||||||
_refDecFrame, _lengthSourceFrame) == true);
|
_refDecFrame, _lengthSourceFrame) == true);
|
||||||
|
@ -181,7 +181,7 @@ class TestVp8Impl : public ::testing::Test {
|
|||||||
|
|
||||||
scoped_ptr<Vp8UnitTestEncodeCompleteCallback> encode_complete_callback_;
|
scoped_ptr<Vp8UnitTestEncodeCompleteCallback> encode_complete_callback_;
|
||||||
scoped_ptr<Vp8UnitTestDecodeCompleteCallback> decode_complete_callback_;
|
scoped_ptr<Vp8UnitTestDecodeCompleteCallback> decode_complete_callback_;
|
||||||
scoped_array<uint8_t> source_buffer_;
|
scoped_ptr<uint8_t[]> source_buffer_;
|
||||||
FILE* source_file_;
|
FILE* source_file_;
|
||||||
I420VideoFrame input_frame_;
|
I420VideoFrame input_frame_;
|
||||||
scoped_ptr<VideoEncoder> encoder_;
|
scoped_ptr<VideoEncoder> encoder_;
|
||||||
|
@ -142,7 +142,7 @@ int SequenceCoder(webrtc::test::CommandLineParser& parser) {
|
|||||||
EXPECT_EQ(0, decoder->InitDecode(&inst, 1));
|
EXPECT_EQ(0, decoder->InitDecode(&inst, 1));
|
||||||
webrtc::I420VideoFrame input_frame;
|
webrtc::I420VideoFrame input_frame;
|
||||||
unsigned int length = webrtc::CalcBufferSize(webrtc::kI420, width, height);
|
unsigned int length = webrtc::CalcBufferSize(webrtc::kI420, width, height);
|
||||||
webrtc::scoped_array<uint8_t> frame_buffer(new uint8_t[length]);
|
webrtc::scoped_ptr<uint8_t[]> frame_buffer(new uint8_t[length]);
|
||||||
|
|
||||||
int half_width = (width + 1) / 2;
|
int half_width = (width + 1) / 2;
|
||||||
// Set and register callbacks.
|
// Set and register callbacks.
|
||||||
|
@ -61,7 +61,7 @@ class RawRtpPacket {
|
|||||||
uint16_t seq_num() const { return seq_num_; }
|
uint16_t seq_num() const { return seq_num_; }
|
||||||
|
|
||||||
private:
|
private:
|
||||||
scoped_array<uint8_t> data_;
|
scoped_ptr<uint8_t[]> data_;
|
||||||
uint32_t length_;
|
uint32_t length_;
|
||||||
int64_t resend_time_ms_;
|
int64_t resend_time_ms_;
|
||||||
uint32_t ssrc_;
|
uint32_t ssrc_;
|
||||||
|
@ -19,7 +19,7 @@ TEST_F(VideoProcessingModuleTest, BrightnessDetection)
|
|||||||
uint32_t frameNum = 0;
|
uint32_t frameNum = 0;
|
||||||
int32_t brightnessWarning = 0;
|
int32_t brightnessWarning = 0;
|
||||||
uint32_t warningCount = 0;
|
uint32_t warningCount = 0;
|
||||||
scoped_array<uint8_t> video_buffer(new uint8_t[frame_length_]);
|
scoped_ptr<uint8_t[]> video_buffer(new uint8_t[frame_length_]);
|
||||||
while (fread(video_buffer.get(), 1, frame_length_, source_file_) ==
|
while (fread(video_buffer.get(), 1, frame_length_, source_file_) ==
|
||||||
frame_length_)
|
frame_length_)
|
||||||
{
|
{
|
||||||
|
@ -39,7 +39,7 @@ TEST_F(VideoProcessingModuleTest, ColorEnhancement)
|
|||||||
ASSERT_TRUE(modFile != NULL) << "Could not open output file.\n";
|
ASSERT_TRUE(modFile != NULL) << "Could not open output file.\n";
|
||||||
|
|
||||||
uint32_t frameNum = 0;
|
uint32_t frameNum = 0;
|
||||||
scoped_array<uint8_t> video_buffer(new uint8_t[frame_length_]);
|
scoped_ptr<uint8_t[]> video_buffer(new uint8_t[frame_length_]);
|
||||||
while (fread(video_buffer.get(), 1, frame_length_, source_file_) ==
|
while (fread(video_buffer.get(), 1, frame_length_, source_file_) ==
|
||||||
frame_length_)
|
frame_length_)
|
||||||
{
|
{
|
||||||
@ -86,7 +86,7 @@ TEST_F(VideoProcessingModuleTest, ColorEnhancement)
|
|||||||
width_, half_width_, half_width_);
|
width_, half_width_, half_width_);
|
||||||
|
|
||||||
// Compare frame-by-frame.
|
// Compare frame-by-frame.
|
||||||
scoped_array<uint8_t> ref_buffer(new uint8_t[frame_length_]);
|
scoped_ptr<uint8_t[]> ref_buffer(new uint8_t[frame_length_]);
|
||||||
while (fread(video_buffer.get(), 1, frame_length_, modFile) ==
|
while (fread(video_buffer.get(), 1, frame_length_, modFile) ==
|
||||||
frame_length_)
|
frame_length_)
|
||||||
{
|
{
|
||||||
@ -114,7 +114,7 @@ TEST_F(VideoProcessingModuleTest, ColorEnhancement)
|
|||||||
// Verify that all color pixels are enhanced, and no luminance values are
|
// Verify that all color pixels are enhanced, and no luminance values are
|
||||||
// altered.
|
// altered.
|
||||||
|
|
||||||
scoped_array<uint8_t> testFrame(new uint8_t[frame_length_]);
|
scoped_ptr<uint8_t[]> testFrame(new uint8_t[frame_length_]);
|
||||||
|
|
||||||
// Use value 128 as probe value, since we know that this will be changed
|
// Use value 128 as probe value, since we know that this will be changed
|
||||||
// in the enhancement.
|
// in the enhancement.
|
||||||
|
@ -23,7 +23,7 @@ TEST_F(VideoProcessingModuleTest, ContentAnalysis) {
|
|||||||
ca__c.Initialize(width_,height_);
|
ca__c.Initialize(width_,height_);
|
||||||
ca__sse.Initialize(width_,height_);
|
ca__sse.Initialize(width_,height_);
|
||||||
|
|
||||||
scoped_array<uint8_t> video_buffer(new uint8_t[frame_length_]);
|
scoped_ptr<uint8_t[]> video_buffer(new uint8_t[frame_length_]);
|
||||||
while (fread(video_buffer.get(), 1, frame_length_, source_file_)
|
while (fread(video_buffer.get(), 1, frame_length_, source_file_)
|
||||||
== frame_length_) {
|
== frame_length_) {
|
||||||
// Using ConvertToI420 to add stride to the image.
|
// Using ConvertToI420 to add stride to the image.
|
||||||
|
@ -43,7 +43,7 @@ TEST_F(VideoProcessingModuleTest, Deflickering)
|
|||||||
"Could not open output file: " << output_file << "\n";
|
"Could not open output file: " << output_file << "\n";
|
||||||
|
|
||||||
printf("\nRun time [us / frame]:\n");
|
printf("\nRun time [us / frame]:\n");
|
||||||
scoped_array<uint8_t> video_buffer(new uint8_t[frame_length_]);
|
scoped_ptr<uint8_t[]> video_buffer(new uint8_t[frame_length_]);
|
||||||
for (uint32_t run_idx = 0; run_idx < NumRuns; run_idx++)
|
for (uint32_t run_idx = 0; run_idx < NumRuns; run_idx++)
|
||||||
{
|
{
|
||||||
TickTime t0;
|
TickTime t0;
|
||||||
|
@ -49,7 +49,7 @@ TEST_F(VideoProcessingModuleTest, DISABLED_ON_ANDROID(Denoising))
|
|||||||
int32_t modifiedPixels = 0;
|
int32_t modifiedPixels = 0;
|
||||||
|
|
||||||
frameNum = 0;
|
frameNum = 0;
|
||||||
scoped_array<uint8_t> video_buffer(new uint8_t[frame_length_]);
|
scoped_ptr<uint8_t[]> video_buffer(new uint8_t[frame_length_]);
|
||||||
while (fread(video_buffer.get(), 1, frame_length_, source_file_) ==
|
while (fread(video_buffer.get(), 1, frame_length_, source_file_) ==
|
||||||
frame_length_)
|
frame_length_)
|
||||||
{
|
{
|
||||||
|
@ -89,7 +89,7 @@ TEST_F(VideoProcessingModuleTest, HandleNullBuffer) {
|
|||||||
|
|
||||||
TEST_F(VideoProcessingModuleTest, HandleBadStats) {
|
TEST_F(VideoProcessingModuleTest, HandleBadStats) {
|
||||||
VideoProcessingModule::FrameStats stats;
|
VideoProcessingModule::FrameStats stats;
|
||||||
scoped_array<uint8_t> video_buffer(new uint8_t[frame_length_]);
|
scoped_ptr<uint8_t[]> video_buffer(new uint8_t[frame_length_]);
|
||||||
ASSERT_EQ(frame_length_, fread(video_buffer.get(), 1, frame_length_,
|
ASSERT_EQ(frame_length_, fread(video_buffer.get(), 1, frame_length_,
|
||||||
source_file_));
|
source_file_));
|
||||||
EXPECT_EQ(0, ConvertToI420(kI420, video_buffer.get(), 0, 0,
|
EXPECT_EQ(0, ConvertToI420(kI420, video_buffer.get(), 0, 0,
|
||||||
@ -129,7 +129,7 @@ TEST_F(VideoProcessingModuleTest, IdenticalResultsAfterReset) {
|
|||||||
I420VideoFrame video_frame2;
|
I420VideoFrame video_frame2;
|
||||||
VideoProcessingModule::FrameStats stats;
|
VideoProcessingModule::FrameStats stats;
|
||||||
// Only testing non-static functions here.
|
// Only testing non-static functions here.
|
||||||
scoped_array<uint8_t> video_buffer(new uint8_t[frame_length_]);
|
scoped_ptr<uint8_t[]> video_buffer(new uint8_t[frame_length_]);
|
||||||
ASSERT_EQ(frame_length_, fread(video_buffer.get(), 1, frame_length_,
|
ASSERT_EQ(frame_length_, fread(video_buffer.get(), 1, frame_length_,
|
||||||
source_file_));
|
source_file_));
|
||||||
EXPECT_EQ(0, ConvertToI420(kI420, video_buffer.get(), 0, 0,
|
EXPECT_EQ(0, ConvertToI420(kI420, video_buffer.get(), 0, 0,
|
||||||
@ -172,7 +172,7 @@ TEST_F(VideoProcessingModuleTest, IdenticalResultsAfterReset) {
|
|||||||
|
|
||||||
TEST_F(VideoProcessingModuleTest, FrameStats) {
|
TEST_F(VideoProcessingModuleTest, FrameStats) {
|
||||||
VideoProcessingModule::FrameStats stats;
|
VideoProcessingModule::FrameStats stats;
|
||||||
scoped_array<uint8_t> video_buffer(new uint8_t[frame_length_]);
|
scoped_ptr<uint8_t[]> video_buffer(new uint8_t[frame_length_]);
|
||||||
ASSERT_EQ(frame_length_, fread(video_buffer.get(), 1, frame_length_,
|
ASSERT_EQ(frame_length_, fread(video_buffer.get(), 1, frame_length_,
|
||||||
source_file_));
|
source_file_));
|
||||||
EXPECT_EQ(0, ConvertToI420(kI420, video_buffer.get(), 0, 0,
|
EXPECT_EQ(0, ConvertToI420(kI420, video_buffer.get(), 0, 0,
|
||||||
@ -242,7 +242,7 @@ TEST_F(VideoProcessingModuleTest, Resampler) {
|
|||||||
vpm_->EnableTemporalDecimation(false);
|
vpm_->EnableTemporalDecimation(false);
|
||||||
|
|
||||||
// Reading test frame
|
// Reading test frame
|
||||||
scoped_array<uint8_t> video_buffer(new uint8_t[frame_length_]);
|
scoped_ptr<uint8_t[]> video_buffer(new uint8_t[frame_length_]);
|
||||||
ASSERT_EQ(frame_length_, fread(video_buffer.get(), 1, frame_length_,
|
ASSERT_EQ(frame_length_, fread(video_buffer.get(), 1, frame_length_,
|
||||||
source_file_));
|
source_file_));
|
||||||
// Using ConvertToI420 to add stride to the image.
|
// Using ConvertToI420 to add stride to the image.
|
||||||
|
@ -563,82 +563,4 @@ bool operator!=(T* p1, const webrtc::scoped_ptr<T, D>& p2) {
|
|||||||
return p1 != p2.get();
|
return p1 != p2.get();
|
||||||
}
|
}
|
||||||
|
|
||||||
namespace webrtc {
|
|
||||||
|
|
||||||
// DEPRECATED: Use scoped_ptr<T[]> instead.
|
|
||||||
// TODO(ajm): Remove scoped_array.
|
|
||||||
//
|
|
||||||
// scoped_array extends scoped_ptr to arrays. Deletion of the array pointed to
|
|
||||||
// is guaranteed, either on destruction of the scoped_array or via an explicit
|
|
||||||
// reset(). Use shared_array or std::vector if your needs are more complex.
|
|
||||||
|
|
||||||
template<typename T>
|
|
||||||
class scoped_array {
|
|
||||||
private:
|
|
||||||
|
|
||||||
T* ptr;
|
|
||||||
|
|
||||||
scoped_array(scoped_array const &);
|
|
||||||
scoped_array & operator=(scoped_array const &);
|
|
||||||
|
|
||||||
public:
|
|
||||||
|
|
||||||
typedef T element_type;
|
|
||||||
|
|
||||||
explicit scoped_array(T* p = NULL) : ptr(p) {}
|
|
||||||
|
|
||||||
~scoped_array() {
|
|
||||||
typedef char type_must_be_complete[sizeof(T)];
|
|
||||||
delete[] ptr;
|
|
||||||
}
|
|
||||||
|
|
||||||
void reset(T* p = NULL) {
|
|
||||||
typedef char type_must_be_complete[sizeof(T)];
|
|
||||||
|
|
||||||
if (ptr != p) {
|
|
||||||
T* arr = ptr;
|
|
||||||
ptr = p;
|
|
||||||
// Delete last, in case arr destructor indirectly results in ~scoped_array
|
|
||||||
delete [] arr;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
T& operator[](ptrdiff_t i) const {
|
|
||||||
assert(ptr != NULL);
|
|
||||||
assert(i >= 0);
|
|
||||||
return ptr[i];
|
|
||||||
}
|
|
||||||
|
|
||||||
T* get() const {
|
|
||||||
return ptr;
|
|
||||||
}
|
|
||||||
|
|
||||||
void swap(scoped_array & b) {
|
|
||||||
T* tmp = b.ptr;
|
|
||||||
b.ptr = ptr;
|
|
||||||
ptr = tmp;
|
|
||||||
}
|
|
||||||
|
|
||||||
T* release() {
|
|
||||||
T* tmp = ptr;
|
|
||||||
ptr = NULL;
|
|
||||||
return tmp;
|
|
||||||
}
|
|
||||||
|
|
||||||
T** accept() {
|
|
||||||
if (ptr) {
|
|
||||||
delete [] ptr;
|
|
||||||
ptr = NULL;
|
|
||||||
}
|
|
||||||
return &ptr;
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
template<class T> inline
|
|
||||||
void swap(scoped_array<T>& a, scoped_array<T>& b) {
|
|
||||||
a.swap(b);
|
|
||||||
}
|
|
||||||
|
|
||||||
} // namespace webrtc
|
|
||||||
|
|
||||||
#endif // WEBRTC_SYSTEM_WRAPPERS_INTERFACE_SCOPED_PTR_H_
|
#endif // WEBRTC_SYSTEM_WRAPPERS_INTERFACE_SCOPED_PTR_H_
|
||||||
|
@ -47,7 +47,7 @@ class FakeNetworkPipeTest : public ::testing::Test {
|
|||||||
}
|
}
|
||||||
|
|
||||||
void SendPackets(FakeNetworkPipe* pipe, int number_packets, int kPacketSize) {
|
void SendPackets(FakeNetworkPipe* pipe, int number_packets, int kPacketSize) {
|
||||||
scoped_array<uint8_t> packet(new uint8_t[kPacketSize]);
|
scoped_ptr<uint8_t[]> packet(new uint8_t[kPacketSize]);
|
||||||
for (int i = 0; i < number_packets; ++i) {
|
for (int i = 0; i < number_packets; ++i) {
|
||||||
pipe->SendPacket(packet.get(), kPacketSize);
|
pipe->SendPacket(packet.get(), kPacketSize);
|
||||||
}
|
}
|
||||||
|
@ -111,8 +111,8 @@ int CalculateMetrics(VideoMetricsType video_metrics_type,
|
|||||||
const size_t frame_length = 3 * width * height >> 1;
|
const size_t frame_length = 3 * width * height >> 1;
|
||||||
I420VideoFrame ref_frame;
|
I420VideoFrame ref_frame;
|
||||||
I420VideoFrame test_frame;
|
I420VideoFrame test_frame;
|
||||||
scoped_array<uint8_t> ref_buffer(new uint8_t[frame_length]);
|
scoped_ptr<uint8_t[]> ref_buffer(new uint8_t[frame_length]);
|
||||||
scoped_array<uint8_t> test_buffer(new uint8_t[frame_length]);
|
scoped_ptr<uint8_t[]> test_buffer(new uint8_t[frame_length]);
|
||||||
|
|
||||||
// Set decoded image parameters.
|
// Set decoded image parameters.
|
||||||
int half_width = (width + 1) / 2;
|
int half_width = (width + 1) / 2;
|
||||||
|
@ -38,7 +38,7 @@ int EditFrames(const string& in_path, int width, int height,
|
|||||||
// Frame size of I420.
|
// Frame size of I420.
|
||||||
int frame_length = CalcBufferSize(kI420, width, height);
|
int frame_length = CalcBufferSize(kI420, width, height);
|
||||||
|
|
||||||
webrtc::scoped_array<uint8_t> temp_buffer(new uint8_t[frame_length]);
|
webrtc::scoped_ptr<uint8_t[]> temp_buffer(new uint8_t[frame_length]);
|
||||||
|
|
||||||
FILE* out_fid = fopen(out_path.c_str(), "wb");
|
FILE* out_fid = fopen(out_path.c_str(), "wb");
|
||||||
|
|
||||||
|
@ -53,8 +53,8 @@ class FrameEditingTest : public ::testing::Test {
|
|||||||
}
|
}
|
||||||
// Compares the frames in both streams to the end of one of the streams.
|
// Compares the frames in both streams to the end of one of the streams.
|
||||||
void CompareToTheEnd(FILE* test_video_fid, FILE* ref_video_fid,
|
void CompareToTheEnd(FILE* test_video_fid, FILE* ref_video_fid,
|
||||||
scoped_array<int>* ref_buffer,
|
scoped_ptr<int[]>* ref_buffer,
|
||||||
scoped_array<int>* test_buffer) {
|
scoped_ptr<int[]>* test_buffer) {
|
||||||
while (!feof(test_video_fid) && !feof(ref_video_fid)) {
|
while (!feof(test_video_fid) && !feof(ref_video_fid)) {
|
||||||
num_bytes_read_ = fread(ref_buffer->get(), 1, kFrameSize, ref_video_fid);
|
num_bytes_read_ = fread(ref_buffer->get(), 1, kFrameSize, ref_video_fid);
|
||||||
if (!feof(ref_video_fid)) {
|
if (!feof(ref_video_fid)) {
|
||||||
@ -78,8 +78,8 @@ class FrameEditingTest : public ::testing::Test {
|
|||||||
FILE* original_fid_;
|
FILE* original_fid_;
|
||||||
FILE* edited_fid_;
|
FILE* edited_fid_;
|
||||||
int num_bytes_read_;
|
int num_bytes_read_;
|
||||||
scoped_array<int> original_buffer_;
|
scoped_ptr<int[]> original_buffer_;
|
||||||
scoped_array<int> edited_buffer_;
|
scoped_ptr<int[]> edited_buffer_;
|
||||||
int num_frames_read_;
|
int num_frames_read_;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -30,7 +30,7 @@ struct Frame {
|
|||||||
memcpy(this->buffer.get(), buffer, buffer_size);
|
memcpy(this->buffer.get(), buffer, buffer_size);
|
||||||
}
|
}
|
||||||
|
|
||||||
webrtc::scoped_array<unsigned char> buffer;
|
webrtc::scoped_ptr<unsigned char[]> buffer;
|
||||||
int buffer_size;
|
int buffer_size;
|
||||||
uint32_t timestamp;
|
uint32_t timestamp;
|
||||||
int64_t render_time;
|
int64_t render_time;
|
||||||
|
@ -535,7 +535,7 @@ void ViECapturer::DeliverI420Frame(I420VideoFrame* video_frame) {
|
|||||||
unsigned int length = CalcBufferSize(kI420,
|
unsigned int length = CalcBufferSize(kI420,
|
||||||
video_frame->width(),
|
video_frame->width(),
|
||||||
video_frame->height());
|
video_frame->height());
|
||||||
scoped_array<uint8_t> video_buffer(new uint8_t[length]);
|
scoped_ptr<uint8_t[]> video_buffer(new uint8_t[length]);
|
||||||
ExtractBuffer(*video_frame, length, video_buffer.get());
|
ExtractBuffer(*video_frame, length, video_buffer.get());
|
||||||
effect_filter_->Transform(length,
|
effect_filter_->Transform(length,
|
||||||
video_buffer.get(),
|
video_buffer.get(),
|
||||||
|
@ -1401,7 +1401,7 @@ int32_t ViEChannel::FrameToRender(
|
|||||||
unsigned int length = CalcBufferSize(kI420,
|
unsigned int length = CalcBufferSize(kI420,
|
||||||
video_frame.width(),
|
video_frame.width(),
|
||||||
video_frame.height());
|
video_frame.height());
|
||||||
scoped_array<uint8_t> video_buffer(new uint8_t[length]);
|
scoped_ptr<uint8_t[]> video_buffer(new uint8_t[length]);
|
||||||
ExtractBuffer(video_frame, length, video_buffer.get());
|
ExtractBuffer(video_frame, length, video_buffer.get());
|
||||||
effect_filter_->Transform(length,
|
effect_filter_->Transform(length,
|
||||||
video_buffer.get(),
|
video_buffer.get(),
|
||||||
|
@ -503,7 +503,7 @@ void ViEEncoder::DeliverFrame(int id,
|
|||||||
unsigned int length = CalcBufferSize(kI420,
|
unsigned int length = CalcBufferSize(kI420,
|
||||||
video_frame->width(),
|
video_frame->width(),
|
||||||
video_frame->height());
|
video_frame->height());
|
||||||
scoped_array<uint8_t> video_buffer(new uint8_t[length]);
|
scoped_ptr<uint8_t[]> video_buffer(new uint8_t[length]);
|
||||||
ExtractBuffer(*video_frame, length, video_buffer.get());
|
ExtractBuffer(*video_frame, length, video_buffer.get());
|
||||||
effect_filter_->Transform(length,
|
effect_filter_->Transform(length,
|
||||||
video_buffer.get(),
|
video_buffer.get(),
|
||||||
|
@ -4279,7 +4279,7 @@ Channel::GetRtpRtcp(RtpRtcp** rtpRtcpModule, RtpReceiver** rtp_receiver) const
|
|||||||
int32_t
|
int32_t
|
||||||
Channel::MixOrReplaceAudioWithFile(int mixingFrequency)
|
Channel::MixOrReplaceAudioWithFile(int mixingFrequency)
|
||||||
{
|
{
|
||||||
scoped_array<int16_t> fileBuffer(new int16_t[640]);
|
scoped_ptr<int16_t[]> fileBuffer(new int16_t[640]);
|
||||||
int fileSamples(0);
|
int fileSamples(0);
|
||||||
|
|
||||||
{
|
{
|
||||||
@ -4349,7 +4349,7 @@ Channel::MixAudioWithFile(AudioFrame& audioFrame,
|
|||||||
{
|
{
|
||||||
assert(mixingFrequency <= 32000);
|
assert(mixingFrequency <= 32000);
|
||||||
|
|
||||||
scoped_array<int16_t> fileBuffer(new int16_t[640]);
|
scoped_ptr<int16_t[]> fileBuffer(new int16_t[640]);
|
||||||
int fileSamples(0);
|
int fileSamples(0);
|
||||||
|
|
||||||
{
|
{
|
||||||
|
@ -53,7 +53,7 @@ class FakeVoEExternalMedia : public VoEExternalMedia {
|
|||||||
int samples_per_channel, int sample_rate_hz,
|
int samples_per_channel, int sample_rate_hz,
|
||||||
int num_channels) {
|
int num_channels) {
|
||||||
const int length = samples_per_channel * num_channels;
|
const int length = samples_per_channel * num_channels;
|
||||||
scoped_array<int16_t> data;
|
scoped_ptr<int16_t[]> data;
|
||||||
if (!audio) {
|
if (!audio) {
|
||||||
data.reset(new int16_t[length]);
|
data.reset(new int16_t[length]);
|
||||||
memset(data.get(), 0, length * sizeof(data[0]));
|
memset(data.get(), 0, length * sizeof(data[0]));
|
||||||
|
@ -1217,7 +1217,7 @@ int32_t TransmitMixer::RecordAudioToFile(
|
|||||||
int32_t TransmitMixer::MixOrReplaceAudioWithFile(
|
int32_t TransmitMixer::MixOrReplaceAudioWithFile(
|
||||||
int mixingFrequency)
|
int mixingFrequency)
|
||||||
{
|
{
|
||||||
scoped_array<int16_t> fileBuffer(new int16_t[640]);
|
scoped_ptr<int16_t[]> fileBuffer(new int16_t[640]);
|
||||||
|
|
||||||
int fileSamples(0);
|
int fileSamples(0);
|
||||||
{
|
{
|
||||||
|
Loading…
x
Reference in New Issue
Block a user