Update FrameCombiner et al to use DeinterleavedView
* FrameCombiner is simpler. No additional channel pointers for buffers.
* Improve consistency in using views in downstream classes.
* Deprecate older methods (some have upstream dependencies).
* Use samples per channel instead of sample rate where the former is
really what's needed.
Bug: chromium:335805780
Change-Id: I0dde8ed7a5a187bbddd18d3b6c649aa0865e6d4a
Reviewed-on: https://webrtc-review.googlesource.com/c/src/+/352582
Commit-Queue: Tomas Gunnarsson <tommi@webrtc.org>
Reviewed-by: Sam Zackrisson <saza@webrtc.org>
Cr-Commit-Position: refs/heads/main@{#42575}
diff --git a/modules/audio_mixer/frame_combiner.cc b/modules/audio_mixer/frame_combiner.cc
index ca79ecd..dfe9511 100644
--- a/modules/audio_mixer/frame_combiner.cc
+++ b/modules/audio_mixer/frame_combiner.cc
@@ -106,22 +106,22 @@
}
}
-void RunLimiter(AudioFrameView<float> mixing_buffer_view, Limiter* limiter) {
- limiter->SetSamplesPerChannel(mixing_buffer_view.samples_per_channel());
- limiter->Process(mixing_buffer_view);
+void RunLimiter(DeinterleavedView<float> deinterleaved, Limiter* limiter) {
+ limiter->SetSamplesPerChannel(deinterleaved.samples_per_channel());
+ limiter->Process(deinterleaved);
}
// Both interleaves and rounds.
-void InterleaveToAudioFrame(AudioFrameView<const float> mixing_buffer_view,
+void InterleaveToAudioFrame(DeinterleavedView<float> deinterleaved,
AudioFrame* audio_frame_for_mixing) {
InterleavedView<int16_t> mixing_data = audio_frame_for_mixing->mutable_data(
- mixing_buffer_view.samples_per_channel(),
- mixing_buffer_view.num_channels());
+ deinterleaved.samples_per_channel(), deinterleaved.num_channels());
// Put data in the result frame.
for (size_t i = 0; i < mixing_data.num_channels(); ++i) {
+ auto channel = deinterleaved[i];
for (size_t j = 0; j < mixing_data.samples_per_channel(); ++j) {
mixing_data[mixing_data.num_channels() * j + i] =
- FloatS16ToS16(mixing_buffer_view.channel(i)[j]);
+ FloatS16ToS16(channel[j]);
}
}
}
@@ -191,21 +191,11 @@
mixing_buffer_.data(), samples_per_channel, number_of_channels);
MixToFloatFrame(mix_list, deinterleaved);
- // Put float data in an AudioFrameView.
- // TODO(tommi): We should be able to just use `deinterleaved` without an
- // additional array of pointers.
- std::array<float*, kMaximumNumberOfChannels> channel_pointers{};
- for (size_t i = 0; i < number_of_channels; ++i) {
- channel_pointers[i] = deinterleaved[i].data();
- }
- AudioFrameView<float> mixing_buffer_view(
- channel_pointers.data(), number_of_channels, samples_per_channel);
-
if (use_limiter_) {
- RunLimiter(mixing_buffer_view, &limiter_);
+ RunLimiter(deinterleaved, &limiter_);
}
- InterleaveToAudioFrame(mixing_buffer_view, audio_frame_for_mixing);
+ InterleaveToAudioFrame(deinterleaved, audio_frame_for_mixing);
}
} // namespace webrtc
diff --git a/modules/audio_processing/agc2/fixed_digital_level_estimator.cc b/modules/audio_processing/agc2/fixed_digital_level_estimator.cc
index a927b9f..73edcf0 100644
--- a/modules/audio_processing/agc2/fixed_digital_level_estimator.cc
+++ b/modules/audio_processing/agc2/fixed_digital_level_estimator.cc
@@ -56,15 +56,15 @@
}
std::array<float, kSubFramesInFrame> FixedDigitalLevelEstimator::ComputeLevel(
- const AudioFrameView<const float>& float_frame) {
+ DeinterleavedView<const float> float_frame) {
RTC_DCHECK_GT(float_frame.num_channels(), 0);
RTC_DCHECK_EQ(float_frame.samples_per_channel(), samples_in_frame_);
// Compute max envelope without smoothing.
std::array<float, kSubFramesInFrame> envelope{};
- for (int channel_idx = 0; channel_idx < float_frame.num_channels();
+ for (size_t channel_idx = 0; channel_idx < float_frame.num_channels();
++channel_idx) {
- const auto channel = float_frame.channel(channel_idx);
+ const auto channel = float_frame[channel_idx];
for (int sub_frame = 0; sub_frame < kSubFramesInFrame; ++sub_frame) {
for (int sample_in_sub_frame = 0;
sample_in_sub_frame < samples_in_sub_frame_; ++sample_in_sub_frame) {
@@ -99,7 +99,7 @@
// Dump data for debug.
RTC_DCHECK(apm_data_dumper_);
- const auto channel = float_frame.channel(0);
+ const auto channel = float_frame[0];
apm_data_dumper_->DumpRaw("agc2_level_estimator_samples",
samples_in_sub_frame_,
&channel[sub_frame * samples_in_sub_frame_]);
diff --git a/modules/audio_processing/agc2/fixed_digital_level_estimator.h b/modules/audio_processing/agc2/fixed_digital_level_estimator.h
index eee6428..af10250 100644
--- a/modules/audio_processing/agc2/fixed_digital_level_estimator.h
+++ b/modules/audio_processing/agc2/fixed_digital_level_estimator.h
@@ -27,7 +27,7 @@
public:
// `samples_per_channel` is expected to be derived from this formula:
// sample_rate_hz * kFrameDurationMs / 1000
- // or
+ // or, for a 10ms duration:
// sample_rate_hz / 100
// I.e. the number of samples for 10ms of the given sample rate. The
// expectation is that samples per channel is divisible by
@@ -46,7 +46,13 @@
// ms of audio produces a level estimates in the same scale. The
// level estimate contains kSubFramesInFrame values.
std::array<float, kSubFramesInFrame> ComputeLevel(
- const AudioFrameView<const float>& float_frame);
+ DeinterleavedView<const float> float_frame);
+
+ [[deprecated(
+ "Use DeinterleavedView variant")]] std::array<float, kSubFramesInFrame>
+ ComputeLevel(const AudioFrameView<const float>& float_frame) {
+ return ComputeLevel(float_frame.view());
+ }
// Rate may be changed at any time (but not concurrently) from the
// value passed to the constructor. The class is not thread safe.
diff --git a/modules/audio_processing/agc2/fixed_digital_level_estimator_unittest.cc b/modules/audio_processing/agc2/fixed_digital_level_estimator_unittest.cc
index d83f380..c76db85 100644
--- a/modules/audio_processing/agc2/fixed_digital_level_estimator_unittest.cc
+++ b/modules/audio_processing/agc2/fixed_digital_level_estimator_unittest.cc
@@ -40,8 +40,8 @@
num_channels, samples_per_channel, input_level_linear_scale);
for (int i = 0; i < 500; ++i) {
- const auto level = level_estimator.ComputeLevel(
- vectors_with_float_frame.float_frame_view());
+ const auto level =
+ level_estimator.ComputeLevel(vectors_with_float_frame.view());
// Give the estimator some time to ramp up.
if (i < 50) {
@@ -74,8 +74,8 @@
// Give the LevelEstimator plenty of time to ramp up and stabilize
float last_level = 0.f;
for (int i = 0; i < 500; ++i) {
- const auto level_envelope = level_estimator.ComputeLevel(
- vectors_with_float_frame.float_frame_view());
+ const auto level_envelope =
+ level_estimator.ComputeLevel(vectors_with_float_frame.view());
last_level = *level_envelope.rbegin();
}
@@ -87,8 +87,8 @@
DbfsToFloatS16(input_level_db - level_reduction_db);
int sub_frames_until_level_reduction = 0;
while (last_level > reduced_level_linear) {
- const auto level_envelope = level_estimator.ComputeLevel(
- vectors_with_zero_float_frame.float_frame_view());
+ const auto level_envelope =
+ level_estimator.ComputeLevel(vectors_with_zero_float_frame.view());
for (const auto& v : level_envelope) {
EXPECT_LT(v, last_level);
sub_frames_until_level_reduction++;
diff --git a/modules/audio_processing/agc2/limiter.cc b/modules/audio_processing/agc2/limiter.cc
index 4191298..270dc13 100644
--- a/modules/audio_processing/agc2/limiter.cc
+++ b/modules/audio_processing/agc2/limiter.cc
@@ -46,22 +46,20 @@
void ComputePerSampleSubframeFactors(
const std::array<float, kSubFramesInFrame + 1>& scaling_factors,
- int samples_per_channel,
- rtc::ArrayView<float> per_sample_scaling_factors) {
- const int num_subframes = scaling_factors.size() - 1;
- const int subframe_size =
- rtc::CheckedDivExact(samples_per_channel, num_subframes);
+ MonoView<float> per_sample_scaling_factors) {
+ const size_t num_subframes = scaling_factors.size() - 1;
+ const int subframe_size = rtc::CheckedDivExact(
+ SamplesPerChannel(per_sample_scaling_factors), num_subframes);
// Handle first sub-frame differently in case of attack.
const bool is_attack = scaling_factors[0] > scaling_factors[1];
if (is_attack) {
InterpolateFirstSubframe(
scaling_factors[0], scaling_factors[1],
- rtc::ArrayView<float>(
- per_sample_scaling_factors.subview(0, subframe_size)));
+ per_sample_scaling_factors.subview(0, subframe_size));
}
- for (int i = is_attack ? 1 : 0; i < num_subframes; ++i) {
+ for (size_t i = is_attack ? 1 : 0; i < num_subframes; ++i) {
const int subframe_start = i * subframe_size;
const float scaling_start = scaling_factors[i];
const float scaling_end = scaling_factors[i + 1];
@@ -74,12 +72,12 @@
}
void ScaleSamples(MonoView<const float> per_sample_scaling_factors,
- AudioFrameView<float> signal) {
+ DeinterleavedView<float> signal) {
const int samples_per_channel = signal.samples_per_channel();
RTC_DCHECK_EQ(samples_per_channel,
SamplesPerChannel(per_sample_scaling_factors));
- for (int i = 0; i < signal.num_channels(); ++i) {
- MonoView<float> channel = signal.channel(i);
+ for (size_t i = 0; i < signal.num_channels(); ++i) {
+ MonoView<float> channel = signal[i];
for (int j = 0; j < samples_per_channel; ++j) {
channel[j] = rtc::SafeClamp(channel[j] * per_sample_scaling_factors[j],
kMinFloatS16Value, kMaxFloatS16Value);
@@ -107,7 +105,10 @@
Limiter::~Limiter() = default;
-void Limiter::Process(AudioFrameView<float> signal) {
+void Limiter::Process(DeinterleavedView<float> signal) {
+ RTC_DCHECK_LE(signal.samples_per_channel(),
+ kMaximalNumberOfSamplesPerChannel);
+
const std::array<float, kSubFramesInFrame> level_estimate =
level_estimator_.ComputeLevel(signal);
@@ -118,13 +119,9 @@
return interp_gain_curve_.LookUpGainToApply(x);
});
- const int samples_per_channel = signal.samples_per_channel();
- RTC_DCHECK_LE(samples_per_channel, kMaximalNumberOfSamplesPerChannel);
-
- auto per_sample_scaling_factors =
- MonoView<float>(&per_sample_scaling_factors_[0], samples_per_channel);
- ComputePerSampleSubframeFactors(scaling_factors_, samples_per_channel,
- per_sample_scaling_factors);
+ MonoView<float> per_sample_scaling_factors(&per_sample_scaling_factors_[0],
+ signal.samples_per_channel());
+ ComputePerSampleSubframeFactors(scaling_factors_, per_sample_scaling_factors);
ScaleSamples(per_sample_scaling_factors, signal);
last_scaling_factor_ = scaling_factors_.back();
diff --git a/modules/audio_processing/agc2/limiter.h b/modules/audio_processing/agc2/limiter.h
index e2db138..f5244be 100644
--- a/modules/audio_processing/agc2/limiter.h
+++ b/modules/audio_processing/agc2/limiter.h
@@ -39,7 +39,12 @@
~Limiter();
// Applies limiter and hard-clipping to `signal`.
- void Process(AudioFrameView<float> signal);
+ void Process(DeinterleavedView<float> signal);
+
+ [[deprecated("Use DeinterleavedView version")]] void Process(
+ AudioFrameView<float> signal) {
+ return Process(signal.view());
+ }
InterpolatedGainCurve::Stats GetGainCurveStats() const;
// Supported values must be
diff --git a/modules/audio_processing/agc2/limiter_unittest.cc b/modules/audio_processing/agc2/limiter_unittest.cc
index 905b6b0..6c72e72 100644
--- a/modules/audio_processing/agc2/limiter_unittest.cc
+++ b/modules/audio_processing/agc2/limiter_unittest.cc
@@ -10,7 +10,8 @@
#include "modules/audio_processing/agc2/limiter.h"
-#include "api/audio/audio_frame.h"
+#include <algorithm>
+
#include "common_audio/include/audio_util.h"
#include "modules/audio_processing/agc2/agc2_common.h"
#include "modules/audio_processing/agc2/agc2_testing_common.h"
@@ -21,40 +22,40 @@
namespace webrtc {
TEST(Limiter, LimiterShouldConstructAndRun) {
- const size_t samples_per_channel = SampleRateToDefaultChannelSize(48000);
+ constexpr size_t kSamplesPerChannel = 480;
ApmDataDumper apm_data_dumper(0);
- Limiter limiter(&apm_data_dumper, samples_per_channel, "");
+ Limiter limiter(&apm_data_dumper, kSamplesPerChannel, "");
- VectorFloatFrame vectors_with_float_frame(1, samples_per_channel,
- kMaxAbsFloatS16Value);
- limiter.Process(vectors_with_float_frame.float_frame_view());
+ std::array<float, kSamplesPerChannel> buffer;
+ buffer.fill(kMaxAbsFloatS16Value);
+ limiter.Process(
+ DeinterleavedView<float>(buffer.data(), kSamplesPerChannel, 1));
}
TEST(Limiter, OutputVolumeAboveThreshold) {
- const size_t samples_per_channel = SampleRateToDefaultChannelSize(48000);
+ constexpr size_t kSamplesPerChannel = 480;
const float input_level =
(kMaxAbsFloatS16Value + DbfsToFloatS16(test::kLimiterMaxInputLevelDbFs)) /
2.f;
ApmDataDumper apm_data_dumper(0);
- Limiter limiter(&apm_data_dumper, samples_per_channel, "");
+ Limiter limiter(&apm_data_dumper, kSamplesPerChannel, "");
+
+ std::array<float, kSamplesPerChannel> buffer;
// Give the level estimator time to adapt.
for (int i = 0; i < 5; ++i) {
- VectorFloatFrame vectors_with_float_frame(1, samples_per_channel,
- input_level);
- limiter.Process(vectors_with_float_frame.float_frame_view());
+ std::fill(buffer.begin(), buffer.end(), input_level);
+ limiter.Process(
+ DeinterleavedView<float>(buffer.data(), kSamplesPerChannel, 1));
}
- VectorFloatFrame vectors_with_float_frame(1, samples_per_channel,
- input_level);
- limiter.Process(vectors_with_float_frame.float_frame_view());
- rtc::ArrayView<const float> channel =
- vectors_with_float_frame.float_frame_view().channel(0);
-
- for (const auto& sample : channel) {
- EXPECT_LT(0.9f * kMaxAbsFloatS16Value, sample);
+ std::fill(buffer.begin(), buffer.end(), input_level);
+ limiter.Process(
+ DeinterleavedView<float>(buffer.data(), kSamplesPerChannel, 1));
+ for (const auto& sample : buffer) {
+ ASSERT_LT(0.9f * kMaxAbsFloatS16Value, sample);
}
}
diff --git a/modules/audio_processing/gain_controller2.cc b/modules/audio_processing/gain_controller2.cc
index 181c401..9b19bc8 100644
--- a/modules/audio_processing/gain_controller2.cc
+++ b/modules/audio_processing/gain_controller2.cc
@@ -258,7 +258,7 @@
// computation in `limiter_`.
fixed_gain_applier_.ApplyGain(float_frame);
- limiter_.Process(float_frame);
+ limiter_.Process(float_frame.view());
// Periodically log limiter stats.
if (++calls_since_last_limiter_log_ == kLogLimiterStatsPeriodNumFrames) {