|  | /* | 
|  | *  Copyright (c) 2013 The WebRTC project authors. All Rights Reserved. | 
|  | * | 
|  | *  Use of this source code is governed by a BSD-style license | 
|  | *  that can be found in the LICENSE file in the root of the source | 
|  | *  tree. An additional intellectual property rights grant can be found | 
|  | *  in the file PATENTS.  All contributing project authors may | 
|  | *  be found in the AUTHORS file in the root of the source tree. | 
|  | */ | 
|  |  | 
|  | #include <memory> | 
|  | #include <string> | 
|  |  | 
|  | #include "modules/audio_coding/codecs/opus/opus_inst.h" | 
|  | #include "modules/audio_coding/codecs/opus/opus_interface.h" | 
|  | #include "modules/audio_coding/neteq/tools/audio_loop.h" | 
|  | #include "rtc_base/checks.h" | 
|  | #include "rtc_base/numerics/safe_conversions.h" | 
|  | #include "test/gtest.h" | 
|  | #include "test/testsupport/file_utils.h" | 
|  |  | 
|  | namespace webrtc { | 
|  |  | 
|  | namespace { | 
|  | // Equivalent to SDP params | 
|  | // {{"channel_mapping", "0,1,2,3"}, {"coupled_streams", "2"}}. | 
|  | constexpr unsigned char kQuadChannelMapping[] = {0, 1, 2, 3}; | 
|  | constexpr int kQuadTotalStreams = 2; | 
|  | constexpr int kQuadCoupledStreams = 2; | 
|  |  | 
|  | constexpr unsigned char kStereoChannelMapping[] = {0, 1}; | 
|  | constexpr int kStereoTotalStreams = 1; | 
|  | constexpr int kStereoCoupledStreams = 1; | 
|  |  | 
|  | constexpr unsigned char kMonoChannelMapping[] = {0}; | 
|  | constexpr int kMonoTotalStreams = 1; | 
|  | constexpr int kMonoCoupledStreams = 0; | 
|  |  | 
|  | void CreateSingleOrMultiStreamEncoder(WebRtcOpusEncInst** opus_encoder, | 
|  | int channels, | 
|  | int application, | 
|  | bool use_multistream, | 
|  | int encoder_sample_rate_hz) { | 
|  | EXPECT_TRUE(channels == 1 || channels == 2 || use_multistream); | 
|  | if (use_multistream) { | 
|  | EXPECT_EQ(encoder_sample_rate_hz, 48000); | 
|  | if (channels == 1) { | 
|  | EXPECT_EQ(0, WebRtcOpus_MultistreamEncoderCreate( | 
|  | opus_encoder, channels, application, kMonoTotalStreams, | 
|  | kMonoCoupledStreams, kMonoChannelMapping)); | 
|  | } else if (channels == 2) { | 
|  | EXPECT_EQ(0, WebRtcOpus_MultistreamEncoderCreate( | 
|  | opus_encoder, channels, application, kStereoTotalStreams, | 
|  | kStereoCoupledStreams, kStereoChannelMapping)); | 
|  | } else if (channels == 4) { | 
|  | EXPECT_EQ(0, WebRtcOpus_MultistreamEncoderCreate( | 
|  | opus_encoder, channels, application, kQuadTotalStreams, | 
|  | kQuadCoupledStreams, kQuadChannelMapping)); | 
|  | } else { | 
|  | EXPECT_TRUE(false) << channels; | 
|  | } | 
|  | } else { | 
|  | EXPECT_EQ(0, WebRtcOpus_EncoderCreate(opus_encoder, channels, application, | 
|  | encoder_sample_rate_hz)); | 
|  | } | 
|  | } | 
|  |  | 
|  | void CreateSingleOrMultiStreamDecoder(WebRtcOpusDecInst** opus_decoder, | 
|  | int channels, | 
|  | bool use_multistream, | 
|  | int decoder_sample_rate_hz) { | 
|  | EXPECT_TRUE(channels == 1 || channels == 2 || use_multistream); | 
|  | if (use_multistream) { | 
|  | EXPECT_EQ(decoder_sample_rate_hz, 48000); | 
|  | if (channels == 1) { | 
|  | EXPECT_EQ(0, WebRtcOpus_MultistreamDecoderCreate( | 
|  | opus_decoder, channels, kMonoTotalStreams, | 
|  | kMonoCoupledStreams, kMonoChannelMapping)); | 
|  | } else if (channels == 2) { | 
|  | EXPECT_EQ(0, WebRtcOpus_MultistreamDecoderCreate( | 
|  | opus_decoder, channels, kStereoTotalStreams, | 
|  | kStereoCoupledStreams, kStereoChannelMapping)); | 
|  | } else if (channels == 4) { | 
|  | EXPECT_EQ(0, WebRtcOpus_MultistreamDecoderCreate( | 
|  | opus_decoder, channels, kQuadTotalStreams, | 
|  | kQuadCoupledStreams, kQuadChannelMapping)); | 
|  | } else { | 
|  | EXPECT_TRUE(false) << channels; | 
|  | } | 
|  | } else { | 
|  | EXPECT_EQ(0, WebRtcOpus_DecoderCreate(opus_decoder, channels, | 
|  | decoder_sample_rate_hz)); | 
|  | } | 
|  | } | 
|  |  | 
|  | int SamplesPerChannel(int sample_rate_hz, int duration_ms) { | 
|  | const int samples_per_ms = rtc::CheckedDivExact(sample_rate_hz, 1000); | 
|  | return samples_per_ms * duration_ms; | 
|  | } | 
|  |  | 
|  | using test::AudioLoop; | 
|  | using ::testing::Combine; | 
|  | using ::testing::TestWithParam; | 
|  | using ::testing::Values; | 
|  |  | 
|  | // Maximum number of bytes in output bitstream. | 
|  | const size_t kMaxBytes = 2000; | 
|  |  | 
|  | class OpusTest | 
|  | : public TestWithParam<::testing::tuple<size_t, int, bool, int, int>> { | 
|  | protected: | 
|  | OpusTest() = default; | 
|  |  | 
|  | void TestDtxEffect(bool dtx, int block_length_ms); | 
|  |  | 
|  | void TestCbrEffect(bool dtx, int block_length_ms); | 
|  |  | 
|  | // Prepare |speech_data_| for encoding, read from a hard-coded file. | 
|  | // After preparation, |speech_data_.GetNextBlock()| returns a pointer to a | 
|  | // block of |block_length_ms| milliseconds. The data is looped every | 
|  | // |loop_length_ms| milliseconds. | 
|  | void PrepareSpeechData(int block_length_ms, int loop_length_ms); | 
|  |  | 
|  | int EncodeDecode(WebRtcOpusEncInst* encoder, | 
|  | rtc::ArrayView<const int16_t> input_audio, | 
|  | WebRtcOpusDecInst* decoder, | 
|  | int16_t* output_audio, | 
|  | int16_t* audio_type); | 
|  |  | 
|  | void SetMaxPlaybackRate(WebRtcOpusEncInst* encoder, | 
|  | opus_int32 expect, | 
|  | int32_t set); | 
|  |  | 
|  | void CheckAudioBounded(const int16_t* audio, | 
|  | size_t samples, | 
|  | size_t channels, | 
|  | uint16_t bound) const; | 
|  |  | 
|  | WebRtcOpusEncInst* opus_encoder_ = nullptr; | 
|  | WebRtcOpusDecInst* opus_decoder_ = nullptr; | 
|  | AudioLoop speech_data_; | 
|  | uint8_t bitstream_[kMaxBytes]; | 
|  | size_t encoded_bytes_ = 0; | 
|  | const size_t channels_{std::get<0>(GetParam())}; | 
|  | const int application_{std::get<1>(GetParam())}; | 
|  | const bool use_multistream_{std::get<2>(GetParam())}; | 
|  | const int encoder_sample_rate_hz_{std::get<3>(GetParam())}; | 
|  | const int decoder_sample_rate_hz_{std::get<4>(GetParam())}; | 
|  | }; | 
|  |  | 
|  | }  // namespace | 
|  |  | 
|  | // Singlestream: Try all combinations. | 
|  | INSTANTIATE_TEST_SUITE_P(Singlestream, | 
|  | OpusTest, | 
|  | testing::Combine(testing::Values(1, 2), | 
|  | testing::Values(0, 1), | 
|  | testing::Values(false), | 
|  | testing::Values(16000, 48000), | 
|  | testing::Values(16000, 48000))); | 
|  |  | 
|  | // Multistream: Some representative cases (only 48 kHz for now). | 
|  | INSTANTIATE_TEST_SUITE_P( | 
|  | Multistream, | 
|  | OpusTest, | 
|  | testing::Values(std::make_tuple(1, 0, true, 48000, 48000), | 
|  | std::make_tuple(2, 1, true, 48000, 48000), | 
|  | std::make_tuple(4, 0, true, 48000, 48000), | 
|  | std::make_tuple(4, 1, true, 48000, 48000))); | 
|  |  | 
|  | void OpusTest::PrepareSpeechData(int block_length_ms, int loop_length_ms) { | 
|  | std::map<int, std::string> channel_to_basename = { | 
|  | {1, "audio_coding/testfile32kHz"}, | 
|  | {2, "audio_coding/teststereo32kHz"}, | 
|  | {4, "audio_coding/speech_4_channels_48k_one_second"}}; | 
|  | std::map<int, std::string> channel_to_suffix = { | 
|  | {1, "pcm"}, {2, "pcm"}, {4, "wav"}}; | 
|  | const std::string file_name = webrtc::test::ResourcePath( | 
|  | channel_to_basename[channels_], channel_to_suffix[channels_]); | 
|  | if (loop_length_ms < block_length_ms) { | 
|  | loop_length_ms = block_length_ms; | 
|  | } | 
|  | const int sample_rate_khz = | 
|  | rtc::CheckedDivExact(encoder_sample_rate_hz_, 1000); | 
|  | EXPECT_TRUE(speech_data_.Init(file_name, | 
|  | loop_length_ms * sample_rate_khz * channels_, | 
|  | block_length_ms * sample_rate_khz * channels_)); | 
|  | } | 
|  |  | 
|  | void OpusTest::SetMaxPlaybackRate(WebRtcOpusEncInst* encoder, | 
|  | opus_int32 expect, | 
|  | int32_t set) { | 
|  | opus_int32 bandwidth; | 
|  | EXPECT_EQ(0, WebRtcOpus_SetMaxPlaybackRate(opus_encoder_, set)); | 
|  | EXPECT_EQ(0, WebRtcOpus_GetMaxPlaybackRate(opus_encoder_, &bandwidth)); | 
|  | EXPECT_EQ(expect, bandwidth); | 
|  | } | 
|  |  | 
|  | void OpusTest::CheckAudioBounded(const int16_t* audio, | 
|  | size_t samples, | 
|  | size_t channels, | 
|  | uint16_t bound) const { | 
|  | for (size_t i = 0; i < samples; ++i) { | 
|  | for (size_t c = 0; c < channels; ++c) { | 
|  | ASSERT_GE(audio[i * channels + c], -bound); | 
|  | ASSERT_LE(audio[i * channels + c], bound); | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | int OpusTest::EncodeDecode(WebRtcOpusEncInst* encoder, | 
|  | rtc::ArrayView<const int16_t> input_audio, | 
|  | WebRtcOpusDecInst* decoder, | 
|  | int16_t* output_audio, | 
|  | int16_t* audio_type) { | 
|  | const int input_samples_per_channel = | 
|  | rtc::CheckedDivExact(input_audio.size(), channels_); | 
|  | int encoded_bytes_int = | 
|  | WebRtcOpus_Encode(encoder, input_audio.data(), input_samples_per_channel, | 
|  | kMaxBytes, bitstream_); | 
|  | EXPECT_GE(encoded_bytes_int, 0); | 
|  | encoded_bytes_ = static_cast<size_t>(encoded_bytes_int); | 
|  | if (encoded_bytes_ != 0) { | 
|  | int est_len = WebRtcOpus_DurationEst(decoder, bitstream_, encoded_bytes_); | 
|  | int act_len = WebRtcOpus_Decode(decoder, bitstream_, encoded_bytes_, | 
|  | output_audio, audio_type); | 
|  | EXPECT_EQ(est_len, act_len); | 
|  | return act_len; | 
|  | } else { | 
|  | int total_dtx_len = 0; | 
|  | const int output_samples_per_channel = input_samples_per_channel * | 
|  | decoder_sample_rate_hz_ / | 
|  | encoder_sample_rate_hz_; | 
|  | while (total_dtx_len < output_samples_per_channel) { | 
|  | int est_len = WebRtcOpus_DurationEst(decoder, NULL, 0); | 
|  | int act_len = WebRtcOpus_Decode(decoder, NULL, 0, | 
|  | &output_audio[total_dtx_len * channels_], | 
|  | audio_type); | 
|  | EXPECT_EQ(est_len, act_len); | 
|  | total_dtx_len += act_len; | 
|  | } | 
|  | return total_dtx_len; | 
|  | } | 
|  | } | 
|  |  | 
|  | // Test if encoder/decoder can enter DTX mode properly and do not enter DTX when | 
|  | // they should not. This test is signal dependent. | 
|  | void OpusTest::TestDtxEffect(bool dtx, int block_length_ms) { | 
|  | PrepareSpeechData(block_length_ms, 2000); | 
|  | const size_t input_samples = | 
|  | rtc::CheckedDivExact(encoder_sample_rate_hz_, 1000) * block_length_ms; | 
|  | const size_t output_samples = | 
|  | rtc::CheckedDivExact(decoder_sample_rate_hz_, 1000) * block_length_ms; | 
|  |  | 
|  | // Create encoder memory. | 
|  | CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_, | 
|  | use_multistream_, encoder_sample_rate_hz_); | 
|  | CreateSingleOrMultiStreamDecoder(&opus_decoder_, channels_, use_multistream_, | 
|  | decoder_sample_rate_hz_); | 
|  |  | 
|  | // Set bitrate. | 
|  | EXPECT_EQ( | 
|  | 0, WebRtcOpus_SetBitRate(opus_encoder_, channels_ == 1 ? 32000 : 64000)); | 
|  |  | 
|  | // Set input audio as silence. | 
|  | std::vector<int16_t> silence(input_samples * channels_, 0); | 
|  |  | 
|  | // Setting DTX. | 
|  | EXPECT_EQ(0, dtx ? WebRtcOpus_EnableDtx(opus_encoder_) | 
|  | : WebRtcOpus_DisableDtx(opus_encoder_)); | 
|  |  | 
|  | int16_t audio_type; | 
|  | int16_t* output_data_decode = new int16_t[output_samples * channels_]; | 
|  |  | 
|  | for (int i = 0; i < 100; ++i) { | 
|  | EXPECT_EQ(output_samples, | 
|  | static_cast<size_t>(EncodeDecode( | 
|  | opus_encoder_, speech_data_.GetNextBlock(), opus_decoder_, | 
|  | output_data_decode, &audio_type))); | 
|  | // If not DTX, it should never enter DTX mode. If DTX, we do not care since | 
|  | // whether it enters DTX depends on the signal type. | 
|  | if (!dtx) { | 
|  | EXPECT_GT(encoded_bytes_, 1U); | 
|  | EXPECT_EQ(0, opus_encoder_->in_dtx_mode); | 
|  | EXPECT_EQ(0, opus_decoder_->in_dtx_mode); | 
|  | EXPECT_EQ(0, audio_type);  // Speech. | 
|  | } | 
|  | } | 
|  |  | 
|  | // We input some silent segments. In DTX mode, the encoder will stop sending. | 
|  | // However, DTX may happen after a while. | 
|  | for (int i = 0; i < 30; ++i) { | 
|  | EXPECT_EQ(output_samples, static_cast<size_t>(EncodeDecode( | 
|  | opus_encoder_, silence, opus_decoder_, | 
|  | output_data_decode, &audio_type))); | 
|  | if (!dtx) { | 
|  | EXPECT_GT(encoded_bytes_, 1U); | 
|  | EXPECT_EQ(0, opus_encoder_->in_dtx_mode); | 
|  | EXPECT_EQ(0, opus_decoder_->in_dtx_mode); | 
|  | EXPECT_EQ(0, audio_type);  // Speech. | 
|  | } else if (encoded_bytes_ == 1) { | 
|  | EXPECT_EQ(1, opus_encoder_->in_dtx_mode); | 
|  | EXPECT_EQ(1, opus_decoder_->in_dtx_mode); | 
|  | EXPECT_EQ(2, audio_type);  // Comfort noise. | 
|  | break; | 
|  | } | 
|  | } | 
|  |  | 
|  | // When Opus is in DTX, it wakes up in a regular basis. It sends two packets, | 
|  | // one with an arbitrary size and the other of 1-byte, then stops sending for | 
|  | // a certain number of frames. | 
|  |  | 
|  | // |max_dtx_frames| is the maximum number of frames Opus can stay in DTX. | 
|  | // TODO(kwiberg): Why does this number depend on the encoding sample rate? | 
|  | const int max_dtx_frames = | 
|  | (encoder_sample_rate_hz_ == 16000 ? 800 : 400) / block_length_ms + 1; | 
|  |  | 
|  | // We run |kRunTimeMs| milliseconds of pure silence. | 
|  | const int kRunTimeMs = 4500; | 
|  |  | 
|  | // We check that, after a |kCheckTimeMs| milliseconds (given that the CNG in | 
|  | // Opus needs time to adapt), the absolute values of DTX decoded signal are | 
|  | // bounded by |kOutputValueBound|. | 
|  | const int kCheckTimeMs = 4000; | 
|  |  | 
|  | #if defined(OPUS_FIXED_POINT) | 
|  | // Fixed-point Opus generates a random (comfort) noise, which has a less | 
|  | // predictable value bound than its floating-point Opus. This value depends on | 
|  | // input signal, and the time window for checking the output values (between | 
|  | // |kCheckTimeMs| and |kRunTimeMs|). | 
|  | const uint16_t kOutputValueBound = 30; | 
|  |  | 
|  | #else | 
|  | const uint16_t kOutputValueBound = 2; | 
|  | #endif | 
|  |  | 
|  | int time = 0; | 
|  | while (time < kRunTimeMs) { | 
|  | // DTX mode is maintained for maximum |max_dtx_frames| frames. | 
|  | int i = 0; | 
|  | for (; i < max_dtx_frames; ++i) { | 
|  | time += block_length_ms; | 
|  | EXPECT_EQ(output_samples, static_cast<size_t>(EncodeDecode( | 
|  | opus_encoder_, silence, opus_decoder_, | 
|  | output_data_decode, &audio_type))); | 
|  | if (dtx) { | 
|  | if (encoded_bytes_ > 1) | 
|  | break; | 
|  | EXPECT_EQ(0U, encoded_bytes_)  // Send 0 byte. | 
|  | << "Opus should have entered DTX mode."; | 
|  | EXPECT_EQ(1, opus_encoder_->in_dtx_mode); | 
|  | EXPECT_EQ(1, opus_decoder_->in_dtx_mode); | 
|  | EXPECT_EQ(2, audio_type);  // Comfort noise. | 
|  | if (time >= kCheckTimeMs) { | 
|  | CheckAudioBounded(output_data_decode, output_samples, channels_, | 
|  | kOutputValueBound); | 
|  | } | 
|  | } else { | 
|  | EXPECT_GT(encoded_bytes_, 1U); | 
|  | EXPECT_EQ(0, opus_encoder_->in_dtx_mode); | 
|  | EXPECT_EQ(0, opus_decoder_->in_dtx_mode); | 
|  | EXPECT_EQ(0, audio_type);  // Speech. | 
|  | } | 
|  | } | 
|  |  | 
|  | if (dtx) { | 
|  | // With DTX, Opus must stop transmission for some time. | 
|  | EXPECT_GT(i, 1); | 
|  | } | 
|  |  | 
|  | // We expect a normal payload. | 
|  | EXPECT_EQ(0, opus_encoder_->in_dtx_mode); | 
|  | EXPECT_EQ(0, opus_decoder_->in_dtx_mode); | 
|  | EXPECT_EQ(0, audio_type);  // Speech. | 
|  |  | 
|  | // Enters DTX again immediately. | 
|  | time += block_length_ms; | 
|  | EXPECT_EQ(output_samples, static_cast<size_t>(EncodeDecode( | 
|  | opus_encoder_, silence, opus_decoder_, | 
|  | output_data_decode, &audio_type))); | 
|  | if (dtx) { | 
|  | EXPECT_EQ(1U, encoded_bytes_);  // Send 1 byte. | 
|  | EXPECT_EQ(1, opus_encoder_->in_dtx_mode); | 
|  | EXPECT_EQ(1, opus_decoder_->in_dtx_mode); | 
|  | EXPECT_EQ(2, audio_type);  // Comfort noise. | 
|  | if (time >= kCheckTimeMs) { | 
|  | CheckAudioBounded(output_data_decode, output_samples, channels_, | 
|  | kOutputValueBound); | 
|  | } | 
|  | } else { | 
|  | EXPECT_GT(encoded_bytes_, 1U); | 
|  | EXPECT_EQ(0, opus_encoder_->in_dtx_mode); | 
|  | EXPECT_EQ(0, opus_decoder_->in_dtx_mode); | 
|  | EXPECT_EQ(0, audio_type);  // Speech. | 
|  | } | 
|  | } | 
|  |  | 
|  | silence[0] = 10000; | 
|  | if (dtx) { | 
|  | // Verify that encoder/decoder can jump out from DTX mode. | 
|  | EXPECT_EQ(output_samples, static_cast<size_t>(EncodeDecode( | 
|  | opus_encoder_, silence, opus_decoder_, | 
|  | output_data_decode, &audio_type))); | 
|  | EXPECT_GT(encoded_bytes_, 1U); | 
|  | EXPECT_EQ(0, opus_encoder_->in_dtx_mode); | 
|  | EXPECT_EQ(0, opus_decoder_->in_dtx_mode); | 
|  | EXPECT_EQ(0, audio_type);  // Speech. | 
|  | } | 
|  |  | 
|  | // Free memory. | 
|  | delete[] output_data_decode; | 
|  | EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_)); | 
|  | EXPECT_EQ(0, WebRtcOpus_DecoderFree(opus_decoder_)); | 
|  | } | 
|  |  | 
|  | // Test if CBR does what we expect. | 
|  | void OpusTest::TestCbrEffect(bool cbr, int block_length_ms) { | 
|  | PrepareSpeechData(block_length_ms, 2000); | 
|  | const size_t output_samples = | 
|  | rtc::CheckedDivExact(decoder_sample_rate_hz_, 1000) * block_length_ms; | 
|  |  | 
|  | int32_t max_pkt_size_diff = 0; | 
|  | int32_t prev_pkt_size = 0; | 
|  |  | 
|  | // Create encoder memory. | 
|  | CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_, | 
|  | use_multistream_, encoder_sample_rate_hz_); | 
|  | CreateSingleOrMultiStreamDecoder(&opus_decoder_, channels_, use_multistream_, | 
|  | decoder_sample_rate_hz_); | 
|  |  | 
|  | // Set bitrate. | 
|  | EXPECT_EQ( | 
|  | 0, WebRtcOpus_SetBitRate(opus_encoder_, channels_ == 1 ? 32000 : 64000)); | 
|  |  | 
|  | // Setting CBR. | 
|  | EXPECT_EQ(0, cbr ? WebRtcOpus_EnableCbr(opus_encoder_) | 
|  | : WebRtcOpus_DisableCbr(opus_encoder_)); | 
|  |  | 
|  | int16_t audio_type; | 
|  | std::vector<int16_t> audio_out(output_samples * channels_); | 
|  | for (int i = 0; i < 100; ++i) { | 
|  | EXPECT_EQ(output_samples, | 
|  | static_cast<size_t>( | 
|  | EncodeDecode(opus_encoder_, speech_data_.GetNextBlock(), | 
|  | opus_decoder_, audio_out.data(), &audio_type))); | 
|  |  | 
|  | if (prev_pkt_size > 0) { | 
|  | int32_t diff = std::abs((int32_t)encoded_bytes_ - prev_pkt_size); | 
|  | max_pkt_size_diff = std::max(max_pkt_size_diff, diff); | 
|  | } | 
|  | prev_pkt_size = rtc::checked_cast<int32_t>(encoded_bytes_); | 
|  | } | 
|  |  | 
|  | if (cbr) { | 
|  | EXPECT_EQ(max_pkt_size_diff, 0); | 
|  | } else { | 
|  | EXPECT_GT(max_pkt_size_diff, 0); | 
|  | } | 
|  |  | 
|  | // Free memory. | 
|  | EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_)); | 
|  | EXPECT_EQ(0, WebRtcOpus_DecoderFree(opus_decoder_)); | 
|  | } | 
|  |  | 
|  | // Test failing Create. | 
|  | TEST(OpusTest, OpusCreateFail) { | 
|  | WebRtcOpusEncInst* opus_encoder; | 
|  | WebRtcOpusDecInst* opus_decoder; | 
|  |  | 
|  | // Test to see that an invalid pointer is caught. | 
|  | EXPECT_EQ(-1, WebRtcOpus_EncoderCreate(NULL, 1, 0, 48000)); | 
|  | // Invalid channel number. | 
|  | EXPECT_EQ(-1, WebRtcOpus_EncoderCreate(&opus_encoder, 257, 0, 48000)); | 
|  | // Invalid applciation mode. | 
|  | EXPECT_EQ(-1, WebRtcOpus_EncoderCreate(&opus_encoder, 1, 2, 48000)); | 
|  | // Invalid sample rate. | 
|  | EXPECT_EQ(-1, WebRtcOpus_EncoderCreate(&opus_encoder, 1, 0, 12345)); | 
|  |  | 
|  | EXPECT_EQ(-1, WebRtcOpus_DecoderCreate(NULL, 1, 48000)); | 
|  | // Invalid channel number. | 
|  | EXPECT_EQ(-1, WebRtcOpus_DecoderCreate(&opus_decoder, 257, 48000)); | 
|  | // Invalid sample rate. | 
|  | EXPECT_EQ(-1, WebRtcOpus_DecoderCreate(&opus_decoder, 1, 12345)); | 
|  | } | 
|  |  | 
|  | // Test failing Free. | 
|  | TEST(OpusTest, OpusFreeFail) { | 
|  | // Test to see that an invalid pointer is caught. | 
|  | EXPECT_EQ(-1, WebRtcOpus_EncoderFree(NULL)); | 
|  | EXPECT_EQ(-1, WebRtcOpus_DecoderFree(NULL)); | 
|  | } | 
|  |  | 
|  | // Test normal Create and Free. | 
|  | TEST_P(OpusTest, OpusCreateFree) { | 
|  | CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_, | 
|  | use_multistream_, encoder_sample_rate_hz_); | 
|  | CreateSingleOrMultiStreamDecoder(&opus_decoder_, channels_, use_multistream_, | 
|  | decoder_sample_rate_hz_); | 
|  | EXPECT_TRUE(opus_encoder_ != NULL); | 
|  | EXPECT_TRUE(opus_decoder_ != NULL); | 
|  | // Free encoder and decoder memory. | 
|  | EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_)); | 
|  | EXPECT_EQ(0, WebRtcOpus_DecoderFree(opus_decoder_)); | 
|  | } | 
|  |  | 
|  | #define ENCODER_CTL(inst, vargs)               \ | 
|  | inst->encoder                                \ | 
|  | ? opus_encoder_ctl(inst->encoder, vargs) \ | 
|  | : opus_multistream_encoder_ctl(inst->multistream_encoder, vargs) | 
|  |  | 
|  | TEST_P(OpusTest, OpusEncodeDecode) { | 
|  | PrepareSpeechData(20, 20); | 
|  |  | 
|  | // Create encoder memory. | 
|  | CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_, | 
|  | use_multistream_, encoder_sample_rate_hz_); | 
|  | CreateSingleOrMultiStreamDecoder(&opus_decoder_, channels_, use_multistream_, | 
|  | decoder_sample_rate_hz_); | 
|  |  | 
|  | // Set bitrate. | 
|  | EXPECT_EQ( | 
|  | 0, WebRtcOpus_SetBitRate(opus_encoder_, channels_ == 1 ? 32000 : 64000)); | 
|  |  | 
|  | // Check number of channels for decoder. | 
|  | EXPECT_EQ(channels_, WebRtcOpus_DecoderChannels(opus_decoder_)); | 
|  |  | 
|  | // Check application mode. | 
|  | opus_int32 app; | 
|  | ENCODER_CTL(opus_encoder_, OPUS_GET_APPLICATION(&app)); | 
|  | EXPECT_EQ(application_ == 0 ? OPUS_APPLICATION_VOIP : OPUS_APPLICATION_AUDIO, | 
|  | app); | 
|  |  | 
|  | // Encode & decode. | 
|  | int16_t audio_type; | 
|  | const int decode_samples_per_channel = | 
|  | SamplesPerChannel(decoder_sample_rate_hz_, /*ms=*/20); | 
|  | int16_t* output_data_decode = | 
|  | new int16_t[decode_samples_per_channel * channels_]; | 
|  | EXPECT_EQ(decode_samples_per_channel, | 
|  | EncodeDecode(opus_encoder_, speech_data_.GetNextBlock(), | 
|  | opus_decoder_, output_data_decode, &audio_type)); | 
|  |  | 
|  | // Free memory. | 
|  | delete[] output_data_decode; | 
|  | EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_)); | 
|  | EXPECT_EQ(0, WebRtcOpus_DecoderFree(opus_decoder_)); | 
|  | } | 
|  |  | 
|  | TEST_P(OpusTest, OpusSetBitRate) { | 
|  | // Test without creating encoder memory. | 
|  | EXPECT_EQ(-1, WebRtcOpus_SetBitRate(opus_encoder_, 60000)); | 
|  |  | 
|  | // Create encoder memory, try with different bitrates. | 
|  | CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_, | 
|  | use_multistream_, encoder_sample_rate_hz_); | 
|  | EXPECT_EQ(0, WebRtcOpus_SetBitRate(opus_encoder_, 30000)); | 
|  | EXPECT_EQ(0, WebRtcOpus_SetBitRate(opus_encoder_, 60000)); | 
|  | EXPECT_EQ(0, WebRtcOpus_SetBitRate(opus_encoder_, 300000)); | 
|  | EXPECT_EQ(0, WebRtcOpus_SetBitRate(opus_encoder_, 600000)); | 
|  |  | 
|  | // Free memory. | 
|  | EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_)); | 
|  | } | 
|  |  | 
|  | TEST_P(OpusTest, OpusSetComplexity) { | 
|  | // Test without creating encoder memory. | 
|  | EXPECT_EQ(-1, WebRtcOpus_SetComplexity(opus_encoder_, 9)); | 
|  |  | 
|  | // Create encoder memory, try with different complexities. | 
|  | CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_, | 
|  | use_multistream_, encoder_sample_rate_hz_); | 
|  |  | 
|  | EXPECT_EQ(0, WebRtcOpus_SetComplexity(opus_encoder_, 0)); | 
|  | EXPECT_EQ(0, WebRtcOpus_SetComplexity(opus_encoder_, 10)); | 
|  | EXPECT_EQ(-1, WebRtcOpus_SetComplexity(opus_encoder_, 11)); | 
|  |  | 
|  | // Free memory. | 
|  | EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_)); | 
|  | } | 
|  |  | 
|  | TEST_P(OpusTest, OpusSetBandwidth) { | 
|  | if (channels_ > 2) { | 
|  | // TODO(webrtc:10217): investigate why multi-stream Opus reports | 
|  | // narrowband when it's configured with FULLBAND. | 
|  | return; | 
|  | } | 
|  | PrepareSpeechData(20, 20); | 
|  |  | 
|  | int16_t audio_type; | 
|  | const int decode_samples_per_channel = | 
|  | SamplesPerChannel(decoder_sample_rate_hz_, /*ms=*/20); | 
|  | std::unique_ptr<int16_t[]> output_data_decode( | 
|  | new int16_t[decode_samples_per_channel * channels_]()); | 
|  |  | 
|  | // Test without creating encoder memory. | 
|  | EXPECT_EQ(-1, | 
|  | WebRtcOpus_SetBandwidth(opus_encoder_, OPUS_BANDWIDTH_NARROWBAND)); | 
|  | EXPECT_EQ(-1, WebRtcOpus_GetBandwidth(opus_encoder_)); | 
|  |  | 
|  | // Create encoder memory, try with different bandwidths. | 
|  | CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_, | 
|  | use_multistream_, encoder_sample_rate_hz_); | 
|  | CreateSingleOrMultiStreamDecoder(&opus_decoder_, channels_, use_multistream_, | 
|  | decoder_sample_rate_hz_); | 
|  |  | 
|  | EXPECT_EQ(-1, WebRtcOpus_SetBandwidth(opus_encoder_, | 
|  | OPUS_BANDWIDTH_NARROWBAND - 1)); | 
|  | EXPECT_EQ(0, | 
|  | WebRtcOpus_SetBandwidth(opus_encoder_, OPUS_BANDWIDTH_NARROWBAND)); | 
|  | EncodeDecode(opus_encoder_, speech_data_.GetNextBlock(), opus_decoder_, | 
|  | output_data_decode.get(), &audio_type); | 
|  | EXPECT_EQ(OPUS_BANDWIDTH_NARROWBAND, WebRtcOpus_GetBandwidth(opus_encoder_)); | 
|  | EXPECT_EQ(0, WebRtcOpus_SetBandwidth(opus_encoder_, OPUS_BANDWIDTH_FULLBAND)); | 
|  | EncodeDecode(opus_encoder_, speech_data_.GetNextBlock(), opus_decoder_, | 
|  | output_data_decode.get(), &audio_type); | 
|  | EXPECT_EQ(encoder_sample_rate_hz_ == 16000 ? OPUS_BANDWIDTH_WIDEBAND | 
|  | : OPUS_BANDWIDTH_FULLBAND, | 
|  | WebRtcOpus_GetBandwidth(opus_encoder_)); | 
|  | EXPECT_EQ( | 
|  | -1, WebRtcOpus_SetBandwidth(opus_encoder_, OPUS_BANDWIDTH_FULLBAND + 1)); | 
|  | EncodeDecode(opus_encoder_, speech_data_.GetNextBlock(), opus_decoder_, | 
|  | output_data_decode.get(), &audio_type); | 
|  | EXPECT_EQ(encoder_sample_rate_hz_ == 16000 ? OPUS_BANDWIDTH_WIDEBAND | 
|  | : OPUS_BANDWIDTH_FULLBAND, | 
|  | WebRtcOpus_GetBandwidth(opus_encoder_)); | 
|  |  | 
|  | // Free memory. | 
|  | EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_)); | 
|  | EXPECT_EQ(0, WebRtcOpus_DecoderFree(opus_decoder_)); | 
|  | } | 
|  |  | 
|  | TEST_P(OpusTest, OpusForceChannels) { | 
|  | // Test without creating encoder memory. | 
|  | EXPECT_EQ(-1, WebRtcOpus_SetForceChannels(opus_encoder_, 1)); | 
|  |  | 
|  | CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_, | 
|  | use_multistream_, encoder_sample_rate_hz_); | 
|  | ASSERT_NE(nullptr, opus_encoder_); | 
|  |  | 
|  | if (channels_ >= 2) { | 
|  | EXPECT_EQ(-1, WebRtcOpus_SetForceChannels(opus_encoder_, 3)); | 
|  | EXPECT_EQ(0, WebRtcOpus_SetForceChannels(opus_encoder_, 2)); | 
|  | EXPECT_EQ(0, WebRtcOpus_SetForceChannels(opus_encoder_, 1)); | 
|  | EXPECT_EQ(0, WebRtcOpus_SetForceChannels(opus_encoder_, 0)); | 
|  | } else { | 
|  | EXPECT_EQ(-1, WebRtcOpus_SetForceChannels(opus_encoder_, 2)); | 
|  | EXPECT_EQ(0, WebRtcOpus_SetForceChannels(opus_encoder_, 1)); | 
|  | EXPECT_EQ(0, WebRtcOpus_SetForceChannels(opus_encoder_, 0)); | 
|  | } | 
|  |  | 
|  | EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_)); | 
|  | } | 
|  |  | 
|  | // Encode and decode one frame, initialize the decoder and | 
|  | // decode once more. | 
|  | TEST_P(OpusTest, OpusDecodeInit) { | 
|  | PrepareSpeechData(20, 20); | 
|  |  | 
|  | // Create encoder memory. | 
|  | CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_, | 
|  | use_multistream_, encoder_sample_rate_hz_); | 
|  | CreateSingleOrMultiStreamDecoder(&opus_decoder_, channels_, use_multistream_, | 
|  | decoder_sample_rate_hz_); | 
|  |  | 
|  | // Encode & decode. | 
|  | int16_t audio_type; | 
|  | const int decode_samples_per_channel = | 
|  | SamplesPerChannel(decoder_sample_rate_hz_, /*ms=*/20); | 
|  | int16_t* output_data_decode = | 
|  | new int16_t[decode_samples_per_channel * channels_]; | 
|  | EXPECT_EQ(decode_samples_per_channel, | 
|  | EncodeDecode(opus_encoder_, speech_data_.GetNextBlock(), | 
|  | opus_decoder_, output_data_decode, &audio_type)); | 
|  |  | 
|  | WebRtcOpus_DecoderInit(opus_decoder_); | 
|  |  | 
|  | EXPECT_EQ(decode_samples_per_channel, | 
|  | WebRtcOpus_Decode(opus_decoder_, bitstream_, encoded_bytes_, | 
|  | output_data_decode, &audio_type)); | 
|  |  | 
|  | // Free memory. | 
|  | delete[] output_data_decode; | 
|  | EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_)); | 
|  | EXPECT_EQ(0, WebRtcOpus_DecoderFree(opus_decoder_)); | 
|  | } | 
|  |  | 
|  | TEST_P(OpusTest, OpusEnableDisableFec) { | 
|  | // Test without creating encoder memory. | 
|  | EXPECT_EQ(-1, WebRtcOpus_EnableFec(opus_encoder_)); | 
|  | EXPECT_EQ(-1, WebRtcOpus_DisableFec(opus_encoder_)); | 
|  |  | 
|  | // Create encoder memory. | 
|  | CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_, | 
|  | use_multistream_, encoder_sample_rate_hz_); | 
|  |  | 
|  | EXPECT_EQ(0, WebRtcOpus_EnableFec(opus_encoder_)); | 
|  | EXPECT_EQ(0, WebRtcOpus_DisableFec(opus_encoder_)); | 
|  |  | 
|  | // Free memory. | 
|  | EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_)); | 
|  | } | 
|  |  | 
|  | TEST_P(OpusTest, OpusEnableDisableDtx) { | 
|  | // Test without creating encoder memory. | 
|  | EXPECT_EQ(-1, WebRtcOpus_EnableDtx(opus_encoder_)); | 
|  | EXPECT_EQ(-1, WebRtcOpus_DisableDtx(opus_encoder_)); | 
|  |  | 
|  | // Create encoder memory. | 
|  | CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_, | 
|  | use_multistream_, encoder_sample_rate_hz_); | 
|  |  | 
|  | opus_int32 dtx; | 
|  |  | 
|  | // DTX is off by default. | 
|  | ENCODER_CTL(opus_encoder_, OPUS_GET_DTX(&dtx)); | 
|  | EXPECT_EQ(0, dtx); | 
|  |  | 
|  | // Test to enable DTX. | 
|  | EXPECT_EQ(0, WebRtcOpus_EnableDtx(opus_encoder_)); | 
|  | ENCODER_CTL(opus_encoder_, OPUS_GET_DTX(&dtx)); | 
|  | EXPECT_EQ(1, dtx); | 
|  |  | 
|  | // Test to disable DTX. | 
|  | EXPECT_EQ(0, WebRtcOpus_DisableDtx(opus_encoder_)); | 
|  | ENCODER_CTL(opus_encoder_, OPUS_GET_DTX(&dtx)); | 
|  | EXPECT_EQ(0, dtx); | 
|  |  | 
|  | // Free memory. | 
|  | EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_)); | 
|  | } | 
|  |  | 
|  | TEST_P(OpusTest, OpusDtxOff) { | 
|  | TestDtxEffect(false, 10); | 
|  | TestDtxEffect(false, 20); | 
|  | TestDtxEffect(false, 40); | 
|  | } | 
|  |  | 
|  | TEST_P(OpusTest, OpusDtxOn) { | 
|  | if (channels_ > 2) { | 
|  | // TODO(webrtc:10218): adapt the test to the sizes and order of multi-stream | 
|  | // DTX packets. | 
|  | return; | 
|  | } | 
|  | TestDtxEffect(true, 10); | 
|  | TestDtxEffect(true, 20); | 
|  | TestDtxEffect(true, 40); | 
|  | } | 
|  |  | 
|  | TEST_P(OpusTest, OpusCbrOff) { | 
|  | TestCbrEffect(false, 10); | 
|  | TestCbrEffect(false, 20); | 
|  | TestCbrEffect(false, 40); | 
|  | } | 
|  |  | 
|  | TEST_P(OpusTest, OpusCbrOn) { | 
|  | TestCbrEffect(true, 10); | 
|  | TestCbrEffect(true, 20); | 
|  | TestCbrEffect(true, 40); | 
|  | } | 
|  |  | 
|  | TEST_P(OpusTest, OpusSetPacketLossRate) { | 
|  | // Test without creating encoder memory. | 
|  | EXPECT_EQ(-1, WebRtcOpus_SetPacketLossRate(opus_encoder_, 50)); | 
|  |  | 
|  | // Create encoder memory. | 
|  | CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_, | 
|  | use_multistream_, encoder_sample_rate_hz_); | 
|  |  | 
|  | EXPECT_EQ(0, WebRtcOpus_SetPacketLossRate(opus_encoder_, 50)); | 
|  | EXPECT_EQ(-1, WebRtcOpus_SetPacketLossRate(opus_encoder_, -1)); | 
|  | EXPECT_EQ(-1, WebRtcOpus_SetPacketLossRate(opus_encoder_, 101)); | 
|  |  | 
|  | // Free memory. | 
|  | EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_)); | 
|  | } | 
|  |  | 
|  | TEST_P(OpusTest, OpusSetMaxPlaybackRate) { | 
|  | // Test without creating encoder memory. | 
|  | EXPECT_EQ(-1, WebRtcOpus_SetMaxPlaybackRate(opus_encoder_, 20000)); | 
|  |  | 
|  | // Create encoder memory. | 
|  | CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_, | 
|  | use_multistream_, encoder_sample_rate_hz_); | 
|  |  | 
|  | SetMaxPlaybackRate(opus_encoder_, OPUS_BANDWIDTH_FULLBAND, 48000); | 
|  | SetMaxPlaybackRate(opus_encoder_, OPUS_BANDWIDTH_FULLBAND, 24001); | 
|  | SetMaxPlaybackRate(opus_encoder_, OPUS_BANDWIDTH_SUPERWIDEBAND, 24000); | 
|  | SetMaxPlaybackRate(opus_encoder_, OPUS_BANDWIDTH_SUPERWIDEBAND, 16001); | 
|  | SetMaxPlaybackRate(opus_encoder_, OPUS_BANDWIDTH_WIDEBAND, 16000); | 
|  | SetMaxPlaybackRate(opus_encoder_, OPUS_BANDWIDTH_WIDEBAND, 12001); | 
|  | SetMaxPlaybackRate(opus_encoder_, OPUS_BANDWIDTH_MEDIUMBAND, 12000); | 
|  | SetMaxPlaybackRate(opus_encoder_, OPUS_BANDWIDTH_MEDIUMBAND, 8001); | 
|  | SetMaxPlaybackRate(opus_encoder_, OPUS_BANDWIDTH_NARROWBAND, 8000); | 
|  | SetMaxPlaybackRate(opus_encoder_, OPUS_BANDWIDTH_NARROWBAND, 4000); | 
|  |  | 
|  | // Free memory. | 
|  | EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_)); | 
|  | } | 
|  |  | 
|  | // Test PLC. | 
|  | TEST_P(OpusTest, OpusDecodePlc) { | 
|  | PrepareSpeechData(20, 20); | 
|  |  | 
|  | // Create encoder memory. | 
|  | CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_, | 
|  | use_multistream_, encoder_sample_rate_hz_); | 
|  | CreateSingleOrMultiStreamDecoder(&opus_decoder_, channels_, use_multistream_, | 
|  | decoder_sample_rate_hz_); | 
|  |  | 
|  | // Set bitrate. | 
|  | EXPECT_EQ( | 
|  | 0, WebRtcOpus_SetBitRate(opus_encoder_, channels_ == 1 ? 32000 : 64000)); | 
|  |  | 
|  | // Check number of channels for decoder. | 
|  | EXPECT_EQ(channels_, WebRtcOpus_DecoderChannels(opus_decoder_)); | 
|  |  | 
|  | // Encode & decode. | 
|  | int16_t audio_type; | 
|  | const int decode_samples_per_channel = | 
|  | SamplesPerChannel(decoder_sample_rate_hz_, /*ms=*/20); | 
|  | int16_t* output_data_decode = | 
|  | new int16_t[decode_samples_per_channel * channels_]; | 
|  | EXPECT_EQ(decode_samples_per_channel, | 
|  | EncodeDecode(opus_encoder_, speech_data_.GetNextBlock(), | 
|  | opus_decoder_, output_data_decode, &audio_type)); | 
|  |  | 
|  | // Call decoder PLC. | 
|  | constexpr int kPlcDurationMs = 10; | 
|  | const int plc_samples = decoder_sample_rate_hz_ * kPlcDurationMs / 1000; | 
|  | int16_t* plc_buffer = new int16_t[plc_samples * channels_]; | 
|  | EXPECT_EQ(plc_samples, | 
|  | WebRtcOpus_Decode(opus_decoder_, NULL, 0, plc_buffer, &audio_type)); | 
|  |  | 
|  | // Free memory. | 
|  | delete[] plc_buffer; | 
|  | delete[] output_data_decode; | 
|  | EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_)); | 
|  | EXPECT_EQ(0, WebRtcOpus_DecoderFree(opus_decoder_)); | 
|  | } | 
|  |  | 
|  | // Duration estimation. | 
|  | TEST_P(OpusTest, OpusDurationEstimation) { | 
|  | PrepareSpeechData(20, 20); | 
|  |  | 
|  | // Create. | 
|  | CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_, | 
|  | use_multistream_, encoder_sample_rate_hz_); | 
|  | CreateSingleOrMultiStreamDecoder(&opus_decoder_, channels_, use_multistream_, | 
|  | decoder_sample_rate_hz_); | 
|  |  | 
|  | // 10 ms. We use only first 10 ms of a 20 ms block. | 
|  | auto speech_block = speech_data_.GetNextBlock(); | 
|  | int encoded_bytes_int = WebRtcOpus_Encode( | 
|  | opus_encoder_, speech_block.data(), | 
|  | rtc::CheckedDivExact(speech_block.size(), 2 * channels_), kMaxBytes, | 
|  | bitstream_); | 
|  | EXPECT_GE(encoded_bytes_int, 0); | 
|  | EXPECT_EQ(SamplesPerChannel(decoder_sample_rate_hz_, /*ms=*/10), | 
|  | WebRtcOpus_DurationEst(opus_decoder_, bitstream_, | 
|  | static_cast<size_t>(encoded_bytes_int))); | 
|  |  | 
|  | // 20 ms | 
|  | speech_block = speech_data_.GetNextBlock(); | 
|  | encoded_bytes_int = | 
|  | WebRtcOpus_Encode(opus_encoder_, speech_block.data(), | 
|  | rtc::CheckedDivExact(speech_block.size(), channels_), | 
|  | kMaxBytes, bitstream_); | 
|  | EXPECT_GE(encoded_bytes_int, 0); | 
|  | EXPECT_EQ(SamplesPerChannel(decoder_sample_rate_hz_, /*ms=*/20), | 
|  | WebRtcOpus_DurationEst(opus_decoder_, bitstream_, | 
|  | static_cast<size_t>(encoded_bytes_int))); | 
|  |  | 
|  | // Free memory. | 
|  | EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_)); | 
|  | EXPECT_EQ(0, WebRtcOpus_DecoderFree(opus_decoder_)); | 
|  | } | 
|  |  | 
|  | TEST_P(OpusTest, OpusDecodeRepacketized) { | 
|  | if (channels_ > 2) { | 
|  | // As per the Opus documentation | 
|  | // https://mf4.xiph.org/jenkins/view/opus/job/opus/ws/doc/html/group__opus__repacketizer.html#details, | 
|  | // multiple streams are not supported. | 
|  | return; | 
|  | } | 
|  | constexpr size_t kPackets = 6; | 
|  |  | 
|  | PrepareSpeechData(20, 20 * kPackets); | 
|  |  | 
|  | // Create encoder memory. | 
|  | CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_, | 
|  | use_multistream_, encoder_sample_rate_hz_); | 
|  | ASSERT_NE(nullptr, opus_encoder_); | 
|  | CreateSingleOrMultiStreamDecoder(&opus_decoder_, channels_, use_multistream_, | 
|  | decoder_sample_rate_hz_); | 
|  | ASSERT_NE(nullptr, opus_decoder_); | 
|  |  | 
|  | // Set bitrate. | 
|  | EXPECT_EQ( | 
|  | 0, WebRtcOpus_SetBitRate(opus_encoder_, channels_ == 1 ? 32000 : 64000)); | 
|  |  | 
|  | // Check number of channels for decoder. | 
|  | EXPECT_EQ(channels_, WebRtcOpus_DecoderChannels(opus_decoder_)); | 
|  |  | 
|  | // Encode & decode. | 
|  | int16_t audio_type; | 
|  | const int decode_samples_per_channel = | 
|  | SamplesPerChannel(decoder_sample_rate_hz_, /*ms=*/20); | 
|  | std::unique_ptr<int16_t[]> output_data_decode( | 
|  | new int16_t[kPackets * decode_samples_per_channel * channels_]); | 
|  | OpusRepacketizer* rp = opus_repacketizer_create(); | 
|  |  | 
|  | size_t num_packets = 0; | 
|  | constexpr size_t kMaxCycles = 100; | 
|  | for (size_t idx = 0; idx < kMaxCycles; ++idx) { | 
|  | auto speech_block = speech_data_.GetNextBlock(); | 
|  | encoded_bytes_ = | 
|  | WebRtcOpus_Encode(opus_encoder_, speech_block.data(), | 
|  | rtc::CheckedDivExact(speech_block.size(), channels_), | 
|  | kMaxBytes, bitstream_); | 
|  | if (opus_repacketizer_cat(rp, bitstream_, | 
|  | rtc::checked_cast<opus_int32>(encoded_bytes_)) == | 
|  | OPUS_OK) { | 
|  | ++num_packets; | 
|  | if (num_packets == kPackets) { | 
|  | break; | 
|  | } | 
|  | } else { | 
|  | // Opus repacketizer cannot guarantee a success. We try again if it fails. | 
|  | opus_repacketizer_init(rp); | 
|  | num_packets = 0; | 
|  | } | 
|  | } | 
|  | EXPECT_EQ(kPackets, num_packets); | 
|  |  | 
|  | encoded_bytes_ = opus_repacketizer_out(rp, bitstream_, kMaxBytes); | 
|  |  | 
|  | EXPECT_EQ(decode_samples_per_channel * kPackets, | 
|  | static_cast<size_t>(WebRtcOpus_DurationEst( | 
|  | opus_decoder_, bitstream_, encoded_bytes_))); | 
|  |  | 
|  | EXPECT_EQ(decode_samples_per_channel * kPackets, | 
|  | static_cast<size_t>( | 
|  | WebRtcOpus_Decode(opus_decoder_, bitstream_, encoded_bytes_, | 
|  | output_data_decode.get(), &audio_type))); | 
|  |  | 
|  | // Free memory. | 
|  | opus_repacketizer_destroy(rp); | 
|  | EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_)); | 
|  | EXPECT_EQ(0, WebRtcOpus_DecoderFree(opus_decoder_)); | 
|  | } | 
|  |  | 
|  | TEST(OpusVadTest, CeltUnknownStatus) { | 
|  | const uint8_t celt[] = {0x80}; | 
|  | EXPECT_EQ(WebRtcOpus_PacketHasVoiceActivity(celt, 1), -1); | 
|  | } | 
|  |  | 
|  | TEST(OpusVadTest, Mono20msVadSet) { | 
|  | uint8_t silk20msMonoVad[] = {0x78, 0x80}; | 
|  | EXPECT_TRUE(WebRtcOpus_PacketHasVoiceActivity(silk20msMonoVad, 2)); | 
|  | } | 
|  |  | 
|  | TEST(OpusVadTest, Mono20MsVadUnset) { | 
|  | uint8_t silk20msMonoSilence[] = {0x78, 0x00}; | 
|  | EXPECT_FALSE(WebRtcOpus_PacketHasVoiceActivity(silk20msMonoSilence, 2)); | 
|  | } | 
|  |  | 
|  | TEST(OpusVadTest, Stereo20MsVadOnSideChannel) { | 
|  | uint8_t silk20msStereoVadSideChannel[] = {0x78 | 0x04, 0x20}; | 
|  | EXPECT_TRUE( | 
|  | WebRtcOpus_PacketHasVoiceActivity(silk20msStereoVadSideChannel, 2)); | 
|  | } | 
|  |  | 
|  | TEST(OpusVadTest, TwoOpusMonoFramesVadOnSecond) { | 
|  | uint8_t twoMonoFrames[] = {0x78 | 0x1, 0x00, 0x80}; | 
|  | EXPECT_TRUE(WebRtcOpus_PacketHasVoiceActivity(twoMonoFrames, 3)); | 
|  | } | 
|  |  | 
|  | TEST(OpusVadTest, DtxEmptyPacket) { | 
|  | const uint8_t dtx[] = {0x78}; | 
|  | EXPECT_FALSE(WebRtcOpus_PacketHasVoiceActivity(dtx, 1)); | 
|  | } | 
|  |  | 
|  | TEST(OpusVadTest, DtxBackgroundNoisePacket) { | 
|  | // DTX sends a frame coding background noise every 20 packets: | 
|  | //   https://tools.ietf.org/html/rfc6716#section-2.1.9 | 
|  | // The packet below represents such a frame and was captured using | 
|  | // Wireshark while disabling encryption. | 
|  | const uint8_t dtx[] = {0x78, 0x07, 0xc9, 0x79, 0xc8, 0xc9, 0x57, 0xc0, 0xa2, | 
|  | 0x12, 0x23, 0xfa, 0xef, 0x67, 0xf3, 0x2e, 0xe3, 0xd3, | 
|  | 0xd5, 0xe9, 0xec, 0xdb, 0x3e, 0xbc, 0x80, 0xb6, 0x6e, | 
|  | 0x2a, 0xb7, 0x8c, 0x83, 0xcd, 0x83, 0xcd, 0x00}; | 
|  | EXPECT_FALSE(WebRtcOpus_PacketHasVoiceActivity(dtx, 35)); | 
|  | } | 
|  |  | 
|  | }  // namespace webrtc |