|  | /* | 
|  | *  Copyright (c) 2016 The WebRTC project authors. All Rights Reserved. | 
|  | * | 
|  | *  Use of this source code is governed by a BSD-style license | 
|  | *  that can be found in the LICENSE file in the root of the source | 
|  | *  tree. An additional intellectual property rights grant can be found | 
|  | *  in the file PATENTS.  All contributing project authors may | 
|  | *  be found in the AUTHORS file in the root of the source tree. | 
|  | */ | 
|  |  | 
|  | #include "modules/audio_coding/codecs/legacy_encoded_audio_frame.h" | 
|  |  | 
|  | #include <algorithm> | 
|  | #include <memory> | 
|  | #include <utility> | 
|  |  | 
|  | #include "rtc_base/checks.h" | 
|  |  | 
|  | namespace webrtc { | 
|  |  | 
|  | LegacyEncodedAudioFrame::LegacyEncodedAudioFrame(AudioDecoder* decoder, | 
|  | rtc::Buffer&& payload) | 
|  | : decoder_(decoder), payload_(std::move(payload)) {} | 
|  |  | 
|  | LegacyEncodedAudioFrame::~LegacyEncodedAudioFrame() = default; | 
|  |  | 
|  | size_t LegacyEncodedAudioFrame::Duration() const { | 
|  | const int ret = decoder_->PacketDuration(payload_.data(), payload_.size()); | 
|  | return (ret < 0) ? 0 : static_cast<size_t>(ret); | 
|  | } | 
|  |  | 
|  | absl::optional<AudioDecoder::EncodedAudioFrame::DecodeResult> | 
|  | LegacyEncodedAudioFrame::Decode(rtc::ArrayView<int16_t> decoded) const { | 
|  | AudioDecoder::SpeechType speech_type = AudioDecoder::kSpeech; | 
|  | const int ret = decoder_->Decode( | 
|  | payload_.data(), payload_.size(), decoder_->SampleRateHz(), | 
|  | decoded.size() * sizeof(int16_t), decoded.data(), &speech_type); | 
|  |  | 
|  | if (ret < 0) | 
|  | return absl::nullopt; | 
|  |  | 
|  | return DecodeResult{static_cast<size_t>(ret), speech_type}; | 
|  | } | 
|  |  | 
|  | std::vector<AudioDecoder::ParseResult> LegacyEncodedAudioFrame::SplitBySamples( | 
|  | AudioDecoder* decoder, | 
|  | rtc::Buffer&& payload, | 
|  | uint32_t timestamp, | 
|  | size_t bytes_per_ms, | 
|  | uint32_t timestamps_per_ms) { | 
|  | RTC_DCHECK(payload.data()); | 
|  | std::vector<AudioDecoder::ParseResult> results; | 
|  | size_t split_size_bytes = payload.size(); | 
|  |  | 
|  | // Find a "chunk size" >= 20 ms and < 40 ms. | 
|  | const size_t min_chunk_size = bytes_per_ms * 20; | 
|  | if (min_chunk_size >= payload.size()) { | 
|  | std::unique_ptr<LegacyEncodedAudioFrame> frame( | 
|  | new LegacyEncodedAudioFrame(decoder, std::move(payload))); | 
|  | results.emplace_back(timestamp, 0, std::move(frame)); | 
|  | } else { | 
|  | // Reduce the split size by half as long as |split_size_bytes| is at least | 
|  | // twice the minimum chunk size (so that the resulting size is at least as | 
|  | // large as the minimum chunk size). | 
|  | while (split_size_bytes >= 2 * min_chunk_size) { | 
|  | split_size_bytes /= 2; | 
|  | } | 
|  |  | 
|  | const uint32_t timestamps_per_chunk = static_cast<uint32_t>( | 
|  | split_size_bytes * timestamps_per_ms / bytes_per_ms); | 
|  | size_t byte_offset; | 
|  | uint32_t timestamp_offset; | 
|  | for (byte_offset = 0, timestamp_offset = 0; byte_offset < payload.size(); | 
|  | byte_offset += split_size_bytes, | 
|  | timestamp_offset += timestamps_per_chunk) { | 
|  | split_size_bytes = | 
|  | std::min(split_size_bytes, payload.size() - byte_offset); | 
|  | rtc::Buffer new_payload(payload.data() + byte_offset, split_size_bytes); | 
|  | std::unique_ptr<LegacyEncodedAudioFrame> frame( | 
|  | new LegacyEncodedAudioFrame(decoder, std::move(new_payload))); | 
|  | results.emplace_back(timestamp + timestamp_offset, 0, std::move(frame)); | 
|  | } | 
|  | } | 
|  |  | 
|  | return results; | 
|  | } | 
|  |  | 
|  | }  // namespace webrtc |