| /* |
| * Copyright (c) 2015 The WebRTC project authors. All Rights Reserved. |
| * |
| * Use of this source code is governed by a BSD-style license |
| * that can be found in the LICENSE file in the root of the source |
| * tree. An additional intellectual property rights grant can be found |
| * in the file PATENTS. All contributing project authors may |
| * be found in the AUTHORS file in the root of the source tree. |
| */ |
| #include "video/video_quality_test.h" |
| |
| #include <stdio.h> |
| |
| #if defined(WEBRTC_WIN) |
| #include <conio.h> |
| #endif |
| |
| #include <algorithm> |
| #include <deque> |
| #include <map> |
| #include <memory> |
| #include <string> |
| #include <vector> |
| |
| #include "api/fec_controller_override.h" |
| #include "api/rtc_event_log_output_file.h" |
| #include "api/task_queue/default_task_queue_factory.h" |
| #include "api/task_queue/task_queue_base.h" |
| #include "api/test/create_frame_generator.h" |
| #include "api/video/builtin_video_bitrate_allocator_factory.h" |
| #include "api/video_codecs/video_encoder.h" |
| #include "call/fake_network_pipe.h" |
| #include "call/simulated_network.h" |
| #include "media/base/media_constants.h" |
| #include "media/engine/adm_helpers.h" |
| #include "media/engine/encoder_simulcast_proxy.h" |
| #include "media/engine/fake_video_codec_factory.h" |
| #include "media/engine/internal_encoder_factory.h" |
| #include "media/engine/webrtc_video_engine.h" |
| #include "modules/audio_device/include/audio_device.h" |
| #include "modules/audio_mixer/audio_mixer_impl.h" |
| #include "modules/video_coding/codecs/h264/include/h264.h" |
| #include "modules/video_coding/codecs/multiplex/include/multiplex_decoder_adapter.h" |
| #include "modules/video_coding/codecs/multiplex/include/multiplex_encoder_adapter.h" |
| #include "modules/video_coding/codecs/vp8/include/vp8.h" |
| #include "modules/video_coding/codecs/vp9/include/vp9.h" |
| #include "modules/video_coding/utility/ivf_file_writer.h" |
| #include "rtc_base/strings/string_builder.h" |
| #include "rtc_base/task_queue_for_test.h" |
| #include "test/platform_video_capturer.h" |
| #include "test/testsupport/file_utils.h" |
| #include "test/video_renderer.h" |
| #include "video/frame_dumping_decoder.h" |
| #ifdef WEBRTC_WIN |
| #include "modules/audio_device/include/audio_device_factory.h" |
| #endif |
| #include "video/config/encoder_stream_factory.h" |
| |
| namespace webrtc { |
| |
| namespace { |
| enum : int { // The first valid value is 1. |
| kAbsSendTimeExtensionId = 1, |
| kGenericFrameDescriptorExtensionId00, |
| kGenericFrameDescriptorExtensionId01, |
| kTransportSequenceNumberExtensionId, |
| kVideoContentTypeExtensionId, |
| kVideoTimingExtensionId, |
| }; |
| |
| constexpr char kSyncGroup[] = "av_sync"; |
| constexpr int kOpusMinBitrateBps = 6000; |
| constexpr int kOpusBitrateFbBps = 32000; |
| constexpr int kFramesSentInQuickTest = 1; |
| constexpr uint32_t kThumbnailSendSsrcStart = 0xE0000; |
| constexpr uint32_t kThumbnailRtxSsrcStart = 0xF0000; |
| |
| constexpr int kDefaultMaxQp = cricket::WebRtcVideoChannel::kDefaultQpMax; |
| |
| const VideoEncoder::Capabilities kCapabilities(false); |
| |
| std::pair<uint32_t, uint32_t> GetMinMaxBitratesBps(const VideoCodec& codec, |
| size_t spatial_idx) { |
| uint32_t min_bitrate = codec.minBitrate; |
| uint32_t max_bitrate = codec.maxBitrate; |
| if (spatial_idx < codec.numberOfSimulcastStreams) { |
| min_bitrate = |
| std::max(min_bitrate, codec.simulcastStream[spatial_idx].minBitrate); |
| max_bitrate = |
| std::min(max_bitrate, codec.simulcastStream[spatial_idx].maxBitrate); |
| } |
| if (codec.codecType == VideoCodecType::kVideoCodecVP9 && |
| spatial_idx < codec.VP9().numberOfSpatialLayers) { |
| min_bitrate = |
| std::max(min_bitrate, codec.spatialLayers[spatial_idx].minBitrate); |
| max_bitrate = |
| std::min(max_bitrate, codec.spatialLayers[spatial_idx].maxBitrate); |
| } |
| max_bitrate = std::max(max_bitrate, min_bitrate); |
| return {min_bitrate * 1000, max_bitrate * 1000}; |
| } |
| |
| class VideoStreamFactory |
| : public VideoEncoderConfig::VideoStreamFactoryInterface { |
| public: |
| explicit VideoStreamFactory(const std::vector<VideoStream>& streams) |
| : streams_(streams) {} |
| |
| private: |
| std::vector<VideoStream> CreateEncoderStreams( |
| int frame_width, |
| int frame_height, |
| const VideoEncoderConfig& encoder_config) override { |
| // The highest layer must match the incoming resolution. |
| std::vector<VideoStream> streams = streams_; |
| streams[streams_.size() - 1].height = frame_height; |
| streams[streams_.size() - 1].width = frame_width; |
| |
| streams[0].bitrate_priority = encoder_config.bitrate_priority; |
| return streams; |
| } |
| |
| std::vector<VideoStream> streams_; |
| }; |
| |
| // This wrapper provides two features needed by the video quality tests: |
| // 1. Invoke VideoAnalyzer callbacks before and after encoding each frame. |
| // 2. Write the encoded frames to file, one file per simulcast layer. |
| class QualityTestVideoEncoder : public VideoEncoder, |
| private EncodedImageCallback { |
| public: |
| QualityTestVideoEncoder(std::unique_ptr<VideoEncoder> encoder, |
| VideoAnalyzer* analyzer, |
| std::vector<FileWrapper> files, |
| double overshoot_factor) |
| : encoder_(std::move(encoder)), |
| overshoot_factor_(overshoot_factor), |
| analyzer_(analyzer) { |
| for (FileWrapper& file : files) { |
| writers_.push_back( |
| IvfFileWriter::Wrap(std::move(file), /* byte_limit= */ 100000000)); |
| } |
| } |
| |
| // Implement VideoEncoder |
| void SetFecControllerOverride( |
| FecControllerOverride* fec_controller_override) { |
| // Ignored. |
| } |
| |
| int32_t InitEncode(const VideoCodec* codec_settings, |
| const Settings& settings) override { |
| codec_settings_ = *codec_settings; |
| return encoder_->InitEncode(codec_settings, settings); |
| } |
| |
| int32_t RegisterEncodeCompleteCallback( |
| EncodedImageCallback* callback) override { |
| callback_ = callback; |
| return encoder_->RegisterEncodeCompleteCallback(this); |
| } |
| |
| int32_t Release() override { return encoder_->Release(); } |
| |
| int32_t Encode(const VideoFrame& frame, |
| const std::vector<VideoFrameType>* frame_types) { |
| if (analyzer_) { |
| analyzer_->PreEncodeOnFrame(frame); |
| } |
| return encoder_->Encode(frame, frame_types); |
| } |
| |
| void SetRates(const RateControlParameters& parameters) override { |
| RTC_DCHECK_GT(overshoot_factor_, 0.0); |
| if (overshoot_factor_ == 1.0) { |
| encoder_->SetRates(parameters); |
| return; |
| } |
| |
| // Simulating encoder overshooting target bitrate, by configuring actual |
| // encoder too high. Take care not to adjust past limits of config, |
| // otherwise encoders may crash on DCHECK. |
| VideoBitrateAllocation overshot_allocation; |
| for (size_t si = 0; si < kMaxSpatialLayers; ++si) { |
| const uint32_t spatial_layer_bitrate_bps = |
| parameters.bitrate.GetSpatialLayerSum(si); |
| if (spatial_layer_bitrate_bps == 0) { |
| continue; |
| } |
| |
| uint32_t min_bitrate_bps; |
| uint32_t max_bitrate_bps; |
| std::tie(min_bitrate_bps, max_bitrate_bps) = |
| GetMinMaxBitratesBps(codec_settings_, si); |
| double overshoot_factor = overshoot_factor_; |
| const uint32_t corrected_bitrate = rtc::checked_cast<uint32_t>( |
| overshoot_factor * spatial_layer_bitrate_bps); |
| if (corrected_bitrate < min_bitrate_bps) { |
| overshoot_factor = min_bitrate_bps / spatial_layer_bitrate_bps; |
| } else if (corrected_bitrate > max_bitrate_bps) { |
| overshoot_factor = max_bitrate_bps / spatial_layer_bitrate_bps; |
| } |
| |
| for (size_t ti = 0; ti < kMaxTemporalStreams; ++ti) { |
| if (parameters.bitrate.HasBitrate(si, ti)) { |
| overshot_allocation.SetBitrate( |
| si, ti, |
| rtc::checked_cast<uint32_t>( |
| overshoot_factor * parameters.bitrate.GetBitrate(si, ti))); |
| } |
| } |
| } |
| |
| return encoder_->SetRates( |
| RateControlParameters(overshot_allocation, parameters.framerate_fps, |
| parameters.bandwidth_allocation)); |
| } |
| |
| void OnPacketLossRateUpdate(float packet_loss_rate) override { |
| encoder_->OnPacketLossRateUpdate(packet_loss_rate); |
| } |
| |
| void OnRttUpdate(int64_t rtt_ms) override { encoder_->OnRttUpdate(rtt_ms); } |
| |
| void OnLossNotification(const LossNotification& loss_notification) override { |
| encoder_->OnLossNotification(loss_notification); |
| } |
| |
| EncoderInfo GetEncoderInfo() const override { |
| EncoderInfo info = encoder_->GetEncoderInfo(); |
| if (overshoot_factor_ != 1.0) { |
| // We're simulating bad encoder, don't forward trusted setting |
| // from eg libvpx. |
| info.has_trusted_rate_controller = false; |
| } |
| return info; |
| } |
| |
| private: |
| // Implement EncodedImageCallback |
| Result OnEncodedImage(const EncodedImage& encoded_image, |
| const CodecSpecificInfo* codec_specific_info) override { |
| if (codec_specific_info) { |
| int simulcast_index; |
| if (codec_specific_info->codecType == kVideoCodecVP9) { |
| simulcast_index = 0; |
| } else { |
| simulcast_index = encoded_image.SpatialIndex().value_or(0); |
| } |
| RTC_DCHECK_GE(simulcast_index, 0); |
| if (analyzer_) { |
| analyzer_->PostEncodeOnFrame(simulcast_index, |
| encoded_image.Timestamp()); |
| } |
| if (static_cast<size_t>(simulcast_index) < writers_.size()) { |
| writers_[simulcast_index]->WriteFrame(encoded_image, |
| codec_specific_info->codecType); |
| } |
| } |
| |
| return callback_->OnEncodedImage(encoded_image, codec_specific_info); |
| } |
| |
| void OnDroppedFrame(DropReason reason) override { |
| callback_->OnDroppedFrame(reason); |
| } |
| |
| const std::unique_ptr<VideoEncoder> encoder_; |
| const double overshoot_factor_; |
| VideoAnalyzer* const analyzer_; |
| std::vector<std::unique_ptr<IvfFileWriter>> writers_; |
| EncodedImageCallback* callback_ = nullptr; |
| VideoCodec codec_settings_; |
| }; |
| |
| #if defined(WEBRTC_WIN) && !defined(WINUWP) |
| void PressEnterToContinue(TaskQueueBase* task_queue) { |
| puts(">> Press ENTER to continue..."); |
| |
| while (!_kbhit() || _getch() != '\r') { |
| // Drive the message loop for the thread running the task_queue |
| SendTask(task_queue, [&]() { |
| MSG msg; |
| if (PeekMessage(&msg, NULL, 0, 0, PM_REMOVE)) { |
| TranslateMessage(&msg); |
| DispatchMessage(&msg); |
| } |
| }); |
| } |
| } |
| #else |
| void PressEnterToContinue(TaskQueueBase* /*task_queue*/) { |
| puts(">> Press ENTER to continue..."); |
| while (getc(stdin) != '\n' && !feof(stdin)) |
| ; // NOLINT |
| } |
| #endif |
| |
| } // namespace |
| |
| std::unique_ptr<VideoDecoder> VideoQualityTest::CreateVideoDecoder( |
| const SdpVideoFormat& format) { |
| std::unique_ptr<VideoDecoder> decoder; |
| if (format.name == "multiplex") { |
| decoder = std::make_unique<MultiplexDecoderAdapter>( |
| decoder_factory_.get(), SdpVideoFormat(cricket::kVp9CodecName)); |
| } else if (format.name == "FakeCodec") { |
| decoder = webrtc::FakeVideoDecoderFactory::CreateVideoDecoder(); |
| } else { |
| decoder = decoder_factory_->CreateVideoDecoder(format); |
| } |
| if (!params_.logging.encoded_frame_base_path.empty()) { |
| rtc::StringBuilder str; |
| str << receive_logs_++; |
| std::string path = |
| params_.logging.encoded_frame_base_path + "." + str.str() + ".recv.ivf"; |
| decoder = CreateFrameDumpingDecoderWrapper( |
| std::move(decoder), FileWrapper::OpenWriteOnly(path)); |
| } |
| return decoder; |
| } |
| |
| std::unique_ptr<VideoEncoder> VideoQualityTest::CreateVideoEncoder( |
| const SdpVideoFormat& format, |
| VideoAnalyzer* analyzer) { |
| std::unique_ptr<VideoEncoder> encoder; |
| if (format.name == "VP8") { |
| encoder = |
| std::make_unique<EncoderSimulcastProxy>(encoder_factory_.get(), format); |
| } else if (format.name == "multiplex") { |
| encoder = std::make_unique<MultiplexEncoderAdapter>( |
| encoder_factory_.get(), SdpVideoFormat(cricket::kVp9CodecName)); |
| } else if (format.name == "FakeCodec") { |
| encoder = webrtc::FakeVideoEncoderFactory::CreateVideoEncoder(); |
| } else { |
| encoder = encoder_factory_->CreateVideoEncoder(format); |
| } |
| |
| std::vector<FileWrapper> encoded_frame_dump_files; |
| if (!params_.logging.encoded_frame_base_path.empty()) { |
| char ss_buf[100]; |
| rtc::SimpleStringBuilder sb(ss_buf); |
| sb << send_logs_++; |
| std::string prefix = |
| params_.logging.encoded_frame_base_path + "." + sb.str() + ".send."; |
| encoded_frame_dump_files.push_back( |
| FileWrapper::OpenWriteOnly(prefix + "1.ivf")); |
| encoded_frame_dump_files.push_back( |
| FileWrapper::OpenWriteOnly(prefix + "2.ivf")); |
| encoded_frame_dump_files.push_back( |
| FileWrapper::OpenWriteOnly(prefix + "3.ivf")); |
| } |
| |
| double overshoot_factor = 1.0; |
| // Match format to either of the streams in dual-stream mode in order to get |
| // the overshoot factor. This is not very robust but we can't know for sure |
| // which stream this encoder is meant for, from within the factory. |
| if (format == |
| SdpVideoFormat(params_.video[0].codec, params_.video[0].sdp_params)) { |
| overshoot_factor = params_.video[0].encoder_overshoot_factor; |
| } else if (format == SdpVideoFormat(params_.video[1].codec, |
| params_.video[1].sdp_params)) { |
| overshoot_factor = params_.video[1].encoder_overshoot_factor; |
| } |
| if (overshoot_factor == 0.0) { |
| // If params were zero-initialized, set to 1.0 instead. |
| overshoot_factor = 1.0; |
| } |
| |
| if (analyzer || !encoded_frame_dump_files.empty() || overshoot_factor > 1.0) { |
| encoder = std::make_unique<QualityTestVideoEncoder>( |
| std::move(encoder), analyzer, std::move(encoded_frame_dump_files), |
| overshoot_factor); |
| } |
| |
| return encoder; |
| } |
| |
| VideoQualityTest::VideoQualityTest( |
| std::unique_ptr<InjectionComponents> injection_components) |
| : clock_(Clock::GetRealTimeClock()), |
| task_queue_factory_(CreateDefaultTaskQueueFactory()), |
| rtc_event_log_factory_(task_queue_factory_.get()), |
| video_decoder_factory_([this](const SdpVideoFormat& format) { |
| return this->CreateVideoDecoder(format); |
| }), |
| video_encoder_factory_([this](const SdpVideoFormat& format) { |
| return this->CreateVideoEncoder(format, nullptr); |
| }), |
| video_encoder_factory_with_analyzer_( |
| [this](const SdpVideoFormat& format) { |
| return this->CreateVideoEncoder(format, analyzer_.get()); |
| }), |
| video_bitrate_allocator_factory_( |
| CreateBuiltinVideoBitrateAllocatorFactory()), |
| receive_logs_(0), |
| send_logs_(0), |
| injection_components_(std::move(injection_components)), |
| num_video_streams_(0) { |
| if (injection_components_ == nullptr) { |
| injection_components_ = std::make_unique<InjectionComponents>(); |
| } |
| if (injection_components_->video_decoder_factory != nullptr) { |
| decoder_factory_ = std::move(injection_components_->video_decoder_factory); |
| } else { |
| decoder_factory_ = std::make_unique<InternalDecoderFactory>(); |
| } |
| if (injection_components_->video_encoder_factory != nullptr) { |
| encoder_factory_ = std::move(injection_components_->video_encoder_factory); |
| } else { |
| encoder_factory_ = std::make_unique<InternalEncoderFactory>(); |
| } |
| |
| payload_type_map_ = test::CallTest::payload_type_map_; |
| RTC_DCHECK(payload_type_map_.find(kPayloadTypeH264) == |
| payload_type_map_.end()); |
| RTC_DCHECK(payload_type_map_.find(kPayloadTypeVP8) == |
| payload_type_map_.end()); |
| RTC_DCHECK(payload_type_map_.find(kPayloadTypeVP9) == |
| payload_type_map_.end()); |
| RTC_DCHECK(payload_type_map_.find(kPayloadTypeGeneric) == |
| payload_type_map_.end()); |
| payload_type_map_[kPayloadTypeH264] = webrtc::MediaType::VIDEO; |
| payload_type_map_[kPayloadTypeVP8] = webrtc::MediaType::VIDEO; |
| payload_type_map_[kPayloadTypeVP9] = webrtc::MediaType::VIDEO; |
| payload_type_map_[kPayloadTypeGeneric] = webrtc::MediaType::VIDEO; |
| |
| fec_controller_factory_ = |
| std::move(injection_components_->fec_controller_factory); |
| network_state_predictor_factory_ = |
| std::move(injection_components_->network_state_predictor_factory); |
| network_controller_factory_ = |
| std::move(injection_components_->network_controller_factory); |
| } |
| |
| VideoQualityTest::InjectionComponents::InjectionComponents() = default; |
| |
| VideoQualityTest::InjectionComponents::~InjectionComponents() = default; |
| |
| void VideoQualityTest::TestBody() {} |
| |
| std::string VideoQualityTest::GenerateGraphTitle() const { |
| rtc::StringBuilder ss; |
| ss << params_.video[0].codec; |
| ss << " (" << params_.video[0].target_bitrate_bps / 1000 << "kbps"; |
| ss << ", " << params_.video[0].fps << " FPS"; |
| if (params_.screenshare[0].scroll_duration) |
| ss << ", " << params_.screenshare[0].scroll_duration << "s scroll"; |
| if (params_.ss[0].streams.size() > 1) |
| ss << ", Stream #" << params_.ss[0].selected_stream; |
| if (params_.ss[0].num_spatial_layers > 1) |
| ss << ", Layer #" << params_.ss[0].selected_sl; |
| ss << ")"; |
| return ss.Release(); |
| } |
| |
| void VideoQualityTest::CheckParamsAndInjectionComponents() { |
| if (injection_components_ == nullptr) { |
| injection_components_ = std::make_unique<InjectionComponents>(); |
| } |
| if (!params_.config && injection_components_->sender_network == nullptr && |
| injection_components_->receiver_network == nullptr) { |
| params_.config = BuiltInNetworkBehaviorConfig(); |
| } |
| RTC_CHECK( |
| (params_.config && injection_components_->sender_network == nullptr && |
| injection_components_->receiver_network == nullptr) || |
| (!params_.config && injection_components_->sender_network != nullptr && |
| injection_components_->receiver_network != nullptr)); |
| for (size_t video_idx = 0; video_idx < num_video_streams_; ++video_idx) { |
| // Iterate over primary and secondary video streams. |
| if (!params_.video[video_idx].enabled) |
| return; |
| // Add a default stream in none specified. |
| if (params_.ss[video_idx].streams.empty()) |
| params_.ss[video_idx].streams.push_back( |
| VideoQualityTest::DefaultVideoStream(params_, video_idx)); |
| if (params_.ss[video_idx].num_spatial_layers == 0) |
| params_.ss[video_idx].num_spatial_layers = 1; |
| |
| if (params_.config) { |
| if (params_.config->loss_percent != 0 || |
| params_.config->queue_length_packets != 0) { |
| // Since LayerFilteringTransport changes the sequence numbers, we can't |
| // use that feature with pack loss, since the NACK request would end up |
| // retransmitting the wrong packets. |
| RTC_CHECK(params_.ss[video_idx].selected_sl == -1 || |
| params_.ss[video_idx].selected_sl == |
| params_.ss[video_idx].num_spatial_layers - 1); |
| RTC_CHECK(params_.video[video_idx].selected_tl == -1 || |
| params_.video[video_idx].selected_tl == |
| params_.video[video_idx].num_temporal_layers - 1); |
| } |
| } |
| |
| // TODO(ivica): Should max_bitrate_bps == -1 represent inf max bitrate, as |
| // it does in some parts of the code? |
| RTC_CHECK_GE(params_.video[video_idx].max_bitrate_bps, |
| params_.video[video_idx].target_bitrate_bps); |
| RTC_CHECK_GE(params_.video[video_idx].target_bitrate_bps, |
| params_.video[video_idx].min_bitrate_bps); |
| int selected_stream = params_.ss[video_idx].selected_stream; |
| if (params_.video[video_idx].selected_tl > -1) { |
| RTC_CHECK_LT(selected_stream, params_.ss[video_idx].streams.size()) |
| << "Can not use --selected_tl when --selected_stream is all streams"; |
| int stream_tl = params_.ss[video_idx] |
| .streams[selected_stream] |
| .num_temporal_layers.value_or(1); |
| RTC_CHECK_LT(params_.video[video_idx].selected_tl, stream_tl); |
| } |
| RTC_CHECK_LE(params_.ss[video_idx].selected_stream, |
| params_.ss[video_idx].streams.size()); |
| for (const VideoStream& stream : params_.ss[video_idx].streams) { |
| RTC_CHECK_GE(stream.min_bitrate_bps, 0); |
| RTC_CHECK_GE(stream.target_bitrate_bps, stream.min_bitrate_bps); |
| RTC_CHECK_GE(stream.max_bitrate_bps, stream.target_bitrate_bps); |
| } |
| // TODO(ivica): Should we check if the sum of all streams/layers is equal to |
| // the total bitrate? We anyway have to update them in the case bitrate |
| // estimator changes the total bitrates. |
| RTC_CHECK_GE(params_.ss[video_idx].num_spatial_layers, 1); |
| RTC_CHECK_LE(params_.ss[video_idx].selected_sl, |
| params_.ss[video_idx].num_spatial_layers); |
| RTC_CHECK( |
| params_.ss[video_idx].spatial_layers.empty() || |
| params_.ss[video_idx].spatial_layers.size() == |
| static_cast<size_t>(params_.ss[video_idx].num_spatial_layers)); |
| if (params_.video[video_idx].codec == "VP8") { |
| RTC_CHECK_EQ(params_.ss[video_idx].num_spatial_layers, 1); |
| } else if (params_.video[video_idx].codec == "VP9") { |
| RTC_CHECK_EQ(params_.ss[video_idx].streams.size(), 1); |
| } |
| RTC_CHECK_GE(params_.call.num_thumbnails, 0); |
| if (params_.call.num_thumbnails > 0) { |
| RTC_CHECK_EQ(params_.ss[video_idx].num_spatial_layers, 1); |
| RTC_CHECK_EQ(params_.ss[video_idx].streams.size(), 3); |
| RTC_CHECK_EQ(params_.video[video_idx].num_temporal_layers, 3); |
| RTC_CHECK_EQ(params_.video[video_idx].codec, "VP8"); |
| } |
| // Dual streams with FEC not supported in tests yet. |
| RTC_CHECK(!params_.video[video_idx].flexfec || num_video_streams_ == 1); |
| RTC_CHECK(!params_.video[video_idx].ulpfec || num_video_streams_ == 1); |
| } |
| } |
| |
| // Static. |
| std::vector<int> VideoQualityTest::ParseCSV(const std::string& str) { |
| // Parse comma separated nonnegative integers, where some elements may be |
| // empty. The empty values are replaced with -1. |
| // E.g. "10,-20,,30,40" --> {10, 20, -1, 30,40} |
| // E.g. ",,10,,20," --> {-1, -1, 10, -1, 20, -1} |
| std::vector<int> result; |
| if (str.empty()) |
| return result; |
| |
| const char* p = str.c_str(); |
| int value = -1; |
| int pos; |
| while (*p) { |
| if (*p == ',') { |
| result.push_back(value); |
| value = -1; |
| ++p; |
| continue; |
| } |
| RTC_CHECK_EQ(sscanf(p, "%d%n", &value, &pos), 1) |
| << "Unexpected non-number value."; |
| p += pos; |
| } |
| result.push_back(value); |
| return result; |
| } |
| |
| // Static. |
| VideoStream VideoQualityTest::DefaultVideoStream(const Params& params, |
| size_t video_idx) { |
| VideoStream stream; |
| stream.width = params.video[video_idx].width; |
| stream.height = params.video[video_idx].height; |
| stream.max_framerate = params.video[video_idx].fps; |
| stream.min_bitrate_bps = params.video[video_idx].min_bitrate_bps; |
| stream.target_bitrate_bps = params.video[video_idx].target_bitrate_bps; |
| stream.max_bitrate_bps = params.video[video_idx].max_bitrate_bps; |
| stream.max_qp = kDefaultMaxQp; |
| stream.num_temporal_layers = params.video[video_idx].num_temporal_layers; |
| stream.active = true; |
| return stream; |
| } |
| |
| // Static. |
| VideoStream VideoQualityTest::DefaultThumbnailStream() { |
| VideoStream stream; |
| stream.width = 320; |
| stream.height = 180; |
| stream.max_framerate = 7; |
| stream.min_bitrate_bps = 7500; |
| stream.target_bitrate_bps = 37500; |
| stream.max_bitrate_bps = 50000; |
| stream.max_qp = kDefaultMaxQp; |
| return stream; |
| } |
| |
| // Static. |
| void VideoQualityTest::FillScalabilitySettings( |
| Params* params, |
| size_t video_idx, |
| const std::vector<std::string>& stream_descriptors, |
| int num_streams, |
| size_t selected_stream, |
| int num_spatial_layers, |
| int selected_sl, |
| InterLayerPredMode inter_layer_pred, |
| const std::vector<std::string>& sl_descriptors) { |
| if (params->ss[video_idx].streams.empty() && |
| params->ss[video_idx].infer_streams) { |
| webrtc::VideoEncoder::EncoderInfo encoder_info; |
| webrtc::VideoEncoderConfig encoder_config; |
| encoder_config.codec_type = |
| PayloadStringToCodecType(params->video[video_idx].codec); |
| encoder_config.content_type = |
| params->screenshare[video_idx].enabled |
| ? webrtc::VideoEncoderConfig::ContentType::kScreen |
| : webrtc::VideoEncoderConfig::ContentType::kRealtimeVideo; |
| encoder_config.max_bitrate_bps = params->video[video_idx].max_bitrate_bps; |
| encoder_config.min_transmit_bitrate_bps = |
| params->video[video_idx].min_transmit_bps; |
| encoder_config.number_of_streams = num_streams; |
| encoder_config.spatial_layers = params->ss[video_idx].spatial_layers; |
| encoder_config.simulcast_layers = std::vector<VideoStream>(num_streams); |
| encoder_config.video_stream_factory = |
| rtc::make_ref_counted<cricket::EncoderStreamFactory>( |
| params->video[video_idx].codec, kDefaultMaxQp, |
| params->screenshare[video_idx].enabled, true, encoder_info); |
| params->ss[video_idx].streams = |
| encoder_config.video_stream_factory->CreateEncoderStreams( |
| params->video[video_idx].width, params->video[video_idx].height, |
| encoder_config); |
| } else { |
| // Read VideoStream and SpatialLayer elements from a list of comma separated |
| // lists. To use a default value for an element, use -1 or leave empty. |
| // Validity checks performed in CheckParamsAndInjectionComponents. |
| RTC_CHECK(params->ss[video_idx].streams.empty()); |
| for (const auto& descriptor : stream_descriptors) { |
| if (descriptor.empty()) |
| continue; |
| VideoStream stream = |
| VideoQualityTest::DefaultVideoStream(*params, video_idx); |
| std::vector<int> v = VideoQualityTest::ParseCSV(descriptor); |
| if (v[0] != -1) |
| stream.width = static_cast<size_t>(v[0]); |
| if (v[1] != -1) |
| stream.height = static_cast<size_t>(v[1]); |
| if (v[2] != -1) |
| stream.max_framerate = v[2]; |
| if (v[3] != -1) |
| stream.min_bitrate_bps = v[3]; |
| if (v[4] != -1) |
| stream.target_bitrate_bps = v[4]; |
| if (v[5] != -1) |
| stream.max_bitrate_bps = v[5]; |
| if (v.size() > 6 && v[6] != -1) |
| stream.max_qp = v[6]; |
| if (v.size() > 7 && v[7] != -1) { |
| stream.num_temporal_layers = v[7]; |
| } else { |
| // Automatic TL thresholds for more than two layers not supported. |
| RTC_CHECK_LE(params->video[video_idx].num_temporal_layers, 2); |
| } |
| params->ss[video_idx].streams.push_back(stream); |
| } |
| } |
| |
| params->ss[video_idx].num_spatial_layers = std::max(1, num_spatial_layers); |
| params->ss[video_idx].selected_stream = selected_stream; |
| |
| params->ss[video_idx].selected_sl = selected_sl; |
| params->ss[video_idx].inter_layer_pred = inter_layer_pred; |
| RTC_CHECK(params->ss[video_idx].spatial_layers.empty()); |
| for (const auto& descriptor : sl_descriptors) { |
| if (descriptor.empty()) |
| continue; |
| std::vector<int> v = VideoQualityTest::ParseCSV(descriptor); |
| RTC_CHECK_EQ(v.size(), 8); |
| |
| SpatialLayer layer = {0}; |
| layer.width = v[0]; |
| layer.height = v[1]; |
| layer.maxFramerate = v[2]; |
| layer.numberOfTemporalLayers = v[3]; |
| layer.maxBitrate = v[4]; |
| layer.minBitrate = v[5]; |
| layer.targetBitrate = v[6]; |
| layer.qpMax = v[7]; |
| layer.active = true; |
| |
| params->ss[video_idx].spatial_layers.push_back(layer); |
| } |
| } |
| |
| void VideoQualityTest::SetupVideo(Transport* send_transport, |
| Transport* recv_transport) { |
| size_t total_streams_used = 0; |
| video_receive_configs_.clear(); |
| video_send_configs_.clear(); |
| video_encoder_configs_.clear(); |
| bool decode_all_receive_streams = true; |
| size_t num_video_substreams = params_.ss[0].streams.size(); |
| RTC_CHECK(num_video_streams_ > 0); |
| video_encoder_configs_.resize(num_video_streams_); |
| std::string generic_codec_name; |
| webrtc::VideoEncoder::EncoderInfo encoder_info; |
| for (size_t video_idx = 0; video_idx < num_video_streams_; ++video_idx) { |
| video_send_configs_.push_back(VideoSendStream::Config(send_transport)); |
| video_encoder_configs_.push_back(VideoEncoderConfig()); |
| num_video_substreams = params_.ss[video_idx].streams.size(); |
| RTC_CHECK_GT(num_video_substreams, 0); |
| for (size_t i = 0; i < num_video_substreams; ++i) |
| video_send_configs_[video_idx].rtp.ssrcs.push_back( |
| kVideoSendSsrcs[total_streams_used + i]); |
| |
| int payload_type; |
| if (params_.video[video_idx].codec == "H264") { |
| payload_type = kPayloadTypeH264; |
| } else if (params_.video[video_idx].codec == "VP8") { |
| payload_type = kPayloadTypeVP8; |
| } else if (params_.video[video_idx].codec == "VP9") { |
| payload_type = kPayloadTypeVP9; |
| } else if (params_.video[video_idx].codec == "multiplex") { |
| payload_type = kPayloadTypeVP9; |
| } else if (params_.video[video_idx].codec == "FakeCodec") { |
| payload_type = kFakeVideoSendPayloadType; |
| } else { |
| RTC_CHECK(generic_codec_name.empty() || |
| generic_codec_name == params_.video[video_idx].codec) |
| << "Supplying multiple generic codecs is unsupported."; |
| RTC_LOG(LS_INFO) << "Treating codec " << params_.video[video_idx].codec |
| << " as generic."; |
| payload_type = kPayloadTypeGeneric; |
| generic_codec_name = params_.video[video_idx].codec; |
| } |
| video_send_configs_[video_idx].encoder_settings.encoder_factory = |
| (video_idx == 0) ? &video_encoder_factory_with_analyzer_ |
| : &video_encoder_factory_; |
| video_send_configs_[video_idx].encoder_settings.bitrate_allocator_factory = |
| video_bitrate_allocator_factory_.get(); |
| |
| video_send_configs_[video_idx].rtp.payload_name = |
| params_.video[video_idx].codec; |
| video_send_configs_[video_idx].rtp.payload_type = payload_type; |
| video_send_configs_[video_idx].rtp.nack.rtp_history_ms = kNackRtpHistoryMs; |
| video_send_configs_[video_idx].rtp.rtx.payload_type = kSendRtxPayloadType; |
| for (size_t i = 0; i < num_video_substreams; ++i) { |
| video_send_configs_[video_idx].rtp.rtx.ssrcs.push_back( |
| kSendRtxSsrcs[i + total_streams_used]); |
| } |
| video_send_configs_[video_idx].rtp.extensions.clear(); |
| if (params_.call.send_side_bwe) { |
| video_send_configs_[video_idx].rtp.extensions.emplace_back( |
| RtpExtension::kTransportSequenceNumberUri, |
| kTransportSequenceNumberExtensionId); |
| } else { |
| video_send_configs_[video_idx].rtp.extensions.emplace_back( |
| RtpExtension::kAbsSendTimeUri, kAbsSendTimeExtensionId); |
| } |
| |
| if (params_.call.generic_descriptor) { |
| video_send_configs_[video_idx].rtp.extensions.emplace_back( |
| RtpExtension::kGenericFrameDescriptorUri00, |
| kGenericFrameDescriptorExtensionId00); |
| } |
| |
| video_send_configs_[video_idx].rtp.extensions.emplace_back( |
| RtpExtension::kVideoContentTypeUri, kVideoContentTypeExtensionId); |
| video_send_configs_[video_idx].rtp.extensions.emplace_back( |
| RtpExtension::kVideoTimingUri, kVideoTimingExtensionId); |
| |
| video_encoder_configs_[video_idx].video_format.name = |
| params_.video[video_idx].codec; |
| |
| video_encoder_configs_[video_idx].video_format.parameters = |
| params_.video[video_idx].sdp_params; |
| |
| video_encoder_configs_[video_idx].codec_type = |
| PayloadStringToCodecType(params_.video[video_idx].codec); |
| |
| video_encoder_configs_[video_idx].min_transmit_bitrate_bps = |
| params_.video[video_idx].min_transmit_bps; |
| |
| video_send_configs_[video_idx].suspend_below_min_bitrate = |
| params_.video[video_idx].suspend_below_min_bitrate; |
| |
| video_encoder_configs_[video_idx].number_of_streams = |
| params_.ss[video_idx].streams.size(); |
| video_encoder_configs_[video_idx].max_bitrate_bps = 0; |
| for (size_t i = 0; i < params_.ss[video_idx].streams.size(); ++i) { |
| video_encoder_configs_[video_idx].max_bitrate_bps += |
| params_.ss[video_idx].streams[i].max_bitrate_bps; |
| } |
| video_encoder_configs_[video_idx].simulcast_layers = |
| std::vector<VideoStream>(params_.ss[video_idx].streams.size()); |
| if (!params_.ss[video_idx].infer_streams) { |
| video_encoder_configs_[video_idx].simulcast_layers = |
| params_.ss[video_idx].streams; |
| } |
| video_encoder_configs_[video_idx].video_stream_factory = |
| rtc::make_ref_counted<cricket::EncoderStreamFactory>( |
| params_.video[video_idx].codec, |
| params_.ss[video_idx].streams[0].max_qp, |
| params_.screenshare[video_idx].enabled, true, encoder_info); |
| |
| video_encoder_configs_[video_idx].spatial_layers = |
| params_.ss[video_idx].spatial_layers; |
| |
| video_encoder_configs_[video_idx].frame_drop_enabled = true; |
| |
| decode_all_receive_streams = params_.ss[video_idx].selected_stream == |
| params_.ss[video_idx].streams.size(); |
| absl::optional<int> decode_sub_stream; |
| if (!decode_all_receive_streams) |
| decode_sub_stream = params_.ss[video_idx].selected_stream; |
| CreateMatchingVideoReceiveConfigs( |
| video_send_configs_[video_idx], recv_transport, |
| params_.call.send_side_bwe, &video_decoder_factory_, decode_sub_stream, |
| true, kNackRtpHistoryMs); |
| |
| if (params_.screenshare[video_idx].enabled) { |
| // Fill out codec settings. |
| video_encoder_configs_[video_idx].content_type = |
| VideoEncoderConfig::ContentType::kScreen; |
| degradation_preference_ = DegradationPreference::MAINTAIN_RESOLUTION; |
| if (params_.video[video_idx].codec == "VP8") { |
| VideoCodecVP8 vp8_settings = VideoEncoder::GetDefaultVp8Settings(); |
| vp8_settings.denoisingOn = false; |
| vp8_settings.numberOfTemporalLayers = static_cast<unsigned char>( |
| params_.video[video_idx].num_temporal_layers); |
| video_encoder_configs_[video_idx].encoder_specific_settings = |
| rtc::make_ref_counted< |
| VideoEncoderConfig::Vp8EncoderSpecificSettings>(vp8_settings); |
| } else if (params_.video[video_idx].codec == "VP9") { |
| VideoCodecVP9 vp9_settings = VideoEncoder::GetDefaultVp9Settings(); |
| vp9_settings.denoisingOn = false; |
| vp9_settings.automaticResizeOn = false; |
| vp9_settings.numberOfTemporalLayers = static_cast<unsigned char>( |
| params_.video[video_idx].num_temporal_layers); |
| vp9_settings.numberOfSpatialLayers = static_cast<unsigned char>( |
| params_.ss[video_idx].num_spatial_layers); |
| vp9_settings.interLayerPred = params_.ss[video_idx].inter_layer_pred; |
| // High FPS vp9 screenshare requires flexible mode. |
| if (params_.ss[video_idx].num_spatial_layers > 1) { |
| vp9_settings.flexibleMode = true; |
| } |
| video_encoder_configs_[video_idx].encoder_specific_settings = |
| rtc::make_ref_counted< |
| VideoEncoderConfig::Vp9EncoderSpecificSettings>(vp9_settings); |
| } |
| } else if (params_.ss[video_idx].num_spatial_layers > 1) { |
| // If SVC mode without screenshare, still need to set codec specifics. |
| RTC_CHECK(params_.video[video_idx].codec == "VP9"); |
| VideoCodecVP9 vp9_settings = VideoEncoder::GetDefaultVp9Settings(); |
| vp9_settings.numberOfTemporalLayers = static_cast<unsigned char>( |
| params_.video[video_idx].num_temporal_layers); |
| vp9_settings.numberOfSpatialLayers = |
| static_cast<unsigned char>(params_.ss[video_idx].num_spatial_layers); |
| vp9_settings.interLayerPred = params_.ss[video_idx].inter_layer_pred; |
| vp9_settings.automaticResizeOn = false; |
| video_encoder_configs_[video_idx].encoder_specific_settings = |
| rtc::make_ref_counted<VideoEncoderConfig::Vp9EncoderSpecificSettings>( |
| vp9_settings); |
| RTC_DCHECK_EQ(video_encoder_configs_[video_idx].simulcast_layers.size(), |
| 1); |
| // Min bitrate will be enforced by spatial layer config instead. |
| video_encoder_configs_[video_idx].simulcast_layers[0].min_bitrate_bps = 0; |
| } else if (params_.video[video_idx].automatic_scaling) { |
| if (params_.video[video_idx].codec == "VP8") { |
| VideoCodecVP8 vp8_settings = VideoEncoder::GetDefaultVp8Settings(); |
| vp8_settings.automaticResizeOn = true; |
| video_encoder_configs_[video_idx].encoder_specific_settings = |
| rtc::make_ref_counted< |
| VideoEncoderConfig::Vp8EncoderSpecificSettings>(vp8_settings); |
| } else if (params_.video[video_idx].codec == "VP9") { |
| VideoCodecVP9 vp9_settings = VideoEncoder::GetDefaultVp9Settings(); |
| // Only enable quality scaler for single spatial layer. |
| vp9_settings.automaticResizeOn = |
| params_.ss[video_idx].num_spatial_layers == 1; |
| video_encoder_configs_[video_idx].encoder_specific_settings = |
| rtc::make_ref_counted< |
| VideoEncoderConfig::Vp9EncoderSpecificSettings>(vp9_settings); |
| } else if (params_.video[video_idx].codec == "H264") { |
| // Quality scaling is always on for H.264. |
| } else if (params_.video[video_idx].codec == cricket::kAv1CodecName) { |
| // TODO(bugs.webrtc.org/11404): Propagate the flag to |
| // aom_codec_enc_cfg_t::rc_resize_mode in Av1 encoder wrapper. |
| // Until then do nothing, specially do not crash. |
| } else { |
| RTC_DCHECK_NOTREACHED() |
| << "Automatic scaling not supported for codec " |
| << params_.video[video_idx].codec << ", stream " << video_idx; |
| } |
| } else { |
| // Default mode. Single SL, no automatic_scaling, |
| if (params_.video[video_idx].codec == "VP8") { |
| VideoCodecVP8 vp8_settings = VideoEncoder::GetDefaultVp8Settings(); |
| vp8_settings.automaticResizeOn = false; |
| video_encoder_configs_[video_idx].encoder_specific_settings = |
| rtc::make_ref_counted< |
| VideoEncoderConfig::Vp8EncoderSpecificSettings>(vp8_settings); |
| } else if (params_.video[video_idx].codec == "VP9") { |
| VideoCodecVP9 vp9_settings = VideoEncoder::GetDefaultVp9Settings(); |
| vp9_settings.automaticResizeOn = false; |
| video_encoder_configs_[video_idx].encoder_specific_settings = |
| rtc::make_ref_counted< |
| VideoEncoderConfig::Vp9EncoderSpecificSettings>(vp9_settings); |
| } else if (params_.video[video_idx].codec == "H264") { |
| video_encoder_configs_[video_idx].encoder_specific_settings = nullptr; |
| } |
| } |
| total_streams_used += num_video_substreams; |
| } |
| |
| // FEC supported only for single video stream mode yet. |
| if (params_.video[0].flexfec) { |
| if (decode_all_receive_streams) { |
| SetSendFecConfig(GetVideoSendConfig()->rtp.ssrcs); |
| } else { |
| SetSendFecConfig({kVideoSendSsrcs[params_.ss[0].selected_stream]}); |
| } |
| |
| CreateMatchingFecConfig(recv_transport, *GetVideoSendConfig()); |
| GetFlexFecConfig()->rtp.transport_cc = params_.call.send_side_bwe; |
| if (params_.call.send_side_bwe) { |
| GetFlexFecConfig()->rtp.extensions.push_back( |
| RtpExtension(RtpExtension::kTransportSequenceNumberUri, |
| kTransportSequenceNumberExtensionId)); |
| } else { |
| GetFlexFecConfig()->rtp.extensions.push_back( |
| RtpExtension(RtpExtension::kAbsSendTimeUri, kAbsSendTimeExtensionId)); |
| } |
| } |
| |
| if (params_.video[0].ulpfec) { |
| SetSendUlpFecConfig(GetVideoSendConfig()); |
| if (decode_all_receive_streams) { |
| for (auto& receive_config : video_receive_configs_) { |
| SetReceiveUlpFecConfig(&receive_config); |
| } |
| } else { |
| SetReceiveUlpFecConfig( |
| &video_receive_configs_[params_.ss[0].selected_stream]); |
| } |
| } |
| } |
| |
| void VideoQualityTest::SetupThumbnails(Transport* send_transport, |
| Transport* recv_transport) { |
| for (int i = 0; i < params_.call.num_thumbnails; ++i) { |
| // Thumbnails will be send in the other way: from receiver_call to |
| // sender_call. |
| VideoSendStream::Config thumbnail_send_config(recv_transport); |
| thumbnail_send_config.rtp.ssrcs.push_back(kThumbnailSendSsrcStart + i); |
| thumbnail_send_config.encoder_settings.encoder_factory = |
| &video_encoder_factory_; |
| thumbnail_send_config.encoder_settings.bitrate_allocator_factory = |
| video_bitrate_allocator_factory_.get(); |
| thumbnail_send_config.rtp.payload_name = params_.video[0].codec; |
| thumbnail_send_config.rtp.payload_type = kPayloadTypeVP8; |
| thumbnail_send_config.rtp.nack.rtp_history_ms = kNackRtpHistoryMs; |
| thumbnail_send_config.rtp.rtx.payload_type = kSendRtxPayloadType; |
| thumbnail_send_config.rtp.rtx.ssrcs.push_back(kThumbnailRtxSsrcStart + i); |
| thumbnail_send_config.rtp.extensions.clear(); |
| if (params_.call.send_side_bwe) { |
| thumbnail_send_config.rtp.extensions.push_back( |
| RtpExtension(RtpExtension::kTransportSequenceNumberUri, |
| kTransportSequenceNumberExtensionId)); |
| } else { |
| thumbnail_send_config.rtp.extensions.push_back( |
| RtpExtension(RtpExtension::kAbsSendTimeUri, kAbsSendTimeExtensionId)); |
| } |
| |
| VideoEncoderConfig thumbnail_encoder_config; |
| thumbnail_encoder_config.codec_type = kVideoCodecVP8; |
| thumbnail_encoder_config.video_format.name = "VP8"; |
| thumbnail_encoder_config.min_transmit_bitrate_bps = 7500; |
| thumbnail_send_config.suspend_below_min_bitrate = |
| params_.video[0].suspend_below_min_bitrate; |
| thumbnail_encoder_config.number_of_streams = 1; |
| thumbnail_encoder_config.max_bitrate_bps = 50000; |
| std::vector<VideoStream> streams{params_.ss[0].streams[0]}; |
| thumbnail_encoder_config.video_stream_factory = |
| rtc::make_ref_counted<VideoStreamFactory>(streams); |
| thumbnail_encoder_config.spatial_layers = params_.ss[0].spatial_layers; |
| |
| thumbnail_encoder_configs_.push_back(thumbnail_encoder_config.Copy()); |
| thumbnail_send_configs_.push_back(thumbnail_send_config.Copy()); |
| |
| AddMatchingVideoReceiveConfigs( |
| &thumbnail_receive_configs_, thumbnail_send_config, send_transport, |
| params_.call.send_side_bwe, &video_decoder_factory_, absl::nullopt, |
| false, kNackRtpHistoryMs); |
| } |
| for (size_t i = 0; i < thumbnail_send_configs_.size(); ++i) { |
| thumbnail_send_streams_.push_back(receiver_call_->CreateVideoSendStream( |
| thumbnail_send_configs_[i].Copy(), |
| thumbnail_encoder_configs_[i].Copy())); |
| } |
| for (size_t i = 0; i < thumbnail_receive_configs_.size(); ++i) { |
| thumbnail_receive_streams_.push_back(sender_call_->CreateVideoReceiveStream( |
| thumbnail_receive_configs_[i].Copy())); |
| } |
| } |
| |
| void VideoQualityTest::DestroyThumbnailStreams() { |
| for (VideoSendStream* thumbnail_send_stream : thumbnail_send_streams_) { |
| receiver_call_->DestroyVideoSendStream(thumbnail_send_stream); |
| } |
| thumbnail_send_streams_.clear(); |
| for (VideoReceiveStreamInterface* thumbnail_receive_stream : |
| thumbnail_receive_streams_) { |
| sender_call_->DestroyVideoReceiveStream(thumbnail_receive_stream); |
| } |
| thumbnail_send_streams_.clear(); |
| thumbnail_receive_streams_.clear(); |
| for (std::unique_ptr<rtc::VideoSourceInterface<VideoFrame>>& video_capturer : |
| thumbnail_capturers_) { |
| video_capturer.reset(); |
| } |
| } |
| |
| void VideoQualityTest::SetupThumbnailCapturers(size_t num_thumbnail_streams) { |
| VideoStream thumbnail = DefaultThumbnailStream(); |
| for (size_t i = 0; i < num_thumbnail_streams; ++i) { |
| auto frame_generator_capturer = |
| std::make_unique<test::FrameGeneratorCapturer>( |
| clock_, |
| test::CreateSquareFrameGenerator(static_cast<int>(thumbnail.width), |
| static_cast<int>(thumbnail.height), |
| absl::nullopt, absl::nullopt), |
| thumbnail.max_framerate, *task_queue_factory_); |
| EXPECT_TRUE(frame_generator_capturer->Init()); |
| thumbnail_capturers_.push_back(std::move(frame_generator_capturer)); |
| } |
| } |
| |
| std::unique_ptr<test::FrameGeneratorInterface> |
| VideoQualityTest::CreateFrameGenerator(size_t video_idx) { |
| // Setup frame generator. |
| const size_t kWidth = 1850; |
| const size_t kHeight = 1110; |
| std::unique_ptr<test::FrameGeneratorInterface> frame_generator; |
| if (params_.screenshare[video_idx].generate_slides) { |
| frame_generator = test::CreateSlideFrameGenerator( |
| kWidth, kHeight, |
| params_.screenshare[video_idx].slide_change_interval * |
| params_.video[video_idx].fps); |
| } else { |
| std::vector<std::string> slides = params_.screenshare[video_idx].slides; |
| if (slides.empty()) { |
| slides.push_back(test::ResourcePath("web_screenshot_1850_1110", "yuv")); |
| slides.push_back(test::ResourcePath("presentation_1850_1110", "yuv")); |
| slides.push_back(test::ResourcePath("photo_1850_1110", "yuv")); |
| slides.push_back(test::ResourcePath("difficult_photo_1850_1110", "yuv")); |
| } |
| if (params_.screenshare[video_idx].scroll_duration == 0) { |
| // Cycle image every slide_change_interval seconds. |
| frame_generator = test::CreateFromYuvFileFrameGenerator( |
| slides, kWidth, kHeight, |
| params_.screenshare[video_idx].slide_change_interval * |
| params_.video[video_idx].fps); |
| } else { |
| RTC_CHECK_LE(params_.video[video_idx].width, kWidth); |
| RTC_CHECK_LE(params_.video[video_idx].height, kHeight); |
| RTC_CHECK_GT(params_.screenshare[video_idx].slide_change_interval, 0); |
| const int kPauseDurationMs = |
| (params_.screenshare[video_idx].slide_change_interval - |
| params_.screenshare[video_idx].scroll_duration) * |
| 1000; |
| RTC_CHECK_LE(params_.screenshare[video_idx].scroll_duration, |
| params_.screenshare[video_idx].slide_change_interval); |
| |
| frame_generator = test::CreateScrollingInputFromYuvFilesFrameGenerator( |
| clock_, slides, kWidth, kHeight, params_.video[video_idx].width, |
| params_.video[video_idx].height, |
| params_.screenshare[video_idx].scroll_duration * 1000, |
| kPauseDurationMs); |
| } |
| } |
| return frame_generator; |
| } |
| |
| void VideoQualityTest::CreateCapturers() { |
| RTC_DCHECK(video_sources_.empty()); |
| video_sources_.resize(num_video_streams_); |
| for (size_t video_idx = 0; video_idx < num_video_streams_; ++video_idx) { |
| std::unique_ptr<test::FrameGeneratorInterface> frame_generator; |
| if (params_.screenshare[video_idx].enabled) { |
| frame_generator = CreateFrameGenerator(video_idx); |
| } else if (params_.video[video_idx].clip_path == "Generator") { |
| frame_generator = test::CreateSquareFrameGenerator( |
| static_cast<int>(params_.video[video_idx].width), |
| static_cast<int>(params_.video[video_idx].height), absl::nullopt, |
| absl::nullopt); |
| } else if (params_.video[video_idx].clip_path == "GeneratorI420A") { |
| frame_generator = test::CreateSquareFrameGenerator( |
| static_cast<int>(params_.video[video_idx].width), |
| static_cast<int>(params_.video[video_idx].height), |
| test::FrameGeneratorInterface::OutputType::kI420A, absl::nullopt); |
| } else if (params_.video[video_idx].clip_path == "GeneratorI010") { |
| frame_generator = test::CreateSquareFrameGenerator( |
| static_cast<int>(params_.video[video_idx].width), |
| static_cast<int>(params_.video[video_idx].height), |
| test::FrameGeneratorInterface::OutputType::kI010, absl::nullopt); |
| } else if (params_.video[video_idx].clip_path == "GeneratorNV12") { |
| frame_generator = test::CreateSquareFrameGenerator( |
| static_cast<int>(params_.video[video_idx].width), |
| static_cast<int>(params_.video[video_idx].height), |
| test::FrameGeneratorInterface::OutputType::kNV12, absl::nullopt); |
| } else if (params_.video[video_idx].clip_path.empty()) { |
| video_sources_[video_idx] = test::CreateVideoCapturer( |
| params_.video[video_idx].width, params_.video[video_idx].height, |
| params_.video[video_idx].fps, |
| params_.video[video_idx].capture_device_index); |
| if (video_sources_[video_idx]) { |
| continue; |
| } else { |
| // Failed to get actual camera, use chroma generator as backup. |
| frame_generator = test::CreateSquareFrameGenerator( |
| static_cast<int>(params_.video[video_idx].width), |
| static_cast<int>(params_.video[video_idx].height), absl::nullopt, |
| absl::nullopt); |
| } |
| } else { |
| frame_generator = test::CreateFromYuvFileFrameGenerator( |
| {params_.video[video_idx].clip_path}, params_.video[video_idx].width, |
| params_.video[video_idx].height, 1); |
| ASSERT_TRUE(frame_generator) << "Could not create capturer for " |
| << params_.video[video_idx].clip_path |
| << ".yuv. Is this file present?"; |
| } |
| ASSERT_TRUE(frame_generator); |
| auto frame_generator_capturer = |
| std::make_unique<test::FrameGeneratorCapturer>( |
| clock_, std::move(frame_generator), params_.video[video_idx].fps, |
| *task_queue_factory_); |
| EXPECT_TRUE(frame_generator_capturer->Init()); |
| video_sources_[video_idx] = std::move(frame_generator_capturer); |
| } |
| } |
| |
| void VideoQualityTest::StartAudioStreams() { |
| audio_send_stream_->Start(); |
| for (AudioReceiveStreamInterface* audio_recv_stream : audio_receive_streams_) |
| audio_recv_stream->Start(); |
| } |
| |
| void VideoQualityTest::StartThumbnails() { |
| for (VideoSendStream* send_stream : thumbnail_send_streams_) |
| send_stream->Start(); |
| for (VideoReceiveStreamInterface* receive_stream : thumbnail_receive_streams_) |
| receive_stream->Start(); |
| } |
| |
| void VideoQualityTest::StopThumbnails() { |
| for (VideoReceiveStreamInterface* receive_stream : thumbnail_receive_streams_) |
| receive_stream->Stop(); |
| for (VideoSendStream* send_stream : thumbnail_send_streams_) |
| send_stream->Stop(); |
| } |
| |
| std::unique_ptr<test::LayerFilteringTransport> |
| VideoQualityTest::CreateSendTransport() { |
| std::unique_ptr<NetworkBehaviorInterface> network_behavior = nullptr; |
| if (injection_components_->sender_network == nullptr) { |
| network_behavior = std::make_unique<SimulatedNetwork>(*params_.config); |
| } else { |
| network_behavior = std::move(injection_components_->sender_network); |
| } |
| return std::make_unique<test::LayerFilteringTransport>( |
| task_queue(), |
| std::make_unique<FakeNetworkPipe>(clock_, std::move(network_behavior)), |
| sender_call_.get(), kPayloadTypeVP8, kPayloadTypeVP9, |
| params_.video[0].selected_tl, params_.ss[0].selected_sl, |
| payload_type_map_, kVideoSendSsrcs[0], |
| static_cast<uint32_t>(kVideoSendSsrcs[0] + params_.ss[0].streams.size() - |
| 1)); |
| } |
| |
| std::unique_ptr<test::DirectTransport> |
| VideoQualityTest::CreateReceiveTransport() { |
| std::unique_ptr<NetworkBehaviorInterface> network_behavior = nullptr; |
| if (injection_components_->receiver_network == nullptr) { |
| network_behavior = std::make_unique<SimulatedNetwork>(*params_.config); |
| } else { |
| network_behavior = std::move(injection_components_->receiver_network); |
| } |
| return std::make_unique<test::DirectTransport>( |
| task_queue(), |
| std::make_unique<FakeNetworkPipe>(clock_, std::move(network_behavior)), |
| receiver_call_.get(), payload_type_map_); |
| } |
| |
| void VideoQualityTest::RunWithAnalyzer(const Params& params) { |
| num_video_streams_ = params.call.dual_video ? 2 : 1; |
| std::unique_ptr<test::LayerFilteringTransport> send_transport; |
| std::unique_ptr<test::DirectTransport> recv_transport; |
| FILE* graph_data_output_file = nullptr; |
| |
| params_ = params; |
| // TODO(ivica): Merge with RunWithRenderer and use a flag / argument to |
| // differentiate between the analyzer and the renderer case. |
| CheckParamsAndInjectionComponents(); |
| |
| if (!params_.analyzer.graph_data_output_filename.empty()) { |
| graph_data_output_file = |
| fopen(params_.analyzer.graph_data_output_filename.c_str(), "w"); |
| RTC_CHECK(graph_data_output_file) |
| << "Can't open the file " << params_.analyzer.graph_data_output_filename |
| << "!"; |
| } |
| |
| if (!params.logging.rtc_event_log_name.empty()) { |
| send_event_log_ = rtc_event_log_factory_.CreateRtcEventLog( |
| RtcEventLog::EncodingType::Legacy); |
| recv_event_log_ = rtc_event_log_factory_.CreateRtcEventLog( |
| RtcEventLog::EncodingType::Legacy); |
| std::unique_ptr<RtcEventLogOutputFile> send_output( |
| std::make_unique<RtcEventLogOutputFile>( |
| params.logging.rtc_event_log_name + "_send", |
| RtcEventLog::kUnlimitedOutput)); |
| std::unique_ptr<RtcEventLogOutputFile> recv_output( |
| std::make_unique<RtcEventLogOutputFile>( |
| params.logging.rtc_event_log_name + "_recv", |
| RtcEventLog::kUnlimitedOutput)); |
| bool event_log_started = |
| send_event_log_->StartLogging(std::move(send_output), |
| RtcEventLog::kImmediateOutput) && |
| recv_event_log_->StartLogging(std::move(recv_output), |
| RtcEventLog::kImmediateOutput); |
| RTC_DCHECK(event_log_started); |
| } else { |
| send_event_log_ = std::make_unique<RtcEventLogNull>(); |
| recv_event_log_ = std::make_unique<RtcEventLogNull>(); |
| } |
| |
| SendTask(task_queue(), [this, ¶ms, &send_transport, &recv_transport]() { |
| Call::Config send_call_config(send_event_log_.get()); |
| Call::Config recv_call_config(recv_event_log_.get()); |
| send_call_config.bitrate_config = params.call.call_bitrate_config; |
| recv_call_config.bitrate_config = params.call.call_bitrate_config; |
| if (params_.audio.enabled) |
| InitializeAudioDevice(&send_call_config, &recv_call_config, |
| params_.audio.use_real_adm); |
| |
| CreateCalls(send_call_config, recv_call_config); |
| send_transport = CreateSendTransport(); |
| recv_transport = CreateReceiveTransport(); |
| }); |
| |
| std::string graph_title = params_.analyzer.graph_title; |
| if (graph_title.empty()) |
| graph_title = VideoQualityTest::GenerateGraphTitle(); |
| bool is_quick_test_enabled = field_trial::IsEnabled("WebRTC-QuickPerfTest"); |
| analyzer_ = std::make_unique<VideoAnalyzer>( |
| send_transport.get(), params_.analyzer.test_label, |
| params_.analyzer.avg_psnr_threshold, params_.analyzer.avg_ssim_threshold, |
| is_quick_test_enabled |
| ? kFramesSentInQuickTest |
| : params_.analyzer.test_durations_secs * params_.video[0].fps, |
| is_quick_test_enabled |
| ? TimeDelta::Millis(1) |
| : TimeDelta::Seconds(params_.analyzer.test_durations_secs), |
| graph_data_output_file, graph_title, |
| kVideoSendSsrcs[params_.ss[0].selected_stream], |
| kSendRtxSsrcs[params_.ss[0].selected_stream], |
| static_cast<size_t>(params_.ss[0].selected_stream), |
| params.ss[0].selected_sl, params_.video[0].selected_tl, |
| is_quick_test_enabled, clock_, params_.logging.rtp_dump_name, |
| task_queue()); |
| |
| SendTask(task_queue(), [&]() { |
| analyzer_->SetCall(sender_call_.get()); |
| analyzer_->SetReceiver(receiver_call_->Receiver()); |
| send_transport->SetReceiver(analyzer_.get()); |
| recv_transport->SetReceiver(sender_call_->Receiver()); |
| |
| SetupVideo(analyzer_.get(), recv_transport.get()); |
| SetupThumbnails(analyzer_.get(), recv_transport.get()); |
| video_receive_configs_[params_.ss[0].selected_stream].renderer = |
| analyzer_.get(); |
| |
| CreateFlexfecStreams(); |
| CreateVideoStreams(); |
| analyzer_->SetSendStream(video_send_streams_[0]); |
| analyzer_->SetReceiveStream( |
| video_receive_streams_[params_.ss[0].selected_stream]); |
| |
| GetVideoSendStream()->SetSource(analyzer_->OutputInterface(), |
| degradation_preference_); |
| SetupThumbnailCapturers(params_.call.num_thumbnails); |
| for (size_t i = 0; i < thumbnail_send_streams_.size(); ++i) { |
| thumbnail_send_streams_[i]->SetSource(thumbnail_capturers_[i].get(), |
| degradation_preference_); |
| } |
| |
| CreateCapturers(); |
| |
| analyzer_->SetSource(video_sources_[0].get(), true); |
| |
| for (size_t video_idx = 1; video_idx < num_video_streams_; ++video_idx) { |
| video_send_streams_[video_idx]->SetSource(video_sources_[video_idx].get(), |
| degradation_preference_); |
| } |
| |
| if (params_.audio.enabled) { |
| SetupAudio(send_transport.get()); |
| StartAudioStreams(); |
| analyzer_->SetAudioReceiveStream(audio_receive_streams_[0]); |
| } |
| StartVideoStreams(); |
| StartThumbnails(); |
| analyzer_->StartMeasuringCpuProcessTime(); |
| }); |
| |
| analyzer_->Wait(); |
| |
| SendTask(task_queue(), [&]() { |
| StopThumbnails(); |
| Stop(); |
| |
| DestroyStreams(); |
| DestroyThumbnailStreams(); |
| |
| if (graph_data_output_file) |
| fclose(graph_data_output_file); |
| |
| send_transport.reset(); |
| recv_transport.reset(); |
| |
| DestroyCalls(); |
| }); |
| analyzer_ = nullptr; |
| } |
| |
| rtc::scoped_refptr<AudioDeviceModule> VideoQualityTest::CreateAudioDevice() { |
| #ifdef WEBRTC_WIN |
| RTC_LOG(LS_INFO) << "Using latest version of ADM on Windows"; |
| // We must initialize the COM library on a thread before we calling any of |
| // the library functions. All COM functions in the ADM will return |
| // CO_E_NOTINITIALIZED otherwise. The legacy ADM for Windows used internal |
| // COM initialization but the new ADM requires COM to be initialized |
| // externally. |
| com_initializer_ = |
| std::make_unique<ScopedCOMInitializer>(ScopedCOMInitializer::kMTA); |
| RTC_CHECK(com_initializer_->Succeeded()); |
| RTC_CHECK(webrtc_win::core_audio_utility::IsSupported()); |
| RTC_CHECK(webrtc_win::core_audio_utility::IsMMCSSSupported()); |
| return CreateWindowsCoreAudioAudioDeviceModule(task_queue_factory_.get()); |
| #else |
| // Use legacy factory method on all platforms except Windows. |
| return AudioDeviceModule::Create(AudioDeviceModule::kPlatformDefaultAudio, |
| task_queue_factory_.get()); |
| #endif |
| } |
| |
| void VideoQualityTest::InitializeAudioDevice(Call::Config* send_call_config, |
| Call::Config* recv_call_config, |
| bool use_real_adm) { |
| rtc::scoped_refptr<AudioDeviceModule> audio_device; |
| if (use_real_adm) { |
| // Run test with real ADM (using default audio devices) if user has |
| // explicitly set the --audio and --use_real_adm command-line flags. |
| audio_device = CreateAudioDevice(); |
| } else { |
| // By default, create a test ADM which fakes audio. |
| audio_device = TestAudioDeviceModule::Create( |
| task_queue_factory_.get(), |
| TestAudioDeviceModule::CreatePulsedNoiseCapturer(32000, 48000), |
| TestAudioDeviceModule::CreateDiscardRenderer(48000), 1.f); |
| } |
| RTC_CHECK(audio_device); |
| |
| AudioState::Config audio_state_config; |
| audio_state_config.audio_mixer = AudioMixerImpl::Create(); |
| audio_state_config.audio_processing = AudioProcessingBuilder().Create(); |
| audio_state_config.audio_device_module = audio_device; |
| send_call_config->audio_state = AudioState::Create(audio_state_config); |
| recv_call_config->audio_state = AudioState::Create(audio_state_config); |
| if (use_real_adm) { |
| // The real ADM requires extra initialization: setting default devices, |
| // setting up number of channels etc. Helper class also calls |
| // AudioDeviceModule::Init(). |
| webrtc::adm_helpers::Init(audio_device.get()); |
| } else { |
| audio_device->Init(); |
| } |
| // Always initialize the ADM before injecting a valid audio transport. |
| RTC_CHECK(audio_device->RegisterAudioCallback( |
| send_call_config->audio_state->audio_transport()) == 0); |
| } |
| |
| void VideoQualityTest::SetupAudio(Transport* transport) { |
| AudioSendStream::Config audio_send_config(transport); |
| audio_send_config.rtp.ssrc = kAudioSendSsrc; |
| |
| // Add extension to enable audio send side BWE, and allow audio bit rate |
| // adaptation. |
| audio_send_config.rtp.extensions.clear(); |
| audio_send_config.send_codec_spec = AudioSendStream::Config::SendCodecSpec( |
| kAudioSendPayloadType, |
| {"OPUS", |
| 48000, |
| 2, |
| {{"usedtx", (params_.audio.dtx ? "1" : "0")}, {"stereo", "1"}}}); |
| |
| if (params_.call.send_side_bwe) { |
| audio_send_config.rtp.extensions.push_back( |
| webrtc::RtpExtension(webrtc::RtpExtension::kTransportSequenceNumberUri, |
| kTransportSequenceNumberExtensionId)); |
| audio_send_config.min_bitrate_bps = kOpusMinBitrateBps; |
| audio_send_config.max_bitrate_bps = kOpusBitrateFbBps; |
| audio_send_config.send_codec_spec->transport_cc_enabled = true; |
| // Only allow ANA when send-side BWE is enabled. |
| audio_send_config.audio_network_adaptor_config = params_.audio.ana_config; |
| } |
| audio_send_config.encoder_factory = audio_encoder_factory_; |
| SetAudioConfig(audio_send_config); |
| |
| std::string sync_group; |
| if (params_.video[0].enabled && params_.audio.sync_video) |
| sync_group = kSyncGroup; |
| |
| CreateMatchingAudioConfigs(transport, sync_group); |
| CreateAudioStreams(); |
| } |
| |
| void VideoQualityTest::RunWithRenderers(const Params& params) { |
| RTC_LOG(LS_INFO) << __FUNCTION__; |
| num_video_streams_ = params.call.dual_video ? 2 : 1; |
| std::unique_ptr<test::LayerFilteringTransport> send_transport; |
| std::unique_ptr<test::DirectTransport> recv_transport; |
| std::unique_ptr<test::VideoRenderer> local_preview; |
| std::vector<std::unique_ptr<test::VideoRenderer>> loopback_renderers; |
| |
| if (!params.logging.rtc_event_log_name.empty()) { |
| send_event_log_ = rtc_event_log_factory_.CreateRtcEventLog( |
| RtcEventLog::EncodingType::Legacy); |
| recv_event_log_ = rtc_event_log_factory_.CreateRtcEventLog( |
| RtcEventLog::EncodingType::Legacy); |
| std::unique_ptr<RtcEventLogOutputFile> send_output( |
| std::make_unique<RtcEventLogOutputFile>( |
| params.logging.rtc_event_log_name + "_send", |
| RtcEventLog::kUnlimitedOutput)); |
| std::unique_ptr<RtcEventLogOutputFile> recv_output( |
| std::make_unique<RtcEventLogOutputFile>( |
| params.logging.rtc_event_log_name + "_recv", |
| RtcEventLog::kUnlimitedOutput)); |
| bool event_log_started = |
| send_event_log_->StartLogging(std::move(send_output), |
| /*output_period_ms=*/5000) && |
| recv_event_log_->StartLogging(std::move(recv_output), |
| /*output_period_ms=*/5000); |
| RTC_DCHECK(event_log_started); |
| } else { |
| send_event_log_ = std::make_unique<RtcEventLogNull>(); |
| recv_event_log_ = std::make_unique<RtcEventLogNull>(); |
| } |
| |
| SendTask(task_queue(), [&]() { |
| params_ = params; |
| CheckParamsAndInjectionComponents(); |
| |
| // TODO(ivica): Remove bitrate_config and use the default Call::Config(), to |
| // match the full stack tests. |
| Call::Config send_call_config(send_event_log_.get()); |
| send_call_config.bitrate_config = params_.call.call_bitrate_config; |
| Call::Config recv_call_config(recv_event_log_.get()); |
| |
| if (params_.audio.enabled) |
| InitializeAudioDevice(&send_call_config, &recv_call_config, |
| params_.audio.use_real_adm); |
| |
| CreateCalls(send_call_config, recv_call_config); |
| |
| // TODO(minyue): consider if this is a good transport even for audio only |
| // calls. |
| send_transport = CreateSendTransport(); |
| |
| recv_transport = CreateReceiveTransport(); |
| |
| // TODO(ivica): Use two calls to be able to merge with RunWithAnalyzer or at |
| // least share as much code as possible. That way this test would also match |
| // the full stack tests better. |
| send_transport->SetReceiver(receiver_call_->Receiver()); |
| recv_transport->SetReceiver(sender_call_->Receiver()); |
| |
| if (params_.video[0].enabled) { |
| // Create video renderers. |
| SetupVideo(send_transport.get(), recv_transport.get()); |
| size_t num_streams_processed = 0; |
| for (size_t video_idx = 0; video_idx < num_video_streams_; ++video_idx) { |
| const size_t selected_stream_id = params_.ss[video_idx].selected_stream; |
| const size_t num_streams = params_.ss[video_idx].streams.size(); |
| if (selected_stream_id == num_streams) { |
| for (size_t stream_id = 0; stream_id < num_streams; ++stream_id) { |
| rtc::StringBuilder oss; |
| oss << "Loopback Video #" << video_idx << " - Stream #" |
| << static_cast<int>(stream_id); |
| loopback_renderers.emplace_back(test::VideoRenderer::Create( |
| oss.str().c_str(), |
| params_.ss[video_idx].streams[stream_id].width, |
| params_.ss[video_idx].streams[stream_id].height)); |
| video_receive_configs_[stream_id + num_streams_processed].renderer = |
| loopback_renderers.back().get(); |
| if (params_.audio.enabled && params_.audio.sync_video) |
| video_receive_configs_[stream_id + num_streams_processed] |
| .sync_group = kSyncGroup; |
| } |
| } else { |
| rtc::StringBuilder oss; |
| oss << "Loopback Video #" << video_idx; |
| loopback_renderers.emplace_back(test::VideoRenderer::Create( |
| oss.str().c_str(), |
| params_.ss[video_idx].streams[selected_stream_id].width, |
| params_.ss[video_idx].streams[selected_stream_id].height)); |
| video_receive_configs_[selected_stream_id + num_streams_processed] |
| .renderer = loopback_renderers.back().get(); |
| if (params_.audio.enabled && params_.audio.sync_video) |
| video_receive_configs_[num_streams_processed + selected_stream_id] |
| .sync_group = kSyncGroup; |
| } |
| num_streams_processed += num_streams; |
| } |
| CreateFlexfecStreams(); |
| CreateVideoStreams(); |
| |
| CreateCapturers(); |
| if (params_.video[0].enabled) { |
| // Create local preview |
| local_preview.reset(test::VideoRenderer::Create( |
| "Local Preview", params_.video[0].width, params_.video[0].height)); |
| |
| video_sources_[0]->AddOrUpdateSink(local_preview.get(), |
| rtc::VideoSinkWants()); |
| } |
| ConnectVideoSourcesToStreams(); |
| } |
| |
| if (params_.audio.enabled) { |
| SetupAudio(send_transport.get()); |
| } |
| |
| Start(); |
| }); |
| |
| PressEnterToContinue(task_queue()); |
| |
| SendTask(task_queue(), [&]() { |
| Stop(); |
| DestroyStreams(); |
| |
| send_transport.reset(); |
| recv_transport.reset(); |
| |
| local_preview.reset(); |
| loopback_renderers.clear(); |
| |
| DestroyCalls(); |
| }); |
| } |
| |
| } // namespace webrtc |