[cleanup] Fix redundant webrtc name specifier

This CL was uploaded by git cl split.

R=hta@webrtc.org


No-Iwyu: LSC
Bug: webrtc:42232595
Change-Id: Ibcffecf0d1e04b73850501f0f2cb88f39384a79e
Reviewed-on: https://webrtc-review.googlesource.com/c/src/+/390620
Reviewed-by: Harald Alvestrand <hta@webrtc.org>
Auto-Submit: Evan Shrubsole <eshr@webrtc.org>
Commit-Queue: Evan Shrubsole <eshr@webrtc.org>
Cr-Commit-Position: refs/heads/main@{#44584}
diff --git a/common_video/bitrate_adjuster_unittest.cc b/common_video/bitrate_adjuster_unittest.cc
index 1ff911c..8660533 100644
--- a/common_video/bitrate_adjuster_unittest.cc
+++ b/common_video/bitrate_adjuster_unittest.cc
@@ -34,7 +34,7 @@
     const size_t frame_size_bytes =
         (bitrate_bps * frame_interval_ms) / (8 * 1000);
     for (size_t i = 0; i < update_frame_interval; ++i) {
-      clock_.AdvanceTime(webrtc::TimeDelta::Millis(frame_interval_ms));
+      clock_.AdvanceTime(TimeDelta::Millis(frame_interval_ms));
       adjuster_.Update(frame_size_bytes);
     }
   }
diff --git a/common_video/libyuv/libyuv_unittest.cc b/common_video/libyuv/libyuv_unittest.cc
index fbaf67c..b0e9794 100644
--- a/common_video/libyuv/libyuv_unittest.cc
+++ b/common_video/libyuv/libyuv_unittest.cc
@@ -97,8 +97,7 @@
       frame_length_(CalcBufferSize(VideoType::kI420, 352, 288)) {}
 
 void TestLibYuv::SetUp() {
-  const std::string input_file_name =
-      webrtc::test::ResourcePath("foreman_cif", "yuv");
+  const std::string input_file_name = test::ResourcePath("foreman_cif", "yuv");
   source_file_ = fopen(input_file_name.c_str(), "rb");
   ASSERT_TRUE(source_file_ != NULL)
       << "Cannot read file: " << input_file_name << "\n";
@@ -106,12 +105,11 @@
   scoped_refptr<I420BufferInterface> buffer(
       test::ReadI420Buffer(width_, height_, source_file_));
 
-  orig_frame_ =
-      std::make_unique<VideoFrame>(VideoFrame::Builder()
-                                       .set_video_frame_buffer(buffer)
-                                       .set_rotation(webrtc::kVideoRotation_0)
-                                       .set_timestamp_us(0)
-                                       .build());
+  orig_frame_ = std::make_unique<VideoFrame>(VideoFrame::Builder()
+                                                 .set_video_frame_buffer(buffer)
+                                                 .set_rotation(kVideoRotation_0)
+                                                 .set_timestamp_us(0)
+                                                 .build());
 }
 
 void TestLibYuv::TearDown() {
@@ -125,7 +123,7 @@
   // Reading YUV frame - testing on the first frame of the foreman sequence
   int j = 0;
   std::string output_file_name =
-      webrtc::test::OutputPath() + "LibYuvTest_conversion.yuv";
+      test::OutputPath() + "LibYuvTest_conversion.yuv";
   FILE* output_file = fopen(output_file_name.c_str(), "wb");
   ASSERT_TRUE(output_file != NULL);
 
@@ -291,7 +289,7 @@
 TEST_F(TestLibYuv, ConvertAlignedFrame) {
   // Reading YUV frame - testing on the first frame of the foreman sequence
   std::string output_file_name =
-      webrtc::test::OutputPath() + "LibYuvTest_conversion.yuv";
+      test::OutputPath() + "LibYuvTest_conversion.yuv";
   FILE* output_file = fopen(output_file_name.c_str(), "wb");
   ASSERT_TRUE(output_file != NULL);
 
diff --git a/common_video/video_frame_unittest.cc b/common_video/video_frame_unittest.cc
index c2611aa..21ddf8d 100644
--- a/common_video/video_frame_unittest.cc
+++ b/common_video/video_frame_unittest.cc
@@ -146,7 +146,7 @@
 template <class T>
 void CheckRotate(int width,
                  int height,
-                 webrtc::VideoRotation rotation,
+                 VideoRotation rotation,
                  const T& rotated) {
   int rotated_width = width;
   int rotated_height = height;
@@ -204,7 +204,7 @@
   VideoFrame frame =
       VideoFrame::Builder()
           .set_video_frame_buffer(I420Buffer::Create(10, 10, 10, 14, 90))
-          .set_rotation(webrtc::kVideoRotation_0)
+          .set_rotation(kVideoRotation_0)
           .set_timestamp_ms(789)
           .build();
   const int valid_value = 10;
@@ -251,10 +251,8 @@
   VideoFrame frame2(frame1);
 
   EXPECT_EQ(frame1.video_frame_buffer(), frame2.video_frame_buffer());
-  const webrtc::I420BufferInterface* yuv1 =
-      frame1.video_frame_buffer()->GetI420();
-  const webrtc::I420BufferInterface* yuv2 =
-      frame2.video_frame_buffer()->GetI420();
+  const I420BufferInterface* yuv1 = frame1.video_frame_buffer()->GetI420();
+  const I420BufferInterface* yuv2 = frame2.video_frame_buffer()->GetI420();
   EXPECT_EQ(yuv1->DataY(), yuv2->DataY());
   EXPECT_EQ(yuv1->DataU(), yuv2->DataU());
   EXPECT_EQ(yuv1->DataV(), yuv2->DataV());
@@ -276,8 +274,8 @@
 }
 
 TEST(TestVideoFrame, TextureInitialValues) {
-  VideoFrame frame = test::FakeNativeBuffer::CreateFrame(
-      640, 480, 100, 10, webrtc::kVideoRotation_0);
+  VideoFrame frame =
+      test::FakeNativeBuffer::CreateFrame(640, 480, 100, 10, kVideoRotation_0);
   EXPECT_EQ(640, frame.width());
   EXPECT_EQ(480, frame.height());
   EXPECT_EQ(100u, frame.rtp_timestamp());
@@ -425,7 +423,7 @@
 TYPED_TEST_SUITE_P(TestPlanarYuvBufferRotate);
 
 TYPED_TEST_P(TestPlanarYuvBufferRotate, Rotates) {
-  for (const webrtc::VideoRotation& rotation : this->RotationParams) {
+  for (const VideoRotation& rotation : this->RotationParams) {
     scoped_refptr<TypeParam> buffer = CreateGradient<TypeParam>(640, 480);
     scoped_refptr<TypeParam> rotated_buffer =
         TypeParam::Rotate(*buffer, rotation);
diff --git a/logging/rtc_event_log/encoder/rtc_event_log_encoder_new_format.cc b/logging/rtc_event_log/encoder/rtc_event_log_encoder_new_format.cc
index d6fce46..38bf7cf 100644
--- a/logging/rtc_event_log/encoder/rtc_event_log_encoder_new_format.cc
+++ b/logging/rtc_event_log/encoder/rtc_event_log_encoder_new_format.cc
@@ -170,19 +170,19 @@
 }
 
 rtclog2::DtlsTransportStateEvent::DtlsTransportState ConvertToProtoFormat(
-    webrtc::DtlsTransportState state) {
+    DtlsTransportState state) {
   switch (state) {
-    case webrtc::DtlsTransportState::kNew:
+    case DtlsTransportState::kNew:
       return rtclog2::DtlsTransportStateEvent::DTLS_TRANSPORT_NEW;
-    case webrtc::DtlsTransportState::kConnecting:
+    case DtlsTransportState::kConnecting:
       return rtclog2::DtlsTransportStateEvent::DTLS_TRANSPORT_CONNECTING;
-    case webrtc::DtlsTransportState::kConnected:
+    case DtlsTransportState::kConnected:
       return rtclog2::DtlsTransportStateEvent::DTLS_TRANSPORT_CONNECTED;
-    case webrtc::DtlsTransportState::kClosed:
+    case DtlsTransportState::kClosed:
       return rtclog2::DtlsTransportStateEvent::DTLS_TRANSPORT_CLOSED;
-    case webrtc::DtlsTransportState::kFailed:
+    case DtlsTransportState::kFailed:
       return rtclog2::DtlsTransportStateEvent::DTLS_TRANSPORT_FAILED;
-    case webrtc::DtlsTransportState::kNumValues:
+    case DtlsTransportState::kNumValues:
       RTC_DCHECK_NOTREACHED();
   }
   RTC_DCHECK_NOTREACHED();
diff --git a/logging/rtc_event_log/rtc_event_log_parser.cc b/logging/rtc_event_log/rtc_event_log_parser.cc
index aff0181..61ab4f5 100644
--- a/logging/rtc_event_log/rtc_event_log_parser.cc
+++ b/logging/rtc_event_log/rtc_event_log_parser.cc
@@ -1123,8 +1123,7 @@
 // is not stored in the event log.
 // TODO(ivoc): Remove this once this mapping is stored in the event log for
 //             audio streams. Tracking bug: webrtc:6399
-webrtc::RtpHeaderExtensionMap
-ParsedRtcEventLog::GetDefaultHeaderExtensionMap() {
+RtpHeaderExtensionMap ParsedRtcEventLog::GetDefaultHeaderExtensionMap() {
   // Values from before the default RTP header extension IDs were removed.
   constexpr int kAudioLevelDefaultId = 1;
   constexpr int kTimestampOffsetDefaultId = 2;
@@ -1136,7 +1135,7 @@
   constexpr int kVideoTimingDefaultId = 8;
   constexpr int kDependencyDescriptorDefaultId = 9;
 
-  webrtc::RtpHeaderExtensionMap default_map(/*extmap_allow_mixed=*/true);
+  RtpHeaderExtensionMap default_map(/*extmap_allow_mixed=*/true);
   default_map.Register<AudioLevelExtension>(kAudioLevelDefaultId);
   default_map.Register<TransmissionOffset>(kTimestampOffsetDefaultId);
   default_map.Register<AbsoluteSendTime>(kAbsSendTimeDefaultId);
diff --git a/video/buffered_frame_decryptor.cc b/video/buffered_frame_decryptor.cc
index 29ec8ae..03fc2eb 100644
--- a/video/buffered_frame_decryptor.cc
+++ b/video/buffered_frame_decryptor.cc
@@ -73,7 +73,7 @@
   }
   // Retrieve the maximum possible size of the decrypted payload.
   const size_t max_plaintext_byte_size =
-      frame_decryptor_->GetMaxPlaintextByteSize(webrtc::MediaType::VIDEO,
+      frame_decryptor_->GetMaxPlaintextByteSize(MediaType::VIDEO,
                                                 frame->size());
   RTC_CHECK_LE(max_plaintext_byte_size, frame->size());
   // Place the decrypted frame inline into the existing frame.
@@ -88,9 +88,8 @@
 
   // Attempt to decrypt the video frame.
   const FrameDecryptorInterface::Result decrypt_result =
-      frame_decryptor_->Decrypt(webrtc::MediaType::VIDEO, /*csrcs=*/{},
-                                additional_data, *frame,
-                                inline_decrypted_bitstream);
+      frame_decryptor_->Decrypt(MediaType::VIDEO, /*csrcs=*/{}, additional_data,
+                                *frame, inline_decrypted_bitstream);
   // Optionally call the callback if there was a change in status
   if (decrypt_result.status != last_status_) {
     last_status_ = decrypt_result.status;
diff --git a/video/config/encoder_stream_factory.cc b/video/config/encoder_stream_factory.cc
index 601c750..8820c98 100644
--- a/video/config/encoder_stream_factory.cc
+++ b/video/config/encoder_stream_factory.cc
@@ -42,7 +42,6 @@
 namespace webrtc {
 namespace {
 
-using ::webrtc::FieldTrialsView;
 
 const int kMinLayerSize = 16;
 
@@ -73,10 +72,10 @@
 }
 
 bool IsTemporalLayersSupported(VideoCodecType codec_type) {
-  return codec_type == webrtc::VideoCodecType::kVideoCodecVP8 ||
-         codec_type == webrtc::VideoCodecType::kVideoCodecVP9 ||
-         codec_type == webrtc::VideoCodecType::kVideoCodecAV1 ||
-         codec_type == webrtc::VideoCodecType::kVideoCodecH265;
+  return codec_type == VideoCodecType::kVideoCodecVP8 ||
+         codec_type == VideoCodecType::kVideoCodecVP9 ||
+         codec_type == VideoCodecType::kVideoCodecAV1 ||
+         codec_type == VideoCodecType::kVideoCodecH265;
 }
 
 size_t FindRequiredActiveLayers(const VideoEncoderConfig& encoder_config) {
@@ -111,14 +110,14 @@
 
 int GetDefaultMaxQp(VideoCodecType codec_type) {
   switch (codec_type) {
-    case webrtc::kVideoCodecH264:
-    case webrtc::kVideoCodecH265:
+    case kVideoCodecH264:
+    case kVideoCodecH265:
       return kDefaultVideoMaxQpH26x;
-    case webrtc::kVideoCodecVP8:
-    case webrtc::kVideoCodecVP9:
-    case webrtc::kVideoCodecGeneric:
+    case kVideoCodecVP8:
+    case kVideoCodecVP9:
+    case kVideoCodecGeneric:
       return kDefaultVideoMaxQpVpx;
-    case webrtc::kVideoCodecAV1:
+    case kVideoCodecAV1:
       return kDefaultVideoMaxQpAv1;
   }
 }
@@ -131,7 +130,7 @@
                            size_t simulcast_layers) {
   int base2_exponent = static_cast<int>(simulcast_layers) - 1;
   const std::optional<int> experimental_base2_exponent =
-      webrtc::NormalizeSimulcastSizeExperiment::GetBase2Exponent(field_trials);
+      NormalizeSimulcastSizeExperiment::GetBase2Exponent(field_trials);
   if (experimental_base2_exponent &&
       (size > (1 << *experimental_base2_exponent))) {
     base2_exponent = *experimental_base2_exponent;
@@ -149,10 +148,10 @@
 
   // Allow an experiment to override the minimum bitrate for the lowest
   // spatial layer. The experiment's configuration has the lowest priority.
-  layers[0].min_bitrate_bps = experimental_min_bitrate
-                                  .value_or(webrtc::DataRate::BitsPerSec(
-                                      webrtc::kDefaultMinVideoBitrateBps))
-                                  .bps<int>();
+  layers[0].min_bitrate_bps =
+      experimental_min_bitrate
+          .value_or(DataRate::BitsPerSec(kDefaultMinVideoBitrateBps))
+          .bps<int>();
 
   const bool temporal_layers_supported =
       IsTemporalLayersSupported(encoder_config.codec_type);
@@ -220,14 +219,14 @@
 
   bool is_highest_layer_max_bitrate_configured =
       encoder_config.simulcast_layers[layers.size() - 1].max_bitrate_bps > 0;
-  bool is_screencast = encoder_config.content_type ==
-                       webrtc::VideoEncoderConfig::ContentType::kScreen;
+  bool is_screencast =
+      encoder_config.content_type == VideoEncoderConfig::ContentType::kScreen;
   if (!is_screencast && !is_highest_layer_max_bitrate_configured &&
       encoder_config.max_bitrate_bps > 0) {
     // No application-configured maximum for the largest layer.
     // If there is bitrate leftover, give it to the largest layer.
-    webrtc::BoostMaxSimulcastLayer(
-        webrtc::DataRate::BitsPerSec(encoder_config.max_bitrate_bps), &layers);
+    BoostMaxSimulcastLayer(DataRate::BitsPerSec(encoder_config.max_bitrate_bps),
+                           &layers);
   }
 
   // Sort the layers by max_bitrate_bps, they might not always be from
@@ -259,7 +258,7 @@
 }  // namespace
 
 EncoderStreamFactory::EncoderStreamFactory(
-    const webrtc::VideoEncoder::EncoderInfo& encoder_info,
+    const VideoEncoder::EncoderInfo& encoder_info,
     std::optional<VideoSourceRestrictions> restrictions)
     : encoder_info_requested_resolution_alignment_(
           encoder_info.requested_resolution_alignment),
@@ -296,7 +295,7 @@
 
   std::vector<VideoStream> streams;
   if (is_simulcast ||
-      webrtc::SimulcastUtility::IsConferenceModeScreenshare(encoder_config)) {
+      SimulcastUtility::IsConferenceModeScreenshare(encoder_config)) {
     streams = CreateSimulcastOrConferenceModeScreenshareStreams(
         trials, frame_width, frame_height, encoder_config,
         experimental_min_bitrate);
@@ -318,8 +317,8 @@
     int height,
     const VideoEncoderConfig& encoder_config,
     const std::optional<DataRate>& experimental_min_bitrate) const {
-  bool is_screencast = encoder_config.content_type ==
-                       webrtc::VideoEncoderConfig::ContentType::kScreen;
+  bool is_screencast =
+      encoder_config.content_type == VideoEncoderConfig::ContentType::kScreen;
 
   // The max bitrate specified by the API.
   // - `encoder_config.simulcast_layers[0].max_bitrate_bps` comes from the first
@@ -346,8 +345,8 @@
 
   int min_bitrate_bps =
       experimental_min_bitrate
-          ? webrtc::saturated_cast<int>(experimental_min_bitrate->bps())
-          : webrtc::kDefaultMinVideoBitrateBps;
+          ? saturated_cast<int>(experimental_min_bitrate->bps())
+          : kDefaultMinVideoBitrateBps;
   if (encoder_config.simulcast_layers[0].min_bitrate_bps > 0) {
     // Use set min bitrate.
     min_bitrate_bps = encoder_config.simulcast_layers[0].min_bitrate_bps;
@@ -388,7 +387,7 @@
         kMinLayerSize);
   }
 
-  if (encoder_config.codec_type == webrtc::VideoCodecType::kVideoCodecVP9) {
+  if (encoder_config.codec_type == VideoCodecType::kVideoCodecVP9) {
     RTC_DCHECK(encoder_config.encoder_specific_settings);
     // Use VP9 SVC layering from codec settings which might be initialized
     // though field trial in ConfigureVideoEncoderSettings.
@@ -407,12 +406,12 @@
         (layer.num_temporal_layers > 1u || num_spatial_layers > 1)) {
       // In SVC mode, the VP9 max bitrate is determined by SvcConfig, instead of
       // GetMaxDefaultVideoBitrateKbps().
-      std::vector<webrtc::SpatialLayer> svc_layers =
-          webrtc::GetSvcConfig(width, height, max_framerate,
-                               /*first_active_layer=*/0, num_spatial_layers,
-                               *layer.num_temporal_layers, is_screencast);
+      std::vector<SpatialLayer> svc_layers =
+          GetSvcConfig(width, height, max_framerate,
+                       /*first_active_layer=*/0, num_spatial_layers,
+                       *layer.num_temporal_layers, is_screencast);
       int sum_max_bitrates_kbps = 0;
-      for (const webrtc::SpatialLayer& spatial_layer : svc_layers) {
+      for (const SpatialLayer& spatial_layer : svc_layers) {
         sum_max_bitrates_kbps += spatial_layer.maxBitrate;
       }
       RTC_DCHECK_GE(sum_max_bitrates_kbps, 0);
@@ -468,9 +467,9 @@
 
   // Use legacy simulcast screenshare if conference mode is explicitly enabled
   // or use the regular simulcast configuration path which is generic.
-  std::vector<VideoStream> layers = webrtc::GetSimulcastConfig(
+  std::vector<VideoStream> layers = GetSimulcastConfig(
       resolutions,
-      webrtc::SimulcastUtility::IsConferenceModeScreenshare(encoder_config),
+      SimulcastUtility::IsConferenceModeScreenshare(encoder_config),
       IsTemporalLayersSupported(encoder_config.codec_type), trials,
       encoder_config.codec_type);
 
@@ -509,8 +508,8 @@
     VideoSinkWants wants;
     wants.is_active = true;
     wants.target_pixel_count = restrictions_->target_pixels_per_frame();
-    wants.max_pixel_count = webrtc::dchecked_cast<int>(
-        restrictions_->max_pixels_per_frame().value_or(
+    wants.max_pixel_count =
+        dchecked_cast<int>(restrictions_->max_pixels_per_frame().value_or(
             std::numeric_limits<int>::max()));
     wants.aggregates.emplace(VideoSinkWants::Aggregates());
     wants.resolution_alignment = encoder_info_requested_resolution_alignment_;
@@ -532,7 +531,7 @@
     int height,
     const VideoEncoderConfig& encoder_config) const {
   std::vector<Resolution> resolutions;
-  if (webrtc::SimulcastUtility::IsConferenceModeScreenshare(encoder_config)) {
+  if (SimulcastUtility::IsConferenceModeScreenshare(encoder_config)) {
     for (size_t i = 0; i < encoder_config.number_of_streams; ++i) {
       resolutions.push_back({.width = width, .height = height});
     }
@@ -540,7 +539,7 @@
     size_t min_num_layers = FindRequiredActiveLayers(encoder_config);
     size_t max_num_layers =
         !encoder_config.HasScaleResolutionDownTo()
-            ? webrtc::LimitSimulcastLayerCount(
+            ? LimitSimulcastLayerCount(
                   min_num_layers, encoder_config.number_of_streams, width,
                   height, trials, encoder_config.codec_type)
             : encoder_config.number_of_streams;
@@ -556,8 +555,8 @@
     if (encoder_config.HasScaleResolutionDownTo() &&
         restrictions_.has_value() &&
         restrictions_->max_pixels_per_frame().has_value()) {
-      int max_pixels = webrtc::dchecked_cast<int>(
-          restrictions_->max_pixels_per_frame().value());
+      int max_pixels =
+          dchecked_cast<int>(restrictions_->max_pixels_per_frame().value());
       int prev_pixel_count =
           encoder_config.simulcast_layers[0]
               .scale_resolution_down_to.value_or(Resolution())
@@ -592,8 +591,7 @@
     }
 
     const bool norm_size_configured =
-        webrtc::NormalizeSimulcastSizeExperiment::GetBase2Exponent(trials)
-            .has_value();
+        NormalizeSimulcastSizeExperiment::GetBase2Exponent(trials).has_value();
     const int normalized_width =
         (default_scale_factors_used || norm_size_configured) &&
                 (width >= kMinLayerSize)
diff --git a/video/config/encoder_stream_factory_unittest.cc b/video/config/encoder_stream_factory_unittest.cc
index a36efaa..5c93527 100644
--- a/video/config/encoder_stream_factory_unittest.cc
+++ b/video/config/encoder_stream_factory_unittest.cc
@@ -42,7 +42,6 @@
 using ::testing::SizeIs;
 using ::testing::TestWithParam;
 using ::testing::Values;
-using ::webrtc::EncoderStreamFactory;
 
 struct CreateVideoStreamParams {
   int width = 0;
@@ -482,7 +481,7 @@
                           .scalability_mode = ScalabilityMode::kL1T2})}}),
             Values(
 #ifdef RTC_ENABLE_H265
-                webrtc::kVideoCodecH265,
+                kVideoCodecH265,
 #endif
                 VideoCodecType::kVideoCodecAV1,
                 VideoCodecType::kVideoCodecVP9)));
diff --git a/video/config/simulcast.cc b/video/config/simulcast.cc
index 66e9005..acc2e15 100644
--- a/video/config/simulcast.cc
+++ b/video/config/simulcast.cc
@@ -37,7 +37,6 @@
 
 namespace {
 
-using ::webrtc::FieldTrialsView;
 
 constexpr char kUseLegacySimulcastLayerLimitFieldTrial[] =
     "WebRTC-LegacySimulcastLayerLimit";
@@ -47,16 +46,16 @@
 // Limits for legacy conference screensharing mode. Currently used for the
 // lower of the two simulcast streams.
 constexpr DataRate kScreenshareDefaultTl0Bitrate =
-    webrtc::DataRate::KilobitsPerSec(200);
+    DataRate::KilobitsPerSec(200);
 constexpr DataRate kScreenshareDefaultTl1Bitrate =
-    webrtc::DataRate::KilobitsPerSec(1000);
+    DataRate::KilobitsPerSec(1000);
 
 // Min/max bitrate for the higher one of the two simulcast stream used for
 // screen content.
 constexpr DataRate kScreenshareHighStreamMinBitrate =
-    webrtc::DataRate::KilobitsPerSec(600);
+    DataRate::KilobitsPerSec(600);
 constexpr DataRate kScreenshareHighStreamMaxBitrate =
-    webrtc::DataRate::KilobitsPerSec(1250);
+    DataRate::KilobitsPerSec(1250);
 
 constexpr int kDefaultNumTemporalLayers = 3;
 constexpr int kScreenshareMaxSimulcastLayers = 2;
@@ -83,62 +82,47 @@
 // simulcast layers at what bitrates (maximum, target, and minimum).
 // Important!! Keep this table from high resolution to low resolution.
 constexpr const SimulcastFormat kSimulcastFormatsVP8[] = {
-    {1920, 1080, 3, webrtc::DataRate::KilobitsPerSec(5000),
-     webrtc::DataRate::KilobitsPerSec(4000),
-     webrtc::DataRate::KilobitsPerSec(800)},
-    {1280, 720, 3, webrtc::DataRate::KilobitsPerSec(2500),
-     webrtc::DataRate::KilobitsPerSec(2500),
-     webrtc::DataRate::KilobitsPerSec(600)},
-    {960, 540, 3, webrtc::DataRate::KilobitsPerSec(1200),
-     webrtc::DataRate::KilobitsPerSec(1200),
-     webrtc::DataRate::KilobitsPerSec(350)},
-    {640, 360, 2, webrtc::DataRate::KilobitsPerSec(700),
-     webrtc::DataRate::KilobitsPerSec(500),
-     webrtc::DataRate::KilobitsPerSec(150)},
-    {480, 270, 2, webrtc::DataRate::KilobitsPerSec(450),
-     webrtc::DataRate::KilobitsPerSec(350),
-     webrtc::DataRate::KilobitsPerSec(150)},
-    {320, 180, 1, webrtc::DataRate::KilobitsPerSec(200),
-     webrtc::DataRate::KilobitsPerSec(150),
-     webrtc::DataRate::KilobitsPerSec(30)},
+    {1920, 1080, 3, DataRate::KilobitsPerSec(5000),
+     DataRate::KilobitsPerSec(4000), DataRate::KilobitsPerSec(800)},
+    {1280, 720, 3, DataRate::KilobitsPerSec(2500),
+     DataRate::KilobitsPerSec(2500), DataRate::KilobitsPerSec(600)},
+    {960, 540, 3, DataRate::KilobitsPerSec(1200),
+     DataRate::KilobitsPerSec(1200), DataRate::KilobitsPerSec(350)},
+    {640, 360, 2, DataRate::KilobitsPerSec(700), DataRate::KilobitsPerSec(500),
+     DataRate::KilobitsPerSec(150)},
+    {480, 270, 2, DataRate::KilobitsPerSec(450), DataRate::KilobitsPerSec(350),
+     DataRate::KilobitsPerSec(150)},
+    {320, 180, 1, DataRate::KilobitsPerSec(200), DataRate::KilobitsPerSec(150),
+     DataRate::KilobitsPerSec(30)},
     // As the resolution goes down, interpolate the target and max bitrates down
     // towards zero. The min bitrate is still limited at 30 kbps and the target
     // and the max will be capped from below accordingly.
-    {0, 0, 1, webrtc::DataRate::KilobitsPerSec(0),
-     webrtc::DataRate::KilobitsPerSec(0),
-     webrtc::DataRate::KilobitsPerSec(30)}};
+    {0, 0, 1, DataRate::KilobitsPerSec(0), DataRate::KilobitsPerSec(0),
+     DataRate::KilobitsPerSec(30)}};
 
 // These tables describe from which resolution we can use how many
 // simulcast layers at what bitrates (maximum, target, and minimum).
 // Important!! Keep this table from high resolution to low resolution.
 constexpr const SimulcastFormat kSimulcastFormatsVP9[] = {
-    {1920, 1080, 3, webrtc::DataRate::KilobitsPerSec(3367),
-     webrtc::DataRate::KilobitsPerSec(3367),
-     webrtc::DataRate::KilobitsPerSec(769)},
-    {1280, 720, 3, webrtc::DataRate::KilobitsPerSec(1524),
-     webrtc::DataRate::KilobitsPerSec(1524),
-     webrtc::DataRate::KilobitsPerSec(481)},
-    {960, 540, 3, webrtc::DataRate::KilobitsPerSec(879),
-     webrtc::DataRate::KilobitsPerSec(879),
-     webrtc::DataRate::KilobitsPerSec(337)},
-    {640, 360, 2, webrtc::DataRate::KilobitsPerSec(420),
-     webrtc::DataRate::KilobitsPerSec(420),
-     webrtc::DataRate::KilobitsPerSec(193)},
-    {480, 270, 2, webrtc::DataRate::KilobitsPerSec(257),
-     webrtc::DataRate::KilobitsPerSec(257),
-     webrtc::DataRate::KilobitsPerSec(121)},
-    {320, 180, 1, webrtc::DataRate::KilobitsPerSec(142),
-     webrtc::DataRate::KilobitsPerSec(142),
-     webrtc::DataRate::KilobitsPerSec(30)},
-    {240, 135, 1, webrtc::DataRate::KilobitsPerSec(101),
-     webrtc::DataRate::KilobitsPerSec(101),
-     webrtc::DataRate::KilobitsPerSec(30)},
+    {1920, 1080, 3, DataRate::KilobitsPerSec(3367),
+     DataRate::KilobitsPerSec(3367), DataRate::KilobitsPerSec(769)},
+    {1280, 720, 3, DataRate::KilobitsPerSec(1524),
+     DataRate::KilobitsPerSec(1524), DataRate::KilobitsPerSec(481)},
+    {960, 540, 3, DataRate::KilobitsPerSec(879), DataRate::KilobitsPerSec(879),
+     DataRate::KilobitsPerSec(337)},
+    {640, 360, 2, DataRate::KilobitsPerSec(420), DataRate::KilobitsPerSec(420),
+     DataRate::KilobitsPerSec(193)},
+    {480, 270, 2, DataRate::KilobitsPerSec(257), DataRate::KilobitsPerSec(257),
+     DataRate::KilobitsPerSec(121)},
+    {320, 180, 1, DataRate::KilobitsPerSec(142), DataRate::KilobitsPerSec(142),
+     DataRate::KilobitsPerSec(30)},
+    {240, 135, 1, DataRate::KilobitsPerSec(101), DataRate::KilobitsPerSec(101),
+     DataRate::KilobitsPerSec(30)},
     // As the resolution goes down, interpolate the target and max bitrates down
     // towards zero. The min bitrate is still limited at 30 kbps and the target
     // and the max will be capped from below accordingly.
-    {0, 0, 1, webrtc::DataRate::KilobitsPerSec(0),
-     webrtc::DataRate::KilobitsPerSec(0),
-     webrtc::DataRate::KilobitsPerSec(30)}};
+    {0, 0, 1, DataRate::KilobitsPerSec(0), DataRate::KilobitsPerSec(0),
+     DataRate::KilobitsPerSec(30)}};
 
 constexpr DataRate Interpolate(const DataRate& a,
                                const DataRate& b,
@@ -154,16 +138,16 @@
 
 int GetDefaultSimulcastTemporalLayers(VideoCodecType codec) {
   switch (codec) {
-    case webrtc::kVideoCodecVP8:
-    case webrtc::kVideoCodecVP9:
-    case webrtc::kVideoCodecAV1:
-    case webrtc::kVideoCodecH264:
-    case webrtc::kVideoCodecGeneric:
+    case kVideoCodecVP8:
+    case kVideoCodecVP9:
+    case kVideoCodecAV1:
+    case kVideoCodecH264:
+    case kVideoCodecGeneric:
       return kDefaultNumTemporalLayers;
     // For codec type that has no software fallback, defaults to L1T1 for
     // initial simulcast setup, as this is the only scalability mode secure to
     // be supported.
-    case webrtc::kVideoCodecH265:
+    case kVideoCodecH265:
       return 1;
   }
   RTC_DCHECK_NOTREACHED() << "Unsupported codec.";
@@ -175,15 +159,15 @@
     VideoCodecType codec) {
   std::vector<SimulcastFormat> formats;
   switch (codec) {
-    case webrtc::kVideoCodecGeneric:
-    case webrtc::kVideoCodecVP8:
-    case webrtc::kVideoCodecH264:
+    case kVideoCodecGeneric:
+    case kVideoCodecVP8:
+    case kVideoCodecH264:
       formats.insert(formats.begin(), std::begin(kSimulcastFormatsVP8),
                      std::end(kSimulcastFormatsVP8));
       break;
-    case webrtc::kVideoCodecVP9:
-    case webrtc::kVideoCodecAV1:
-    case webrtc::kVideoCodecH265:
+    case kVideoCodecVP9:
+    case kVideoCodecAV1:
+    case kVideoCodecH265:
       formats.insert(formats.begin(), std::begin(kSimulcastFormatsVP9),
                      std::end(kSimulcastFormatsVP9));
       break;
@@ -323,7 +307,7 @@
   layers[0].width = width;
   layers[0].height = height;
   layers[0].max_framerate = 5;
-  layers[0].min_bitrate_bps = webrtc::kDefaultMinVideoBitrateBps;
+  layers[0].min_bitrate_bps = kDefaultMinVideoBitrateBps;
   layers[0].target_bitrate_bps = kScreenshareDefaultTl0Bitrate.bps();
   layers[0].max_bitrate_bps = kScreenshareDefaultTl1Bitrate.bps();
   layers[0].num_temporal_layers = temporal_layers_supported ? 2 : 1;
@@ -340,7 +324,7 @@
       // layers were enabled.
       max_bitrate_bps = static_cast<int>(
           kScreenshareHighStreamMaxBitrate.bps() *
-          webrtc::SimulcastRateAllocator::GetTemporalRateAllocation(
+          SimulcastRateAllocator::GetTemporalRateAllocation(
               kScreenshareTemporalLayers, 0, base_heavy_tl3_rate_alloc));
     } else {
       // Experimental temporal layer mode used, use increased max bitrate.
@@ -378,8 +362,8 @@
     // than configured `max_ratio`. pixels_down is the selected index in
     // kSimulcastFormats based on pixels.
     FieldTrialOptional<double> max_ratio("max_ratio");
-    webrtc::ParseFieldTrial({&max_ratio},
-                            trials.Lookup("WebRTC-SimulcastLayerLimitRoundUp"));
+    ParseFieldTrial({&max_ratio},
+                    trials.Lookup("WebRTC-SimulcastLayerLimitRoundUp"));
 
     size_t reduced_num_layers =
         std::max(min_num_layers,
@@ -413,14 +397,14 @@
 
 DataRate GetTotalMaxBitrate(const std::vector<VideoStream>& layers) {
   if (layers.empty())
-    return webrtc::DataRate::Zero();
+    return DataRate::Zero();
 
   int total_max_bitrate_bps = 0;
   for (size_t s = 0; s < layers.size() - 1; ++s) {
     total_max_bitrate_bps += layers[s].target_bitrate_bps;
   }
   total_max_bitrate_bps += layers.back().max_bitrate_bps;
-  return webrtc::DataRate::BitsPerSec(total_max_bitrate_bps);
+  return DataRate::BitsPerSec(total_max_bitrate_bps);
 }
 
 std::vector<VideoStream> GetSimulcastConfig(
diff --git a/video/config/simulcast_unittest.cc b/video/config/simulcast_unittest.cc
index 94b48fe..628e4ea 100644
--- a/video/config/simulcast_unittest.cc
+++ b/video/config/simulcast_unittest.cc
@@ -73,7 +73,7 @@
 std::vector<Resolution> CreateResolutions(int max_width,
                                           int max_height,
                                           int num_streams) {
-  std::vector<webrtc::Resolution> resolutions(num_streams);
+  std::vector<Resolution> resolutions(num_streams);
   for (int i = 0; i < num_streams; ++i) {
     resolutions[i].width = max_width >> (num_streams - i - 1);
     resolutions[i].height = max_height >> (num_streams - i - 1);
@@ -107,10 +107,10 @@
   streams[1].target_bitrate_bps = 200000;
   streams[2].max_bitrate_bps = 400000;
 
-  const webrtc::DataRate one_bps = webrtc::DataRate::BitsPerSec(1);
+  const DataRate one_bps = DataRate::BitsPerSec(1);
 
   // No bitrate above the total max to give to the highest stream.
-  const webrtc::DataRate max_total_bitrate = GetTotalMaxBitrate(streams);
+  const DataRate max_total_bitrate = GetTotalMaxBitrate(streams);
   BoostMaxSimulcastLayer(max_total_bitrate, &streams);
   EXPECT_EQ(400000, streams[2].max_bitrate_bps);
   EXPECT_EQ(max_total_bitrate, GetTotalMaxBitrate(streams));
@@ -129,7 +129,7 @@
   const size_t kMaxLayers = 3;
   std::vector<VideoStream> streams =
       GetSimulcastConfig(CreateResolutions(1280, 720, kMaxLayers),
-                         !kScreenshare, true, trials, webrtc::kVideoCodecVP8);
+                         !kScreenshare, true, trials, kVideoCodecVP8);
 
   ASSERT_THAT(streams, SizeIs(kMaxLayers));
   EXPECT_EQ(320u, streams[0].width);
@@ -157,7 +157,7 @@
   const size_t kMaxLayers = 3;
   std::vector<VideoStream> streams =
       GetSimulcastConfig(CreateResolutions(1280, 720, kMaxLayers),
-                         !kScreenshare, true, trials, webrtc::kVideoCodecH265);
+                         !kScreenshare, true, trials, kVideoCodecH265);
 
   ASSERT_THAT(streams, SizeIs(kMaxLayers));
   EXPECT_EQ(320u, streams[0].width);
@@ -188,7 +188,7 @@
   const size_t kMaxLayers = 3;
   std::vector<VideoStream> streams =
       GetSimulcastConfig(CreateResolutions(1280, 720, kMaxLayers),
-                         !kScreenshare, true, trials, webrtc::kVideoCodecVP8);
+                         !kScreenshare, true, trials, kVideoCodecVP8);
 
   ASSERT_THAT(streams, SizeIs(kMaxLayers));
   EXPECT_EQ(kExpected[0].min_bitrate_bps, streams[0].min_bitrate_bps);
@@ -209,7 +209,7 @@
   const size_t kMaxLayers = 2;
   std::vector<VideoStream> streams =
       GetSimulcastConfig(CreateResolutions(1280, 720, kMaxLayers),
-                         !kScreenshare, true, trials, webrtc::kVideoCodecVP8);
+                         !kScreenshare, true, trials, kVideoCodecVP8);
 
   ASSERT_THAT(streams, SizeIs(kMaxLayers));
   EXPECT_EQ(640u, streams[0].width);
@@ -224,7 +224,7 @@
       std::vector<Resolution>{{.width = 1400, .height = 800},
                               {.width = 1400, .height = 800},
                               {.width = 1400, .height = 800}},
-      kScreenshare, true, trials, webrtc::kVideoCodecVP8);
+      kScreenshare, true, trials, kVideoCodecVP8);
 
   EXPECT_THAT(streams, SizeIs(2));
   for (size_t i = 0; i < streams.size(); ++i) {
@@ -244,7 +244,7 @@
   ExplicitKeyValueConfig trials("");
   std::vector<VideoStream> streams = GetSimulcastConfig(
       std::vector<Resolution>{{.width = 1400, .height = 800}}, kScreenshare,
-      true, trials, webrtc::kVideoCodecVP8);
+      true, trials, kVideoCodecVP8);
   EXPECT_THAT(streams, SizeIs(1));
 }
 
@@ -252,7 +252,7 @@
   ExplicitKeyValueConfig trials("");
   std::vector<VideoStream> streams =
       GetSimulcastConfig(std::vector<Resolution>{{.width = 900, .height = 800}},
-                         !kScreenshare, true, trials, webrtc::kVideoCodecVP8);
+                         !kScreenshare, true, trials, kVideoCodecVP8);
 
   ASSERT_THAT(streams, SizeIs(1));
   EXPECT_EQ(900u, streams[0].width);
@@ -273,7 +273,7 @@
 
   std::vector<VideoStream> streams =
       GetSimulcastConfig(CreateResolutions(kWidth, kHeight, kMaxLayers),
-                         !kScreenshare, true, trials, webrtc::kVideoCodecVP8);
+                         !kScreenshare, true, trials, kVideoCodecVP8);
 
   ASSERT_THAT(streams, SizeIs(kMaxLayers));
   EXPECT_EQ(kWidth, streams[2].width);
@@ -296,18 +296,18 @@
   const int kMaxLayers = 3;
 
   size_t num_layers = LimitSimulcastLayerCount(kMinLayers, kMaxLayers, 960, 540,
-                                               trials, webrtc::kVideoCodecVP8);
+                                               trials, kVideoCodecVP8);
   EXPECT_EQ(num_layers, 3u);
   // <960x540: 2 layers
   num_layers = LimitSimulcastLayerCount(kMinLayers, kMaxLayers, 960, 539,
-                                        trials, webrtc::kVideoCodecVP8);
+                                        trials, kVideoCodecVP8);
   EXPECT_EQ(num_layers, 2u);
   num_layers = LimitSimulcastLayerCount(kMinLayers, kMaxLayers, 480, 270,
-                                        trials, webrtc::kVideoCodecVP8);
+                                        trials, kVideoCodecVP8);
   EXPECT_EQ(num_layers, 2u);
   // <480x270: 1 layer
   num_layers = LimitSimulcastLayerCount(kMinLayers, kMaxLayers, 480, 269,
-                                        trials, webrtc::kVideoCodecVP8);
+                                        trials, kVideoCodecVP8);
   EXPECT_EQ(num_layers, 1u);
 }
 
@@ -318,24 +318,24 @@
   const int kMaxLayers = 3;
 
   size_t num_layers = LimitSimulcastLayerCount(kMinLayers, kMaxLayers, 960, 540,
-                                               trials, webrtc::kVideoCodecVP8);
+                                               trials, kVideoCodecVP8);
   EXPECT_EQ(num_layers, 3u);
   // Lowest cropped height where max layers from higher resolution is used.
   num_layers = LimitSimulcastLayerCount(kMinLayers, kMaxLayers, 960, 512,
-                                        trials, webrtc::kVideoCodecVP8);
+                                        trials, kVideoCodecVP8);
   EXPECT_EQ(num_layers, 3u);
   num_layers = LimitSimulcastLayerCount(kMinLayers, kMaxLayers, 960, 508,
-                                        trials, webrtc::kVideoCodecVP8);
+                                        trials, kVideoCodecVP8);
   EXPECT_EQ(num_layers, 2u);
   num_layers = LimitSimulcastLayerCount(kMinLayers, kMaxLayers, 480, 270,
-                                        trials, webrtc::kVideoCodecVP8);
+                                        trials, kVideoCodecVP8);
   EXPECT_EQ(num_layers, 2u);
   // Lowest cropped height where max layers from higher resolution is used.
   num_layers = LimitSimulcastLayerCount(kMinLayers, kMaxLayers, 480, 256,
-                                        trials, webrtc::kVideoCodecVP8);
+                                        trials, kVideoCodecVP8);
   EXPECT_EQ(num_layers, 2u);
   num_layers = LimitSimulcastLayerCount(kMinLayers, kMaxLayers, 480, 254,
-                                        trials, webrtc::kVideoCodecVP8);
+                                        trials, kVideoCodecVP8);
   EXPECT_EQ(num_layers, 1u);
 }
 
@@ -347,14 +347,14 @@
   const int kMaxLayers = 3;
 
   size_t num_layers = LimitSimulcastLayerCount(kMinLayers, kMaxLayers, 480, 270,
-                                               trials, webrtc::kVideoCodecVP8);
+                                               trials, kVideoCodecVP8);
   EXPECT_EQ(num_layers, 2u);
   // Lowest cropped height where max layers from higher resolution is used.
   num_layers = LimitSimulcastLayerCount(kMinLayers, kMaxLayers, 480, 252,
-                                        trials, webrtc::kVideoCodecVP8);
+                                        trials, kVideoCodecVP8);
   EXPECT_EQ(num_layers, 2u);
   num_layers = LimitSimulcastLayerCount(kMinLayers, kMaxLayers, 480, 250,
-                                        trials, webrtc::kVideoCodecVP8);
+                                        trials, kVideoCodecVP8);
   EXPECT_EQ(num_layers, 1u);
 }
 
@@ -366,7 +366,7 @@
   const size_t kMaxLayers = 3;
   std::vector<VideoStream> streams = GetSimulcastConfig(
       CreateResolutions(/*max_width=*/960, /*max_height=*/540, kMaxLayers),
-      !kScreenshare, true, trials, webrtc::kVideoCodecVP8);
+      !kScreenshare, true, trials, kVideoCodecVP8);
 
   ASSERT_THAT(streams, SizeIs(kMaxLayers));
   EXPECT_EQ(240u, streams[0].width);
@@ -383,7 +383,7 @@
 
   std::vector<VideoStream> streams =
       GetSimulcastConfig(std::vector<Resolution>{{.width = 1, .height = 1}},
-                         !kScreenshare, true, trials, webrtc::kVideoCodecVP8);
+                         !kScreenshare, true, trials, kVideoCodecVP8);
 
   ASSERT_THAT(streams, SizeIs(1));
   EXPECT_EQ(1u, streams[0].width);
@@ -401,7 +401,7 @@
   const size_t kMaxLayers = 3;
   std::vector<VideoStream> streams = GetSimulcastConfig(
       CreateResolutions(/*max_width=*/960, /*max_height=*/540, kMaxLayers),
-      !kScreenshare, true, trials, webrtc::kVideoCodecVP8);
+      !kScreenshare, true, trials, kVideoCodecVP8);
 
   ASSERT_THAT(streams, SizeIs(kMaxLayers));
   EXPECT_EQ(240u, streams[0].width);
@@ -419,10 +419,10 @@
 };
 
 using BitrateLimitsTest =
-    TestWithParam<std::tuple<webrtc::VideoCodecType, BitrateLimitsTestParams>>;
+    TestWithParam<std::tuple<VideoCodecType, BitrateLimitsTestParams>>;
 
 TEST_P(BitrateLimitsTest, VerifyBitrateLimits) {
-  const auto codec_type = std::get<webrtc::VideoCodecType>(GetParam());
+  const auto codec_type = std::get<VideoCodecType>(GetParam());
   const auto test_params = std::get<BitrateLimitsTestParams>(GetParam());
   ExplicitKeyValueConfig trials("");
   std::vector<VideoStream> streams = GetSimulcastConfig(
@@ -442,7 +442,7 @@
 INSTANTIATE_TEST_SUITE_P(
     Vp8H264,
     BitrateLimitsTest,
-    Combine(Values(webrtc::kVideoCodecVP8, webrtc::kVideoCodecH264),
+    Combine(Values(kVideoCodecVP8, kVideoCodecH264),
             Values(BitrateLimitsTestParams{.width = 1920,
                                            .height = 1080,
                                            .expected_min_bitrate_kbps{150, 350,
@@ -460,8 +460,7 @@
                        .expected_min_bitrate_kbps{30, 150, 350},
                        .expected_max_bitrate_kbps{200, 450, 1200}})),
     [](const TestParamInfo<BitrateLimitsTest::ParamType>& info) {
-      return CodecTypeToPayloadString(
-                 std::get<webrtc::VideoCodecType>(info.param)) +
+      return CodecTypeToPayloadString(std::get<VideoCodecType>(info.param)) +
              std::to_string(
                  std::get<BitrateLimitsTestParams>(info.param).height);
     });
@@ -472,10 +471,10 @@
     Combine(
         Values(
 #ifdef RTC_ENABLE_H265
-            webrtc::kVideoCodecH265,
+            kVideoCodecH265,
 #endif
-            webrtc::kVideoCodecAV1,
-            webrtc::kVideoCodecVP9),
+            kVideoCodecAV1,
+            kVideoCodecVP9),
         Values(
             BitrateLimitsTestParams{.width = 1920,
                                     .height = 1080,
@@ -491,8 +490,7 @@
                 .expected_min_bitrate_kbps{30, 121, 337},
                 .expected_max_bitrate_kbps{101, 257, 879}})),
     [](const TestParamInfo<BitrateLimitsTest::ParamType>& info) {
-      return CodecTypeToPayloadString(
-                 std::get<webrtc::VideoCodecType>(info.param)) +
+      return CodecTypeToPayloadString(std::get<VideoCodecType>(info.param)) +
              std::to_string(
                  std::get<BitrateLimitsTestParams>(info.param).height);
     });
@@ -506,7 +504,7 @@
   const size_t kMaxLayers = 3;
   std::vector<VideoStream> streams =
       GetSimulcastConfig(CreateResolutions(1280, 720, kMaxLayers),
-                         !kScreenshare, true, trials, webrtc::kVideoCodecH265);
+                         !kScreenshare, true, trials, kVideoCodecH265);
 
   ASSERT_THAT(streams, SizeIs(kMaxLayers));
   for (size_t i = 0; i < streams.size(); ++i) {
diff --git a/video/corruption_detection/frame_instrumentation_generator_unittest.cc b/video/corruption_detection/frame_instrumentation_generator_unittest.cc
index d4362c0..fff4cb4 100644
--- a/video/corruption_detection/frame_instrumentation_generator_unittest.cc
+++ b/video/corruption_detection/frame_instrumentation_generator_unittest.cc
@@ -685,7 +685,7 @@
       VideoCodecType::kVideoCodecVP8);
 
   bool frames_destroyed[4] = {};
-  class TestBuffer : public webrtc::I420Buffer {
+  class TestBuffer : public I420Buffer {
    public:
     TestBuffer(int width, int height, bool* frame_destroyed_indicator)
         : I420Buffer(width, height),
diff --git a/video/end_to_end_tests/bandwidth_tests.cc b/video/end_to_end_tests/bandwidth_tests.cc
index d327a76..2f6ec09 100644
--- a/video/end_to_end_tests/bandwidth_tests.cc
+++ b/video/end_to_end_tests/bandwidth_tests.cc
@@ -327,7 +327,7 @@
 TEST_F(BandwidthEndToEndTest, ReportsSetEncoderRates) {
   // If these fields trial are on, we get lower bitrates than expected by this
   // test, due to the packetization overhead and encoder pushback.
-  webrtc::test::ScopedFieldTrials field_trials(
+  test::ScopedFieldTrials field_trials(
       std::string(field_trial::GetFieldTrialString()) +
       "WebRTC-VideoRateControl/bitrate_adjuster:false/");
   class EncoderRateStatsTest : public test::EndToEndTest,
diff --git a/video/end_to_end_tests/resolution_bitrate_limits_tests.cc b/video/end_to_end_tests/resolution_bitrate_limits_tests.cc
index a8503d9..786535a 100644
--- a/video/end_to_end_tests/resolution_bitrate_limits_tests.cc
+++ b/video/end_to_end_tests/resolution_bitrate_limits_tests.cc
@@ -228,7 +228,7 @@
 };
 
 TEST_P(ResolutionBitrateLimitsTest, LimitsApplied) {
-  webrtc::test::ScopedFieldTrials field_trials(
+  test::ScopedFieldTrials field_trials(
       "WebRTC-GetEncoderInfoOverride/"
       "frame_size_pixels:921600,"
       "min_start_bitrate_bps:0,"
@@ -272,7 +272,7 @@
 }
 TEST_F(ResolutionBitrateLimitsWithScalabilityModeTest,
        OneStreamLimitsAppliedForOneSpatialLayer) {
-  webrtc::test::ScopedFieldTrials field_trials(
+  test::ScopedFieldTrials field_trials(
       "WebRTC-GetEncoderInfoOverride/"
       "frame_size_pixels:921600,"
       "min_start_bitrate_bps:0,"
@@ -291,7 +291,7 @@
 
 TEST_F(ResolutionBitrateLimitsWithScalabilityModeTest,
        OneStreamLimitsNotAppliedForMultipleSpatialLayers) {
-  webrtc::test::ScopedFieldTrials field_trials(
+  test::ScopedFieldTrials field_trials(
       "WebRTC-GetEncoderInfoOverride/"
       "frame_size_pixels:230400|921600,"
       "min_start_bitrate_bps:0|0,"
@@ -324,7 +324,7 @@
 }
 
 TEST_P(ResolutionBitrateLimitsTest, IntersectionApplied) {
-  webrtc::test::ScopedFieldTrials field_trials(
+  test::ScopedFieldTrials field_trials(
       "WebRTC-GetEncoderInfoOverride/"
       "frame_size_pixels:921600,"
       "min_start_bitrate_bps:0,"
@@ -343,7 +343,7 @@
 }
 
 TEST_P(ResolutionBitrateLimitsTest, LimitsAppliedMiddleActive) {
-  webrtc::test::ScopedFieldTrials field_trials(
+  test::ScopedFieldTrials field_trials(
       "WebRTC-GetEncoderInfoOverride/"
       "frame_size_pixels:230400|921600,"
       "min_start_bitrate_bps:0|0,"
@@ -360,7 +360,7 @@
 }
 
 TEST_P(ResolutionBitrateLimitsTest, EncodingMinMaxBitrateAppliedMiddleActive) {
-  webrtc::test::ScopedFieldTrials field_trials(
+  test::ScopedFieldTrials field_trials(
       "WebRTC-GetEncoderInfoOverride/"
       "frame_size_pixels:230400|921600,"
       "min_start_bitrate_bps:0|0,"
@@ -383,7 +383,7 @@
 }
 
 TEST_P(ResolutionBitrateLimitsTest, MinBitrateNotAboveEncodingMax) {
-  webrtc::test::ScopedFieldTrials field_trials(
+  test::ScopedFieldTrials field_trials(
       "WebRTC-GetEncoderInfoOverride/"
       "frame_size_pixels:230400|921600,"
       "min_start_bitrate_bps:0|0,"
@@ -404,7 +404,7 @@
 }
 
 TEST_P(ResolutionBitrateLimitsTest, MaxBitrateNotBelowEncodingMin) {
-  webrtc::test::ScopedFieldTrials field_trials(
+  test::ScopedFieldTrials field_trials(
       "WebRTC-GetEncoderInfoOverride/"
       "frame_size_pixels:230400|921600,"
       "min_start_bitrate_bps:0|0,"
@@ -464,7 +464,7 @@
 }
 
 TEST_P(ResolutionBitrateLimitsTest, LimitsAppliedHighestActive) {
-  webrtc::test::ScopedFieldTrials field_trials(
+  test::ScopedFieldTrials field_trials(
       "WebRTC-GetEncoderInfoOverride/"
       "frame_size_pixels:230400|921600,"
       "min_start_bitrate_bps:0|0,"
@@ -481,7 +481,7 @@
 }
 
 TEST_P(ResolutionBitrateLimitsTest, EncodingMinMaxBitrateAppliedHighestActive) {
-  webrtc::test::ScopedFieldTrials field_trials(
+  test::ScopedFieldTrials field_trials(
       "WebRTC-GetEncoderInfoOverride/"
       "frame_size_pixels:230400|921600,"
       "min_start_bitrate_bps:0|0,"
@@ -506,7 +506,7 @@
 }
 
 TEST_P(ResolutionBitrateLimitsTest, LimitsNotAppliedLowestActive) {
-  webrtc::test::ScopedFieldTrials field_trials(
+  test::ScopedFieldTrials field_trials(
       "WebRTC-GetEncoderInfoOverride/"
       "frame_size_pixels:230400|921600,"
       "min_start_bitrate_bps:0|0,"
@@ -527,7 +527,7 @@
 
 TEST_F(ResolutionBitrateLimitsWithScalabilityModeTest,
        LimitsAppliedForVp9OneSpatialLayer) {
-  webrtc::test::ScopedFieldTrials field_trials(
+  test::ScopedFieldTrials field_trials(
       "WebRTC-GetEncoderInfoOverride/"
       "frame_size_pixels:230400|921600,"
       "min_start_bitrate_bps:0|0,"
@@ -547,7 +547,7 @@
 
 TEST_F(ResolutionBitrateLimitsWithScalabilityModeTest,
        LimitsNotAppliedForVp9MultipleSpatialLayers) {
-  webrtc::test::ScopedFieldTrials field_trials(
+  test::ScopedFieldTrials field_trials(
       "WebRTC-GetEncoderInfoOverride/"
       "frame_size_pixels:230400|921600,"
       "min_start_bitrate_bps:0|0,"
@@ -570,7 +570,7 @@
 
 TEST_F(ResolutionBitrateLimitsWithScalabilityModeTest,
        OneStreamLimitsAppliedForAv1OneSpatialLayer) {
-  webrtc::test::ScopedFieldTrials field_trials(
+  test::ScopedFieldTrials field_trials(
       "WebRTC-GetEncoderInfoOverride/"
       "frame_size_pixels:921600,"
       "min_start_bitrate_bps:0,"
@@ -589,7 +589,7 @@
 
 TEST_F(ResolutionBitrateLimitsWithScalabilityModeTest,
        LimitsAppliedForAv1SingleSpatialLayer) {
-  webrtc::test::ScopedFieldTrials field_trials(
+  test::ScopedFieldTrials field_trials(
       "WebRTC-GetEncoderInfoOverride/"
       "frame_size_pixels:230400|921600,"
       "min_start_bitrate_bps:0|0,"
@@ -609,7 +609,7 @@
 
 TEST_F(ResolutionBitrateLimitsWithScalabilityModeTest,
        EncodingMinMaxBitrateAppliedForAv1SingleSpatialLayer) {
-  webrtc::test::ScopedFieldTrials field_trials(
+  test::ScopedFieldTrials field_trials(
       "WebRTC-GetEncoderInfoOverride/"
       "frame_size_pixels:921600,"
       "min_start_bitrate_bps:0,"
@@ -631,7 +631,7 @@
 
 TEST_F(ResolutionBitrateLimitsWithScalabilityModeTest,
        LimitsNotAppliedForAv1MultipleSpatialLayers) {
-  webrtc::test::ScopedFieldTrials field_trials(
+  test::ScopedFieldTrials field_trials(
       "WebRTC-GetEncoderInfoOverride/"
       "frame_size_pixels:230400|921600,"
       "min_start_bitrate_bps:0|0,"
@@ -653,7 +653,7 @@
 }
 
 TEST_P(ResolutionBitrateLimitsTest, LimitsNotAppliedSimulcast) {
-  webrtc::test::ScopedFieldTrials field_trials(
+  test::ScopedFieldTrials field_trials(
       "WebRTC-GetEncoderInfoOverride/"
       "frame_size_pixels:230400|921600,"
       "min_start_bitrate_bps:0|0,"
diff --git a/video/end_to_end_tests/rtp_rtcp_tests.cc b/video/end_to_end_tests/rtp_rtcp_tests.cc
index 712d4a8..1e0f1d7 100644
--- a/video/end_to_end_tests/rtp_rtcp_tests.cc
+++ b/video/end_to_end_tests/rtp_rtcp_tests.cc
@@ -362,7 +362,7 @@
         // to wait for a second for periodic Sender Report.
         rtcp::RapidResyncRequest force_send_sr_back_request;
         Buffer packet = force_send_sr_back_request.Build();
-        static_cast<webrtc::Transport*>(receive_transport_.get())
+        static_cast<Transport*>(receive_transport_.get())
             ->SendRtcp(packet, /*packet_options=*/{});
       }
       CreateFrameGeneratorCapturer(30, 1280, 720);
diff --git a/video/frame_cadence_adapter_unittest.cc b/video/frame_cadence_adapter_unittest.cc
index b40171e..285b917 100644
--- a/video/frame_cadence_adapter_unittest.cc
+++ b/video/frame_cadence_adapter_unittest.cc
@@ -1295,7 +1295,7 @@
 
 TEST_F(ZeroHertzQueueOverloadTest,
        QueueOverloadIsDisabledForZeroHerzWhenKillSwitchIsEnabled) {
-  webrtc::test::ScopedKeyValueConfig field_trials(
+  test::ScopedKeyValueConfig field_trials(
       field_trials_, "WebRTC-ZeroHertzQueueOverload/Disabled/");
   adapter_.reset();
   adapter_ = CreateAdapter(field_trials, time_controller_.GetClock());
diff --git a/video/frame_decode_timing.cc b/video/frame_decode_timing.cc
index 04b8e22..9871ba9 100644
--- a/video/frame_decode_timing.cc
+++ b/video/frame_decode_timing.cc
@@ -22,8 +22,7 @@
 
 namespace webrtc {
 
-FrameDecodeTiming::FrameDecodeTiming(Clock* clock,
-                                     webrtc::VCMTiming const* timing)
+FrameDecodeTiming::FrameDecodeTiming(Clock* clock, VCMTiming const* timing)
     : clock_(clock), timing_(timing) {
   RTC_DCHECK(clock_);
   RTC_DCHECK(timing_);
diff --git a/video/frame_decode_timing_unittest.cc b/video/frame_decode_timing_unittest.cc
index e4359b5..da7a098 100644
--- a/video/frame_decode_timing_unittest.cc
+++ b/video/frame_decode_timing_unittest.cc
@@ -35,10 +35,10 @@
 
 namespace {
 
-class FakeVCMTiming : public webrtc::VCMTiming {
+class FakeVCMTiming : public VCMTiming {
  public:
   explicit FakeVCMTiming(Clock* clock, const FieldTrialsView& field_trials)
-      : webrtc::VCMTiming(clock, field_trials) {}
+      : VCMTiming(clock, field_trials) {}
 
   Timestamp RenderTime(uint32_t frame_timestamp, Timestamp now) const override {
     RTC_DCHECK(render_time_map_.contains(frame_timestamp));
diff --git a/video/frame_encode_metadata_writer_unittest.cc b/video/frame_encode_metadata_writer_unittest.cc
index bb584b8..4580caa 100644
--- a/video/frame_encode_metadata_writer_unittest.cc
+++ b/video/frame_encode_metadata_writer_unittest.cc
@@ -386,8 +386,7 @@
   bitrate_allocation.SetBitrate(0, 0, 500000);
   encode_timer.OnSetRates(bitrate_allocation, 30);
 
-  webrtc::ColorSpace color_space =
-      CreateTestColorSpace(/*with_hdr_metadata=*/true);
+  ColorSpace color_space = CreateTestColorSpace(/*with_hdr_metadata=*/true);
   image.SetRtpTimestamp(static_cast<uint32_t>(kTimestampMs * 90));
   VideoFrame frame = VideoFrame::Builder()
                          .set_timestamp_ms(kTimestampMs)
diff --git a/video/full_stack_tests.cc b/video/full_stack_tests.cc
index 8245e6f..f8c809c 100644
--- a/video/full_stack_tests.cc
+++ b/video/full_stack_tests.cc
@@ -978,7 +978,7 @@
 #if !defined(WEBRTC_MAC)
 
 TEST(FullStackTest, Vp9ksvc_3sl_High) {
-  webrtc::test::ScopedFieldTrials override_trials(
+  test::ScopedFieldTrials override_trials(
       AppendFieldTrials("WebRTC-Vp9IssueKeyFrameOnLayerDeactivation/Enabled/"));
   auto fixture = CreateVideoQualityTestFixture();
   ParamsWithLogging simulcast;
@@ -993,7 +993,7 @@
 }
 
 TEST(FullStackTest, Vp9ksvc_3sl_Low) {
-  webrtc::test::ScopedFieldTrials override_trials(
+  test::ScopedFieldTrials override_trials(
       AppendFieldTrials("WebRTC-Vp9IssueKeyFrameOnLayerDeactivation/Enabled/"));
   auto fixture = CreateVideoQualityTestFixture();
   ParamsWithLogging simulcast;
@@ -1008,7 +1008,7 @@
 }
 
 TEST(FullStackTest, Vp9ksvc_3sl_Low_Bw_Limited) {
-  webrtc::test::ScopedFieldTrials override_trials(
+  test::ScopedFieldTrials override_trials(
       AppendFieldTrials("WebRTC-Vp9IssueKeyFrameOnLayerDeactivation/Enabled/"));
   auto fixture = CreateVideoQualityTestFixture();
   ParamsWithLogging simulcast;
@@ -1024,7 +1024,7 @@
 }
 
 TEST(FullStackTest, Vp9ksvc_3sl_Medium_Network_Restricted) {
-  webrtc::test::ScopedFieldTrials override_trials(
+  test::ScopedFieldTrials override_trials(
       AppendFieldTrials("WebRTC-Vp9IssueKeyFrameOnLayerDeactivation/Enabled/"));
   auto fixture = CreateVideoQualityTestFixture();
   ParamsWithLogging simulcast;
@@ -1042,7 +1042,7 @@
 
 // TODO(webrtc:9722): Remove when experiment is cleaned up.
 TEST(FullStackTest, Vp9ksvc_3sl_Medium_Network_Restricted_Trusted_Rate) {
-  webrtc::test::ScopedFieldTrials override_trials(
+  test::ScopedFieldTrials override_trials(
       AppendFieldTrials("WebRTC-Vp9IssueKeyFrameOnLayerDeactivation/Enabled/"));
   auto fixture = CreateVideoQualityTestFixture();
   ParamsWithLogging simulcast;
@@ -1087,7 +1087,7 @@
   simulcast.ss[0] = {
       streams, 2, 1, 0, InterLayerPredMode::kOn, std::vector<SpatialLayer>(),
       true};
-  webrtc::test::ScopedFieldTrials override_trials(AppendFieldTrials(
+  test::ScopedFieldTrials override_trials(AppendFieldTrials(
       "WebRTC-ForceSimulatedOveruseIntervalMs/1000-50000-300/"));
   fixture->RunWithAnalyzer(simulcast);
 }
diff --git a/video/pc_full_stack_tests.cc b/video/pc_full_stack_tests.cc
index 841f961..f0c4c0e 100644
--- a/video/pc_full_stack_tests.cc
+++ b/video/pc_full_stack_tests.cc
@@ -39,14 +39,14 @@
 
 namespace webrtc {
 
-using ::webrtc::webrtc_pc_e2e::AudioConfig;
-using ::webrtc::webrtc_pc_e2e::EmulatedSFUConfig;
-using ::webrtc::webrtc_pc_e2e::PeerConfigurer;
-using ::webrtc::webrtc_pc_e2e::RunParams;
-using ::webrtc::webrtc_pc_e2e::ScreenShareConfig;
-using ::webrtc::webrtc_pc_e2e::VideoCodecConfig;
-using ::webrtc::webrtc_pc_e2e::VideoConfig;
-using ::webrtc::webrtc_pc_e2e::VideoSimulcastConfig;
+using webrtc_pc_e2e::AudioConfig;
+using webrtc_pc_e2e::EmulatedSFUConfig;
+using webrtc_pc_e2e::PeerConfigurer;
+using webrtc_pc_e2e::RunParams;
+using webrtc_pc_e2e::ScreenShareConfig;
+using webrtc_pc_e2e::VideoCodecConfig;
+using webrtc_pc_e2e::VideoConfig;
+using webrtc_pc_e2e::VideoSimulcastConfig;
 
 namespace {
 
@@ -1235,7 +1235,7 @@
 #if defined(RTC_ENABLE_VP9)
 
 TEST(PCFullStackTest, Pc_Screenshare_Slides_Vp9_3sl_High_Fps) {
-  webrtc::test::ScopedFieldTrials override_trials(
+  test::ScopedFieldTrials override_trials(
       AppendFieldTrials("WebRTC-Vp9InterLayerPred/"
                         "Enabled,inter_layer_pred_mode:on/"));
   std::unique_ptr<NetworkEmulationManager> network_emulation_manager =
@@ -1269,7 +1269,7 @@
 }
 
 TEST(PCFullStackTest, Pc_Vp9svc_3sl_High) {
-  webrtc::test::ScopedFieldTrials override_trials(
+  test::ScopedFieldTrials override_trials(
       AppendFieldTrials("WebRTC-Vp9InterLayerPred/"
                         "Enabled,inter_layer_pred_mode:on/"));
   std::unique_ptr<NetworkEmulationManager> network_emulation_manager =
@@ -1302,7 +1302,7 @@
 }
 
 TEST(PCFullStackTest, Pc_Vp9svc_3sl_Low) {
-  webrtc::test::ScopedFieldTrials override_trials(
+  test::ScopedFieldTrials override_trials(
       AppendFieldTrials("WebRTC-Vp9InterLayerPred/"
                         "Enabled,inter_layer_pred_mode:on/"));
   std::unique_ptr<NetworkEmulationManager> network_emulation_manager =
@@ -1434,7 +1434,7 @@
 #define MAYBE_Pc_Simulcast_HD_High Pc_Simulcast_HD_High
 #endif
 TEST(PCFullStackTest, MAYBE_Pc_Simulcast_HD_High) {
-  webrtc::test::ScopedFieldTrials override_trials(AppendFieldTrials(
+  test::ScopedFieldTrials override_trials(AppendFieldTrials(
       "WebRTC-ForceSimulatedOveruseIntervalMs/1000-50000-300/"));
   std::unique_ptr<NetworkEmulationManager> network_emulation_manager =
       CreateNetworkEmulationManager();
diff --git a/video/receive_statistics_proxy.cc b/video/receive_statistics_proxy.cc
index 22d46e4..b89ae42 100644
--- a/video/receive_statistics_proxy.cc
+++ b/video/receive_statistics_proxy.cc
@@ -614,7 +614,7 @@
                                             VideoContentType content_type,
                                             VideoFrameType frame_type) {
   TimeDelta processing_delay = TimeDelta::Zero();
-  webrtc::Timestamp current_time = clock_->CurrentTime();
+  Timestamp current_time = clock_->CurrentTime();
   // TODO(bugs.webrtc.org/13984): some tests do not fill packet_infos().
   TimeDelta assembly_time = TimeDelta::Zero();
   if (frame.packet_infos().size() > 0) {
diff --git a/video/receive_statistics_proxy_unittest.cc b/video/receive_statistics_proxy_unittest.cc
index a198cd6..5e06231 100644
--- a/video/receive_statistics_proxy_unittest.cc
+++ b/video/receive_statistics_proxy_unittest.cc
@@ -128,7 +128,7 @@
 
 TEST_F(ReceiveStatisticsProxyTest, OnDecodedFrameIncreasesFramesDecoded) {
   EXPECT_EQ(0u, statistics_proxy_->GetStats().frames_decoded);
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   for (uint32_t i = 1; i <= 3; ++i) {
     statistics_proxy_->OnDecodedFrame(frame, std::nullopt, TimeDelta::Zero(),
                                       VideoContentType::UNSPECIFIED,
@@ -141,7 +141,7 @@
   const Frequency kFps = Frequency::Hertz(20);
   const int kRequiredSamples =
       TimeDelta::Seconds(metrics::kMinRunTimeInSeconds) * kFps;
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   for (int i = 0; i < kRequiredSamples; ++i) {
     statistics_proxy_->OnDecodedFrame(frame, std::nullopt, TimeDelta::Zero(),
                                       VideoContentType::UNSPECIFIED,
@@ -159,7 +159,7 @@
   const Frequency kFps = Frequency::Hertz(20);
   const int kRequiredSamples =
       TimeDelta::Seconds(metrics::kMinRunTimeInSeconds) * kFps;
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   for (int i = 0; i < kRequiredSamples - 1; ++i) {
     statistics_proxy_->OnDecodedFrame(frame, std::nullopt, TimeDelta::Zero(),
                                       VideoContentType::UNSPECIFIED,
@@ -174,7 +174,7 @@
 TEST_F(ReceiveStatisticsProxyTest,
        OnDecodedFrameWithQpDoesNotResetFramesDecodedOrTotalDecodeTime) {
   EXPECT_EQ(0u, statistics_proxy_->GetStats().frames_decoded);
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   TimeDelta expected_total_decode_time = TimeDelta::Zero();
   unsigned int expected_frames_decoded = 0;
   for (uint32_t i = 1; i <= 3; ++i) {
@@ -204,7 +204,7 @@
 TEST_F(ReceiveStatisticsProxyTest, OnDecodedFrameIncreasesProcessingDelay) {
   const TimeDelta kProcessingDelay = TimeDelta::Millis(10);
   EXPECT_EQ(0u, statistics_proxy_->GetStats().frames_decoded);
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   TimeDelta expected_total_processing_delay = TimeDelta::Zero();
   unsigned int expected_frames_decoded = 0;
   // We set receive time fixed and increase the clock by 10ms
@@ -242,7 +242,7 @@
 TEST_F(ReceiveStatisticsProxyTest, OnDecodedFrameIncreasesAssemblyTime) {
   const TimeDelta kAssemblyTime = TimeDelta::Millis(7);
   EXPECT_EQ(0u, statistics_proxy_->GetStats().frames_decoded);
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   TimeDelta expected_total_assembly_time = TimeDelta::Zero();
   unsigned int expected_frames_decoded = 0;
   unsigned int expected_frames_assembled_from_multiple_packets = 0;
@@ -316,7 +316,7 @@
 
 TEST_F(ReceiveStatisticsProxyTest, OnDecodedFrameIncreasesQpSum) {
   EXPECT_EQ(std::nullopt, statistics_proxy_->GetStats().qp_sum);
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   statistics_proxy_->OnDecodedFrame(frame, 3u, TimeDelta::Zero(),
                                     VideoContentType::UNSPECIFIED,
                                     VideoFrameType::kVideoFrameKey);
@@ -329,7 +329,7 @@
 
 TEST_F(ReceiveStatisticsProxyTest, OnDecodedFrameIncreasesTotalDecodeTime) {
   EXPECT_EQ(std::nullopt, statistics_proxy_->GetStats().qp_sum);
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   statistics_proxy_->OnDecodedFrame(frame, 3u, TimeDelta::Millis(4),
                                     VideoContentType::UNSPECIFIED,
                                     VideoFrameType::kVideoFrameKey);
@@ -343,7 +343,7 @@
 TEST_F(ReceiveStatisticsProxyTest, ReportsContentType) {
   const std::string kRealtimeString("realtime");
   const std::string kScreenshareString("screen");
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   EXPECT_EQ(kRealtimeString, videocontenttypehelpers::ToString(
                                  statistics_proxy_->GetStats().content_type));
   statistics_proxy_->OnDecodedFrame(frame, 3u, TimeDelta::Zero(),
@@ -359,7 +359,7 @@
 }
 
 TEST_F(ReceiveStatisticsProxyTest, ReportsMaxInterframeDelay) {
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   const TimeDelta kInterframeDelay1 = TimeDelta::Millis(100);
   const TimeDelta kInterframeDelay2 = TimeDelta::Millis(200);
   const TimeDelta kInterframeDelay3 = TimeDelta::Millis(100);
@@ -390,7 +390,7 @@
 }
 
 TEST_F(ReceiveStatisticsProxyTest, ReportInterframeDelayInWindow) {
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   const TimeDelta kInterframeDelay1 = TimeDelta::Millis(900);
   const TimeDelta kInterframeDelay2 = TimeDelta::Millis(750);
   const TimeDelta kInterframeDelay3 = TimeDelta::Millis(700);
@@ -428,7 +428,7 @@
   EXPECT_EQ(0u, stats.freeze_count);
   EXPECT_FALSE(stats.total_freezes_duration_ms);
 
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   for (size_t i = 0; i < VideoQualityObserver::kMinFrameSamplesToDetectFreeze;
        ++i) {
     time_controller_.AdvanceTime(TimeDelta::Millis(30));
@@ -449,7 +449,7 @@
   ASSERT_EQ(0u, stats.pause_count);
   ASSERT_EQ(0u, stats.total_pauses_duration_ms);
 
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   statistics_proxy_->OnRenderedFrame(MetaData(frame));
 
   // Pause.
@@ -467,7 +467,7 @@
   ASSERT_EQ(0u, stats.pause_count);
   ASSERT_EQ(0u, stats.total_pauses_duration_ms);
 
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
 
   // Pause -> Frame -> Pause
   time_controller_.AdvanceTime(TimeDelta::Seconds(5));
@@ -489,7 +489,7 @@
   VideoReceiveStreamInterface::Stats stats = statistics_proxy_->GetStats();
   ASSERT_EQ(0.0, stats.total_inter_frame_delay);
 
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
 
   // Emulate delay before first frame is rendered. This is needed to ensure
   // that frame duration only covers time since first frame is rendered and
@@ -508,7 +508,7 @@
   VideoReceiveStreamInterface::Stats stats = statistics_proxy_->GetStats();
   ASSERT_EQ(0.0, stats.total_squared_inter_frame_delay);
 
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   for (int i = 0; i <= 10; ++i) {
     time_controller_.AdvanceTime(TimeDelta::Millis(30));
     statistics_proxy_->OnRenderedFrame(MetaData(frame));
@@ -522,7 +522,7 @@
 }
 
 TEST_F(ReceiveStatisticsProxyTest, OnDecodedFrameWithoutQpQpSumWontExist) {
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   EXPECT_EQ(std::nullopt, statistics_proxy_->GetStats().qp_sum);
   statistics_proxy_->OnDecodedFrame(frame, std::nullopt, TimeDelta::Zero(),
                                     VideoContentType::UNSPECIFIED,
@@ -531,7 +531,7 @@
 }
 
 TEST_F(ReceiveStatisticsProxyTest, OnDecodedFrameWithoutQpResetsQpSum) {
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   EXPECT_EQ(std::nullopt, statistics_proxy_->GetStats().qp_sum);
   statistics_proxy_->OnDecodedFrame(frame, 3u, TimeDelta::Zero(),
                                     VideoContentType::UNSPECIFIED,
@@ -545,7 +545,7 @@
 
 TEST_F(ReceiveStatisticsProxyTest, OnRenderedFrameIncreasesFramesRendered) {
   EXPECT_EQ(0u, statistics_proxy_->GetStats().frames_rendered);
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   for (uint32_t i = 1; i <= 3; ++i) {
     statistics_proxy_->OnRenderedFrame(MetaData(frame));
     EXPECT_EQ(i, statistics_proxy_->GetStats().frames_rendered);
@@ -695,7 +695,7 @@
 TEST_F(ReceiveStatisticsProxyTest, GetStatsReportsFrameCounts) {
   const int kKeyFrames = 3;
   const int kDeltaFrames = 22;
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   for (int i = 0; i < kKeyFrames; i++) {
     statistics_proxy_->OnDecodedFrame(frame, std::nullopt, TimeDelta::Zero(),
                                       VideoContentType::UNSPECIFIED,
@@ -774,7 +774,7 @@
   time_controller_.AdvanceTime(kLifetime);
   // Need at least one decoded frame to report stream lifetime.
 
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   statistics_proxy_->OnCompleteFrame(true, 1000, VideoContentType::UNSPECIFIED);
   statistics_proxy_->OnDecodedFrame(
       frame, std::nullopt, TimeDelta::Millis(1000),
@@ -920,7 +920,7 @@
        KeyFrameHistogramNotUpdatedForTooFewSamples) {
   const bool kIsKeyFrame = false;
   const int kFrameSizeBytes = 1000;
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
 
   for (int i = 0; i < kMinRequiredSamples - 1; ++i) {
     statistics_proxy_->OnCompleteFrame(kIsKeyFrame, kFrameSizeBytes,
@@ -945,7 +945,7 @@
        KeyFrameHistogramUpdatedForMinRequiredSamples) {
   const bool kIsKeyFrame = false;
   const int kFrameSizeBytes = 1000;
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
 
   for (int i = 0; i < kMinRequiredSamples; ++i) {
     statistics_proxy_->OnCompleteFrame(kIsKeyFrame, kFrameSizeBytes,
@@ -970,7 +970,7 @@
 
 TEST_F(ReceiveStatisticsProxyTest, KeyFrameHistogramIsUpdated) {
   const int kFrameSizeBytes = 1000;
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
 
   for (int i = 0; i < kMinRequiredSamples; ++i) {
     statistics_proxy_->OnCompleteFrame(true, kFrameSizeBytes,
@@ -1058,7 +1058,7 @@
 
 TEST_F(ReceiveStatisticsProxyTest, DoesNotReportStaleFramerates) {
   const Frequency kDefaultFps = Frequency::Hertz(30);
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
 
   for (int i = 0; i < kDefaultFps.hertz(); ++i) {
     // Since OnRenderedFrame is never called the fps in each sample will be 0,
@@ -1137,7 +1137,7 @@
 }
 
 TEST_F(ReceiveStatisticsProxyTest, ZeroDelayReportedIfFrameNotDelayed) {
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   statistics_proxy_->OnDecodedFrame(frame, std::nullopt, TimeDelta::Zero(),
                                     VideoContentType::UNSPECIFIED,
                                     VideoFrameType::kVideoFrameKey);
@@ -1160,7 +1160,7 @@
 
 TEST_F(ReceiveStatisticsProxyTest,
        DelayedFrameHistogramsAreNotUpdatedIfMinRuntimeHasNotPassed) {
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   statistics_proxy_->OnDecodedFrame(frame, std::nullopt, TimeDelta::Zero(),
                                     VideoContentType::UNSPECIFIED,
                                     VideoFrameType::kVideoFrameKey);
@@ -1182,7 +1182,7 @@
 
 TEST_F(ReceiveStatisticsProxyTest,
        DelayedFramesHistogramsAreNotUpdatedIfNoRenderedFrames) {
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   statistics_proxy_->OnDecodedFrame(frame, std::nullopt, TimeDelta::Zero(),
                                     VideoContentType::UNSPECIFIED,
                                     VideoFrameType::kVideoFrameKey);
@@ -1199,7 +1199,7 @@
 }
 
 TEST_F(ReceiveStatisticsProxyTest, DelayReportedIfFrameIsDelayed) {
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   statistics_proxy_->OnDecodedFrame(frame, std::nullopt, TimeDelta::Zero(),
                                     VideoContentType::UNSPECIFIED,
                                     VideoFrameType::kVideoFrameKey);
@@ -1224,7 +1224,7 @@
 }
 
 TEST_F(ReceiveStatisticsProxyTest, AverageDelayOfDelayedFramesIsReported) {
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
   statistics_proxy_->OnDecodedFrame(frame, std::nullopt, TimeDelta::Zero(),
                                     VideoContentType::UNSPECIFIED,
                                     VideoFrameType::kVideoFrameKey);
@@ -1342,7 +1342,7 @@
 TEST_P(ReceiveStatisticsProxyTestWithFreezeDuration, FreezeDetection) {
   VideoReceiveStreamInterface::Stats stats = statistics_proxy_->GetStats();
   EXPECT_EQ(0u, stats.freeze_count);
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
 
   // Add a very long frame. This is need to verify that average frame
   // duration, which is supposed to be calculated as mean of durations of
@@ -1364,9 +1364,9 @@
 
 class ReceiveStatisticsProxyTestWithContent
     : public ReceiveStatisticsProxyTest,
-      public ::testing::WithParamInterface<webrtc::VideoContentType> {
+      public ::testing::WithParamInterface<VideoContentType> {
  protected:
-  const webrtc::VideoContentType content_type_{GetParam()};
+  const VideoContentType content_type_{GetParam()};
 };
 
 INSTANTIATE_TEST_SUITE_P(ContentTypes,
@@ -1376,7 +1376,7 @@
 
 TEST_P(ReceiveStatisticsProxyTestWithContent, InterFrameDelaysAreReported) {
   const TimeDelta kInterFrameDelay = TimeDelta::Millis(33);
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
 
   for (int i = 0; i < kMinRequiredSamples; ++i) {
     statistics_proxy_->OnDecodedFrame(frame, std::nullopt, TimeDelta::Zero(),
@@ -1413,7 +1413,7 @@
        InterFrameDelaysPercentilesAreReported) {
   const TimeDelta kInterFrameDelay = TimeDelta::Millis(33);
   const int kLastFivePercentsSamples = kMinRequiredSamples * 5 / 100;
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
 
   for (int i = 0; i <= kMinRequiredSamples - kLastFivePercentsSamples; ++i) {
     time_controller_.AdvanceTime(kInterFrameDelay);
@@ -1451,7 +1451,7 @@
 TEST_P(ReceiveStatisticsProxyTestWithContent,
        MaxInterFrameDelayOnlyWithValidAverage) {
   const TimeDelta kInterFrameDelay = TimeDelta::Millis(33);
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
 
   for (int i = 0; i < kMinRequiredSamples; ++i) {
     statistics_proxy_->OnDecodedFrame(frame, std::nullopt, TimeDelta::Zero(),
@@ -1475,7 +1475,7 @@
 
 TEST_P(ReceiveStatisticsProxyTestWithContent, MaxInterFrameDelayOnlyWithPause) {
   const TimeDelta kInterFrameDelay = TimeDelta::Millis(33);
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
 
   for (int i = 0; i <= kMinRequiredSamples; ++i) {
     statistics_proxy_->OnDecodedFrame(frame, std::nullopt, TimeDelta::Zero(),
@@ -1563,7 +1563,7 @@
   const TimeDelta kFreezeDelay = TimeDelta::Millis(200);
   const TimeDelta kCallDuration =
       kMinRequiredSamples * kInterFrameDelay + kFreezeDelay;
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
 
   for (int i = 0; i < kMinRequiredSamples; ++i) {
     VideoFrameMetaData meta = MetaData(frame);
@@ -1612,7 +1612,7 @@
   const TimeDelta kPauseDuration = TimeDelta::Seconds(10);
   const TimeDelta kCallDuration =
       kMinRequiredSamples * kFrameDuration + kFreezeDuration + kPauseDuration;
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
 
   for (int i = 0; i < kMinRequiredSamples; ++i) {
     time_controller_.AdvanceTime(kFrameDuration);
@@ -1661,7 +1661,7 @@
 TEST_P(ReceiveStatisticsProxyTestWithContent, PausesAreIgnored) {
   const TimeDelta kInterFrameDelay = TimeDelta::Millis(33);
   const TimeDelta kPauseDuration = TimeDelta::Seconds(10);
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
 
   for (int i = 0; i <= kMinRequiredSamples; ++i) {
     VideoFrameMetaData meta = MetaData(frame);
@@ -1706,7 +1706,7 @@
 TEST_P(ReceiveStatisticsProxyTestWithContent, ManyPausesAtTheBeginning) {
   const TimeDelta kInterFrameDelay = TimeDelta::Millis(33);
   const TimeDelta kPauseDuration = TimeDelta::Seconds(10);
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
 
   for (int i = 0; i <= kMinRequiredSamples; ++i) {
     statistics_proxy_->OnDecodedFrame(frame, std::nullopt, TimeDelta::Zero(),
@@ -1736,8 +1736,8 @@
 
 TEST_P(ReceiveStatisticsProxyTestWithContent, TimeInHdReported) {
   const TimeDelta kInterFrameDelay = TimeDelta::Millis(20);
-  webrtc::VideoFrame frame_hd = CreateFrame(1280, 720);
-  webrtc::VideoFrame frame_sd = CreateFrame(640, 360);
+  VideoFrame frame_hd = CreateFrame(1280, 720);
+  VideoFrame frame_sd = CreateFrame(640, 360);
 
   // HD frames.
   for (int i = 0; i < kMinRequiredSamples; ++i) {
@@ -1777,7 +1777,7 @@
   const TimeDelta kInterFrameDelay = TimeDelta::Millis(20);
   const int kHighQp = 80;
   const int kLowQp = 30;
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
 
   // High quality frames.
   for (int i = 0; i < kMinRequiredSamples; ++i) {
@@ -1821,9 +1821,9 @@
   // To ensure long enough call duration.
   const TimeDelta kInterFrameDelay = TimeDelta::Seconds(2);
 
-  webrtc::VideoFrame frame_hd = CreateFrame(1280, 720);
-  webrtc::VideoFrame frame_sd = CreateFrame(640, 360);
-  webrtc::VideoFrame frame_ld = CreateFrame(320, 180);
+  VideoFrame frame_hd = CreateFrame(1280, 720);
+  VideoFrame frame_sd = CreateFrame(640, 360);
+  VideoFrame frame_ld = CreateFrame(320, 180);
 
   // Call once to pass content type.
   statistics_proxy_->OnDecodedFrame(frame_hd, std::nullopt, TimeDelta::Zero(),
@@ -1854,7 +1854,7 @@
   const int kLowQp = 30;
   const TimeDelta kDecodeTime = TimeDelta::Millis(7);
 
-  webrtc::VideoFrame frame = CreateFrame(kWidth, kHeight);
+  VideoFrame frame = CreateFrame(kWidth, kHeight);
 
   for (int i = 0; i < kMinRequiredSamples; ++i) {
     statistics_proxy_->OnDecodedFrame(frame, kLowQp, kDecodeTime, content_type_,
diff --git a/video/rtp_video_stream_receiver2.cc b/video/rtp_video_stream_receiver2.cc
index bdd2cc2..3e41758 100644
--- a/video/rtp_video_stream_receiver2.cc
+++ b/video/rtp_video_stream_receiver2.cc
@@ -393,7 +393,7 @@
 void RtpVideoStreamReceiver2::AddReceiveCodec(
     uint8_t payload_type,
     VideoCodecType video_codec,
-    const webrtc::CodecParameterMap& codec_params,
+    const CodecParameterMap& codec_params,
     bool raw_payload) {
   RTC_DCHECK_RUN_ON(&packet_sequence_checker_);
   if (codec_params.count(kH264FmtpSpsPpsIdrInKeyframe) > 0 ||
diff --git a/video/rtp_video_stream_receiver2_unittest.cc b/video/rtp_video_stream_receiver2_unittest.cc
index 6a8f963..3440601 100644
--- a/video/rtp_video_stream_receiver2_unittest.cc
+++ b/video/rtp_video_stream_receiver2_unittest.cc
@@ -348,7 +348,7 @@
   received_packet_generator.SetColorSpace(kColorSpace);
 
   // Prepare the receiver for VP9.
-  webrtc::CodecParameterMap codec_params;
+  CodecParameterMap codec_params;
   rtp_video_stream_receiver_->AddReceiveCodec(kVp9PayloadType, kVideoCodecVP9,
                                               codec_params,
                                               /*raw_payload=*/false);
@@ -476,7 +476,7 @@
   const std::vector<uint8_t> kDeltaFramePayload = {5, 6, 7, 8, 9};
 
   // Prepare the receiver for VP9.
-  webrtc::CodecParameterMap codec_params;
+  CodecParameterMap codec_params;
   rtp_video_stream_receiver_->AddReceiveCodec(kVp9PayloadType, kVideoCodecVP9,
                                               codec_params,
                                               /*raw_payload=*/false);
@@ -553,7 +553,7 @@
   const std::vector<uint8_t> kDeltaFramePayload = {5, 6, 7, 8, 9};
 
   // Prepare the receiver for VP9.
-  webrtc::CodecParameterMap codec_params;
+  CodecParameterMap codec_params;
   rtp_video_stream_receiver_->AddReceiveCodec(kVp9PayloadType, kVideoCodecVP9,
                                               codec_params,
                                               /*raw_payload=*/false);
@@ -799,7 +799,7 @@
 
 TEST_P(RtpVideoStreamReceiver2TestH264, InBandSpsPps) {
   constexpr int kH264PayloadType = 98;
-  webrtc::CodecParameterMap codec_params;
+  CodecParameterMap codec_params;
   rtp_video_stream_receiver_->AddReceiveCodec(kH264PayloadType, kVideoCodecH264,
                                               codec_params,
                                               /*raw_payload=*/false);
@@ -856,7 +856,7 @@
 
 TEST_P(RtpVideoStreamReceiver2TestH264, OutOfBandFmtpSpsPps) {
   constexpr int kH264PayloadType = 99;
-  webrtc::CodecParameterMap codec_params;
+  CodecParameterMap codec_params;
   // Example parameter sets from https://tools.ietf.org/html/rfc3984#section-8.2
   // .
   codec_params.insert({kH264FmtpSpropParameterSets, "Z0IACpZTBYmI,aMljiA=="});
@@ -903,7 +903,7 @@
 }
 
 TEST_P(RtpVideoStreamReceiver2TestH264, ForceSpsPpsIdrIsKeyframe) {
-  webrtc::CodecParameterMap codec_params;
+  CodecParameterMap codec_params;
   // Forcing can be done either with field trial or codec_params.
   if (!env_.field_trials().IsEnabled("WebRTC-SpsPpsIdrIsH264Keyframe")) {
     codec_params.insert({kH264FmtpSpsPpsIdrInKeyframe, ""});
@@ -1063,7 +1063,7 @@
   rtp_packet_slice.Parse(raw_rtp_slice.data(), raw_rtp_slice.size());
 
   // Prepare the receiver for H264.
-  webrtc::CodecParameterMap codec_params;
+  CodecParameterMap codec_params;
   rtp_video_stream_receiver_->AddReceiveCodec(kH264PayloadType, kVideoCodecH264,
                                               codec_params, false);
   rtp_video_stream_receiver_->StartReceive();
@@ -1770,7 +1770,7 @@
 
 TEST_F(RtpVideoStreamReceiver2TestH265, H265Bitstream) {
   constexpr int kH265PayloadType = 98;
-  webrtc::CodecParameterMap codec_params;
+  CodecParameterMap codec_params;
   rtp_video_stream_receiver_->AddReceiveCodec(kH265PayloadType, kVideoCodecH265,
                                               codec_params,
                                               /*raw_payload=*/false);
diff --git a/video/send_statistics_proxy.cc b/video/send_statistics_proxy.cc
index 38873b4f..e6858df 100644
--- a/video/send_statistics_proxy.cc
+++ b/video/send_statistics_proxy.cc
@@ -147,7 +147,7 @@
 }
 
 std::optional<int> GetFallbackMaxPixelsIfFieldTrialEnabled(
-    const webrtc::FieldTrialsView& field_trials) {
+    const FieldTrialsView& field_trials) {
   std::string group = field_trials.Lookup(kVp8ForcedFallbackEncoderFieldTrial);
   return (absl::StartsWith(group, "Enabled"))
              ? GetFallbackMaxPixels(group.substr(7))
@@ -155,7 +155,7 @@
 }
 
 std::optional<int> GetFallbackMaxPixelsIfFieldTrialDisabled(
-    const webrtc::FieldTrialsView& field_trials) {
+    const FieldTrialsView& field_trials) {
   std::string group = field_trials.Lookup(kVp8ForcedFallbackEncoderFieldTrial);
   return (absl::StartsWith(group, "Disabled"))
              ? GetFallbackMaxPixels(group.substr(8))
diff --git a/video/video_analyzer.cc b/video/video_analyzer.cc
index 94f06ee..159bd8c 100644
--- a/video/video_analyzer.cc
+++ b/video/video_analyzer.cc
@@ -78,10 +78,10 @@
 namespace webrtc {
 namespace {
 
-using ::webrtc::test::GetGlobalMetricsLogger;
-using ::webrtc::test::ImprovementDirection;
-using ::webrtc::test::Metric;
-using ::webrtc::test::Unit;
+using test::GetGlobalMetricsLogger;
+using test::ImprovementDirection;
+using test::Metric;
+using test::Unit;
 
 constexpr TimeDelta kSendStatsPollingInterval = TimeDelta::Seconds(1);
 constexpr size_t kMaxComparisons = 10;
diff --git a/video/video_loopback.cc b/video/video_loopback.cc
index b8659d1..f2ab825 100644
--- a/video/video_loopback.cc
+++ b/video/video_loopback.cc
@@ -442,9 +442,9 @@
   // InitFieldTrialsFromString stores the char*, so the char array must outlive
   // the application.
   const std::string field_trials = absl::GetFlag(FLAGS_force_fieldtrials);
-  webrtc::field_trial::InitFieldTrialsFromString(field_trials.c_str());
+  field_trial::InitFieldTrialsFromString(field_trials.c_str());
 
-  webrtc::test::RunTest(webrtc::Loopback);
+  test::RunTest(Loopback);
   return 0;
 }
 }  // namespace webrtc
diff --git a/video/video_quality_test.cc b/video/video_quality_test.cc
index 66a8a25..d4e1754 100644
--- a/video/video_quality_test.cc
+++ b/video/video_quality_test.cc
@@ -340,7 +340,7 @@
     const SdpVideoFormat& format) {
   std::unique_ptr<VideoDecoder> decoder;
   if (format.name == "FakeCodec") {
-    decoder = webrtc::FakeVideoDecoderFactory::CreateVideoDecoder();
+    decoder = FakeVideoDecoderFactory::CreateVideoDecoder();
   } else {
     decoder = decoder_factory_->Create(env, format);
   }
@@ -644,14 +644,14 @@
     const std::vector<std::string>& sl_descriptors) {
   if (params->ss[video_idx].streams.empty() &&
       params->ss[video_idx].infer_streams) {
-    webrtc::VideoEncoder::EncoderInfo encoder_info;
-    webrtc::VideoEncoderConfig encoder_config;
+    VideoEncoder::EncoderInfo encoder_info;
+    VideoEncoderConfig encoder_config;
     encoder_config.codec_type =
         PayloadStringToCodecType(params->video[video_idx].codec);
     encoder_config.content_type =
         params->screenshare[video_idx].enabled
-            ? webrtc::VideoEncoderConfig::ContentType::kScreen
-            : webrtc::VideoEncoderConfig::ContentType::kRealtimeVideo;
+            ? VideoEncoderConfig::ContentType::kScreen
+            : VideoEncoderConfig::ContentType::kRealtimeVideo;
     encoder_config.max_bitrate_bps = params->video[video_idx].max_bitrate_bps;
     encoder_config.min_transmit_bitrate_bps =
         params->video[video_idx].min_transmit_bps;
@@ -737,7 +737,7 @@
   RTC_CHECK(num_video_streams_ > 0);
   video_encoder_configs_.resize(num_video_streams_);
   std::string generic_codec_name;
-  webrtc::VideoEncoder::EncoderInfo encoder_info;
+  VideoEncoder::EncoderInfo encoder_info;
   for (size_t video_idx = 0; video_idx < num_video_streams_; ++video_idx) {
     VideoSendStream::Config config(send_transport);
     config.rtp.extmap_allow_mixed = true;
@@ -1423,7 +1423,7 @@
     // The real ADM requires extra initialization: setting default devices,
     // setting up number of channels etc. Helper class also calls
     // AudioDeviceModule::Init().
-    webrtc::adm_helpers::Init(audio_device.get());
+    adm_helpers::Init(audio_device.get());
   } else {
     audio_device->Init();
   }
@@ -1448,8 +1448,8 @@
 
   if (params_.call.send_side_bwe) {
     audio_send_config.rtp.extensions.push_back(
-        webrtc::RtpExtension(webrtc::RtpExtension::kTransportSequenceNumberUri,
-                             kTransportSequenceNumberExtensionId));
+        RtpExtension(RtpExtension::kTransportSequenceNumberUri,
+                     kTransportSequenceNumberExtensionId));
     audio_send_config.min_bitrate_bps = kOpusMinBitrateBps;
     audio_send_config.max_bitrate_bps = kOpusBitrateFbBps;
     // Only allow ANA when send-side BWE is enabled.
diff --git a/video/video_receive_stream2.cc b/video/video_receive_stream2.cc
index 3aced70..c7c1c30 100644
--- a/video/video_receive_stream2.cc
+++ b/video/video_receive_stream2.cc
@@ -132,11 +132,11 @@
     return buffer_;
   }
 
-  std::optional<webrtc::ColorSpace> color_space() const override {
+  std::optional<ColorSpace> color_space() const override {
     return color_space_;
   }
 
-  std::optional<webrtc::VideoRotation> video_rotation() const override {
+  std::optional<VideoRotation> video_rotation() const override {
     return video_rotation_;
   }
 
@@ -156,8 +156,8 @@
   VideoCodecType codec_;
   bool is_key_frame_;
   EncodedResolution resolution_;
-  std::optional<webrtc::ColorSpace> color_space_;
-  std::optional<webrtc::VideoRotation> video_rotation_;
+  std::optional<ColorSpace> color_space_;
+  std::optional<VideoRotation> video_rotation_;
 };
 
 RenderResolution InitialDecoderResolution(const FieldTrialsView& field_trials) {
@@ -174,21 +174,21 @@
 
 // Video decoder class to be used for unknown codecs. Doesn't support decoding
 // but logs messages to LS_ERROR.
-class NullVideoDecoder : public webrtc::VideoDecoder {
+class NullVideoDecoder : public VideoDecoder {
  public:
   bool Configure(const Settings& settings) override {
     RTC_LOG(LS_ERROR) << "Can't initialize NullVideoDecoder.";
     return true;
   }
 
-  int32_t Decode(const webrtc::EncodedImage& input_image,
+  int32_t Decode(const EncodedImage& input_image,
                  int64_t render_time_ms) override {
     RTC_LOG(LS_ERROR) << "The NullVideoDecoder doesn't support decoding.";
     return WEBRTC_VIDEO_CODEC_OK;
   }
 
   int32_t RegisterDecodeCompleteCallback(
-      webrtc::DecodedImageCallback* callback) override {
+      DecodedImageCallback* callback) override {
     RTC_LOG(LS_ERROR)
         << "Can't register decode complete callback on NullVideoDecoder.";
     return WEBRTC_VIDEO_CODEC_OK;
@@ -710,7 +710,7 @@
                                          frame_meta.decode_timestamp);
       }));
 
-  webrtc::MutexLock lock(&pending_resolution_mutex_);
+  MutexLock lock(&pending_resolution_mutex_);
   if (pending_resolution_.has_value()) {
     if (!pending_resolution_->empty() &&
         (video_frame.width() != static_cast<int>(pending_resolution_->width) ||
@@ -729,7 +729,7 @@
 }
 
 void VideoReceiveStream2::SetFrameDecryptor(
-    scoped_refptr<webrtc::FrameDecryptorInterface> frame_decryptor) {
+    scoped_refptr<FrameDecryptorInterface> frame_decryptor) {
   rtp_video_stream_receiver_.SetFrameDecryptor(std::move(frame_decryptor));
 }
 
@@ -973,7 +973,7 @@
       RTC_LOG(LS_ERROR) << "About to halt recordable encoded frame output due "
                            "to too many buffered frames.";
 
-    webrtc::MutexLock lock(&pending_resolution_mutex_);
+    MutexLock lock(&pending_resolution_mutex_);
     if (IsKeyFrameAndUnspecifiedResolution(*frame_ptr) &&
         !pending_resolution_.has_value())
       pending_resolution_.emplace();
@@ -1001,7 +1001,7 @@
     {
       // Fish out `pending_resolution_` to avoid taking the mutex on every lap
       // or dispatching under the mutex in the flush loop.
-      webrtc::MutexLock lock(&pending_resolution_mutex_);
+      MutexLock lock(&pending_resolution_mutex_);
       if (pending_resolution_.has_value())
         pending_resolution = *pending_resolution_;
     }
@@ -1134,7 +1134,7 @@
   }
 }
 
-std::vector<webrtc::RtpSource> VideoReceiveStream2::GetSources() const {
+std::vector<RtpSource> VideoReceiveStream2::GetSources() const {
   RTC_DCHECK_RUN_ON(&worker_sequence_checker_);
   return source_tracker_.GetSources();
 }
diff --git a/video/video_receive_stream2_unittest.cc b/video/video_receive_stream2_unittest.cc
index 61e1872..b19e001 100644
--- a/video/video_receive_stream2_unittest.cc
+++ b/video/video_receive_stream2_unittest.cc
@@ -257,12 +257,10 @@
       video_receive_stream_ = nullptr;
     }
     timing_ = new VCMTiming(&env_.clock(), env_.field_trials());
-    video_receive_stream_ =
-        std::make_unique<webrtc::internal::VideoReceiveStream2>(
-            env_, &fake_call_, kDefaultNumCpuCores, &packet_router_,
-            config_.Copy(), &call_stats_, absl::WrapUnique(timing_),
-            &nack_periodic_processor_,
-            UseMetronome() ? &decode_sync_ : nullptr);
+    video_receive_stream_ = std::make_unique<internal::VideoReceiveStream2>(
+        env_, &fake_call_, kDefaultNumCpuCores, &packet_router_, config_.Copy(),
+        &call_stats_, absl::WrapUnique(timing_), &nack_periodic_processor_,
+        UseMetronome() ? &decode_sync_ : nullptr);
     video_receive_stream_->RegisterWithTransport(
         &rtp_stream_receiver_controller_);
     if (state)
@@ -283,7 +281,7 @@
   test::RtcpPacketParser rtcp_packet_parser_;
   PacketRouter packet_router_;
   RtpStreamReceiverController rtp_stream_receiver_controller_;
-  std::unique_ptr<webrtc::internal::VideoReceiveStream2> video_receive_stream_;
+  std::unique_ptr<internal::VideoReceiveStream2> video_receive_stream_;
   VCMTiming* timing_;
   test::FakeMetronome fake_metronome_;
   DecodeSynchronizer decode_sync_;
@@ -595,7 +593,7 @@
 }
 
 TEST_P(VideoReceiveStream2Test, PassesRotation) {
-  const webrtc::VideoRotation kRotation = webrtc::kVideoRotation_180;
+  const VideoRotation kRotation = kVideoRotation_180;
   std::unique_ptr<test::FakeEncodedFrame> test_frame =
       test::FakeFrameBuilder()
           .Id(0)
diff --git a/video/video_send_stream_impl.cc b/video/video_send_stream_impl.cc
index 9f093f5..c0c8dff 100644
--- a/video/video_send_stream_impl.cc
+++ b/video/video_send_stream_impl.cc
@@ -228,10 +228,9 @@
                               const FieldTrialsView& field_trials) {
   int priority_bitrate = 0;
   if (PayloadStringToCodecType(codec_name) == VideoCodecType::kVideoCodecAV1) {
-    webrtc::FieldTrialParameter<int> av1_priority_bitrate("bitrate", 0);
-    webrtc::ParseFieldTrial(
-        {&av1_priority_bitrate},
-        field_trials.Lookup("WebRTC-AV1-OverridePriorityBitrate"));
+    FieldTrialParameter<int> av1_priority_bitrate("bitrate", 0);
+    ParseFieldTrial({&av1_priority_bitrate},
+                    field_trials.Lookup("WebRTC-AV1-OverridePriorityBitrate"));
     priority_bitrate = av1_priority_bitrate;
   }
   return priority_bitrate;
@@ -299,9 +298,8 @@
 VideoStreamEncoder::BitrateAllocationCallbackType
 GetBitrateAllocationCallbackType(const VideoSendStream::Config& config,
                                  const FieldTrialsView& field_trials) {
-  if (webrtc::RtpExtension::FindHeaderExtensionByUri(
-          config.rtp.extensions,
-          webrtc::RtpExtension::kVideoLayersAllocationUri,
+  if (RtpExtension::FindHeaderExtensionByUri(
+          config.rtp.extensions, RtpExtension::kVideoLayersAllocationUri,
           config.crypto_options.srtp.enable_encrypted_rtp_header_extensions
               ? RtpExtension::Filter::kPreferEncryptedExtension
               : RtpExtension::Filter::kDiscardEncryptedExtension)) {
@@ -349,7 +347,7 @@
     VideoStreamEncoder::BitrateAllocationCallbackType
         bitrate_allocation_callback_type,
     Metronome* metronome,
-    webrtc::VideoEncoderFactory::EncoderSelectorInterface* encoder_selector) {
+    VideoEncoderFactory::EncoderSelectorInterface* encoder_selector) {
   std::unique_ptr<TaskQueueBase, TaskQueueDeleter> encoder_queue =
       env.task_queue_factory().CreateTaskQueue(
           "EncoderQueue", TaskQueueFactory::Priority::NORMAL);
@@ -542,7 +540,7 @@
 }
 
 void VideoSendStreamImpl::SetSource(
-    VideoSourceInterface<webrtc::VideoFrame>* source,
+    VideoSourceInterface<VideoFrame>* source,
     const DegradationPreference& degradation_preference) {
   RTC_DCHECK_RUN_ON(&thread_checker_);
   video_stream_encoder_->SetSource(source, degradation_preference);
diff --git a/video/video_send_stream_impl_unittest.cc b/video/video_send_stream_impl_unittest.cc
index 7a11b78..f0023f1 100644
--- a/video/video_send_stream_impl_unittest.cc
+++ b/video/video_send_stream_impl_unittest.cc
@@ -126,8 +126,7 @@
   MOCK_METHOD(void, SetEncodingData, (size_t, size_t, size_t), (override));
   MOCK_METHOD(std::vector<RtpSequenceNumberMap::Info>,
               GetSentRtpPacketInfos,
-              (uint32_t ssrc,
-               webrtc::ArrayView<const uint16_t> sequence_numbers),
+              (uint32_t ssrc, ArrayView<const uint16_t> sequence_numbers),
               (const, override));
 
   MOCK_METHOD(void, SetFecAllowed, (bool fec_allowed), (override));
@@ -217,7 +216,7 @@
 
  protected:
   GlobalSimulatedTimeController time_controller_;
-  webrtc::test::ScopedKeyValueConfig field_trials_;
+  test::ScopedKeyValueConfig field_trials_;
   NiceMock<MockTransport> transport_;
   NiceMock<MockRtpTransportControllerSend> transport_controller_;
   NiceMock<MockBitrateAllocator> bitrate_allocator_;
diff --git a/video/video_send_stream_tests.cc b/video/video_send_stream_tests.cc
index 69e0c8e..d73601e 100644
--- a/video/video_send_stream_tests.cc
+++ b/video/video_send_stream_tests.cc
@@ -126,7 +126,7 @@
 namespace test {
 class VideoSendStreamPeer {
  public:
-  explicit VideoSendStreamPeer(webrtc::VideoSendStream* base_class_stream)
+  explicit VideoSendStreamPeer(VideoSendStream* base_class_stream)
       : internal_stream_(
             static_cast<internal::VideoSendStreamImpl*>(base_class_stream)) {}
   std::optional<float> GetPacingFactorOverride() const {
@@ -1754,8 +1754,8 @@
     }
 
    private:
-    webrtc::SequenceChecker module_process_thread_;
-    webrtc::SequenceChecker task_queue_thread_;
+    SequenceChecker module_process_thread_;
+    SequenceChecker task_queue_thread_;
     TaskQueueBase* const task_queue_;
     RtpHeaderExtensionMap extensions_;
     Call* call_ RTC_GUARDED_BY(task_queue_thread_);
@@ -1771,7 +1771,7 @@
   static const int kStartBitrateBps = 300000;
   static const int kRelayBandwidthCapBps = 800000;
   static const int kMinPacketsToSend = 100;
-  webrtc::test::ScopedKeyValueConfig field_trials(
+  test::ScopedKeyValueConfig field_trials(
       field_trials_, "WebRTC-Bwe-NetworkRouteConstraints/relay_cap:" +
                          std::to_string(kRelayBandwidthCapBps) + "bps/");
 
@@ -1858,8 +1858,8 @@
     }
 
    private:
-    webrtc::SequenceChecker module_process_thread_;
-    webrtc::SequenceChecker task_queue_thread_;
+    SequenceChecker module_process_thread_;
+    SequenceChecker task_queue_thread_;
     TaskQueueBase* const task_queue_;
     Call* call_ RTC_GUARDED_BY(task_queue_thread_);
     int packets_sent_ RTC_GUARDED_BY(task_queue_thread_);
@@ -2049,11 +2049,11 @@
   }
 
  private:
-  webrtc::SequenceChecker task_queue_thread_;
+  SequenceChecker task_queue_thread_;
   Call* call_ RTC_GUARDED_BY(task_queue_thread_) = nullptr;
   VideoSendStream::Config send_stream_config_{nullptr};
   VideoEncoderConfig encoder_config_;
-  webrtc::SequenceChecker module_process_thread_;
+  SequenceChecker module_process_thread_;
   uint32_t packets_sent_ RTC_GUARDED_BY(task_queue_thread_) = 0;
   bool running_without_padding_ RTC_GUARDED_BY(task_queue_thread_);
   T* const stream_resetter_;
@@ -2729,7 +2729,7 @@
   // TODO(bugs.webrtc.org/12058): If these fields trial are on, we get lower
   // bitrates than expected by this test, due to encoder pushback and subtracted
   // overhead.
-  webrtc::test::ScopedKeyValueConfig field_trials(
+  test::ScopedKeyValueConfig field_trials(
       field_trials_, "WebRTC-VideoRateControl/bitrate_adjuster:false/");
 
   class EncoderBitrateThresholdObserver : public test::SendTest,
@@ -2919,11 +2919,11 @@
 
     int num_rate_allocator_creations_;
     int num_encoder_initializations_;
-    webrtc::Call* call_;
-    webrtc::VideoSendStream* send_stream_;
+    Call* call_;
+    VideoSendStream* send_stream_;
     test::VideoEncoderProxyFactory encoder_factory_;
     std::unique_ptr<VideoBitrateAllocatorFactory> bitrate_allocator_factory_;
-    webrtc::VideoEncoderConfig encoder_config_;
+    VideoEncoderConfig encoder_config_;
   } test(env(), task_queue());
 
   RunBaseTest(&test);
@@ -3380,7 +3380,7 @@
   test::FunctionVideoEncoderFactory encoder_factory_;
   const Vp9TestParams params_;
   VideoCodecVP9 vp9_settings_;
-  webrtc::VideoEncoderConfig encoder_config_;
+  VideoEncoderConfig encoder_config_;
   bool last_packet_marker_ = false;
   std::optional<uint16_t> last_packet_sequence_number_;
   std::optional<uint32_t> last_packet_timestamp_;
@@ -3979,7 +3979,7 @@
           internal_send_peer.GetPacingFactorOverride().value_or(0.0f);
       float expected_pacing_factor = 1.1;  // Strict pacing factor.
       VideoSendStream::Stats stats = send_stream_->GetStats();
-      if (stats.content_type == webrtc::VideoContentType::SCREENSHARE) {
+      if (stats.content_type == VideoContentType::SCREENSHARE) {
         expected_pacing_factor = 1.0f;  // Currently used pacing factor in ALR.
       }
 
@@ -4123,7 +4123,7 @@
         VideoSendStream::Config* send_config,
         std::vector<VideoReceiveStreamInterface::Config>* receive_configs,
         VideoEncoderConfig* encoder_config) override {
-      webrtc::VideoEncoder::EncoderInfo encoder_info;
+      VideoEncoder::EncoderInfo encoder_info;
       send_config->encoder_settings.encoder_factory = encoder_factory_;
       send_config->rtp.payload_name = payload_name_;
       send_config->rtp.payload_type =
diff --git a/video/video_source_sink_controller_unittest.cc b/video/video_source_sink_controller_unittest.cc
index bc4d263..603c726 100644
--- a/video/video_source_sink_controller_unittest.cc
+++ b/video/video_source_sink_controller_unittest.cc
@@ -43,13 +43,9 @@
 
   MOCK_METHOD(void,
               AddOrUpdateSink,
-              (webrtc::VideoSinkInterface<VideoFrame>*,
-               const webrtc::VideoSinkWants&),
+              (VideoSinkInterface<VideoFrame>*, const VideoSinkWants&),
               (override));
-  MOCK_METHOD(void,
-              RemoveSink,
-              (webrtc::VideoSinkInterface<VideoFrame>*),
-              (override));
+  MOCK_METHOD(void, RemoveSink, (VideoSinkInterface<VideoFrame>*), (override));
   MOCK_METHOD(void, RequestRefreshFrame, (), (override));
 };
 
diff --git a/video/video_stream_encoder.cc b/video/video_stream_encoder.cc
index 8623944..212108e 100644
--- a/video/video_stream_encoder.cc
+++ b/video/video_stream_encoder.cc
@@ -346,9 +346,9 @@
         encoder_config.VP9().interLayerPred == InterLayerPredMode::kOn;
     layers_allocation.resolution_and_frame_rate_is_valid = true;
 
-    std::vector<DataRate> aggregated_spatial_bitrate(
-        webrtc::kMaxTemporalStreams, DataRate::Zero());
-    for (int si = 0; si < webrtc::kMaxSpatialLayers; ++si) {
+    std::vector<DataRate> aggregated_spatial_bitrate(kMaxTemporalStreams,
+                                                     DataRate::Zero());
+    for (int si = 0; si < kMaxSpatialLayers; ++si) {
       layers_allocation.resolution_and_frame_rate_is_valid = true;
       if (!target_bitrate.IsSpatialLayerUsed(si) ||
           target_bitrate.GetSpatialLayerSum(si) == 0) {
@@ -692,10 +692,9 @@
     const VideoStreamEncoderSettings& settings,
     std::unique_ptr<OveruseFrameDetector> overuse_detector,
     std::unique_ptr<FrameCadenceAdapterInterface> frame_cadence_adapter,
-    std::unique_ptr<webrtc::TaskQueueBase, webrtc::TaskQueueDeleter>
-        encoder_queue,
+    std::unique_ptr<TaskQueueBase, TaskQueueDeleter> encoder_queue,
     BitrateAllocationCallbackType allocation_cb_type,
-    webrtc::VideoEncoderFactory::EncoderSelectorInterface* encoder_selector)
+    VideoEncoderFactory::EncoderSelectorInterface* encoder_selector)
     : env_(env),
       worker_queue_(TaskQueueBase::Current()),
       number_of_cores_(number_of_cores),
@@ -1012,7 +1011,7 @@
 
       ReconfigureEncoder();
     } else {
-      webrtc::InvokeSetParametersCallback(callback, webrtc::RTCError::OK());
+      InvokeSetParametersCallback(callback, RTCError::OK());
     }
   });
 }
@@ -1485,7 +1484,7 @@
   stream_resource_manager_.ConfigureQualityScaler(info);
   stream_resource_manager_.ConfigureBandwidthQualityScaler(info);
 
-  webrtc::RTCError encoder_configuration_result = webrtc::RTCError::OK();
+  RTCError encoder_configuration_result = RTCError::OK();
 
   if (!encoder_initialized_) {
     RTC_LOG(LS_WARNING) << "Failed to initialize "
@@ -1497,14 +1496,13 @@
       RequestEncoderSwitch();
     } else {
       encoder_configuration_result =
-          webrtc::RTCError(RTCErrorType::UNSUPPORTED_OPERATION);
+          RTCError(RTCErrorType::UNSUPPORTED_OPERATION);
     }
   }
 
   if (!encoder_configuration_callbacks_.empty()) {
     for (auto& callback : encoder_configuration_callbacks_) {
-      webrtc::InvokeSetParametersCallback(callback,
-                                          encoder_configuration_result);
+      InvokeSetParametersCallback(callback, encoder_configuration_result);
     }
     encoder_configuration_callbacks_.clear();
   }